Taking Aim at Cancer’s Heart


Cancer is a unique paradox. At one level it’s as easy as can be to describe: damage to DNA (aka mutations) drives cells to behave abnormally — to make more of themselves when they shouldn’t.

But we all know that cancer’s fiendishly complicated — at least at the level of fine detail. Over the last decade or so the avalanche of sequenced DNA has revealed that every cell in a tumour is different: compare one cell to its neighbour and you’ll find variations in the individual units (the bases A, C, G & T) that make up the chains of DNA.

It’s a nightmare: every cancer is different so we need an infinite number of treatments to control or cure each one. Time to give up and retire to the pub.

Drivers and passengers

Not quite. DNA sequencing has also revealed that, amongst all the genetic mayhem, some mutations are more important than others. The movers and shakers have been dubbed ‘drivers’: those that come along for the ride are ‘passengers’. The hangers-on are heavily in the majority but, even so, several hundred drivers (i.e. mutated genes that give rise to abnormal proteins) have been identified. As it needs a group of drivers to work together to make a cancer we still have the problem that the number of critical combinations that can arise is essentially infinite.

One way of reducing the scale of the problem has been to look at what ‘driver’ proteins do in cells and to target those acting at key points to push cell proliferation beyond the normal.

Playing games

Just recently Giulio CaravagnaAndrea Sottoriva and colleagues at the Institute of Cancer Research, London and the University of Edinburgh have come up with a different approach. The idea goes back to the 1950s when a clever chap from Kansas by the name of Arthur Samuel came up with a program for IBM’s first commercial computer so that it could play draughts (checkers as our American friends call it) in its spare time. The program defined the patterns that could be formed by the pieces on the chequerboard so that, given enough of these, IBM 701 could indicate the optimal moves. Samuel called this machine learning, a precursor of the idea of artificial intelligence.

Perhaps the most famous moment in this saga came in 1997 when a later IBM computer, Deep Blue, beat the then world chess champion Garry Kasparov. Unsurprisingly, Kasparov was a bit miffed and accused IBM of cheating — to wit, getting a human to tell the machine what to do. Let’s hope that in the end he came to terms with the fact that Deep Blue could crank through 200 million positions per second and, however many games Grandmasters have in their heads, they can’t compete with that.

The cancer team realized that the mutations driving the evolution of cancer cells emerge as patterns in the sequence of DNA as a cell moves towards becoming independent of normal controls. Think of each cancer as a family tree of mutations, the key question being which branch leads to the most potent combination.

To pick out these patterns they applied a machine-learning approach known as transfer learning to the DNA sequences from a large number of cancers. They called this ‘repeated evolution in cancer’ — REVOLVER — aimed at picking out mutation patterns at the heart of cancer that foreshadow future genetic changes and can be used to predict how they will evolve.

Identifying patterns of mutation common to different tumours.

Samples are taken from different regions of a patient’s tumour (represented by the coloured dots). Their DNA sequences will have multiple variations that can mask underlying patterns of driver mutations present in some subgroups. The five trees show mutations picked up in those patients. REVOLVER uses transfer learning to screen the sequence data from many patients and pull out evolutionary trajectories shared by subgroups. The dotted red lines highlight common patterns that are represented in the lower strip. From Caravagna et al. 2018.

REVOLVER was applied to sequences from lung, breast, kidney and bowel cancers but there’s no reason it shouldn’t work with other tumours. The big attraction is that if these mini-sequence mutation patterns can be associated generally with how a given tumour develops they should help to inform treatment options and predict survival.

We have in the past referred to the ways cancers evolve as ‘genetic roulette’ — so perhaps it’s appropriate if game-playing computer programs turn out to be useful in teasing out behavioural clues.


Caravagna, G. et al. (2018). Detecting repeated cancer evolution from multi-region tumor sequencing data. Nature Methods 15, 707–714.


Turning Ourselves On


It may seem a bit tasteless but we have to admit that cancer’s a very ‘trendy’ field. That is, there’s always a current fad — something for which either the media or cancer scientists themselves have the hots. Inevitable I suppose, given the importance of cancer to pretty well everyone and the fact that something’s always happening.

If you had to pick the front-running trends of late I guess most of us might go for ‘personalized medicine’ and ‘immunotherapy.’ The first means tailoring treatment to the individual patient, the second is boosting the innate power of the immune system to fight cancer.

Few things are trendier than this blog so it goes without saying that we’ve done endless pieces on these topics (e.g. Fantastic Stuff, Outsourcing the Immune Response, Self-Help – Part 2, bla, bla, bla).

How considerate then of Krijn Dijkstra, Hans Clevers, Emile Voest and colleagues from the Netherlands Cancer Institute to have neatly combined the two in their recent paper.

Simple really

What they did was did was easy — in principle. They grew fresh tumour tissue from patients in dishes in the laboratory. Although it doesn’t work every time, most of the main types of cancer have been grown in this way to give 3D cultures called tumour organoids — tumours-in-a-dish. That’s the ‘personalized’ bit.

Then they took blood from the patient and grew the lymphocytes therein in a dish to expand the T cells that were specific for the patient’s tumour. That’s the ‘‘immuno’ bit.

Growing tumour tissue (from non-small-cell lung cancer (NSCLC) and colorectal cancers [CRC] in culture as tumour organoids. This permits the expansion of T cells from peripheral blood to give an enlarged population of cells that will kill those tumours. From Dijkstra et al. 2018.

And the results?

They were able to show that enriched populations of tumour-reactive T cells could kill tumour organoids and, importantly, that organoids formed from healthy tissue were not attacked by these T cells.

Stained organoids (left) and original tissue (right) from two colorectal cancers (CRC-2 & CRC-5) showing how the organoids grow to have an architecture similar to the original tumour. From Dijkstra et al. 2018.

Their method worked for both bowel tumours and non-small-cell lung cancer but there’s no reason to suppose it can’t be extended to other types of cancer.

Some of their videos showing tumour organoids being chomped up by enriched killer T cells are quite dramatic. Cells labelled green that can be seen in this video are dying.

So there you have it: DIY tumour therapy!


Dijkstra, K.K. et al. (2018). Generation of Tumor-Reactive T Cells
by Co-culture of Peripheral Blood Lymphocytes and Tumor Organoids. Cell 174, 1–13.

Now wash your hands!


You must have spent the last 20 years on a distant planet if you’re unaware that we’re heading for Antibiotic Armaggedon — the rise of “Superbugs”, i.e., bacteria resistant to once-successful medication. Microbes resistant to multiple antimicrobials are called multidrug resistant. It’s a desperate matter because it means trivial infections may become fatal and currently safe surgical procedures may become dangerous.

Time-line of the discovery of different antibiotic classes in clinical use. The key point is that the last antibiotic class to become a successful treatment was discovered in 1987.

What’s the problem?
It’s 30 years since we came up a new class of antibiotics. The golden age launched by Fleming’s celebrated discovery of penicillin is long gone and while the discovery curve has drifted ever downwards since 1960 the bugs have been busy.

Just how busy a bug can be was shown by a large-scale experiment carried out by Roy Kishony and friends. They built a “Mega-Plate” — a Petri Dish 2 ft by 4 ft filled with a jelly for the bacteria to grow in. The bugs were seeded into channels at either end so they would grow towards the middle. The only thing stopping them was four channels dosed with antibiotic at increasing concentrations — 10 times more in each successive channel.

The bugs grow until they hit a wall of antibiotic. There they pause for a think — and, after a bit, an intrepid little group start to make their way into the higher dose of drug. Gradually the number of groups expand until a tidal wave sweeps over that barrier. This is repeated at each new ‘wall’ — four times until the whole tray is a bug fest.

When they pause at each new ‘wall’ they’re not ‘thinking’ of course. They’re just picking up random mutations in their DNA until they are able to advance into the high drug environment. So this experiment is a fantastic visual display of bugs becoming drug-resistant. And it’s terrifying because it takes about 11 days for them to overcome four levels of drug. It’s even more scary in the speeded-up movie as that lasts less than two minutes.

Sound familiar?
It should do as this is a cancer column and readers will know that cancers arise by picking up mutations. To highlight the similarities the picture below is the left-hand half of the bug tray with new colonies shown as linked dots. You could perfectly well think of these as early stage cancer cells acquiring mutations in ‘driver’ genes that push them towards tumour formation.

So that’s pretty scary too and the only good news is that animal cells reproduce much more slowly than bacteria. The fastest they can manage is about 48 hours to grow and divide into two new cells and for many it’s much slower than that. Bugs, on the other hand, can do it in 20 minutes if you feed them enough of the right stuff.

Which is why we don’t all get zonked by cancer at an early age.

The evolution of bacteria on a “Mega-Plate” Petri Dish. The vertical red lines mark the boundaries of increasing antibiotic concentrations. You could equally think of each dot that represents a new bacterial colony being early stage cancer cells acquiring mutations in ‘driver’ genes (white arrows) that push them towards tumour formation. From Roy Kishony’s Laboratory at Harvard Medical School.

Enough of that!
But for once I don’t want to talk about cancer but about a really fascinating piece of work that caught my eye in the journal Cell Reports. It’s by Gianni Panagiotou, Kang Kang and colleagues from The University of Hong Kong and The Hans Knöll Institute, Jena, Germany and it’s all about their travels on the Hong Kong MTR (Mass Transit Railway). This is the network of over 200 km of railway lines with 159 stations that serves the urbanised areas of Hong Kong IslandKowloon, and the New Territories and has a cross- border connection to the neighboring city of Shenzhen in mainland China.

An MTR train on the Tung Chung line that links Lantau Island with Hong Kong Island.

Being scientists of course they weren’t just having a day out. They wanted to know the contents of the microbiome that they and their fellow travellers picked up on the palms of their hands when riding the rails. ‘Microbiome’ means all of the collection of microorganisms — though in practice this is almost entirely bacteria. So they swabbed the palms of volunteers and then threw the full power of modern DNA sequencing and genetic analysis at what they’d scraped off. Or, as they put it: “We conducted a metagenomic study of the Hong Kong MTR system.”

And if you’re thinking it might be possible to take a trip on the Hong Kong Metro without grabbing a handrail or otherwise engaging in what on the London Underground used to be called ‘strap-hanging’ you clearly haven’t tried it!

Hong Kong MTR.


The MTR System and Sampling Procedure. Left: The eight urban lines studied: the Airport Express line and Disneyland Resort branch were excluded. The Central-Hong Kong station and the cross-border rail stations connecting with the MTR and the Shenzhen metro system are labeled. Right: The sampling procedure included handwashing, handrail touching for 30 min and swabbing. From Kang et al. 2018.

Hold very tight please! 

It’s going to become a seriously bumpy ride. The major findings were:

  1. Four groups (phyla) of bacteria dominated: Actinobacteria [51%], Proteobacteria [27%], Firmicutes [11%] and Bacteroidetes [2%]. Followers of this blog will be delighted to spot the last two (B & F) as we’ve met them several times before (in Hitchhiker Or Driver?, Fast Food Fix Focuses on Fibre, Our Inner Self, The Best Laid Plans In Mice and Men, and, of course, in it’s a small world) — that’s how important they are in the context of cancer.
  2. The dominant organism (29% of the community) was P. acnes (one of the Actinobacteria — it’s the bug linked to the skin condition of acne).
  3. Some non-human-associated species (e.g., soil organisms) also popped up that varied enormously in amount from day to day — perhaps because of weather conditions (e.g., humidity).
  4. Variation in the make-up of the microbial communities picked up depended, more than anything else, on the time of day. There was a marked decrease in diversity in afternoon samples compared with those taken in the morning.
  5. Specific species of bacteria associated with individual metro lines. That is, sets of bug types are relatively abundant on a given line compared with all other lines, giving a kind of line-specific signature — though the distinction declines from morning to afternoon. The most physically isolated line, MOS (Ma On Shan), had a greater number of signature species. The MOS runs entirely above ground alongside the Shing Mun Channel, a polluted brackish river, and its ‘signature’ includes bacteria found in sewage.
  6. All of which brings us to bugs with antibiotic resistance genes (ARGs). Across the network 136 ARG families were detected including 24 that are clinically important. Strikingly, lines closer to Shenzhen (ER (East Rail) and MOS) tend to have higher ARG input during the day. Critically, the ER line a.m. signatures become p.m.-enriched in all MTR lines far from Shenzhen — that is, these ARG families spread over the network during the day.

Simplified map of the Hong Kong MTR indicating how antibiotic resistance genes spread during the day from the ER and MOS lines to the entire network. Tetracycline resistance genes: tetA, tetO, tetRRPP and tetMWOS; vancomycin resistance genes: vanC, vanX. From Kang et al. 2018.

These results clearly suggest that the ER line, the only cross-border line linked to mainland China, may be a source of clinically important ARGs, especially against tetracycline, a commonly used antibiotic in China’s swine feedlots. Antibiotics, including tetracycline, can be detected in the soil in the Pearl River Delta area where the cities of Hong Kong and Shenzhen are located.

It should be said that this is by no means the first survey of bugs on rails. Notable ones have looked at the New York and Boston metro systems and they too revealed the potential health risks of the bug communities found on trains and in the stations, including the presence of pathogens and antibiotic resistance. The Boston survey highlighted that different types of materials have surfaces that are preferred by different microbes with high variation in functional capacity and pathogenic potential.

One obvious suggestion from these studies is that world-wide we could do a lot to improve sanitation, e.g., by having hand sanitizer dispensers in all sensible places (at the exits of metro, railway and bike-sharing stations and airports and of course in hospitals). The Hong Kong data are seriously frightening and most people seem blissfully unaware that the invisible world they reveal carries the potential for the destruction of us all.

But, as ever, there’s two sides to the matter. We’ve evolved over millions of years to live with bugs and they with us. However you wash your hands you won’t get rid of every bug and anyway, as what’s-his-name almost says, “They’ll be back!” We all carry around micro-organisms that can be fatal if they get to the wrong place. But, if you’re reasonably fit, there’s a lot to be said for simply following sensible, basic hygiene rules with a philosophy of ‘live and let live.’

Have a nice day commuters, wherever you are!


Kang K., et al. (2018). The Environmental Exposures and Inner- and Intercity Traffic Flows of the Metro System May Contribute to the Skin Microbiome and Resistome. Cell Reports 24, 1190–1202.

Wu, N., Qiao, M., Zhang, B., Cheng, W.D., and Zhu, Y.G. (2010). Abundance and diversity of tetracycline resistance genes in soils adjacent to representative swine feedlots in China. Environ. Sci. Technol. 44, 6933–6939.

Li, Y.W., Wu, X.L., Mo, C.H., Tai, Y.P., Huang, X.P., and Xiang, L. (2011). Investigation of sulfonamide, tetracycline, and quinolone antibiotics in vegetable farmland soil in the Pearl River Delta area, southern China. J. Agric. Food Chem. 59, 7268–7276.

Leung, M.H., Wilkins, D., Li, E.K., Kong, F.K., and Lee, P.K. (2014). Indoor-air microbiome in an urban subway network: diversity and dynamics. Appl. Environ. Microbiol. 80, 6760–6770.

Robertson, C.E., Baumgartner, L.K., Harris, J.K., Peterson, K.L., Stevens, M.J., Frank, D.N., and Pace, N.R. (2013). Culture-independent analysis of aerosol microbiology in a metropolitan subway system. Appl. Environ. Microbiol. 79, 3485–3493.

Afshinnekoo, E., Meydan, C., Chowdhury, S., Jaroudi, D., Boyer, C., Bernstein, N., Maritz, J.M., Reeves, D., Gandara, J., Chhangawala, S., et al. (2015). Geospatial Resolution of Human and Bacterial Diversity with City-Scale Metagenomics. Cell Syst 1, 72–87.

Hsu, T., Joice, R., Vallarino, J., Abu-Ali, G., Hartmann, E.M., Shafquat, A., Du- Long, C., Baranowski, C., Gevers, D., Green, J.L., et al. (2016). Urban Transit System Microbial Communities Differ by Surface Type and Interaction with Humans and the Environment. mSystems 1, e00018–e00016.

Keeping Up With Cancer


Cancer enthusiasts will know that there are zillions of web sites giving info on cancer stats — incidence, mortality, etc. — around the world. Notable is the World Health Organization’s Globocan, an amazing compilation of data on all cancers from every country. The Global Burden of Disease Cancer Collaboration audits diagnosis rates and deaths for 29 types of cancer around the world each year. Needless to say, this too is a vast undertaking involving hundreds of scientists around the world. The organizing genius is Dr. Christina Fitzmaurice of the Institute for Health Metrics and Evaluation in the University of Washington, Seattle and, under her guidance, their update for 2016 has just come out.

What’s new?

In 2016 there were 17.2 million people diagnosed with cancer. 8.9 million died from cancers. By 2030 the number of new cancer cases per year is expected to reach 24 million. Well, you knew the numbers were going to be big — almost incomprehensibly so. But here’s the real shaker: the 17.2 M is 28% up on the 2006 figure — yes, that’s a rise of more than one quarter.

Yearly global cancer deaths from 1990 to 2016.

The green line is total deaths per 100,000 people.

Red line: Cancer death rates taking account of the increase in world population.

Blue line: Age-standardized death rates: these are corrected for population size and age structure. Age-standardization therefore gives a better indication of the prevalence and incidence of underlying cancer risk factors between countries and with time without the influence of demographic and population structure changes.

The numbers on the vertical axis are deaths per 100,000 people. From Our World in Data.

Any real surprises?

No. An increase of more than one quarter in cancer cases does indeed make you think but the grim numbers are only what you would predict from looking at the trends over the last 40 years. The graph shows death rates that, of course, reflect incidence. The total figure (top) shows starkly how the rise in the population of the world and our increasing life-span is steadily pushing up the overall cancer burden.

So it’s a mega-problem but the trends are smooth and gradual. There’s been no drastic upheaval.

Global trends

Yearly global cancer deaths from 1990 to 2016 for the four major cancer types.

The numbers on the vertical axis are age-standardized death rates per 100,000 people.

Because age-standardization assumes a constant population age & structure it permits comparisons between countries and over time without the effects of a changing age distribution within a population. From Our World in Data.

The global trends in deaths from the four major cancers look mildly encouraging (above). However, these should not cheer us up too much. In the developed world there are some positives. In the USA, for example, over the last 17 years deaths from prostate are down from 31.6 to 18.9, for breast from 26.6 to 20.3 and for lung from 55.4 to 40.6 per 100,000 people. For bowel cancer there’s been a slight increase (4.1 to 4.8).

In the wider world, however, the really dispiriting thing shown by the latest figures is that the increases in incidence and deaths are greatest in low- and middle-income countries.

What can we do?

Lung cancer (includes cancers of the trachea and bronchi) remains the world’s biggest cancer killer, accounting for 20% of all deaths in 2016. Over 90% of these were caused by tobacco. In the UK and the USA lung cancer deaths in men have markedly declined as a result of widespread smoking bans, as the graph below shows, and the female figures have started to show a sight decline.

Lung cancer deaths per 100,000 by sex from 1950 to 2002 for the UK and the USA. From Our World in Data.

Asia contributes over half the global burden of cancer but the incidence in Asia is about half that in North America. However, the ratio of cancer deaths to the number of new cancer cases in Asia is double that in North America. Although the leading cause of death world-wide is heart disease, in China it is cancer. Every year more than four million Chinese are diagnosed with the disease and nearly three million die from it. Overall, tobacco smoking is responsible for about one-quarter of all cancer deaths in China. Nevertheless, Chinese smoking rates continue to rise and air pollution in the major cities is fuelling the problem.

The under-developed world, however, continues to be targeted by the tobacco industry and the successful promotion of their products means that there is no end in sight to one of mankind’s more bizarre and revolting forms of self-destruction.

There are, of course, other things within our control that contribute significantly to the global cancer burden. If only we could give everyone clean water to drink, restrict our red meat and processed food consumption and control our exposure to uv in sunlight we would cut cancer by at least one half.

If only …


The Global Burden of Disease Cancer Collaboration (2018). Global, Regional, and National Cancer Incidence, Mortality, Years of Life Lost, Years Lived With Disability, and Disability-Adjusted Life-Years for 29 Cancer Groups, 1990 to 2016: A Systematic Analysis for the Global Burden of Disease Study. JAMA Oncol. Published online June 2, 2018. doi:10.1001/jamaoncol.2018.2706.

Same Again Please


It’s often said that every family has its secret — Uncle Fred’s fondness for the horses, Cousin Bertha’s promiscuity, etc. — whatever it is that ‘we don’t talk about.’ If that’s true the scientific community is no exception. For us the unutterable is reproducibility — meaning you’ve done an experiment, new in some way, but the key questions are: ‘Can you do it again with the same result?’ and, even more important: ‘Can someone else repeat it?’

Once upon a time in my lab we had a standing joke: whoever came bounding along shouting about a new result would be asked ‘How reproducible is it?’ Reply: ‘100%!’ Next question: ‘How often have you done the experiment?’ Reply: ‘Once!!’ Boom, boom!!!

Not a rib-tickler but it did point to the knottiest problem in biological science namely that, when you start tinkering with living systems, you’re never fully in control.

How big is the problem?

But, as dear old Bob once put it, The Times They Are a-Changin’. Our problem was highlighted in the cancer field by the Californian biotechnology company Amgen who announced in 2012 that, over a 10 year period, they’d selected 53 ‘landmark’ cancer papers — and failed to replicate 47 of them! Around the same time a study by Bayer HealthCare found that only about one in four of published studies on potential drug targets was sufficiently strong to be worth following up.

More recently the leading science journal Nature found that almost three quarters of over 1,500 research scientists surveyed had tried to replicate someone else’s experiment and failed. It gets worse! More than half of them owned up to having failed to repeat one of their own experiments! Hooray! We have a result!! If you can’t repeat your own experiment either you’re sloppy (i.e., you haven’t done exactly what you did the first time) or you’ve highlighted the biological variability in the system you’re studying.

If you want an example of biological variation you need look no further than human conception and live births. Somewhere in excess of 50% of fertilized human eggs don’t make it to birth. In other words, if you do a ‘thought experiment’ in which a group of women carry some sort of gadget that flags when one of their eggs is fertilized, only between one in two and one in five of those ‘flagged’ will actually produce an offspring.

However you look at it, whether it’s biological variation, incompetence or plain fraud, we have a problem and Nature’s survey revealed that, to their credit, the majority of scientists agreed that there was a ‘significant crisis.’

The results of the survey by Nature from Baker 2016.

Predictably, but disturbingly for us in the biomedical fields, the greatest confidence in published results was shown by the chemists and physicists whereas only 50% of data in medicine were thought to be reproducible. Oh dear!

Tackling the problem in cancer

The Reproducibility Project: Cancer Biology, launched in 2013, is a collaboration between the Center for Open Science and Science Exchange.

The idea was to take 50 cancer papers published in leading journals and to attempt to replicate their key findings in the most rigorous manner. The number was reduced from 50 to 29 papers due to financial constraints and other factors but the aim remains to find out what affects the robustness of experimental results in preclinical cancer research.

It is a formidable project. Before even starting an experiment, the replication teams devised detailed plans, based on the original reports and, as the result of many hours effort, came up with a strategy that both they and the original experimenters considered was the best they could carry out. The protocols were then peer reviewed and the replication plans were published before the studies began.

Just to give an idea of the effort involved, a typical replication plan comprises many pages of detailed protocols describing reagents, cells and (where appropriate) animals to be used, statistical analysis and any other relevant items, as well as incorporating the input from referees.

The whole endeavor is, in short, a demonstration of scientific practice at its best.

To date ten of these replication studies have been published.

How are we doing?

The critical numbers are that 6 of the 10 replications ‘substantially reproduced’ the original findings, although in 4 of these some results could not be replicated. In 4 of the 10 replications the original findings were not reproduced.

The first thing to say is that a 60% rate of ‘substantial’ successful replication is a major improvement on the 11% to 25% obtained by the biotech companies. The most obvious explanation is that the massive, collaborative effort to tighten up the experimental procedures paid dividends.

The second point to note is that even when a replication attempt fails it cannot be concluded that the original data were wrong. The discrepancy may merely have highlighted how fiendishly tricky biological experimentation can be. The problem is that with living systems, be they cells or animals, you never have complete control. Ask anyone who has a cat.

More likely, however, than biological variation as a cause of discrepancies between experiments is human variation, aka personal bias.

This may come as a surprise to some but, rather than being ‘black and white’ much of scientific interpretation is subjective. Try as I might, can I be sure that in, say, counting stained cells I don’t include some marginal ones because that fits my model? OK: the solution to that is get someone else to do the count ‘blind’ — but I suspect that quite often that’s not done. However, there are even trickier matters. I do half a dozen repeats of an experiment and one gives an odd result (i.e., differs from the other five). Only I can really go through everything involved (from length of coffee breaks to changes in reagent stocks) and decide if there are strong enough grounds to ignore it. I do my best to avoid personal bias but … scientists are only human (fact!).

A closer look at failure

One of the failed replications is a particularly useful illustration for this blog. The replication study tackled a 2012 report that bacterial infection (specifically a bacterium, Fusobacterium nucleatum, that occurs naturally in the human oral cavity) is present in human colon cancers but not in non-cancerous colon tissues. It hit the rocks. They couldn’t detect F. nucleatum in most tumour samples and, when they did, the number of bugs was not significantly different to that in adjacent normal tissue.

Quite by chance, a few months ago, I described some more recent research into this topic in Hitchhiker or Driver?

I thought this was interesting because it showed that not only was F. nucleatum part of the microbiome of bowel cancer but that when tumour cells spread to distant sites (i.e., underwent metastasis) the bugs went along for the ride — raising the key question of whether they actually helped the critical event of metastasis.

So this latest study was consistent with the earlier result and extended it — indeed they actually showed that antibiotic treatment to kill the bugs slowed the growth of human tumour cells in mice.

Where does that leave us?

Well, helpfully, the Reproducibility Project also solicits comments from independent experts to help us make sense of what’s going on. Step forward Cynthia Sears of The Johns Hopkins Hospital. She takes the view that, although the Replication Study didn’t reproduce the original results, the fact that numerous studies have already found an association between F. nucleatum and human colon cancer means there probably is one — consistent with the work described in Hitchhiker or Driver?

One possible explanation for the discrepancy is that the original report studied colon tissue pairs (i.e., tumour and tumour-adjacent tissues) from colon cancer patients but did not report possibly relevant factors like age, sex and ethnicity of patients. In contrast, the replication effort included samples from patients with cancer (tumour and adjacent tissue) and non-diseased control tissue samples from age, sex and ethnicity matched individuals.

So we now know, as Dr. Sears helpfully remarks, that the association between F. nucleatum bugs and human colon cancer is more complicated first appeared! Mmm. And, just in case you were in any doubt, she points out that we need to know more about the who (which Fusobacterium species: there are 12 of them known), the where (where in the colon, where in the world) and the how (the disease mechanisms).

Can we do better?

In the light of all that the obvious question is: what can we do about the number of pre-clinical studies that are difficult if not impossible to reproduce? Answer, I think: not much. Rather than defeatist this seems to me a realistic response. There’s no way we could put in place the rigorous scrutiny of the Reproducibility Project across even a fraction of cancer research projects. The best we can do is make researchers as aware as possible of the problems and encourage them towards the very best practices — and assume that, in the end, the solid results will emerge and the rest will fall by the wayside.

Looking at the sharp end, it’s worth noting that, if you accept that some of the variability in pre-clinical experiments is down to the biological variation we mentioned above, it would at least be consistent with the wide range of patient responses to some cancer treatments. The reason for that, as Cynthia Sears didn’t quite put it, is that we just don’t know enough about how the humans we’re tinkering with actually work.


Baker, M. (2016). Is There a Reproducibility Crisis? Nature 533, 452-454.

Jarvis, G.E. (2017). Early embryo mortality in natural human reproduction: What the data say [version 2; referees: 1 approved, 2 approved with reservations] F1000Research 2017, 5:2765 (doi: 10.12688/f1000research.8937.2).

Monya Baker & Elie Dolgin (2017). Cancer reproducibility project releases first results. Nature 541, 269–270. doi:10.1038/541269a.

Begley, C.G. and Ellis, L.M. (2012). Drug development: Raise standards for preclinical cancer research. Nature 483, 531–533.

Prinz,F., Schlange,T. and Asadullah, K. (2011). Believe it or not: how much can we rely on published data on potential drug targets? NatureRev. Drug Discov. 10, 712.

Caveat Emptor


It must be unprecedented for publication of a scientific research paper to make a big impact on a significant sector of the stock market. But, in these days of ‘spin-off’ companies and the promise of unimaginable riches from the application of molecular biology to every facet of medicine and biology, perhaps it was only a matter of time. Well, the time came with a bang this June when the journal Nature Medicine published two papers from different groups describing essentially the same findings. Result: three companies (CRISPR Therapeutics, Editas Medicine and Intellia) lost about 10% of their stock market value.

I should say that a former student of mine, Anthony Davies, who runs the Californian company Dark Horse Consulting Inc., mentioned these papers to me before I’d spotted them.

What on earth had they found that so scared the punters?

Well, they’d looked in some detail at CRISPR/Cas9, a method for specifically altering genes within organisms (that we described in Re-writing the Manual of Life).

Over the last five years it’s become the most widely used form of gene editing (see, e.g., Seeing a New World and Making Movies in DNA) and, as one of the hottest potatoes in science, the subject of fierce feuding over legal rights, who did what and who’s going to get a Nobel Prize. Yes, scientists do squabbling as well as anyone when the stakes are high.

Nifty though CRISPR/Cas9 is, it has not worked well in stem cells — these are the cells that can keep on making more of themselves and can turn themselves in other types of cell (i.e., differentiate — which is why they’re sometimes called pluripotent stem cells). And that’s a bit of a stumbling block because, if you want to correct a genetic disease by replacing a defective gene with one that’s OK, stem cells are a very attractive target.

Robert Ihry and colleagues at the Novartis Institutes for Biomedical Research got over this problem by modifying the Cas9 DNA construct so that it was incorporated into over 80% of stem cells and, moreover, they could switch it on by the addition of a drug. Turning on the enzyme Cas9 to make double-strand breaks in DNA in such a high proportion of cells revealed very clearly that this killed most of them.

When cells start dying the prime suspect is always P53, a so-called tumour suppressor gene, switched on in response to DNA damage. The p53 protein can activate a programme of cell suicide if the DNA cannot be adequately repaired, thereby preventing the propagation of mutations and the development of cancer. Sure enough, Ihry et al. showed that in stem cells a single cut is enough to turn on P53 — in other words, these cells are extremely sensitive to DNA damage.

Gene editing by Cas9 turns on P53 expression. Left: control cells with no activation of double strand DNA breaks; right: P53 expression (green fluorescence) several days after switching on expression of the Cas9 enzyme. Scale bar = 100 micrometers. From Ihry et al., 2018.

In a corresponding study Emma Haapaniemi and colleagues from the Karolinska Institute and the University of Cambridge, using a different type of cell (a mutated line that keeps on proliferating), showed that blocking P53 (hence preventing the damage response) improves the efficiency of genome editing. Good if you want precision genome editing by risky as it leaves the cell vulnerable to tumour-promoting mutations.

Time to buy?!

As ever, “Let the buyer beware” and this certainly isn’t a suggestion that you get on the line to your stockbroker. These results may have hit share prices but they really aren’t a surprise. What would you expect when you charge uninvited into a cell with a molecular bomb — albeit one as smart as CRISPR/Cas9. The cell responds to the DNA damage as it’s evolved to do — and we’ve known for a long time that P53 activation is exquisitely sensitive: one double-strand break in DNA is enough to turn it on. If the damage can’t be repaired P53’s job is to drive the cell to suicide — a perfect system to prevent mutations accumulating that might lead to cancer. The high sensitivity of stem cells may have evolved because they can develop into every type of cell — thus any fault could be very serious for the organism.

It’s nearly 40 years since P53 was discovered but for all the effort (over 45,000 research papers with P53 in the title) we’re still remarkably ignorant of how this “Guardian of the Genome” really works. By comparison gene editing, and CRISPR/Cas9 in particular, is in its infancy. It’s a wonderful technique and it may yet be possible to get round the problem of the DNA damage response. It may even turn out that DNA can be edited without making double strand breaks.

So maybe don’t rush to buy gene therapy shares — or to sell them. As the Harvard geneticist George Church put it “The stock market isn’t a reflection of the future.” Mind you, as a founder of Editas Medicine he’d certainly hope not.


Ihry, R.J. et al. (2018). p53 inhibits CRISPR–Cas9 engineering in human pluripotent stem cells. Nature Medicine, 1–8.

Haapaniemi, E. et al. (2018). CRISPR–Cas9 genome editing induces a p53-mediated DNA damage response. Nature Medicine (2018) 11 June 2018.

Now You See It


In the pages of this blog we’ve often highlighted the power of fluorescent tags to track molecules and see what they’re up to. It’s a method largely pioneered by the late Roger Tsien and it has revolutionized cell biology over the last 20 years.

In parallel with molecular tagging has come genetic engineering that permits novel genes, usually carried by viruses, to be introduced to cells and animals. As we saw in Gosh! Wonderful GOSH and Blowing Up Cancer, various ‘virotherapy’ approaches have been used with some success to treat leukemias and skin cancers and a trial is underway in China treating metastatic non-small cell lung cancer.

A major aim of genetic engineering is to be able to control the expression of novel genes (i.e. protein production from the encoding DNA sequence) that have been introduced into an animal — in the jargon, to ‘switch’ on or off at will. That can be done but only by administering a drug or some other regulator, either in drinking water, by injection or squirting directly into the lungs. An ideal would be something that’s more controlled and less invasive. How about shining a light on the relevant spot?!

Wacky or what?

That may sound as though we’re veering towards science fiction but reflect for a moment that every animal with vision, however rudimentary, sees by transforming light entering the eyes into electrical signals that the brain turns into a picture of the world around them. This relies on photoreceptor proteins that span the membranes of retinal cells.

How vision works. Light passes through the lens and falls on the retina at the back of the eye. The photoreceptor cells it activates are rod cells (that respond to low light levels — there’s about 100 million of them) and cone cells (stimulated by bright light). Sitting across the membranes of these cells are photoreceptor proteins — rhodopsin in rods and photopsin in cones. Photoreceptor proteins change shape when light falls on them — the driver for this being a small chemical attached to the proteins called retinal, one of the many forms of vitamin A. This shape change allows the proteins to ‘talk’ to the inside of the cell, i.e. to interact with other proteins to switch on enzymes and change the level of ions (sodium and calcium). The upshot is that the signal is passed through neural cells in the optic nerve to the brain where the incoming light signals are processed into the images that we perceive. © Arizona Board of Regents / ASU Ask A Biologist.

The seemingly far-fetched notion of controlling genes by light was floated by Francis Crick in 1999. The field was launched in 2002 by Boris Zemelman and Gero Miesenböck who engineered neurons to express one form of rhodopsin. This gave birth to the subject of optogenetics — using light to control cells in living tissues that have been genetically modified to express light-sensitive ion channels such as rhodopsin. By 2010 optogenetics had advanced to being the ‘Method of the Year’ according to the research journal Nature Methods.

Dropping like flies

One of the most dramatic demonstrations of the power of optogenetics has come from Robert Kittel and colleagues in Würzburg and Göttingen who made a mutant form of a protein called channelrhodopsin-1 (found in green algae) and expressed it in fruit flies (Drosophila melanogaster). The mutant protein (ChR2-XXL) carries very large photocurrents of ions (critically sodium and calcium) with the result that photostimulation can drastically change the behaviour of freely moving flies.

Light-induced stimulation of motor neurons in adult flies expressing a mutant form of rhodopsin ChR2-XXL. Click to run movie.

Left hand tube: Activation of ChR2-XXL in motor neurons with white light LEDs caused reversible immobilization of adult flies. In contrast (right hand tube) flies expressing normal (wild-type) channelrhodopsin-2 showed no response. From Dawydow et al., 2014.

Other optogenetic experiments on flies can be viewed on You Tube, e.g., the TED talk of Gero Miesenböck and the Manchester Fly Facility video of fly maggots, engineered to have a channel protein (channelrhodopsin) in their neurons, responding to blue light.

Of flies … and mice … and men

This is stunning science and it’s opened a new vista in neurobiology. But what about the things we’re concerned with in these pages — treating diseases like diabetes and cancer?

Scheme showing how genetic engineering can make the release of insulin from cells controllable by light. Normally cells of the pancreas (beta cells) take up glucose when its level in the circulation rises (via a glucose transporter protein). The rise in glucose triggers ATP production in the cell. This in turn causes potassium channels in the membrane to close (called depolarization) and this opens calcium channels. The increase in calcium in the cell drives insulin secretion. From Kushibiki et al., 2015.

The left-hand scheme above shows how glucose triggers the pancreas to produce the hormone insulin. Diabetes occurs when either the pancreas doesn’t make enough insulin or when cells of the body don’t respond properly to insulin by taking up glucose.

As a first step to see whether optogenetic regulation of calcium levels in pancreatic cells could trigger insulin release, Toshihiro Kushibiki and colleagues at the National Defense Medical College in Saitama, Japan engineered the channelrhodopsin-1 protein into mouse cells and hit them with laser light of the appropriate frequency. An hour after a short burst of light (a few seconds) the insulin levels had doubled.

The photo below shows a clump of these cells: the nuclei are blue and the channel protein (yellow) can be seen sitting across the cell membranes.


Cells expressing a fluorescently tagged channelrhodopsin protein (yellow). Nuclei are blue. From Kushibiki et al., 2015.



To show that this could work in animals they suspended the engineered cells in a gel and inoculated blobs of the goo under the skin of diabetic mice. Laser burst again: blood glucose levels fell and they showed this was due to the irradiated, implanted cells producing insulin.

Fast forward three years

Those brilliant results highlighted the potential of optogenetic technology as a completely novel approach to a disease that afflicts over 300 million people worldwide.

Scheme showing a Smartphone can be used to regulate the release of insulin from engineered cells implanted in a mouse with diabetes. The key events in the cell are that the light-activated receptor turns on an enzyme (BphS) that in turn controls a transcription regulator (FRTA) that binds to a DNA construct to switch on the Gene Of Interest (GOI) — in this case encoding insulin. (shGLP1, short human glucagon-like peptide 1, is a hormone that has the opposite effect to insulin). From Shao et al., 2017.

In a remarkable confluence of technologies Jiawei Shao and colleagues from a number of institutes in Shanghai, including the Shanghai Academy of Spaceflight Technology, and from ETH Zürich have recently published work that takes the application of optogenetics well and truly into the twenty-first century.

They figured that, as these days nearly everyone lives with their smartphone, the world could use a diabetes app. Essentially they designed a home server SmartController to process wireless signals so that a smartphone could control insulin production by cells in gel capsules implanted in mice. There are differences in the genetic engineering of these cells from those used by Kushibiki’s group but the critical point is unchanged: laser light stimulates insulin release. The capsules carry wirelessly powered LEDs.

The only other thing needed is to know glucose levels. Because mice are only little and they’ve already got their gel capsule, rather than implanting a monitor they took a drop of blood from the tail and used a glucometer. However, looking ahead to human applications, continuous glucose monitors are now available that, placed under the skin, can transmit a radio signal to the controller and, ultimately, it will be possible for the gel capsules to have a built-in battery plus glucose sensor and the whole thing could work automatically.

Any chance of illuminating cancer?

This science is so breathtaking it seems cheeky to ask but, well, I’d say ‘yes but not just yet.’ So long as the ‘drug’ you wish to use can be made biologically (i.e. from DNA by the machinery of the cell), rather than by chemical synthesis, Shao’s Smartphone set-up can readily be adapted to deliver anti-cancer drugs. This might be hugely preferable to the procedures currently in use and would offer an additional advantage by administering drugs in short bursts of lower concentration — a regimen that in some mouse cancer models at least is more effective.


Dawydow, A., Kittel, R.J. et al., 2014. Channelrhodopsin-2–XXL, a powerful optogenetic tool for low-light applications. PNAS 111, 13972-13977.

Kushibiki et al., (2015). Optogenetic control of insulin secretion by pancreatic beta-cells in vitro and in vivo. Gene Therapy 22, 553-559.

Shao, J. et al., 2017. Smartphone-controlled optogenetically engineered cells enable semiautomatic glucose homeostasis in diabetic mice. Science Translational Medicine 9, Issue 387, eaal2298.

Another Fine Mess


Did you guess from the title that this short piece is about the seeming inability of the British Government to run well, most things but especially IT programmes? Of course you did! Provoked by the latest National Health Service furore. In case you’ve been away with the fairies for a bit, a major cock-up in its computer system has just come to light whereby, between 2009 and 2018, it failed to invite 450,000 women between the ages of 68 and 71 for breast screening. Secretary of State for Health, Jeremy Hunt (our man usually on hand with a can of gasoline when there’s a fire), told Parliament that “there may be between 135 and 270 women who had their lives shortened”. Cue: uproar, headlines: HUNDREDS of British women have died of breast cancer (Daily Express), etc.

Logo credit: Breast Cancer Action

I’ve been reluctant to join in because I’ve said all I think is worth saying about breast cancer screening in two earlier pieces (Risk Assessment and Behind the Screen). Reading them again I thought they were a reasonable summary and I don’t think there’s anything new to add. However, this is  a cancer blog and it’s a story that’s made big headlines so I feel honour-bound to offer a brief comment — in addition to sympathizing with the women and families who have been caused much distress.

My reaction was that Hunt was misguided in mentioning specific numbers — not only because he was asking for trouble from the press but mainly because the evidence that screening itself saves lives is highly questionable. For an expert view on this my Cambridge colleague David Spiegelhalter, who is Professor for the Public Understanding of Risk, has analysed the facts behind breast screening with characteristic clarity in the New Scientist.

Anything to add?

I was relieved on re-reading Risk Assessment to see that I’d given considerable coverage to the report that had just come out (2014) from The Swiss Medical Board.  They’d reviewed the history of mammography screening, concluded that systematic screening might prevent about one breast cancer death for every 1000 women screened, noted that there was no evidence that overall mortality was affected and pointed out that false positive test results presented the risk of overdiagnosis.

In the USA, for example, over a 10-year course of annual screening beginning at 50 years of age, one breast-cancer death would have been prevented whilst between 490 and 670 women would have had a false positive mammogram calling for a repeat examination, 70 to 100 an unnecessary biopsy and between 3 and 14 would have been diagnosed with a cancer that would never have become a problem.

Needless to say, this landed the Swiss Big Cheeses in very hot water because there’s an awful lot of vested interests in screening and it’s sort of instinctive that it must be a good thing. But what’s great about science is that you can do experiments — here actually analysing the results of screening programmes — and quite often the results turn to be completely unexpected, as it did in this case where the bottom line was that mammography does more harm than good.

This has led to the recommendation that the current programmes in Switzerland should be phased out and not replaced.

So we’re all agreed then?

Of course not. In England the NHS recommendation remains that women aged 50 to 70 are offered mammography every three years — which is just as well or we’d have Hunt explaining the recent debacle as new initiative. The American Cancer Society “strongly” recommends regular screening mammography starting at age 45 and the National Cancer Institute refers to “experts” that recommend mammography every year starting at age 25 for women with mutations in their BRCA1 or BRCA2 genes.

The latter is really incredible because a study published in the British Medical Journal in 2012 found that these mutations made the carriers much more vulnerable to radiation-induced cancer. Specifically, women with BRCA 1/2 mutations who were exposed to diagnostic radiation (i.e. mammography) before the age of 30 were twice as likely to develop breast cancer, compared to those with normal BRCA genes.

They are susceptible to radiation that would not normally be considered dangerous because the two BRCA genes encode proteins involved in the repair of damaged DNA — and if that is defective you have a recipe for cancer.


So it’s probably true that the only undisputed fact is that we need much better ways for detecting cancers at an early stage of development. The best hope at the moment seems to be the liquid biopsy approach we described in Seeing the Invisible: A Cancer Early Warning System? but that’s still a long way from solving a general cancer problem, well illustrated by breast mammography.

No It Isn’t!


It’s great that newspapers carry the number of science items they do but, as regular readers will know, there’s nothing like the typical cancer headline to get me squawking ‘No it isn’t!” Step forward The Independent with the latest: “Major breakthrough in cancer care … groundbreaking international collaboration …”

Let’s be clear: the subject usually is interesting. In this case it certainly is and it deserves better headlines.

So what has happened?

A big flurry of research papers has just emerged from a joint project of the National Cancer Institute and the National Human Genome Research Institute to make something called The Cancer Genome Atlas (TCGA). This massive initiative is, of course, an offspring of the Human Genome Project, the first full sequencing of the 3,000 million base-pairs of human DNA, completed in 2003. The intervening 15 years have seen a technical revolution, perhaps unparalled in the history of science, such that now genomes can be sequenced in an hour or two for a few hundred dollars. TCGA began in 2006 with the aim of providing a genetic data-base for three cancer types: lung, ovarian, and glioblastoma. Such was its success that it soon expanded to a vast, comprehensive dataset of more than 11,000 cases across 33 tumor types, describing the variety of molecular changes that drive the cancers. The upshot is now being called the Pan-Cancer Atlas — PanCan Atlas, for short.

What do we need to know?

Fortunately not much of the humungous amounts of detail but the scheme below gives an inkling of the scale of this wonderful endeavour — it’s from a short, very readable summary by Carolyn Hutter and Jean Claude Zenklusen.

TCGA by numbers. The scale of the effort and output from The Cancer Genome Atlas. From Hutter and Zenklusen, 2018.

The first point is obvious: sequencing 11,000 paired tumour and normal tissue samples produced mind-boggling masses of data. 2.5 petabytes, in fact. If you have to think twice about your gigas and teras, 1 PB = 1,000,000,000,000,000 B, i.e. 1015 B or 1000 terabytes. A PB is sometimes called, apparently, a quadrillion — and, as the scheme helpfully notes, you’d need over 200,000 DVDs to store it.

The 33 different tumour types included all the common cancers (breast, bowel, lung, prostate, etc.) and 10 rare types.

The figure of seven data types refers to the variety of information accumulated in these studies (e.g., mutations that affect genes, epigenetic changes (DNA methylation), RNA and protein expression, duplication or deletion of stretches of DNA (copy number variation), etc.

After which it’s worth pausing for a moment to contemplate the effort and organization involved in collecting 11,000 paired samples, sequencing them and analyzing the output. It’s true that sequencing itself is now fairly routine, but that’s still an awful lot of experiments. But think for even longer about what’s gone into making some kind of sense of the monstrous amount of data generated.

And it’s important because?

The findings confirm a trend that has begun to emerge over the last few years, namely that the classification of cancers is being redefined. Traditionally they have been grouped on the basis of the tissue of origin (breast, bowel, etc.) but this will gradually be replaced by genetic grouping, reflecting the fact that seemingly unrelated cancers can be driven by common pathways.

The most encouraging thing to come out of the genetic changes driving these tumours is that for about half of them potential treatments are already available. That’s quite a surprise but it doesn’t mean that hitting those targets will actually work as anti-cancer strategies. Nevertheless, it’s a cheering point that the output of this phenomenal project may, as one of the papers noted, serve as a launching pad for real benefit in the not too distant future.

What should science journalists do to stop upsetting me?

Read the papers they comment on rather than simply relying on press releases, never use the words ‘breakthrough’ or ‘groundbreaking’ and grasp the point that science proceeds in very small steps, not always forward, governed by available methods. This work is quite staggering for it is on a scale that is close to unimaginable and, in the end, it will lead to treatments that will affect the lives of almost everyone — but it is just another example of science doing what science does.


Hutter, C. and Zenklusen, J.C. (2018). The Cancer Genome Atlas: Creating Lasting Value beyond Its Data. Cell 173, 283–285.

Hoadley, K.A. et al. (2018). Cell-of-Origin Patterns Dominate the Molecular Classification of 10,000 Tumors from 33 Types of Cancer. Cell 173, 291–304.

Hoadley, K.A. et al. (2014). Multiplatform Analysis of 12 Cancer Types Reveals Molecular Classification within and across Tissues of Origin. Cell 158, 929–944.

Hitchhiker Or Driver?


It’s a little while since we talked about what you might call our hidden self — the vast army of bugs that colonises our nooks and crannies, especially our intestines, and that is essential to our survival.

In Our Inner Self we noted that these little guys outnumber the human cells that make up the body by about ten to one. Actually that estimate has recently been revised — downwards you might be relieved to hear — to about 1.3 bacterial cells per human cell but it doesn’t really matter. They are a major part of what’s called the microbiome — a vast army of microorganisms that call our bodies home but on which we also depend for our very survival.

In our personal army there’s something like 700 different species of bacteria, with thirty or forty making up the majority. We upset them at our peril. Artificial sweeteners, widely used as food additives, can change the proportions of types of gut bacteria. Some antibiotics that kill off bacteria can make mice obese — and they probably do the same to us. Obese humans do indeed have reduced numbers of bugs and obesity itself is associated with increased cancer risk.

In it’s a small world we met two major bacterial sub-families, Bacteroidetes and Firmicutes, and noted that their levels appear to affect the development of liver and bowel cancers. Well, the Bs & Fs are still around you’ll be glad to know but in a recent piece of work the limelight has been taken by another bunch of Fs — a sub-group (i.e. related to the Bs & Fs) called Fusobacterium.

It’s been known for a few years that human colon cancers carry enriched levels of these bugs compared to non-cancerous colon tissues — suggesting, though not proving, that Fusobacteria may be pro-tumorigenic. In the latest, pretty amazing, installment Susan Bullman and colleagues from Harvard, Yale and Barcelona have shown that not merely is Fusobacterium part of the microbiome that colonises human colon cancers but that when these growths spread to distant sites (i.e. metastasise) the little Fs tag along for the ride! 

Bacteria in a primary human bowel tumour.  The arrows show tumour cells infected with Fusobacteria (red dots).

Bacteria in a liver metastasis of the same bowel tumour.  Though more difficult to see, the  red dot (arrow) marks the presence of bacteria from the original tumour. From Bullman et al., 2017.

In other words, when metastasis kicks in it’s not just the tumour cells that escape from the primary site but a whole community of host cells and bugs that sets sail on the high seas of the circulatory system.

But doesn’t that suggest that these bugs might be doing something to help the growth and spread of these tumours? And if so might that suggest that … of course it does and Bullman & Co did the experiment. They tried an antibiotic that kills Fusobacteria (metronidazole) to see if it had any effect on F–carrying tumours. Sure enough it reduced the number of bugs and slowed the growth of human tumour cells in mice.

Growth of human tumour cells in mice. The antibiotic metronidazole slows the growth of these tumour by about 30%. From Bullman et al., 2017.

We’re still a long way from a human therapy but it is quite a startling thought that antibiotics might one day find a place in the cancer drug cabinet.


Bullman, S. et al. (2017). Analysis of Fusobacterium persistence and antibiotic response in colorectal cancer. Science  358, 1443-1448. DOI: 10.1126/science.aal5240