You’re probably wrong

I want to begin this post with a simple thought experiment. It is not original with me, and unfortunately, I don’t recall where I first heard it, but I think that it is a useful exercise. I want you to think about all of the thousands of different views, beliefs, ideas, etc. that you hold. Think about all of the things that you think are true. Now, estimate the percentage of them that are actually true. In other words, if you are honest with yourself, you’ll have to admit that you are almost certainly wrong some of the time. It’s extremely unlikely that you are right 100% of the time. So how often do you think that you are right? 80% of the time? 90% of the time? 95%? Are you presumptuous enough to go to 99%? Regardless of what number you settled on, it was surely something less than 100% (unless you are hopelessly arrogant). You’re not correct 100% of the time, and neither am I, and neither is Stephen Hawking, and neither was Einstein, etc.

Now, having established that you are wrong about at least some of your views, I want you to divide your views into things you’re right about and things you’re wrong about. Go ahead and try it. You can’t do it, and neither can I. You inherently don’t know which of your views are wrong, because if you knew that a view was false, then you wouldn’t hold it as a true belief (as least I hope that is the case). In other words, we all hold the views that we hold because we think that they are true. So we inherently think that all of our views are right even though, if we are honest, we have to admit that at least some of them are almost certainly false.

At this point, we have established two fundamental and critically important points that should act as guiding concepts for you.

  1. Some of your views are wrong
  2. You don’t know which ones are wrong

The implication of those two facts is obvious and inescapable: you must always be willing to consider the possibility that you are wrong about any given topic. This is one of the guiding principle of skepticism, and it is one that everyone should take to heart. It does not matter who you are, it doesn’t matter how smart you are, and it doesn’t matter how well educated you are, you could still be wrong about any of your views. Therefore, you must always consider contrary evidence when you are presented with it (see note at the bottom).

In concept, this sounds simple enough, but actually practicing it is admittedly difficult. We like to think that we are right. We take comfort in thinking that we understand the way that the world works, and being faced with evidence that disagrees with our views is unpleasant. We don’t like being wrong. As a result, we tend to cling to any source that agrees with us, and blindly reject any source that disagrees with us. More often than not, when we are presented with a source or argument that disagrees with one of our views, we simply write it off as “fake news,” we claim that the author had an agenda, we invent elaborate conspiracy theories, we cherry-pick evidence to hurl back at it, etc. We are willing to do anything that it takes to avoid facing the harsh truth that we might be wrong. This is called motivated reasoning, and it is a very serious problem because it prevents you from actually considering evidence and making rational decisions. Fortunately, it is a problem that you can overcome.

You can train yourself to be skeptical. The first step in that is simply to admit that you might be wrong, and, following that, force yourself to actually consider contrary evidence. When you read an article that you like, don’t instantly assume that it is true and hit the “share” button. Rather, stop and think carefully about it. Look at the arguments that it is making, look at the sources that it is citing, fact check the article, read the opposing articles, talk to people who hold the opposing view, etc. Similarly, when you see an article that you disagree with, don’t automatically assume that it is false. Don’t assume that it is “propaganda” or “fake news.” Again, fact check it. There are lots of great fact checking websites out there, fact checking fake newsand you should use them. Try as hard as you can to find information from good sources that supports the opposing. In other words, you should try to disprove your own position. What most people do is simply look for information that supports their view, rather than looking for information that opposes it, and that’s a serious problem. Indeed, it leads to what is known as a confirmation bias. Most people “fact check” by looking for other articles that agree with them. They, of course, easily find them; then they instantly assume that those articles are true, and the existence of those articles further bolsters their view. Do you see why that is a problem? It is a feedback loop that will always leave you with the same view that you started with. There will always be other people who think you’re right, so if you only look at what they have to say, you will only ever see arguments that agree with you, and you will never have to actually consider the possibility that you are wrong. That is a monumental problem.

On that note, try to avoid echo chambers. Don’t unfriend everyone who disagrees with you, and don’t block sources that say you’re wrong. Rather, you should engage with those people and sources. See what they have to say, listen to their arguments, and always be willing to be wrong. I don’t want to get political here, but it politics present a very good example. In recent months, I have seen a large number of people reach the conclusion that only one major news company is trustworthy, and all of the others are agenda driven liars. Thus, they now get their news from only one source. That would be problematic regardless of which source they chose, but it is especially problematic since their chosen source (which won’t be named) is notoriously biased towards one extreme of the political spectrum. Thus, these people have created an echo chamber for themselves, where all of their “news” comes from people who agree with them, and any source that disagrees with them is automatically considered to be false. That is a tremendously huge problem regardless of which end of the political spectrum it falls along, and you should strive to avoid it. Get your information from multiple sources, fact check, and actually read/watch and consider opposing arguments. Also, make sure that the sources that you are using are actually legitimate sources. Although you shouldn’t automatically assume that a source is fake news just because it disagrees with you, there are, nevertheless, a lot of fake news sources out there, and you should be careful to avoid them. Similarly, be leery of sources that are known to have an agenda, because they tend to be biased.

This applies to everyone, regardless of what views you hold and what degrees and qualifications you posses.

This applies to everyone, regardless of what views you hold and what degrees and qualifications you posses.

In short, you should always consider the possibility that you are wrong, carefully fact check everything (especially if it agrees with you), listen to opposing arguments, avoid echo chambers, get information from multiple sources, and even try to discredit your own views rather than trying to support them. I realize that this isn’t easy to do. It is against our nature to be that willing to dismiss our views, but in our modern age of misinformation, this type of skepticism is invaluable and you should strive for it in all areas of life (science, politics, religion, ethics, etc.). I doubt that anyone has ever achieved perfect skepticism (I certainly haven’t), and we are all prone to confirmation biases (myself included), but this is something that you can get better at. You can train yourself to be a skeptic, and as with most things, the more that you do it, the easier that it will become. Make fact checking a habit, rather than something that you only do when you want to disprove someone, and above all else, be willing to be wrong.

Note: Some people may challenge my thought experiment by pointing out that not being able to know which of your views is false is not the same thing as not being able to know that some of your views are true. In other words, you could argue that it is possible to divide your views into things that you are 100% certain are true and things that you are less than 100% certain about. I agree that you can be more certain about some things than others, but I disagree that 100% certainty is ever possible. There are lots of philosophical reasons for this, some of which I discussed here (for example, look up Descartes, the brain in a vat argument, etc.). More practically, however, I think that claiming 100% certainty is the epitome of arrogance. By claiming that you are 100% certain about something, you are saying that you are so knowledgeable, intelligent, and infallible that it is impossible for you to be wrong on a given topic, and that is a rather extreme claim to make. I don’t care who you are, you aren’t perfect enough to achieve 100% certainty. It is always possible that some piece of evidence has alluded you or some cognitive bias has crept in. Thus, you should always be willing to be wrong. To be clear, I’m not suggesting that you can’t be highly certain about some things. You can reach strong conclusions based on the available evidence, but you should always be willing to challenge those conclusions when presented with new evidence (to be clear, I mean actual, proper scientific evidence, not anecdotes, assumptions, conjecture, conspiracy theories, etc.).

Note 2: As several people have pointed out in comments, mathematics are a legitimate exception to my statements regarding absolute certainty (as are the rules of logic) because they are abstract and do not rely on observations of the physical universe. In other words, 2+2 must always equal 4 even if we are living in the matrix, being deceived by an evil genius, etc. Nevertheless, that is clearly quite different from the types of views and beliefs that I was addressing in the post.

Related posts

Posted in Uncategorized | Tagged | 8 Comments

Debunking “The doctors case for homeopathy” by WDDTY: A case study in critically evaluating internet articles

I am very pleased to publish my blog’s first guest post. A reader contacted me a little while ago to see if I would take a look at an essay he had written debunking some of the nonsense that What Doctors Don’t Tell You had written about homeopathy, and although I don’t generally do guest posts, I thought that his piece would be useful to others as an exercise in spotting false information, a rebuke of WDDTY, and a reference for the flaws in many arguments in favor of homeopathy. Therefore, I offered to post it here. Enjoy.

vial of homeopathic medicineI want to explain the problems with an article found in What Doctors Don’t Tell You (WDDTY) titled “The doctors’ case for homeopathy.” This piece is full of misinformation, and it serves as an ideal case study of what I’ve observed to be WDDTY’s typical behavior. Here’s the full list of problems I found with the article, which I will elaborate on below:

  • Cherry picking throughout
  • Making arguments that WDDTY itself admits are invalid
  • Huge over-reliance on anecdotes
  • Outright lies
  • Referencing bad data/bad science
  • Straw man fallacies
  • Deceitful rhetoric, ad hominems, and hyperbole
  • Missing references (that is, unsubstantiated claims)
  • Non sequiturs coupled with hypocrisy
  • Massive internal contradictions
  • Misrepresenting information and quotes from trusted sources

…but all written in a manner that at first appears to be balanced, transparent, and thorough. It is none.

What follows should serve as:

  1. A clear demonstration that the evidence WDDTY provides in favor of homeopathy is unreliable or wrong
  2. A thorough indictment of WDDTY
  3. A cursory lesson in recognising dishonest or fallacious arguments
  4. A cursory lesson in good scientific practice, and good scientific writing

Let’s take a look at each of the problems:


Making arguments that WDDTY itself admits are invalid

The broad pattern of the article is:

  1. Make a lengthy argument
  2. Swiftly concede that the argument made is invalid
  3. Move on to another (or, even, repeat the same) bad lengthy argument
  4. Back to step 2.

Most of the article’s content is rendered useless by these two admissions alone:

  • “Such stories can be dismissed as examples of a particularly dramatic placebo effect or even coincidence”
  • “of course, the fact that pharmaceutical medicine has a poor evidence base doesn’t necessarily make a stronger case for homeopathy.”

With these two remarks the writers have dismissed most of their own arguments. But they know full well that readers will ingest the broad points they’ve made, and gloss over the admissions of their worthlessness. In fact, the subtle admission only makes them (falsely) appear so much more reasonable and balanced. Clever.

This article also claims that scientific studies are thoroughly unreliable, before trying to make a case for homeopathy, using… scientific studies! It’s a neat example of the kind of shameless cherry picking that WDDTY repeatedly uses.


Overreliance on anecdotes

“GPs who prescribe homeopathy acknowledge that their medicines’ mode of action is difficult to explain scientifically, but they cannot deny the often startling evidence of their own eyes.”

Yes, we can deny the evidence of our own eyes. We have to question such evidence, because we know how susceptible it is to (conscious or unconscious) bias, error and misinterpretation. Anyone who’s read Daniel Kahneman’s “Thinking, Fast and Slow” will know that our own interpretations of our own limited experiences are unreliable. This is why we perform blinded, randomized, controlled trials. The “evidence of our own eyes” regarding homeopathy is perfectly well explained by proven phenomena: the placebo effect, external confounders, and regression to the mean (a phenomenon whose named absence in a lengthy article pretending to be balanced about homeopathy is notable). WDDTY’s assertion here shows just how happy the magazine is to accept completely unreliable information — anecdotes — so long as it supports WDDTY’s position.

Remember, the anecdotes that make up most of this article “can be dismissed as examples of a particularly dramatic placebo effect or even coincidence”. So including them was just dishonest.

“…but they’re not uncommon.”

It’s true that anecdotes are not uncommon. They’re everywhere, and in support of anything you can name. Commonality of a belief doesn’t make it any more true.


Outright lies

Sadly, this is where WDDTY comes into its own.

“from the very beginning, there’s been a turf war between two opposing medical systems: the pharmaceutical doctors and the homeopathic doctors.”

There are no such things as “pharmaceutical doctors.” GPs use an extensive array of different treatments (including no treatment); some of those treatments are drugs. (Those same doctors would embrace homeopathic treatments if only the evidence supported them.) The obsessively black-and-white approach that WDDTY takes is insidious and only makes its advocates look simple.

“[In Europe] studies have shown that primary-care practices offering homeopathy have better patient outcomes than pharma-only practices, and often at lower cost.”

To repeat: there are no “pharma-only practices”.

“the monoculture of pharmaceutical drugs”

To repeat: there is no “monoculture of pharmaceutical drugs”.

“In 2010, [the British] Science and Technology Committee report recommended … banning any further homeopathic research”

Here’s what the report actually said:

“We would challenge Professor Harper’s comment that research funding should be directed towards exploring theories that are not scientifically plausible. Research funding is limited and highly competitive. The Government should continue its policy of funding the highest quality applications for important scientific research determined on the basis of peer review.” Then: “We recommend that the Government Chief Scientific Adviser and Professor Harper, Chief Scientist at the DH, get together to see if they can reach an agreed position on the question of whether there is any merit in research funding being directed towards the claimed modes of action of homeopathy.” Also: “There has been enough testing of homeopathy and plenty of evidence showing that it is not efficacious. Competition for research funding is fierce and we cannot see how further research on the efficacy of homeopathy is justified in the face of competing priorities. It is also unethical to enter patients into trials to answer questions that have been settled already.”

In the entire 275-page report, there is absolutely no mention of a research “ban”. WDDTY has taken the report’s clear, balanced and rational recommendation — to direct what limited funding is available towards new research that has at least a chance of yielding positive results — and turned it into “banning any further homeopathic research”: not just spin, but a complete lie.

(By the way, the report is a very good read for anyone sincerely interested in a clear, fair and logical analysis of homeopathy in the UK)

“It is depressing to see the interests of patients being threatened by a small posse of poorly informed and discourteous critics, who mix a little science with denigration and abuse”

The interests of patients are benefited, not threatened, by people highlighting that a popular treatment is ineffective. It’s no “small posse” of people who recognize the failure of homeopathy, and they’re not “poorly informed” – it’s practically the entire community of scientific and medical experts. And they don’t use “a little science” – they refer to the overwhelming consensus of the best available evidence.

WDDTY then claims that studies on animals and even plants have shown homeopathy treatments working better than placebos.

“All of this non-human evidence inherently rules out the dismissive ‘placebo-effect’ explanation.” / “homeopathic remedies have repeatedly been shown to have measurable effects on non-human biological systems, which inherently rules out the placebo effect”

No, it doesn’t. There are plenty of reasons to believe that placebo effects are in evidence in animals; not least that the effects of treatments are recorded by placebo-susceptible humans.

As for plants, this is arguably the low point of the article: WDDTY’s single stated reason to believe homeopathy benefits plants is a study in which 26 out of 30 homeopathic products failed; then when the test was repeated, three of the four remaining homeopathic products failed as well (in one of the preparations, the disease even got worse). And the single successful treatment (with “a small to medium effect size”)? A product for which the study’s authors state, “the exact composition … is not known” – in fact, it appears not to actually be homeopathic at all (a homeopathic version of that same product was one of the 29 that failed). WDDTY fails to mention any of these truths (as well as failing to mention that the journal the study is published in is notoriously shoddy). Citing this as favourable evidence is called p-hacking, and it’s notoriously deceitful (un)scientific behaviour.

To be as clear as possible: the data from this study suggest that homeopathy doesn’t work on plants, and WDDTY has used it (and as its sole reference) to claim that homeopathy does work on plants. It’s clear that WDDTY’s researchers are either completely incompetent or thoroughly dishonest, both to reference this journal and to misconstrue the information within it. Is this demonstrably dreadful magazine really a source that anyone would want to refer to or rely on?

“The war is not about evidence”.

The “war” – such as it is – is all about evidence. The evidence overwhelmingly shows that homeopathic remedies are nothing but placebos – if you’re prepared to look at all of the evidence, and separate the reliable evidence from the unreliable, rather than just referencing the (bad) studies that (do or can be spun to) support homeopathy.

“Because homeopathic medicines often contain not a single molecule of an active ingredient, opponents mock them as an affront to rationality—and indeed, a threat to the whole of science.”

Homeopathy isn’t “a threat to the whole of science” – the whole of science is a threat to homeopathy. If the evidence supported homeopathy, science would evolve to accommodate it. The evidence does not support homeopathy.(Here, for curiosity, is a non-exhaustive list of occasions when the scientific consensus moved on because of new evidence)

“Opponents make much of the ‘consistent failure to demonstrate effect beyond placebo’ when trials of homeopathy are studied; this is untrue.”

It only appears untrue if you accept poor evidence from bad studies, ignore good studies, and pretend anomalies are less statistically likely in good studies than they are. Which is to say, the claim that it’s untrue is itself fundamentally dishonest.

“However, after 70 years of being part of the NHS, homeopathy now faces extinction.”

Come on, no it doesn’t. Homeopathy is a booming business. (I admit this is speculation – I’ve found it extremely hard to come by reliable data on revenue or use of homeopathy; but I see no evidence that homeopathy is facing “extinction” or anything close to it. Boiron alone made a profit of €73.9 million in 2015)

“Clearly, both homeopathy and conventional medicine are on a par with each other when it comes to evidence-based medicine”

See the “bad data” section below for detail on why this is a lie.

“the NHS needs more homeopathy, say Britain’s 400 GPs who currently prescribe it.”

In this article, four – not 400 – GPs are quoted appearing to support homeopathy. As you’ll see in the “unsubstantiated claims” section below, GP referrals for homeopathy are “almost always at the patient’s request rather than as the result of a clinical decision”. It’s yet another outright lie to claim that “Britain’s 400 GPs who currently prescribe it [say] the NHS needs more homeopathy.”

This is WDDTY’s USP: helping people who believe in things disproven by science to convince themselves that they’re actually in good company, and that experts, evidence and science support their beliefs. In reality it’s lie after lie after lie.

“such stories [are] not uncommon”

It seems they are uncommon, since in a 3,700-word piece, WDDTY has only managed to find four GPs apparently prepared to tell such stories. (We know now that that’s because most of the 400 GPs WDDTY mentions probably don’t have such stories at all – they are, in fact, uncommon.) On top of this, I’m giving WDDTY the benefit of the doubt and assuming that the quotes from GPs are real; as you’ll see in the footnote at the end of this document, WDDTY is perfectly happy to make up quotes to support the most despicable lies.

“homeopathy—with its zero side-effects and arguably adequate reliability”

A serious examination shows that, despite trying really hard to do so, this article completely fails to demonstrate that homeopathic remedies have “arguably adequate reliability.” Indeed, the fact that the writers have to resort to the lies, misinterpretations, ad hominems and fallacies that they do suggests (correctly, as it happens) that there isn’t any good evidence that homeopathy has even arguably adequate reliability.


Referencing bad data/bad science

“An analysis of the responses to the homeopathic medicines revealed…”

The audit cited here (Tom Robinson in 2006) bears little resemblance to a scientific study. There’s no control, no placebo, no blinding. It’s not a study at all, just another series of anecdotes.

“Only 3 per cent of his [homeopathy] patients found their condition worsened following treatment”

Pointing out that ineffective remedies don’t usually make things worse is redundant. It’s not a point in favor of homeopathy.

“[Critics’] attacks echo those of the medieval Vatican against Galileo: it cannot be true, so it’s not.”

“It cannot be true, so it’s not” is a perfectly sound argument. It was wrong in the case of Galileo (if anyone even said it to him) because “it cannot be true” was a false claim. In the case of homeopathy: it cannot be true. So it’s not. (Also, and far more significantly: when it’s tested, it’s shown not to be true.)

By the way, the “Galileo Gambit” is a famous (and famously bad) debating technique.

“[Homeopathy’s] ultra-low doses are prepared by ‘serial dilution’ in water—adding a drop of the original medicine to 100 drops of water … homeopaths not only say such extreme dilutions work, but actually work better than the original medicine.”

Even WDDTY’s own writers appear not to have understood the basic theory of homeopathy. There is no “original medicine” to be diluted; only a substance that causes similar symptoms to those the patient is suffering.

“By the end of 2014, homeopathy had been tested in 104 RCTs for 61 different medical conditions: 41 per cent were positive; 5 per cent were negative; and 54 per cent were inconclusive. This track record is strikingly similar to conventional medicine’s. A 2007 analysis of a “large random sample” of RCTs of conventional treatments revealed that 44 per cent were ‘likely to be beneficial’, 7 per cent were ‘likely to be harmful’ and 49 per cent were inconclusive”

This statement about homeopathy studies is not, however, from a proper scientific review. It’s a claim from the Faculty of Homeopathy. WDDTY is comparing a peer-reviewed paper in the prestigious Journal of Evaluation in Clinical Practice to a dubious claim from the Faculty of Homeopathy. Putting the two in the same space and trying to draw a comparison is amateurish. Luckily the Lancet has performed a fair, direct comparison. Homeopathy came out worse.

Indeed, when serious meta-analyses are performed, it’s shown that “there are no health conditions for which there is reliable evidence that homeopathy is effective.”

Most of the studies WDDTY references appear to be any of:

  1. not scientific studies at all, just more anecdotes
  2. lacking controls, lacking blinding, lacking placebos
  3. performed on worthlessly small sample sizes
  4. published in poor quality journals – or even not published in a journal at all
  5. demonstrating different (sometimes opposite!) results to the results WDDTY claims.

As ever, they’re all cherry picked to support WDDTY’s position, ignoring the bigger picture: that the results of hundreds of studies – the vast majority of good studies – overwhelmingly show homeopathy doesn’t work.

And a word about animal studies: animal studies, like case reports and case series (and like anecdotes), are first steps. They’re dipping your toes in the water to see if it’s warm. You use their results to determine whether further, human studies are worth running (and safe to do so). To reference an animal study with positive results when human studies have already shown negative results is thoroughly dishonest (just as referencing individual studies when there are meta-analyses demonstrating the opposite results is dishonest). It’s cherry picking, again. WDDTY’s writers are either incompetent not to know this, or dishonest to ignore it.

More on the hierarchy of scientific evidence can be found here.


Straw man fallacies

“Only a very few critics confine themselves to what they regard as scientific principles—people who believe that science knows everything about everything, and nothing remains to be explained—scientific ‘fundamentalists’, perhaps.”

There is literally nobody in existence “who believes that science knows everything about everything, and nothing remains to be explained”. The threat of “scientific fundamentalism” continues to be massively overstated. Pretending that the people highlighting homeopathy’s flaws are simple-minded “fundamentalists” is a cheap straw man. See the rhetoric section below for more on WDDTY’s attempt to paint a childish caricature of its opponents.


Deceitful rhetoric, ad hominems, and hyperbole

WDDTY pretty shameless in their use of absurdly melodramatic language (some of these are quotes that WDDTY simply features, but quotes that they fail to challenge):

  • “propaganda campaign”
  • “The Good Thinking Society (a name eerily close to Orwell’s dystopian “Ministry of Truth”)”
  • “attacks echo those of the medieval Vatican”
  • “tide of hostility”
  • “vitriolic opponents”
  • “the supposedly respectable British Medical Journal”
  • “that arch-critic of homeopathy”
  • “a small posse of poorly informed and discourteous critics”
  • “scientific ‘fundamentalists’”

This shouldn’t be necessary in a serious scientific publication that can rely on the strength of its arguments. But when you strip out the misinformation, this childish rhetoric is all WDDTY has left. Of course, painting a paranoid picture of people who don’t share your beliefs as one-dimensional, conspiratorial villains makes it far easier to dismiss those people and the valid points they make.


Missing references (that is, unsubstantiated claims)

This is just bad practice: another reason to steer clear of the magazine (which on its Facebook page calls itself a journal; another lie):

  • “Homeopathic medicines appear particularly effective at the extremes of age…”
  • “Many GPs prescribe antibiotics for such conditions”
  • “[In Europe] studies have shown that primary-care practices offering homeopathy have better patient outcomes than pharma-only practices”
  • “One area where homeopathy has scientifically proved more effective than conventional drugs is with upper respiratory tract infections like flu, coughs, colds and sore throats.”
  • “an appropriate homeopathic remedy, now shown to be effective according to ‘outcome studies’”
  • “Britain’s 400 GPs who currently prescribe [homeopathy]”

I sought out the source of that final statistic, that 400 GPs (that’s all but 99.3% of them, by the way) currently prescribe homeopathy. The stat turned out to be yet another unsubstantiated claim from the Faculty of Homeopathy. Along my journey, I discovered that “a review carried out by West Kent NHS Primary Care Trust in 2007 found that [a referral for homeopathy] was almost always at the patient’s request rather than as the result of a clinical decision” (quote taken from Wikipedia; not the source, as the link was broken at time of writing). Even WDDTY’s attempts at appeal-to-authority fallacies are failures.


Non sequiturs coupled with hypocrisy and massive internal contradictions

“But over the past decade, conventional medicine itself has come under fire for having an equally poor evidence base…”

Conventional medicine allegedly having a poor evidence base is not an argument in favour of homeopathy (as the article itself admits moments later, thus demonstrating that this section was a waste of everyone’s time – except, again, for those people who will soak up the points presented and ignore the admission that they’re irrelevant). Ironically, WDDTY uses this section to suggest that scientific studies are practically worthless, then goes on to reference a load of scientific studies in homeopathy’s favor (two typical behaviors of proponents of flawed arguments: claiming evidence is unreliable except when it appears to be favorable; and trying to deflect attention to the flaws of others in order to avoid scrutiny of itself).

“Why should homeopathy be singled out?” asks Dr Kaplan. “To use evidence-based medicine to attack homeopathy exclusively and call for legislation against it, while huge swathes of conventional medicine lack evidence, is rather strange behaviour,” he says. “Double standards? That would be putting it euphemistically.””

Sure, if other treatments lack evidence, we should be acting against them too (and in many cases, we do, in contradiction of Kaplan’s lie that homeopathy is “exclusively” “attacked”). That’s not a reason to give homeopathy a free pass. Homeopathy gets so much attention focused on it because it’s so prominent. It gets attention focused on it because of all these anecdotes!


Misrepresenting information and quotes

“As a result, says Horton, “much of the scientific literature, perhaps half, may simply be untrue”.”

Logic of Science has addressed misuse of this quote specifically. Ironically the kind of papers that Horton suggests are “simply untrue” are the very ones that WDDTY repeatedly references. The unreliability of scientific papers is the very reason we strip out bad data using meta-analyses and systematic reviews – which are the very evidence that WDDTY is careful to ignore.

“That view is shared by the late Professor David Sackett, widely regarded as one of the fathers of clinical epidemiology: “Evidence-based medicine is not restricted to RCTs and meta-analyses,” he wrote 20 years ago. “It involves tracking down the best external evidence with which to answer our clinical questions.””

The best evidence. Not poor evidence cherry picked because it appears to support a position. And when the best evidence is RCTs and meta-analyses, we pay attention to them, rather than ignoring them in favor of poor quality studies and anecdotes. See my earlier link on hierarchy of evidence.

“Who said this? Ironically, no less a figure than that arch-critic of homeopathy, Edzard Ernst.”

(Interestingly in a related piece Ernst suggests that for some problems Acupuncture, Aromatherapy and Hypnosis can do more good than harm. Homeopathy isn’t on the list.)

These quotes are from people whom WDDTY vilifies when they disagree; but the magazine is happy to quote them when their words do (or can be spun to) support their position. Cherry picking. It’s cheap and dishonest and a mark of the gutter press.


The truth at last

“Opponents of homeopathy claim [outcome] studies are worthless as they cannot exclude either the placebo effect or the fact that many conditions are ‘self-limiting’—in time, patients get better anyway. The only way to establish whether homeopathy really works, they say, is through randomized controlled trials (RCTs) testing a homeopathic pill against an identical-looking placebo. According to British pressure group Sense About Science, “over 150 clinical trials have failed to show that homeopathy works. Some small-scale studies have yielded positive results, but this is due to poor methodologies or random effects.””

At last, WDDTY speaks the truth (broadly, with some exaggerated turns of phrase thrown in). These 90 words are just about the only content of any real value in the whole article.



In this article, WDDTY admits (correctly) that anecdotes don’t constitute good evidence, but spends most of its word count using anecdotes to back up its claims. It claims (falsely) that scientific studies are unreliable, then tries to back up its claims using scientific studies. The studies it does reference are awful. Sometimes those studies even draw the very opposite conclusions to that which WDDTY claims. WDDTY highlights an (alleged) poor evidence base for modern medicine in defense of homeopathy, then admits that an (alleged) poor evidence base for modern medicine does not constitute a defence of homeopathy. It entirely ignores the best available evidence, because that evidence doesn’t suit WDDTY’s argument — indeed, the best available evidence shows WDDTY’s position to be wrong. Finally, WDDTY makes a series of claims that are outright lies, a further series of claims that are unsubstantiated, and a series of suggestions that are childish, hyperbolic and absurd.

WDDTY doesn’t exist to tell the truth. It exists to tell its readers what they want to hear, and pretends to have evidence to back it up. With the deepest irony, I paraphrase their own accusation: “it’s not about evidence, but ideology.”

WDDTY’s staff are very good writers, who know how to cherry pick any apparent evidence that suits them (and omit that which doesn’t), dress it up to appear more reliable and balanced than it is – or even to make claims that it doesn’t really suggest at all – and throw in some incendiary hyperbole to excite its readers. I wish I could say the writers at WDDTY are just incompetent – but the high quality of their writing, and the fact that they admit their arguments are invalid and publish them anyway, suggests to me that the lies that fill their magazine are quite intentional. It doesn’t really matter — either way, it’s a thoroughly unreliable source.

There’s no global media conspiracy: if something in the media is important and true, you should almost always be able to find it cited by a credible source. If your best or only source is What Doctors Don’t Tell You (or Joseph Mercola, or Green Med Info, or Andrew Wakefield, or Natural News, all of whom use similar tactics), it’s not because those people and publications are beacons of truth trailblazing in a world of media corruption and deceit; it’s almost certainly because they’re wrong. And time after time, when their claims are scrutinized, they’re shown to be exactly that: wrong.

For a demonstrably rubbish article like this, WDDTY deserves to be avoided. When they publish vile, dangerous lies about the safety and effectiveness of lifesaving interventions like vaccines and cancer treatments (which they do, using the same cheap tricks that I’ve demonstrated above)*, anyone should be embarrassed and ashamed to have anything to do with them.


*Here’s one example. WDDTY, Jan 2013: “Chemotherapy isn’t only useless against cancer–it even encourages the tumour to grow, researchers have discovered… They say that chemotherapy is ‘completely worthless’ and that cancer sufferers would do better by avoiding the drugs altogether.”

The “completely worthless” quote is a fiction; the paper doesn’t recommend avoiding chemotherapy drugs; and when a skeptical reader subsequently contacted the paper’s authors for comment, they said: “The paper says nothing of the sort… Our study has been misquoted and misinterpreted – I believe on purpose…”

Posted in Vaccines/Alternative Medicine | Tagged , , , , | Leave a comment

Climate change and a flat earth (I can’t believe we’re still dealing with this argument)

I'm not sure where this meme originated. if anyone knows, please tell me so that I can give credit.

Image via Terry Pratchett’s “Discovery World.”

When I started this blog, I wanted it to be entirely about science and the fundamental concepts of logic. As such, I decided that topics like religion and politics where off limits. However, anytime that a politician makes a statement about science, they enter my territory and I feel obliged to deal with their comments (by the same token, I only talk about religion when people use it to make claims about science). One of the most common and irritating ways that this occurs is when politicians try to compare modern science (usually climate research) to the idea that the earth is flat and/or we are the canter of the universe. This argument is used ad nauseum and I have dealt with it before. Nevertheless, it is once again featuring prominently in our political dialogue, so I feel the need to explain why it is utter nonsense yet again.

I am, of course, referring to the recent comments by Trump advisor Anthony Scaramucci on CNN. When asked about the scientific consensus on climate change, he made several hopelessly ignorant comments like, “There was an overwhelming science that the Earth was flat, and there was an overwhelming science that we were the center of the world.” Statements like that are, however, absurd for numerous reasons.

First, and most obviously, they are blatantly false. Science is an extremely careful, empirical, and systematic process of making observations, forming hypotheses, and empirically testing those hypotheses by using repeatable, controlled* experiments, and that system is an extremely recent one. It has only existed in its modern form for the last few centuries. In contrast, we have known that the earth was round since the ancient Greeks! So there can’t possibly have been a scientific consensus that the earth was flat because science as we know it did not even exist then! Similarly, the idea that we are the center of the universe was discredited long before modern science, and the idea was largely based on religion. Indeed, although Galileo slightly predates modern science, when he demonstrated that the earth was not the center of the universe, he did so by applying the same basic type of systematic observations and reasoning that scientists use today. Thus, although modern science never said that the earth was the center of the universe, scientific reasoning was the thing that discredited that idea.

I want to elaborate on that last point for a second, because I want to be clear that I am not suggesting that there were no scientists before the modern era. There certainly were, but science was not the formalized discipline that it is now. In other words, although there were people who were making observations, testing hypotheses, etc., there was not a standard for the type of reasoning that qualified as science, nor was there the vast body of philosophical thought about the nature of science that we have today. My point is that the “science” that said that the earth is flat and the earth is the center of the universe is not actually the same thing that we know as science today, thus the comparison is invalid (i.e., it is an equivocation fallacy because it uses the word “science” to mean two different things).

Further, so far I have only been talking about the general concept of what science is, but we also have to take into account the tools that are available to scientists. For example, the statistics and mathematical formulas that we use today have only been around for about 100 years (sometimes much less than that). Those mathematics are extremely important because they are what allow us to rigorously examine our data and assign confidence values to our results. Similarly, today we can use computers to analyze millions of data points, run massive simulations, and do analyses that were unimaginable to scientists just a few decades ago. Further, our ability to accurately make the measurements that we use in those analyses has also increased astronomically. For example, we now have satellites that can continuously collect very precise data from all of the world. Tools like that give us incredible power for understanding our planet and the universe in general, and they give modern science a major advantage over past scientific endeavors.

There is also another problem here, and it is a more fundamental one. If this argument actually worked, then you could use it any time that you wanted, rather than simply for climate change. In other words, if the fact that people used to think that the earth was flat actually gave you cate blanche to reject science, then you would never have to accept any scientific result that you didn’t like. For example, I could say, “scientists say that washing your hands helps prevent the spread of disease, but they were wrong about the earth being flat, so I don’t have to listen to them when they say that we should wash our hands.” That is obviously absurd, but that is exactly the same thing that Scaramucci did. His argument was logically identical, which means that if either of those arguments worked, then both should work, and the fact that the hand washing example is clearly absurd means that the climate change argument must also be absurd. That’s how logic works.

Nevertheless, you might be tempted to think, “but scientists aren’t completely, 100% sure that we are causing climate change, so it is always possible that they are wrong.” Technically, that is true, but here is the important qualifier: science never gives us 100% certainty. It is inherently a skeptical process, and it only tells us what is most likely true given the current evidence, not what is absolutely true (details here). So sure, it is true that I can’t be 100% sure that we are causing climate change, but I also can’t 100% sure that the earth is round. Indeed, I can’t be 100% sure that you even exist, nor can I be 100% sure that Donald Trump isn’t an alien squid from another universe. However, the fact that I can’t be 100% sure about those things does not mean that I get to have any form of practical doubt about them. It would clearly be ridiculous for me to actually think that Trump is an alien squid, and politicians should not make laws based on that notion. Even so, the fact that we can’t be 100% sure about climate change does not mean that you get to have any practical doubt about it, nor should we be basing laws on that doubt.

In technical terms, what I have been describing is known as an argument from ignorance fallacy. It can occur in many ways, but one of its forms happens when you say, “we can’t be sure that X is true, therefore X is false” or “we can’t be sure that X is true, therefore it is reasonable to believe that it isn’t true.” This is not logically valid. The fact that we can’t be completely certain of a conclusion does not mean that you can assume that all of the evidence supporting that conclusion is false. This comes back to one of my favorite topics: the burden of proof. The person making a claim is always responsible for providing the evidence to support that claim. In other words, if you want to claim that we aren’t causing climate change, then you must provide actual evidence that we aren’t. You can’t simply assert that we can’t be 100% certain, the models might be wrong, etc. If you want to claim that all of the scientific evidence is wrong, then you have to provide evidence to support that claim, not speculations. Further, although it is true that science never proves anything with 100% certainty, there are some topics that are so well supported by so many lines of evidence that it is almost unthinkable that they could actually be totally wrong, and anthropogenic climate change is one of those topics (others include atomic theory, the germ theory of diseases, etc.).

This brings me to the final topic that I want to discuss here: the actual evidence for climate change. You see, so far I have been talking in pretty general terms about science, but Scaramucci’s statements are particularly absurd and misguided when we focus in on climate change, because there is so much evidence supporting it. As I have explained in more detail elsewhere, we have tested the sun, volcanoes, and other natural drivers of climate change, and they can’t explain the current warming by themselves (Meehl, et al. 2004; Hansen et al. 2005; Wild et al. 2007; Lockwood and Frohlich 2007, 2008; Lean and Rind 2008; Foster and Rahmstorf 2011; Imbers et al. 2014). However, including our greenhouse gases in the calculations does explain the warming trend (Stott et al. 2001; Meehl et al. 2004; Allen et al. 2006; Lean and Rind 2008; Imbers et al. 2014). In other words, we know that our emissions are driving the warming, because we have tested the natural drivers of climate change, and none of them (even when combined) can account for the current increase in temperatures. Further, we know that CO2 traps heat and is largely responsible for both our climate and past climate changes (Lorius et al. 1990; Tripati et al. 2009; Shakun et al. 2012), we know that we have greatly increased the CO2 in the atmosphere (Bohm et al. 2002; Ghosh and Brand 2003;Wei et al. 2009), and we have used satellites to directly measure the amount of heat energy that CO2 in our atmosphere is trapping, and (just as expected) it is trapping an increasing amount of heat (Harries et al. 2001; Griggs and Harries 2007). In other words, we have direct evidence that our CO2 is causing the earth to trap more heat (i.e., warm). The evidence is overwhelming, and when you look at the scientific literature, you will find thousands of papers saying that we are causing climate change, and only a tiny handful that disagree. The scientific consensus on this topic is incredible (details here), and the consensus among scientists exists because of the evidence. Scientists do not “believe” in climate change (as Scaramucci erroneously asserts), rather we accept it as fact because of the overwhelming evidence. There is simply no serious debate on this topic among actual scientists.

To illustrate this a different way, a few weeks ago, I attended the annual meeting of the Ecological Society of Australia, and there were lots of talks about the effects that climate change is having and will have on plants and animals, but not one of those talks hedged their comments with statements like, “if climate change is real.” Rather, all of them treated anthropogenic climate change as a scientific fact because that is what it is, and it is absolutely crazy to suggest that we should act as if it isn’t settled just because some non-scientists over 2000 years ago thought that the earno matter what crackpot notion you believeth was flat! Further, the fact that a handful of scientists today disagree doesn’t help that argument either because there will always be a few who disagree on any topic (this is an appeal to authority fallacy). Having an advanced degree does not guarantee that you know what you are talking about, and you can sadly find plenty of crackpots even among scientists (note: before you accuse me of committing an appeal to authority fallacy by citing the consensus, please read this post on the difference between appealing to authority and deferring to experts).

To be clear, there is disagreement about the extent of climate change that we are going to cause. There are lots of variables that have to be considered, and there are admittedly still topics that we are actively studying in order to refine out models. However, although the models may be imprecise about the extent of the warming, it is almost unthinkable that we are wrong about the warming itself. Indeed, the idea that we are causing the climate to change is not based on the predictive models, rather it is based on research like the papers I presented earlier. Further, out models have actually done a pretty good job of predicting the current warming, so although they aren’t perfect, they probably aren’t that far off (Hansen et al. 2006; Frame and Stone 2012; Rahmstorf et al. 2012; Cowtan et al. 2015; Marotzke and Firster 2015).

In short, the science that said that the earth is flat or that we are the center of the universe was not the same thing as what we know as science today. Further, the past few decades have furnished us with a fantastic array of amazing tools for collecting and analyzing massive amounts of data, and when we apply those tools to the topic of climate change, we get an extremely consistent result: it is real and we are the main cause of it. Because of that evidence, there is no serious debate about this among scientists. Yes, there are a few dissenting voices, but those will always be there. Further, yes, it is always technically possible that we are wrong, but it is also technically possible that Donald Trump is a space squid, and the fact that something is technically possible does not mean that it is rational to think that it is actually true.

*Note: When I said that science uses controlled experiments, I was not referring specifically to randomized controlled trials (RCTs). Although they are a particularly powerful design, they are far from the only one, and they often aren’t applicable in fields outside of medicine. In climate research, for example, we generally can’t do RCTs, but we can still control our experiments and analyses by measuring all of the confounding factors and including those factors in out analyses.

Related posts

Literature cited

  • Allen et al. 2006. Quantifying anthropogenic influence on recent near-surface temperature change. Surveys in Geophysics 27:491–544.
  • Bohm et al. 2002. Evidence for preindustrial variations in the marine surface water carbonate system from coralline sponges. Geochemistry, Geophysics, Geosystems 3:1–13.
  • Cowtan et al. 2015. Robust comparison of climate models with observations using blended land air and ocean sea surface temperatures. Geophysical Research Letters 42:6526–6534.
  • Foster and Rahmstorf 2011. Global temperature evolution 1979–2010. Environmental Research Letters 7:011002.
  • Frame and Stone 2012. Assessment of the first consensus prediction on climate change. Nature Climate Change 3:357–359.
  • Ghosh and Brand. 2003. Stable isotope ratio mass spectrometry in global climate change research. International Journal of Mass Spectrometry 228:1–33.
  • Griggs and Harries. 2007. Comparison of spectrally resolved outgoing longwave radiation over the tropical Pacific between 1970 and 2003 Using IRIS, IMG, and AIRS. Journal of Climate 20:3982-4001.
  • Hansen et al. 2005. Earth’s energy imbalance: confirmation and implications. Science 308:1431.
  • Hansen et al. 2006. Global temperature change. Proceedings of the National Academy of Sciences 10314288–14293.
  • Harries et al. 2001. Increases in greenhouse forcing inferred from the outgoing longwave radiation spectra of the Earth in 1970 and 1997. Nature 410:355–357.
  • Imbers et al. 2014. Sensitivity of climate change detection and attribution to the characterization of internal climate variability. Journal of Climate 27:3477–3491.
  • Lean and Rind. 2008. How natural and anthropogenic influences alter global and regional surface temperatures: 1889 to 2006. Geophysical Research Letters 35:L18701.
  • Lockwood and Frohlich. 2007. Recently oppositely directed trends in solar climate forcings and the global mean surface air temperature. Proceedings of the National Academy of Sciences 463:2447–2460.
  • Lockwood and Frohlich. 2008. Recently oppositely directed trends in solar climate forcings and the global mean surface air temperature. II. Different reconstructions of the total solar irradiance variation and dependence on response time scale. Proceedings of the National Academy of Sciences 464:1367–1385.
    Lorius et al. 1990. The ice-core record: climate sensitivity and future greenhouse warming. Nature 139–145.
  • Marotzke and Firster 2015. Forcing, feedback and internal variability in global temperature trends. Nature 517:565–570.
  • Meehl, et al. 2004. Combinations of natural and anthropogenic forcings in the twentieth-century climate. Journal of Climate 17:3721–3727.
  • Rahmstorf et al. 2012. Comparing climate projections to observations up to 2011. Environmental Research Letters 7:044035.
  • Shakun et al. 2012. Global warming preceded by increasing carbon dioxide
  • Stott et al. 2001. Attribution of twentieth century temperature change to natural and anthropogenic causes. Climate Dynamics17:1–21.
  • Tripati et al. 2009. Coupling CO2 and ice sheet stability over major climate transitions of the last 20 million years. Science 326:1394–1397.
  • Wei et al. 2009. Evidence for ocean acidification in the Great Barrier Reef of Australia. Geochimica et Cosmochimica Acta 73:2332–2346.
  • Wild et al. 2007. Impact of global dimming and brightening on global warming. Geophysical Research Letters

Posted in Global Warming | Tagged , , , , | 18 Comments

Evolutionary mechanisms part 7: Gene flow

So far in this series, we have talked about natural selection, sexual selection, genetic drift, and mutations, but that still leaves us with one final and very important evolutionary mechanism: gene flow. This simply refers to the movement of genetic information among populations (migration, for example, causes gene flow). At the surface, that may sound quite simple and boring, but it is actually fascinating, complex, and immensely important. For example, gene flow often plays a critical role in determining whether or not an organism will speciate (i.e., split into two species). Similarly, gene flow interacts with genetic drift and selection, and low levels of gene flow may predispose a population for extinction, whereas high levels may be equally harmful by preventing a population from adapting. Therefore, in the final post of this series, I want to talk about gene flow and explain how it interacts with other evolutionary mechanisms and determines the fate of many populations.

Maintaining connectivity

When we talk about gene flow, we are generally talking about the movement of genetic information from one population to another. This can occur either by individuals moving among populations (and mating) or by mechanisms that just transfer genetic material (e.g., wind and animals moving the seeds and pollen of plants).

One of the most powerful traits of gene flow is its ability to prevent populations from splitting apart. If you recall from previous posts, selection and genetic drift are constantly shifting the allele frequencies of populations, but which allele becomes common can vary from one population to the next. Genet drift is, of course, random. Thus, for a given genetic locus, one allele might become common in one population, while becoming quite rare in another. Similarly, selection (which is non-random) simply adapts populations for their current environment. As such, we expect that two populations that are living in different habitats would be under different selection pressures and thus evolve in different directions. Finally, mutations randomly create new genetic information, so if a mutation occurs in population A but not in population B, that will result in yet another genetic difference between the two populations.

The result of all of this is that when two populations are isolated, they both slowly accumulated genetic differences. Different mutations will arise in each one, different alleles will be selected for, and different alleles will be affected by genetic drift. Over many generations, these populations will eventually accumulate so many differences that we would consider them to be different species. This is called allopatric speciation (i.e., the formation of two new species from isolated populations).

Gene flow can, however, counteract that process by moving genetic information from one population into another. Imagine, for example, that a new mutation arises in population A but not population B. However, in each generation, a large number of individuals migrate between the two populations. As a result, some of the individuals who have the mutation carry it with them into population B where they mate with individuals of that population and pass the mutation on. Similar patterns also occur with selection and drift (more on that in a minute). Gene flow constantly shuffles the genetic information around and causes alleles to be exchanged among populations. Thus, it mixes the populations together and can prevent them from splitting apart.

Note: Although low levels of gene flow can be a very important driver of speciation, they are not required. There is also sympatric speciation which occurs when a connected population evolves in two separate directions and becomes two separate species (differences in mating patterns are a good way for this to happen).

 Interactions with genetic drift

Gene flow is often extremely important for the long-term survival of a population, and maintaining gene flow is often a conservation priority. This is because gene flow is a powerful way to counteract genetic drift. If you remember from previous posts, genetic drift is constantly removing variation from a population, but that variation is important because it is what allows populations to cope with future changes to the environment, new diseases or predators, etc. Indeed, it is well established that genetically diverse populations are superior to low-diversity populations (Hedrick and Kalinowski 2000, Keller and Waller 2002, Reed and Frankham 2003, Spielman et al. 2004, Whiteman et al. 2006, Hughes et al. 2008). Additionally, in very small populations, genetic drift can actually overpower selection and cause harmful alleles to rise to prominence (Lacy, 1987; Whitlock, 2000; see this post).

This shows gene flow's ability to counteract genetic drift. In the top panel (A) there is no gene flow, and alleles became fixed in each population by 200 generations. In contrast, in the lower panel (B) there is gene flow, and alleles never stay fixed. Thus, gene flow allowed the populations to maintain a higher diversity.

This shows gene flow’s ability to counteract genetic drift. In the top panel (A) there is no gene flow, and alleles became fixed in each population by 200 generations. In contrast, in the lower panel (B) there is gene flow, and alleles never stay fixed. Thus, gene flow allowed the populations to maintain a higher diversity.

Gene flow can, however, counteract this by constantly restocking the genetic diversity of a population. Consider the examples illustrated on the left. In the top example, I simulated the frequency of a given allele in10 populations of 25 individuals. At the start of each simulation, the allele had a frequency of 50% and there was no selection. In the first set of 10 populations, there was no gene flow, and you can see that the frequency of that allele changed randomly (sometimes increasing, sometimes decreasing), and by 200 generations, in each population it had either become fixed (i.e. everyone had it and the alternative allele was lost) or it was totally removed from the population. In other words, in every case, diversity was lost because of genetic drift. Now, look at the second set of 10 populations (bottom). That simulation was set up identically, but at the start of each generation, I had one individual immigrate from a large neighboring population that had an allele frequency of 50% for the allele in question. As you can see, neither allele ever becomes fixed for more than a few generations, because even if one is lost, gene flow reintroduces it. Thus, the populations maintain a higher diversity than they would without gene flow.

Gene flow can also cause something called the “rescue effect.” Many things such as disease outbreaks, over-harvesting, natural disasters, etc. can cause populations to go through large genetic bottlenecks where they lose a large portion of their genetic diversity. That is very bad for them because, as mentioned earlier, diversity is important for the long-term survival of a population. Gene flow can, however, rescue those populations by resupplying them with some of the diversity that they had lost (Wright 1931, Slatkin 1985, 1987, Whiteley et al. 2015). In other words, if you have two neighboring populations, and one goes through a bottleneck while the other does not, the population that did not go through a bottleneck can rescue the bottlenecked population via gene flow.

Interactions with selection

When gene flow interacts with genetic drift, it is usually beneficial, but when it interacts with selection, it can be very harmful, because it can introduce alleles that aren’t locally adaptive, and those alleles can swamp selection (García-Ramos and Kirkpatrick 1997, Lenormand 2002, Kawecki and Ebert 2004, Foster et al. 2007, Funk et al. 2012). Remember, selection just adapts populations for their current environment. So selection for a give trait may be strong in one population, but weak, non-existent, or even reversed in a neighboring population. As a result, selection may be acting to remove an allele from one population, while gene flow is constantly reintroducing that allele from a neighboring population where the allele is actually beneficial (or at least not harmful).

This shows gene flow's ability to counteract natural selection. In the top panel (A), there is no gene flow, and selection removes one allele from the population as expected. In contrast, in the lower panel (B) there is a high level of gene flow, and selection never manages to remove the harmful recessive allele from the populations.

This shows gene flow’s ability to counteract natural selection. In the top panel (A), there is no gene flow, and selection removes one allele from the population as expected. In contrast, in the lower panel (B) there is a high level of gene flow, and selection never manages to remove the harmful recessive allele from the populations.

I simulated this and have presented the results on the right. The populations (top) were set up in exactly the same way as for my gene flow example, but this time, I added selection. Individuals that had the dominant phenotype had a 100% chance of surviving and reproducing, whereas individuals with the recessive phenotype only had a 90% chance (i.e., there was selection against the recessive allele). As you can see, selection quickly removed the recessive allele (thus the dominant allele had a 100% frequency after only a few generations). If I showed you the survivorship for these populations, you would also see that their survivorship has increased (i.e., they adapted).

Now, let’s do the exact same thing, but let’s have one immigrant per generation, just like in the gene flow example (bottom). As with the gene flow, neither allele ever became fixed. The frequency of the beneficial dominant allele did shift up a bit from the starting point on average, but it never stayed at 100% like it did without gene flow because the gene flow was constantly reintroducing the harmful recessive allele. As a result, these populations could not adapt as well as the populations without gene flow, and they had a lower survivorship. This can be very problematic for populations, especially if they are being faced with something like an introduced disease. It is also a big concern for conservation efforts that move individuals among populations (thus creating artificial gene flow). The flip side of that, however, is that you can, in concept, do targeted genw flow where you take individuals from a population that has a beneficial allele and move those into a population that would benefit from that allele (e.g., take individuals from a population that has adapted to a disease and move them into a population that is still declining from the disease; Kelly and Phillips 2015).


In short, gene flow simply refers to the movement of genetic information among populations. It is an incredibly important evolutionary mechanism because it can prevent populations from splitting apart and it can counteract genetic drift and rescue low diversity populations. However, in some situations it can also counteract natural selection and flood a population with maladaptive alleles. Thus, although it is often beneficial, gene flow can be harmful in some situations.

Related posts

Literature cited (I ran out of time to look up the hyperlinks for all of the citations, sorry. I’ll try to add them in later, but in the mean time you should have no trouble finding them if you just past the article title into Google Scholar)

  • Foster, J. T., B. L. Woodworth, L. E. Eggert, P. J. Hart, D. Palmer, D. C. Duffy, and R. C. Fleischer. 2007. Genetic structure and evolved malaria resistance in Hawaiian honeycreepers. Molecular Ecology 16:4738–4746.
  • Funk, W. C., J. K. McKay, P. A. Hohenlohe, and F. W. Allendorf. 2012. Harnessing genomics for delineating conservation units. Trends in Ecology and Evolution 27:489–496.
  • García-Ramos, G., and M. Kirkpatrick. 1997. Genetic models of adaptation and gene flow in peripheral populations. Evolution 51:21–28.
  • Hedrick, P. W., and S. T. Kalinowski. 2000. Inbreeding depression in conservation biology. Annual Review of Ecology and Systematics 31:139–162.
  • Hughes, A. R., B. D. Inouye, M. T. J. Johnson, N. Underwood, and M. Vellend. 2008. Ecological consequences of genetic diversity. Ecology Letters 11:609–623.
  • Kawecki, T. J., and D. Ebert. 2004. Conceptual issues in local adaptation. Ecology Letters 7:1225–1241.
  • Keller, L. F., and D. M. Waller. 2002. Inbreeding effects in wild populations. Trends in Ecology and Evolution 17:230–241.
  • Kelly, E., and B. L. Phillips. 2015. Targeted gene flow for conservation. Conservation Biology 30:259–267.
  • Lacy, R. C. 1987. Loss of genetic diversity from managed populations: interacting effects of drift, mutation, immigration, selection, and population subdivision. Conservation Biology 1:143–158.
  • Lenormand, T. 2002. Gene flow and the limits to natural selection. Trends in Ecology and Evolution 17:183–189.
  • Reed, D. H., and R. Frankham. 2003. Correlation between fitness and genetic diversity. Conservation Biology 17:230–237.
  • Slatkin, M. 1985. Gene flow in natural populations. Annual Review of Ecology and Systematics 16:393–430.
  • Slatkin, M. 1987. Gene flow and the geographic structure of natural populations. Science 236:787–792.
  • Spielman, D., B. W. Brook, D. A. Briscoe, and R. Frankham. 2004. Does inbreeding and loss of genetic diversity decrease disease resistance? Conservation Genetics 5:439–448.
  • Whiteley, A. R., S. W. Fitzpatrick, W. C. Funk, and D. A. Tallmon. 2015. Genetic rescue to the rescue. Trends in Ecology and Evolution 30:42–49.
  • Whiteman, N. K., K. D. Matson, J. L. Bollmer, and P. G. Parker. 2006. Disease ecology in the Galapagos hawk (Buteo galapagoensis): host genetic diversity, parasite load and natural antibodies. Proceedings of the Royal Society of London B: Biological Sciences 273:797–804.
  • Whitlock, M. C. 2000. Fixation of new alleles and the extinction of small populations: drift load, beneficial alleles, and sexual selection. Evolution 54:1855–1861.
  • Wright, S. 1931. Evolution in Mendelian populations. Genetics 16:97–159.
Posted in Uncategorized | Tagged , , , , | 2 Comments

No, homeopathic remedies can’t “detox” you from exposure to Roundup: Examining Séralini’s latest rat study

Image by Mark Philpott, shared via the Creative Commons license.

Image by Mark Philpott, shared via the Creative Commons license.

One of my main goals for this blog is to help people learn how to evaluate scientific studies. To that end, I have written several posts that dissect papers and explain either why they are robust or why they are untrustworthy (for example, see my posts on Splenda, GMOs, and vaccines). These posts have the dual goals of debunking bad science and helping people think critically, and the time has come for me to write another one of these posts. Earlier this week, someone showed me a recent study which they claimed proved that detoxing is a real thing and there are natural remedies that help your body rid itself of toxins. The study in question is, “Dig1 protects against locomotor and biochemical dysfunctions provoked by Roundup.” As you might imagine, it is less than an exceptional paper. Indeed, it was such a blatantly horrible paper that I thought it would make a good teaching tool to illustrate some of the things that you should watch out for in scientific studies. I’ll summarize the main points below, but I encourage you to read the paper for yourself and see if you can spot the problems with it before you read any further.

I have organized this post in a progression starting with problems that are concerning, but not fatal, then moving to issues that limit the papers conclusions, and ending with problems that completely nullify the paper. I have chosen this order because it is also the progression of knowledge required to spot the problems. Most people should be able to see the red flags that I will start with, so even if you don’t have the statistical knowledge to spot the more technical problems, you can still use those early warning signs as clues that the paper should be scrutinized closely before accepting it.

Authors and conflicts of interest

It is always a good idea to look at both the authors who wrote the paper and the funding sources. Some scientists have reputations for publishing crappy or even fraudulent research, and you should be wary of them. Similarly, financial conflicts of interest should make you more skeptical of a study. Having said that, I want to be absolutely, 100% clear that you cannot claim that a study is wrong simply because of the people who wrote it or their funding sources. Those things are red flags that should make you cautious and should make you look at a paper more closely, but they are not in and of themselves enough to sink a paper (i.e., using them as the basis for outright rejection is an ad hominem/genetic fallacy). Let me put it this way, if I have a study that has some sections that are unclear, but it was written by reputable scientists and did not have any conflicts of interest, then I will probably give the authors the benefit of the doubt. In contrast, if that same paper had been written by notoriously second-rate scientists and/or had serious conflicts of interest, I would be far less willing to give the authors a pass. Another consideration is the general body of literature surrounding the paper. Extraordinary claims require extraordinary evidence, and it is always suspicious when a paper that conflicts with a large body of literature was also written by a fringe scientist and funded by people who stand to benefit from the paper.

Now that all of that has been said, let’s look at the paper itself. The first thing that jumps out is the fact that the final author on this paper is Gilles-Éric Séralini (the last author position is usually reserved for the most senior scientist who was in charge of the overall project). Séralini, for anyone who doesn’t know, is infamous for publishing low-quality, fringe studies in opposition to biotechnology (specifically GMOs). Indeed, he was the author on the infamous rat study which purported to show that GMOs caused cancer in rats, but actually only showed that Séralini doesn’t understand the concept of a proper control. Indeed, the study was so horrible that it was retracted, at which point, Séralini re-published it through a minor and questionable journal that didn’t even bother to send the paper out for review (hardly the actions of a proper scientist).

We aren’t off to a good start, but things get even worse when we look at the funding. The paper is about the supposed benefits of a homeopathic product known as Digeodren (aka Dig1), but it was funded by the company that produces Digeodren (Sevene Pharma). The authors try to get around this by saying, “The authors declare that they have no competing interests. The development of Dig1 by Sevene Pharma was performed completely independently of its assessment,” but that is just a red herring. The fact that the development and testing of Digeodren were separate is completely irrelevant. The point is that the study was funded by the same company that both produces Digeodren and stands to benefit from it. That is, by any reasonable definition, a serious conflict of interest.

Again, to be 100% clear, I am not saying that the study is invalid because it was funded by Sevene Pharma, nor am I saying that it is invalid because it was conducted by Séralini, but, both of those things are serious red flags, and the rest of the study will need to be impeccable if we are going to overlook them.

The journal that published the paper

Another quick and easy thing to look at is the quality of the journal that published the paper. You need to be careful when using this tool, however, because there is plenty of good research that is published in minor journals simply because it is not of a “broad enough scope” or “wide enough impact” to interest major journals. So you need to judge journal quality against the claims being made in a paper. In other words, when a paper is making extraordinary claims but was published in a minor journal, you should be skeptical. As with the authors and conflicts of interest, however, this is not enough to sink a paper, but it is a red flag to watch out for.

So how does our paper do? Well, it is claiming not only that a homeopathic remedy works (more on that in a minute), but also that it can help to remove toxins from your body. Both of those are extraordinary claims that fly in the face of a large body of literature. In other words, if those claims were well supported, then this paper would be of extremely wide interest and should be published in a top journal. Therefore, the fact that it showed up in a fringe journal (BMC Complementary and Alternative Medicine) is yet another warning sign that something is seriously wrong with it.

Extraordinary claims require extraordinary evidence

As alluded to earlier, you should always consider the a prior plausibility of the claims being made in a paper (i.e., how likely are they to be true given the results of other studies). In other words, if a paper is simply reporting something that dozens of other papers have reported, then you don’t need to be too critical (you should still evaluate it, but it requires less scrutiny). In contrast, when a paper is reporting something extraordinary that conflicts with numerous other papers, then the paper needs to present extraordinary evidence to support its claims.

In this case, the claims of the paper are in fact quite extraordinary. First, it is testing a homeopathic remedy. I explained the folly of homeopathy in more detail here, but in short, it relies on the utterly absurd notions that diluting something makes it stronger, like cures like, and water has memory. In other words, homeopathy violates several of our most fundamental scientific concepts. Again, that does not automatically mean that it is wrong because it is always technically possible (albeit very unlikely) that those concepts are in fact flawed. However, if you want to claim that they are flawed, you need to provide some extraordinary evidence, and in the case of homeopathy, that evidence is nowhere to be found. Indeed, systematic reviews of the literature show that homeopathy is nothing more than a placebo. Similarly, detox supplements, shakes, foot baths, etc. are scams. Your body already does a very good job of keeping potentially harmful chemicals at safe levels, and no natural remedies have been shown to actually remove toxins.

Given the weight of evidence against the claims being made by this paper, it would need to be an outstanding study to be convincing. It would need enormous sample sizes, extremely rigorous controls, careful statistics, etc. In other words, it would need to meet an extremely high bar, but as I will demonstrate, it fails to do that.

The importance of the introduction

You can often tell a lot about a paper by its introduction (called the “Background” in this journal). This is where authors are supposed to review the current state of knowledge on the topic of the paper and make the case for why their study is interesting and important. When authors fail to do that convincingly, it is often a sign of underlying problems with the study.

In this case, the introduction is quite short and has several irregularities. First, multiple of the papers that were cited were other Séralini studies. That is not a good sign. There is a lot of other relevant literature out there that should have been included (much of which disagrees with Séralini’s studies). Similarly, several of the cited studies are questionable at best. Indeed, one of their central arguments hangs on a citation to the aforementioned GMO rat study that was so flawed that it was retracted.

Further, the authors cited several relevant papers about the properties of the active ingredients in Digeodren, but they totally failed to mention that Digeodren is a homeopathic remedy and those active ingredients are only present at extremely low concentrations (in this case about 1–10 parts in 100,000). They did mention this briefly in the methods, but its omission from the introduction is extremely troubling. If you are going to do a study on homeopathy, you had dang well better discuss the existing literature on that topic.

Methods: animal trials

Now we finally get to the core of the paper itself, and the first thing that jumps out is the fact that this was an animal trial. As I explained in more detail here, humans have a different physiology than other animals. As such, animal studies have a fairly limited applicability to humans. Therefore, they should be used to identify treatments that are good candidates for human trials, but you cannot jump from saying that something works in rats to saying that it works in humans. To be clear, I am not saying that the results of animal studies are wrong. Indeed, in many cases, the drug in question does in fact work in the species that was being tested, but the fact that it worked in that species does not automatically mean that it will work in humans. As a result, you need to be careful about applying the results of animal studies to humans.

Methods: experimental design

Their experimental design was pretty simple. They took a group of 160 rats and randomly divided them into four groups of 40. One group was kept as a control and did not receive any form of treatment, one group received Roundup in its water, one group received Digeodren in its water, and one group received both Digeodren and Roundup in its water. That’s not a terrible design, but it is also not a great design. A much better approach would have been to include a blocking element.

Imagine, for example, there was a slight thermal gradient in the lab where the rats were housed, and the cage rack containing the control mice ended up being on the warm end, while the cage rack with the Roundup mice ended up being on the cool end. That introduces a new variable and can have dramatic effects on the study. You’d be surprised how much a little thing like that can skew a result. Thus, a much better approach is to do what is known as “blocking.” To do this, instead of having four sets of cages, with each set containing a different group, you have members of each treatment group in each set of cages. In other words, for each set of cages, you randomly select 10 cages from each of your four treatment groups, that way, each set (what we would call a “block”) has 10 individuals from each treatment group (the position of the cages within each block should also be randomized). Now, if there is a thermal gradient (or any other confounding factor), it balances out because it affects all of your treatment groups equally. Further, you can (and should) including that blocking variable in your analyses to actually test for confounding factors across your sets of cages. Failing to block the experiment like that is not always fatal to an experiment (depending on the type of experiment), but it does make me far less confident in the results, and remember, to accept this particular paper, it needs to be an extraordinarily good paper.

This shows three different setups for the same experiment comparing four groups of 40 rats (I am assuming one rat per cage). On the far left, you have what seems to be being described by this study (each experimental group is separate). This is a weak design. A better design is what you see in the middle where you have representatives from each experimental group within each "block." The best design is then to randomize the location of the cages within those blocks (as seen on the far right).

This shows three different setups for the same experiment comparing four groups of 40 rats (I am assuming one rat per cage). On the far left, you have what seems to be being described by this study (each experimental group is separate). This is a weak design. A better design is what you see in the middle where you have representatives from each experimental group within each “block.” The best design is then to randomize the location of the cages within those blocks (as seen on the far right).

A second issue is that this experiment wasn’t blinded. In other words, the researchers knew which rats were in each treatment group. That makes it very easy for their biases to inadvertently influence the experiment, especially given that one of the researchers has a reputation for publishing agenda-driven papers (again, even a slight difference in how the rats were treated could have affected things).

Note: the authors were a bit vague about how their cages were set up, so it is not clear how many rats were in each cage or how many sets of cages there were. However, it is clear that they did not use a proper blocking design.

Methods: The doses

Anytime that you are looking at a toxicology study, you have to look at the doses to see if they are reasonable. Remember, everything (even water) is toxic at a high enough dose. So when a study is looking at an environmental chemical like Roundup, it is important that they use a dose that you would realistically be exposed to in the environment. Otherwise, the study has no real applicability.

In this study, the rats in the Roundup group were given 135mg/kg of Roundup daily. After converting that to a human dose, we find that it is the equivalent of a human consuming 21.9 mg/kg daily. That is an insanely high dose. The exact allowable daily intake (ADI) for glyphosate (i.e. Roundup) varies by country, but it is much lower than that. In Australia, for example, it is 0.3 mg/kg, whereas the WHO sets it as 1 mg/kg. The dose in the experiment is also well above the levels that people are normally exposed to. Even if I want to be generous, and assume the questionable estimates put forth by the “detox project” are correct and people in the US are eating up to 3 mg/kg of glyphosate daily, the dose that the rats received is still seven times that!

To put it simply, this study is worthless because the dose is so unrealistic. Even if the authors had successfully demonstrated that Digeodren did something useful when faced with those levels, that would not in any way shape or form indicate that it does anything useful when exposed to normal levels of Roundup.

Methods: Statistics

Finally, we get to the biggest problem with this study (IMO), and this one would sink it even if it was the only thing wrong with the paper. It is a problem that I write about a lot on this blog, so you may already know where I am going with this. The problem is multiple comparisons. In technical terms, the authors failed to control the family-wise type 1 error rate. In laymen’s terms, this was a statistical fishing trip. They simply did enough comparisons that they eventually got a few that were “significant” just by chance.

I’ve previously written lengthy posts about this, but to be brief, standard statistical tests like what the authors used rely on probabilities for determining statistical significance. In other words, the report “P values” that show you the probability of getting a result of the same effect size or greater than the effect size that you observed if there is not actually an effect. It’s not technically correct, but you can think of this as the probability that you could get your result just by chance. To apply this to our study, they were looking for differences among their groups, so the P values were the probabilities of getting differences as large or greater than the differences that they observed if the treatments don’t actually cause a difference. To actually determine if something is “statistically significant” we compare it to a pre-defined threshold known as “alpha.” In biology, the alpha value is usually 0.05, so any P value less than that is considered significant. What a P value of 0.05 really means, however, is that there is a 5% chance that you could get a difference that large or larger just by chance. This is really important, because it means that you will occasional get “significant” results that arose just by chance, and we call those statistical flukes type 1 errors.

Following all of that, it should make intuitive sense that as you make more comparisons, the odds of getting at least one false positive increase. In other words, if you do enough comparisons, you will eventually find some results that are statistically significant just by chance. So your error rate across all of your tests is actually much higher than 0.05. This is what we call the family-wise type 1 error rate, and it is extremely important. To compensate for it, you should do two things. First, at the outset of your study, you should have a clear prediction of what you expect to be true if your hypothesis is correct, and you should only make the comparisons that are necessary for testing that predication. You should not make a whole bunch of haphazard comparisons and hope that something significant comes out. Second, if you end up using multiple tests to answer the same question (e.g., does drug X work?) then you need to control the family-wise error rate by adjusting your alpha value (this is usually done through a Bonferroni correction). In its simplest terms, this makes the alpha more stringent as you increase the number of comparisons that you do.

So, how did our intrepid scientists do? In short, not well. They made a whopping 29 comparisons, only 8 of which showed any form of significance, and only 6 of which showed significance in a direction that would suggest that Digeodren does anything useful. Further, they did not control the error rate among these tests. In other words, they did exactly the opposite of what you are supposed to do. They went on a fishing trip looking for significance rather than only testing a small set of pre-defined expectations. They made so many comparisons that they got some statistically significant results just by chance. To put this another way, if I set up the exact same experiment with four groups of rats, but I did not give any of them Digeodren or Roundup, and I made the same 29 comparisons among those four groups, I would expect to get several significant results, even though I treated all four groups exactly the same. Their results are statistical flukes, nothing more.

Finally, they did not report their P values for each comparison, which means that we can’t even properly assess their results (see the note below). If they had reported a table of P values like they should have, we could do the Bonferroni correction ourselves, but since they failed to do that, we have nothing to go on.

To be clear, in most cases, the fact that an author did not control their error rate would not automatically mean that their results were statistical flukes, but it would mean that we should consider their paper untrustworthy and reject it. However, in this particular case, there is another important factor to consider. Namely, all of the existing evidence that homeopathy doesn’t work. When you consider that evidence, and the low quality of the experimental design of this particular study, the most rational conclusion is that the results are wrong rather than simply untrustworthy.

Note: If you read the paper, you will see a reference to a Bonferroni test as well as P values, but they only used those within a test rather than across tests. In other words, the tests that they were using (ANOVA and Kruskal-Wallis) make comparisons among several groups (in this case the four treatment groups) and report a single P value that tells you whether or not at least one significant difference exists among those groups. Then, if you get a significant result, you make pairwise comparisons among all of your groups and get individual P values for each comparison. So they reported the P values and controlled the error rates for those individual comparisons within each ANOVA, but I am talking about the P values across ANOVAs, because you should never even do the individual comparisons unless the ANOVA itself is significant, and if you don’t control the error rate across ANOVAs (as they didn’t), a lot of your ANOVAs will be false positives. In other words, they did 29 ANOVAs/Kruskal-Wallis tests, each of which compared four groups, and they controlled the error rates for the post-hoc comparisons of the four groups, but not for the ANOVAs themselves.


In summary, this paper is riddled with problems and is little more than a steaming pile of crap. It had major conflicts of interest, was written by an author with a reputation for publishing shoddy, agenda-driven studies, it was published in a fringe journal, it made inadequate references to the relevant literature, the experimental design was sub-par and failed to incorporate blinding procedures, and (most importantly) it made an astounding 29 comparisons without bothering to control the error rate. This paper is a statistical fishing trip. The authors simply made so many comparisons that they eventually got a few that were significant just by chance. This is a common tactic that is frequently employed by pseudoscientists (and sometimes legitimate researchers as well) and you should learn to spot it.    

Posted in Nature of Science | Tagged , , , , , | 13 Comments

Don’t mistake an assumption for a fact

Carl Sagan quote extraordinary evidence claimsI want you to imagine for a minute that you have been selected for jury duty, and you are sitting in the courtroom listening to the evidence. As you watch, the prosecution calls an expert forensic scientist to the stand, and they carefully explain the facts concerning the forensic evidence, all of which point towards the defendant’s guilt. Then, the defendant’s lawyer stands up and shouts, “Objection, this witness has been paid off to lie about my client!” After a moment of shocked silence, the judge says, “That is quite a claim. What evidence do you have to support it?” The lawyer then responds simply by saying, “only someone who had been bought off would say things like that against my client.” Now, what do you think is going to happen next? Is that a reasonable defense that the judge will accept? Obviously not! The defense is making an extraordinary assumption, and it is clearly invalid to do so.

In an example like that, the problem is obvious. You can’t just make things up to dismiss facts that are inconvenient for you. Indeed, when a person’s fate hangs in the balance, we all want the arguments and evidence to be based on facts, not assumptions. Nevertheless, when it comes to science and many aspects of our daily lives (such as politics), people are often more than happy to accept assumptions, and people frequently state them as if they were facts. Therefore, I am going to provide several examples of this flawed line of reasoning, and explain why it not only doesn’t work, but often commits a logical fallacy.

The first example is probably the most common one that I encounter, and it is highly analogous to my courtroom example. I recently engaged in a Facebook debate with someone who made the bold claim that there is no empirical evidence for climate change. I responded to that assertion by providing him with multiple papers that did in fact provide empirical evidence for climate change. How do you think he responded? You probably guessed it. He simply claimed that all of the scientists involved in those studies had been paid off (just like my fictional lawyer did). That argument was clearly fallacious in my courtroom example, and it is equally flawed here. You can’t just assume that scientists have been paid off any time that they publish a result that you don’t like. You have to provide actual evidence of corruption, otherwise you are making a baseless assumption. In fact, journals actually require scientists to declare any conflicts of interests, and failure to do so is a serious offense (it can result in loss of funding, papers being retracted, etc.). So there is no need to make assumptions about conflicts of interest, because you can easily check and see if they exist. Nevertheless, this is one of the most common arguments among those who reject the results of modern science, and I constantly encounter it among anti-vaccers, those who oppose GMOs, climate change deniers, people who use alternative medicine and/or reject modern medicine, etc.

A very closely related line of attack is simply to accuse the person that you are debating of being a shill, astroturfer, etc.  Indeed, I constantly have to deal with people accusing me of having been bought off by big pharma, Monsanto, the government, lizard people, etc. In every case, however, it is a complete assumption (and a faulty one, as I actually pay money out of my own pocket to maintain this blog). You can’t just go around assuming that everyone who disagrees with you has been bought off. That’s not rational.

I could continue giving copious examples of this type of reasoning (e.g. assuming that “big pharma” has bought of the FDA and CDC), but I think that you get the basic picture, so I want to move on to explaining a bit more about why this type of argument is invalid. There are actually two logically fallacies being committed here. First, all of the specific examples that I have given so far commit a genetic fallacy (sometimes specifically an ad hominem fallacy). These fallacies occur when you make an irrelevant or unjustified attack against the source of the information, rather than addressing the information itself (even a flawed source can sometimes be correct). I have previously talked about these fallacies at length (see previous hyperlinks), so I want to focus instead on the second fallacy. This one is more general and can occur even when a source is not being attacked. It is what is known as an ad hoc fallacy. It generally occurs as a response to an argument, rather than an argument itself, and it has the defining characteristic of proposing a solution that you would never accept unless you were already convinced by the claim that was being defended. In other words, an ad hoc fallacy is a response that is not based on evidence and whose sole purpose is to “solves” a problem in an argument that you are fond of.

That may have seemed a bit complicated, so let me explain by applying it to one of the examples that I used previously. In the example of the climate change denier, his claim was that there is no empirical evidence of climate change (that is the position that he was defending). I presented him with actual evidence, thus clearly contradicting his claim. He then responded to that evidence by asserting that all of those scientists were really paid shills. However, there is no evidence to support that assertion, and there is no reason to think that the assertion is true other than a desire to believe the initial claim. So he was simply assuming that the scientists had been paid off, even though he had no evidence to support that claim. Do you see how this works? It is an assumption that is stated as if it was a fact, but there is no actual reason to think that the assumption is true. It serves no function other than patching the hole in the argument, and you wouldn’t accept it unless you were already convinced that the initial claim was true.

Nevertheless, not all assumptions are ad hoc fallacies, but that doesn’t make them any less flawed when they are used in debates. You simply cannot state assumptions as if they are facts. The hidden cancer cure conspiracies provide a great example of these, because they involve assumptions stacked on other assumptions. In order to argue that pharmaceutical companies are hiding a cure for cancer, first have to assume that a cure actually exists, but that is a meritless assumption. Can you provide me with actual evidence that pharmaceutical companies have a cure hidden on their shelves? No, you can’t, which means that you don’t get to state the claim that they have a cure as if that claim is an evidence-based fact. Similarly, you also have to assume a rather large conspiracy in which everyone involved is willing to sit back and watch their friends and family members die. This conspiracy would have to involve not just the CEOs of companies, but also people on the ethics committee who approved the research, people in the FDA who approve the trials, accountants who handled the budgets, lab techs, the scientists themselves, etc. That’s an awful lot of people to be involved, and you cannot just go around inventing conspiracies like that without providing actual hard evidence of their existence. This conspiracy theory (and the vast majority of conspiracy theories) relies heavily on unsubstantiated assumptions, which is a big part of why it is irrational.

Indeed, if it was permissible to substitute assumptions for facts, then we could all do this anytime that we wanted.  I could, for example, claim that smoking actually has health benefits, and the tobacco industry scientists were actually correct about the safety of cigarettes, but big pharma paid off a ton of scientists to falsify data showing that cigarettes were dangerous because big pharma wanted to profit off of sick people. Now, you probably think, “that’s ridiculous!” and indeed it is, but it is no more ridiculous than claiming that pharmaceutical companies paid off scientists to falsify information on vaccines, claiming that marijuana is only illegal because legalizing it would hurt pharmaceutical companies, claiming that climate scientists are falsifying data, claiming that Monsanto controls the world’s agriculture scientists, etc. All of these are baseless assumptions, and you cannot present them as if they are facts.

As I approach the end of this post, I want to remind everyone of one of the most important concepts in debates and rational thinking: the burden of proof. This states that the person making the claim is always required to provide legitimate evidence to support it. In other words, if you want to claim that scientists have been paid off, then it is your duty to provide actual evidence to support that claim, and if you cannot do that, then you are stating an assumption, not a fact, and your argument is illegitimate. Similarly, if you want to claim that companies are hiding cures, a conspiracy is afoot, etc., you must provide evidence to substantiate those claims. You simply cannot assume things that haven’t been verified, because if you could, then we could all dismiss every single argument that we don’t like simply by assuming the existence of some contrary evidence. Also, it is worth explicitly stating that you have to show the evidence, not the other way around. In other words, if you are claiming that a conspiracy exists, you have to provide evidence that it exists, whereas I do not have to provide evidence that it doesn’t exist. That’s the way that the burden of proof works. In fact, saying “you can’t prove that it doesn’t exist, therefore it is valid to think that it does exist” is a logical fallacy known as an argument from ignorance.

Finally, although I have been speaking specifically in the context of debates, you really should apply this to your own views (regardless of which side of a topic you stand on). Think through your arguments and make sure that you can provide proper evidence from high-quality sources to back up each component of your reasoning. Seek out your assumptions, test them against the evidence, and if they cannot be verified or strongly supported, reject them. Admitting that, “I don’t know” is far, far better than insisting that you do know, when in fact you are simply making an assumption.

Note: I fully expect someone to read this and say, “but we use assumptions all of the time in our daily lives and they are useful.” That is true, but it is irrelevant for several reasons. First, the assumptions that we make in our daily lives are very often flawed and frequently get us into trouble. They are, in fact, unreliable, which is why they aren’t valid substitutes for evidence and can’t be used in scientific debates. Second, many of the things in daily life that we often call “assumptions” are actually evidence based conclusions. For example, if my PhD adviser emails me and asks me to meet him in his office at 3:00, I am going to “assume” that he means 3:00PM, but that is not really an assumption (at least not in the way that I have been using the word in this post). Rather, it is a probabilistic conclusion based on evidence. I say that, because every single time that my adviser has ever asked to meet with me, it has been a meeting during daylight hours. Thus, there is no rational reason to think that he actually means 3:00AM. Situations like that are clearly extremely different from creating conspiracy theories or assuming corruption just because you don’t like what a scientist published.

Posted in Uncategorized | Tagged , , , , , , , , , | 7 Comments

Debunking the creationist myth that mutations don’t produce new and useful information

Genetic variation is a fundamental requirement for evolution, but many of the evolutionary mechanism (such as selection and genetic drift) actually remove variation from populations. Therefore, evolution is entirely reliant on the formation of new genetic information, and without it, evolution would grind to a halt. Creationists often seize this fact, and erroneously claim that evolution is impossible because we have never found a mechanism that is capable of creating new genetic information. This claim is, however, completely false, because mutations do, in fact, create new genetic information. Nevertheless, many creationists respond to that fact by insisting that mutations simply “rearrange” existing genetic information, rather than creating “new” information. Therefore, I want to briefly explain why this argument is fundamentally flawed.

Hopefully everyone recognizes this, but if not, it's Mr. DNA from Jurassic Park

Hopefully everyone recognizes this, but if not, it’s Mr. DNA from Jurassic Park

Before I can explain the problems with this argument, we need to be clear about the basics, and you need to have at least a rudimentary understanding about how DNA works. DNA (or deoxyribonucleic acid if you prefer) consists of four bases adenine (A) guanine (G) thymine (T) and cytosine (C). These four bases get arranged into groups of three, and each group of three codes for an amino acid. Those amino acids then get strung together to form a protein, and those proteins combine to form tissues. Thus, your DNA is your body’s blue-print, and it tells your body which amino acids to make, and how to combine them to make proteins, tissues, organs, etc.

Please note that although what I have presented here is the most fundamental concept for you to understand, actual DNA also includes stop codons, start codons, and many other complexities that are irrelevant for this post.

 Now that you understand the basics of DNA, we can talk about mutations. I explained them in far greater detail here, but in short, they are simply random changes to the genetic code (the ones that are important for evolution usually occur during the formation of egg and sperm cells). There are many different types of mutations such as inversions (which flip a segment of DNA), deletions (which remove base pairs), insertions (which add extra base pairs), substitutions (which insert the wrong base into a chain [e.g. a T instead of a C]), duplications (which duplicate a segment of DNA), and several others. In every case, however, they change the genetic code, and by changing the code, they can change the amino acids and ultimately the proteins that are produced (note: there is redundancy in the way that amino acids are coded, so not all mutations result in changes downstream).

At this point, we can examine creationists’ claim that mutations simply rearrange information rather than producing new information. The most obvious problem with this is simply that some mutations (like insertions) do actually insert entirely new base pairs. In other words, they don’t rearrange the code that is already there. Rather, they add new bases (i.e., new information) to that code.

The second problem is really the more important and fundamental one. Namely, this argument seems to be using the word “new” in a rather peculiar sense, because rearranging the existing bases does in fact produce a new code that often results in the production of different amino acids and new proteins. Let me illustrate. A coding strand of DNA that contains the sequence CTT would code for the amino acid Leucine. However, if a mutation rearranged those three bases so that they were TCT, that strand of DNA would code for the amino acid Serine instead of Leucine. Thus, by simply rearranging the existing bases, we created new information which produced a different amino acid. In other words, we added genetic variation to the population, because the individuals who receive that mutation will produce Serine, while the result of the population is producing Leucine. Objecting to that by claiming that “no new information has been created” is really quite silly because the fact remains that individuals with that mutation are producing a different amino acid than everyone else. Something that codes for the production of an entirely different amino acid is, by any reasonable definition, “new information” (i.e., it is information that was not there before).

A useful analogy to help you conceptualize this is to think about letters in the alphabet. The English language has 26 letters (bases), and we combine those letters to from words (amino acids). We then arrange those words into sentences (proteins), and we arrange those sentences into paragraphs (tissues). Ultimately, we can use those paragraphs to make books, essays, etc. (organisms). Now, according to creationists’ reasoning, it should be impossible to make any new information by simply rearranging those 26 letters, but that is clearly absurd. We can arrange them one way and produce the works of Shakespeare. We can arrange them another way and produce “The Origin of the Species.” We can rearrange them yet again and produce the script to a Stargate episode. Indeed, there are a virtually infinite set of possibilities, each of which contains different information, and the situation is no different for DNA. We can arrange the bases one way and get a dinosaur, and we can rearrange that code (via mutations) and get a chicken. We can rearrange it yet again and get a whale, human, tree, bacteria, mushroom, etc. This notion that making new arrangements of the four existing bases doesn’t produce new information is absurd because everyone agrees that different arrangements of those bases produce very different organisms.

A big part of the problem here once again comes back to the definition of the word “new.” Creationists seem to think that evolution requires something that is completely and totally novel, such as a new base pair or, at the very least, an entirely new amino acid that has never existed anywhere before, but that is a straw man fallacy. Evolution does not require something that has never previously existed anywhere. Rather, it simply needs to have variation. Thus, any change to the genetic code is “new information” in an evolutionary sense, because it provides variation. A useful way to think about this is that evolution doesn’t need “new” information. Rather, it needs “different” information. In other words, all that it needs is a code that is different than the one that was there before.

To further illustrate what I mean by this, it is worth mentioning that even mutations that remove bases can actually produce new information in an evolutionary sense. Going back to the alphabet example, imagine that the document in question is a recipe, and imagine that the instructions get “mutated” by the random deletion of one of the steps. That actually provides new information because the end-product will be something different than the intended product (i.e., there will be variation for the trait). The same thing happens with DNA. If you remove the bases for a particular amino acid, then the final protein product will often be different from the one that was originally coded for, and that variation is all that evolution needs, regardless of whether or not you want to describe it as “new.” To give an actual example of this, the virus that causes HIV typically uses the CKR5 protein to enter macrophages and complete its lifecycle. However, people who have deletions (a type of mutation that removes bases) on both copies of their genes for CKR5 are resistant to HIV because those mutations alter the protein, and HIV cannot bind to this new, modified protein (Dean et al. 1996; Sullivan et al. 2001). Thus, a loss of genetic material results in a new set of instructions, which causes cells to produce a protein with a new modification, and that new modification provides a beneficial function. Creationists may try to say that this example doesn’t illustrate the formation of “new” information because an existing protein was simply modified, but that is, once again, a straw man fallacy, because evolution just needs variation, and this mutation provides that (thus, from an evolutionary perspective, it is new information).

Finally, you might be tempted to protest to all of this on the grounds that, “mutations are nearly always harmful,” but that claim is a myth. Most mutations are actually neutral at the time that they occur (i.e., they are neither beneficial nor harmful; Nachman and Crowell 2000; Eyre-Walker et al. 2007). Further, although harmful mutations certainly do exist, they are selected against, so they are really irrelevant. In other words, even if there were 10,000 times as many harmful mutations as beneficial mutations (which there aren’t), that wouldn’t preclude evolution, because nature would select against the harmful ones and for the beneficial ones. On that note, I want to be absolutely, 100% clear that beneficial mutations do exist and have been well documented (Newcomb et al. 1997; Dean et al. 1996; Sullivan et al. 2001; Shaw et al. 2002, 2003; Joseph and Hall 2004; Perfeito et al. 2007; see Halligan and Keightley 2009 for a good review). Indeed, we have done experiments with bacteria were we monitored populations for many generations, and observed the formation of novel, beneficial mutations that provided the bacteria with new information that allowed them to perform a novel function that they were previously incapable of (Blount et al. 2008; more detailed explanation in the citrate section of this post).

From the Star Trek TOS episode "I, Mudd"

From the Star Trek TOS episode “I, Mudd”

In short, this notion that mutations can’t produce new genetic information is laughably absurd. DNA is simply the code that tells organisms what proteins and structures to make, and modifications to that code result in novel modifications, proteins and structures. Indeed, if you accept that dogs and cats are different because of differences in their DNA, then you have already accepted that rearranging genetic codes can in fact produce new information and result in vastly different organisms. Indeed, saying that rearranging DNA can’t produce new information is no different from saying that rearranging the letters of the alphabet can’t produce new information. Obviously it can, as is evidenced by this new blog post that I am writing. Additionally, when we say that evolution needs “new information,” we don’t mean that it needs something completely novel and totally different from anything else that has ever existed. Rather, we simply mean that there needs to be variation for traits. In other words, any slight modification to an existing trait qualifies as “new information” when talking about evolution. Finally, we have experimentally documented that mutations can produce that variation, and we have demonstrated that beneficial mutations do in fact occur and result in novel proteins that perform novel functions (i.e., the mutations created new information).

Literature Cited

Posted in Science of Evolution | Tagged , , , , | 12 Comments