Cherry picking explained

Cherry picking, suppressing evidence, or the fallacy of incomplete evidence is the act of pointing to individual cases or data that seem to confirm a particular position while ignoring a significant portion of related and similar cases or data that may contradict that position. Cherry picking may be committed intentionally or unintentionally.[1]

The term is based on the perceived process of harvesting fruit, such as cherries. The picker would be expected to select only the ripest and healthiest fruits. An observer who sees only the selected fruit may thus wrongly conclude that most, or even all, of the tree's fruit is in a likewise good condition. This can also give a false impression of the quality of the fruit (since it is only a sample and is not a representative sample). A concept sometimes confused with cherry picking is the idea of gathering only the fruit that is easy to harvest, while ignoring other fruit that is higher up on the tree and thus more difficult to obtain (see low-hanging fruit).

Cherry picking has a negative connotation as the practice neglects, overlooks or directly suppresses evidence that could lead to a complete picture.

Cherry picking can be found in many logical fallacies. For example, the "fallacy of anecdotal evidence" tends to overlook large amounts of data in favor of that known personally, "selective use of evidence" rejects material unfavorable to an argument, while a false dichotomy picks only two options when more are available. Some scholars classify cherry-picking as a fallacy of selective attention, the most common example of which is the confirmation bias.[2] Cherry picking can refer to the selection of data or data sets so a study or survey will give desired, predictable results which may be misleading or even completely contrary to reality.[3]

History

A story about the 5th century BCE atheist philosopher Diagoras of Melos says how, when shown the votive gifts of people who had supposedly escaped death by shipwreck by praying to gods, he pointed out that many people had died at sea in spite of their prayers, yet these cases were not likewise commemorated[4] (this is an example of survivorship bias). Michel de Montaigne (1533–1592) in his essay on prophecies comments on people willing to believe in the validity of supposed seers:

In science

Cherry picking is one of the epistemological characteristics of denialism and widely used by different science denialists to seemingly contradict scientific findings. For example, it is used in climate change denial, evolution denial by creationists, denial of the negative health effects of consuming tobacco products and passive smoking.

In medicine

In a 2002 study, a review of previous medical data found cherry picking in tests of anti-depression medication:

[researchers] reviewed 31 antidepressant efficacy trials to identify the primary exclusion criteria used in determining eligibility for participation. Their findings suggest that patients in current antidepressant trials represent only a minority of patients treated in routine clinical practice for depression. Excluding potential clinical trial subjects with certain profiles means that the ability to generalize the results of antidepressant efficacy trials lacks empirical support, according to the authors.[5]

In argumentation

In argumentation, the practice of "quote mining" is a form of cherry picking, in which the debater selectively picks some quotes supporting a position (or exaggerating an opposing position) while ignoring those that moderate the original quote or put it into a different context. Cherry picking in debates is a large problem as the facts themselves are true but need to be put in context. Because research cannot be done live and is often untimely, cherry-picked facts or quotes usually stick in the public mainstream and, even when corrected, lead to widespread misrepresentation of groups targeted.

One-sided argument

A one-sided argument (also known as card stacking, stacking the deck, ignoring the counterevidence, slanting, and suppressed evidence)[6] is an informal fallacy that occurs when only the reasons supporting a proposition are supplied, while all reasons opposing it are omitted.

Philosophy professor Peter Suber has written:

The one-sidedness fallacy does not make an argument invalid. It may not even make the argument unsound. The fallacy consists in persuading readers, and perhaps ourselves, that we have said enough to tilt the scale of evidence and therefore enough to justify a judgment. If we have been one-sided, though, then we haven't yet said enough to justify a judgment. The arguments on the other side may be stronger than our own. We won't know until we examine them. So the one-sidedness fallacy doesn't mean that your premises are false or irrelevant, only that they are incomplete.[7]

With rational messages, you need to decide if you want to use a one-sided argument or a two-sided argument. A one-sided argument presents only the pro side of the argument, while a two-sided argument presents both sides. Which one you use will depend on which one meets your needs and the type of audience. Generally, one-sided arguments are better with audiences already favorable to your message. Two-sided arguments are best with audiences who are opposed to your argument, are better educated or have already been exposed to counter arguments.

Card stacking is a propaganda technique that seeks to manipulate audience perception of an issue by emphasizing one side and repressing another.[8] Such emphasis may be achieved through media bias or the use of one-sided testimonials, or by simply censoring the voices of critics. The technique is commonly used in persuasive speeches by political candidates to discredit their opponents and to make themselves seem more worthy.[9]

The term originates from the magician's gimmick of "stacking the deck", which involves presenting a deck of cards that appears to have been randomly shuffled but which is, in fact, 'stacked' in a specific order. The magician knows the order and is able to control the outcome of the trick. In poker, cards can be stacked so that certain hands are dealt to certain players.[10]

The phenomenon can be applied to any subject and has wide applications. Whenever a broad spectrum of information exists, appearances can be rigged by highlighting some facts and ignoring others. Card stacking can be a tool of advocacy groups or of those groups with specific agendas.[11] For example, an enlistment poster might focus upon an impressive picture, with words such as "travel" and "adventure", while placing the words, "enlist for two to four years" at the bottom in a smaller and less noticeable point size.[12]

Notes and References

  1. Web site: Klass. Gary. Just Plain Data Analysis: Common Statistical Fallacies in Analyses of Social Indicator Data. Department of Politics and Government, Illinois State University. statlit.org. ~2008. March 25, 2014. dead. https://web.archive.org/web/20140325173711/http://www.statlit.org/pdf/2008KlassASA.pdf. March 25, 2014.
  2. Web site: Fallacies | Internet Encyclopedia of Philosophy.
  3. Book: Goldacre, Ben. Ben Goldacre. Bad Science. 2008. HarperCollins Publishers. 978-0-00-728319-4. 97–99.
  4. Book: Hecht, Jennifer Michael. Jennifer Michael Hecht. Doubt: A History. Harper San Francisco. 2003. Whatever Happened to Zeus and Hera?, 600 BCE–1 CE. 9–10. 0-06-009795-7. https://books.google.com/books?id=QJb16_AAePkC&pg=PA10.
  5. Typical Depression Patients Excluded from Drug Trials; exclusion criteria: is it 'cherry pickin'? . The Brown University Psychopharmacology Update . Wiley Periodicals . 13 . 5 . May 2002 . 1–3. 1068-5308. Based on the studies:
    • Posternak . MA . Zimmerman . M . Keitner . GI . Miller . IW . A reevaluation of the exclusion criteria used in antidepressant efficacy trials . The American Journal of Psychiatry . February 2002 . 159 . 2 . 191–200 . 11823258 . 10.1176/appi.ajp.159.2.191.
    • Zimmerman . M . Mattia . JI . Posternak . MA . Are subjects in pharmacological treatment trials of depression representative of patients in routine clinical practice? . The American Journal of Psychiatry . March 2002 . 159 . 3 . 469–73 . 11870014 . 10.1176/appi.ajp.159.3.469.
  6. Web site: One-Sidedness - The Fallacy Files. 14 October 2014.
  7. Web site: The One-Sidedness Fallacy . Peter Suber . 25 September 2012.
  8. Book: The fine art of propaganda: a study of Father Coughlin's s=Institute for Propaganda Analysis . 1939 . Harcourt Brace and Company . 95–101 . November 24, 2010.
  9. Book: C. S. Kim, John . The art of creative critical thinking . 1993 . University Press of America . 9780819188472. 317–318 . November 24, 2010.
  10. Book: Ruchlis, Hyman . Clear thinking: a practical introduction . Sandra Oddo . 1990 . Prometheus Books . 9780879755942. 195–196 . November 24, 2010.
  11. Book: James, Walene . Immunization: the reality behind the myth, Volume 3 . 1995 . Greenwood Publishing Group . 9780897893596. 193–194 . November 24, 2010.
  12. Book: Shabo, Magedah . Techniques of Propaganda and Persuasion . 2008 . Prestwick House Inc . 9781580498746. 24–29 . November 24, 2010.