The illusion of explanatory depth (IOED) is cognitive bias or an illusion where people tend to believe they understand a topic better than they actually do.[1] The term was coined by Yale researchers Leonid Rozenblit and Frank Keil in 2002.[1] [2] The effect was observed in only one type of knowledge called explanatory knowledge, in this case defined as "knowledge that involves complex causal patterns" (see causal reasoning). The effect has not been observed in procedural, narrative, or factual (descriptive) knowledge.[3] [4] Evidence of the IOED occurring has been found in everyday mechanical and electrical devices such as bicycles, in addition to mental disorders, natural phenomena, folk theories, and politics, with the most studied effect of IOED being in politics in the form of political polarization.[5]
The illusion is related to the Dunning–Kruger effect, differing in that the IOED examines explanatory knowledge as opposed to ability.[6] Limited evidence exists suggesting that the effects of the IOED are less significant in subject matter experts,[7] but it is believed to affect almost everyone, compared to the Dunning–Kruger effect which is usually defined to apply only to those of low to moderate competence.[8] The IOED is more significant for historical knowledge, in cases when knowing about the topic is perceived as socially desirable.[9]
Another description of the IOED is that "we mistake our familiarity with a situation for an understanding of how it works".[10] IOED has also been suggested to explain the perception that psychology as a field is "simple" or "obvious".
There is evidence to support the theory that the IOED is a contributing factor to increased political polarization in the United States.[11] A 2018 study with participants recruited in the context of the 2016 United States presidential election found that higher levels of IOED about political topics is associated with increased support in conspiracy theories.[12]
It is thought that the effects of IOED, especially in politics, can be reduced by asking people to explain the topic rather than only asking people to provide reasons for their beliefs. The specific ways in which people are asked to explain the topic are important, as they may backfire. This was found in research that showed when people are asked to "justify their position", people's beliefs become more extreme. Asking for "reasons" may lead people to strengthen their beliefs by selectively thinking of support for their position, while asking for "explanations" may lead them to confront their lack of knowledge.
The term for IOED was coined by Yale researchers Leonid Rozenblit and Frank Keil in 2002. One inspiration for the IOED concept was research in change blindness suggesting at the time that people grossly overestimated their own spatial memory.[13]
In the experiment they conducted with 16 Yale undergraduate students, they asked them to rate their understanding of devices and simple items. They were then asked to generate a detailed explanation of how they worked and then rerate their understanding of that item. Consistently, ratings were lower after generating an explanation, suggesting they then began to understand that they lacked understanding of that item after attempting to explain. Rozenblit and Keil concluded that having to explain basic concepts or mechanisms, confronts people with the reality that they may not understand the subject as much as they think they do.