Skip to main content

Epistemic Benefits of Delusions (1)


This is the first in a series of two posts by Phil Corlett (pictured above) and Sarah Fineberg (pictured below). Phil and Sarah are both based in the Department of Psychiatry at Yale University. In this post and the next they discuss the adaptive value of delusional beliefs via their predictive coding model of the mind, and the potential delusions have for epistemic benefits (see their recent paper 'The Doxastic Shear Pin: Delusions as Errors of Learning and Memory', in Cognitive Neuropsychiatry). Phil presented a version of the arguments below at the Royal College of Psychiatrists' Annual Meeting in Birmingham in 2015, as part of a session on delusions sponsored by project PERFECT.


The predictive coding model of mind and brain function and dysfunction seems to be committed to veracity; at its heart is an error correcting, plastic, learning mechanism that ought to maximize future rewards, and minimize future punishments like the agents of traditional microeconomics—so called econs (Padoa-Schioppa forthcoming). This seems at odds with predictive coding models of psychopathology and in particular psychotic symptoms like hallucinations and delusions (Corlett et al. 2010). Put simply, if delusions result from a noisy maladaptive learning mechanism, why do individuals learn anything at all—let alone the complex and strongly held beliefs that characterize psychotic illness? We know from behavioural economists like Kahneman, Tversky, and, Thaler that humans can depart from econ-like responding. Can predictive coding depart likewise? And does it depart in interesting ways that are relevant to delusions?

We think so. Bayesian models of cognition and behaviour need not necessarily optimize expected value. For example, Bayesian models of message passing in crowds can recapitulate the rumors and panic that characterize communication after a salient world event (Butts 1998). With regards to delusions, we would like to re-consider Daniel Dennett and Ryan McKay’s assessment of adaptive mis-beliefs (McKay and Dennett 2009). McKay and Dennett explored the existence of misbeliefs—incorrect beliefs that, despite being wrong, nevertheless confer some advantage on the adherent. They argued that only positive illusions—beliefs that one is more competent, more attractive, less biased than in reality, etc.—were evolved, adaptive, misbeliefs (McKay and Dennett 2009). We (and others) think delusions might confer such a function (Hagen 2008).

McKay and Dennett introduced an interesting concept to which we return presently—the doxastic shear-pin (McKay and Dennett 2009). A shear-pin is a concept from engineering. Shear-pins are built into systems in order to disable a machine in trouble so that continued functioning does not destroy the whole machine (an electrical fuse functions similarly). A broken shear-pin allows a machine to continue to function, albeit at an attenuated level. We argue that delusions form when a doxastic shear-pin breaks. The shear-pin has much in common with psychological defenses and biases—ways of believing that perhaps to do not fully reflect reality but rather the reality that an individual desires (McKay and Dennett 2009).

This sounds maladaptive, but there are situations in which wrong beliefs (like overconfidence in ones abilities) can confer an adaptive advantage e.g. when the benefits of contested resources outweigh the costs of competition (Johnson and Fowler 2011). We argue that a doxastic shear pin breaking to produce delusions is not adaptive in this sense (although we note that people with attenuated odd beliefs may be more fecund (Nettle and Clegg 2006)), but rather it allows some continued instrumental engagement with the world (and exploitation of resources) rather than no action at all (Mishara 2009).

We explain the shear-pin in terms of aberrant reinforcement learning and memory mechanisms. These mechanisms are themselves normally adaptive, allowing organisms to exploit the reward contingencies in their environment and respond with flexibility when those contingencies change. Over time (and with repeated exposure) this normally plastic learning mechanism gives way to an inflexible belief, akin to a stimulus response habit that, through overtraining, becomes resistant to contradictory evidence. In the context of the learning model, delusions are adaptive misbeliefs because they provide a means through which patients can continue to engage with their environment (Mishara 2009).

These ideas—of the doxastic shear-pin—raise the notion of the epistemic innocence of delusions; whether some delusions are imperfect cognitions that can be wrong but may still confer some epistemic benefit (Bortolotti 2015). We argue that at least some cases of delusion are epistemically innocent—because they provide an explanation for ineffable, salient, and frightening experiences (Corlett et al. 2010). In our next post we will outline our argument for this claim.

Popular posts from this blog

Delusions in the DSM 5

This post is by Lisa Bortolotti. How has the definition of delusions changed in the DSM 5? Here are some first impressions. In the DSM-IV (Glossary) delusions were defined as follows: Delusion. A false belief based on incorrect inference about external reality that is firmly sustained despite what almost everyone else believes and despite what constitutes incontrovertible and obvious proof or evidence to the contrary. The belief is not one ordinarily accepted by other members of the person's culture or subculture (e.g., it is not an article of religious faith). When a false belief involves a value judgment, it is regarded as a delusion only when the judgment is so extreme as to defy credibility.

Rationalization: Why your intelligence, vigilance and expertise probably don't protect you

Today's post is by Jonathan Ellis , Associate Professor of Philosophy and Director of the Center for Public Philosophy at the University of California, Santa Cruz, and Eric Schwitzgebel , Professor of Philosophy at the University of California, Riverside. This is the first in a two-part contribution on their paper "Rationalization in Moral and Philosophical thought" in Moral Inferences , eds. J. F. Bonnefon and B. Trémolière (Psychology Press, 2017). We’ve all been there. You’re arguing with someone – about politics, or a policy at work, or about whose turn it is to do the dishes – and they keep finding all kinds of self-serving justifications for their view. When one of their arguments is defeated, rather than rethinking their position they just leap to another argument, then maybe another. They’re rationalizing –coming up with convenient defenses for what they want to believe, rather than responding even-handedly to the points you're making. Yo

A co-citation analysis of cross-disciplinarity in the empirically-informed philosophy of mind

Today's post is by  Karen Yan (National Yang Ming Chiao Tung University) on her recent paper (co-authored with Chuan-Ya Liao), " A co-citation analysis of cross-disciplinarity in the empirically-informed philosophy of mind " ( Synthese 2023). Karen Yan What drives us to write this paper is our curiosity about what it means when philosophers of mind claim their works are informed by empirical evidence and how to assess this quality of empirically-informedness. Building on Knobe’s (2015) quantitative metaphilosophical analyses of empirically-informed philosophy of mind (EIPM), we investigated further how empirically-informed philosophers rely on empirical research and what metaphilosophical lessons to draw from our empirical results.  We utilize scientometric tools and categorization analysis to provide an empirically reliable description of EIPM. Our methodological novelty lies in integrating the co-citation analysis tool with the conceptual resources from the philosoph