Skip to main content

Brain Architecture and Biased Beliefs


Today's post is by Christina Moutsiana (pictured above), who is Lecturer in Cognitive Neuroscience at the University of Westminster. Her research focuses on perceptual and cognitive biases which she investigates using a variety of neuroscience tools, behavioural paradigms, and psychophysics. In this post she discusses some of her recent research on belief updating and brain architecture (see Moutsiana et al. 2015 for more details).

We make decisions every day, some of which are critical to our health and wellbeing. These decisions are driven by the beliefs we hold. But do we form beliefs in an accurate manner that allow us to avoid harm and maximize reward? Cognition is not impervious to error, especially when emotion comes into play. One of the most robust cognitive biases is asymmetric belief updating; the tendency for healthy individuals to alter beliefs about their future to a greater extent in response to good information compared with bad information.

Decision neuroscience research suggests that updating beliefs involves not only brain regions known to perform complex cognitive functions, but also key structures for emotion. I therefore wanted to know if individual differences in brain connectivity between these structures relate to biased belief updating. To investigate that I exploited DTI (diffusion tensor imaging) that enables tracking of neural pathways, together with the belief update paradigm (Sharot et al., 2011). The belief update paradigm allows us to compute individual differences in updating self-relevant beliefs about possible life events.

Let's imagine you hold a belief about the likelihood of having a bike accident. I provide you with evidence that this event is more likely than you had initially expected (bad news). I later ask you again how likely you are of have a bike accident, so that I can assess whether you had integrated the information I gave you to adjust your belief. I do the same with your friend. I can then calculate how you differ in your ability to update your beliefs and learn relative to your friend. Replicating previous findings, I found that people learnt better in response to good news than bad news.

Performing probabilistic tracking on the DTI images enabled me to identify neural paths in each individual’s brain and calculate how strong these paths are between areas relevant to belief updating. I focused on paths that include specific frontal regions (the inferior frontal gyrus (IFG) and the medial frontal cortex), which have been found to be associated with valence-dependent learning asymmetries (Sharot et al., 2011). Next, I looked whether connectivity in those networks was related to asymmetric beliefs. Indeed the results identified a neural system of interconnected brain regions for which the strength of white matter connectivity was associated with belief updating as a function of valence.

In particular, the results highlighted an opposing relationship between left frontal–subcortical white matter connectivity and updating in response to good and bad news. For good news (e.g., learning you are less likely to have a bike accident than you thought), stronger connectivity within this system, particularly between the left IFG and key regions for emotion such as left amygdala, hippocampus, pallidum, was associated with greater changes in belief across individuals. Conversely, for bad news, stronger connectivity between the left IFG and left putamen, pallidum, and insular cortex was related to reduced change in beliefs. Thus, left frontal–subcortical anatomical connectivity was associated with belief change differently based on the valence of the information.

Together with past findings (Sharot et al., 2011, 2012), these results characterize a neural architecture that may allow valence to be incorporated into self-relevant belief updating. 

Popular posts from this blog

Delusions in the DSM 5

This post is by Lisa Bortolotti. How has the definition of delusions changed in the DSM 5? Here are some first impressions. In the DSM-IV (Glossary) delusions were defined as follows: Delusion. A false belief based on incorrect inference about external reality that is firmly sustained despite what almost everyone else believes and despite what constitutes incontrovertible and obvious proof or evidence to the contrary. The belief is not one ordinarily accepted by other members of the person's culture or subculture (e.g., it is not an article of religious faith). When a false belief involves a value judgment, it is regarded as a delusion only when the judgment is so extreme as to defy credibility.

Rationalization: Why your intelligence, vigilance and expertise probably don't protect you

Today's post is by Jonathan Ellis , Associate Professor of Philosophy and Director of the Center for Public Philosophy at the University of California, Santa Cruz, and Eric Schwitzgebel , Professor of Philosophy at the University of California, Riverside. This is the first in a two-part contribution on their paper "Rationalization in Moral and Philosophical thought" in Moral Inferences , eds. J. F. Bonnefon and B. Trémolière (Psychology Press, 2017). We’ve all been there. You’re arguing with someone – about politics, or a policy at work, or about whose turn it is to do the dishes – and they keep finding all kinds of self-serving justifications for their view. When one of their arguments is defeated, rather than rethinking their position they just leap to another argument, then maybe another. They’re rationalizing –coming up with convenient defenses for what they want to believe, rather than responding even-handedly to the points you're making. Yo

A co-citation analysis of cross-disciplinarity in the empirically-informed philosophy of mind

Today's post is by  Karen Yan (National Yang Ming Chiao Tung University) on her recent paper (co-authored with Chuan-Ya Liao), " A co-citation analysis of cross-disciplinarity in the empirically-informed philosophy of mind " ( Synthese 2023). Karen Yan What drives us to write this paper is our curiosity about what it means when philosophers of mind claim their works are informed by empirical evidence and how to assess this quality of empirically-informedness. Building on Knobe’s (2015) quantitative metaphilosophical analyses of empirically-informed philosophy of mind (EIPM), we investigated further how empirically-informed philosophers rely on empirical research and what metaphilosophical lessons to draw from our empirical results.  We utilize scientometric tools and categorization analysis to provide an empirically reliable description of EIPM. Our methodological novelty lies in integrating the co-citation analysis tool with the conceptual resources from the philosoph