Psychology
Our own choices generate biases for subsequent decisions
Humans like to think of their judgments as ‘rational’, solely based on objective information. Instead, we have found that people interpret decision-relevant information in a way that is distorted by their previous judgments. This mechanism can account for many important real-life biases, and it may be a natural consequence of the architecture of the brain.
Human judgment and decision-making is strongly shaped by biases. Intriguingly, some of those biases result from the choices we have made in the past. Having committed a categorical judgment, we no longer interpret new information neutrally but are biased to confirm our initial judgment. You may have experienced this: After stating in a poll that you will vote for a political candidate, you feel confirmed by the candidate's good policy proposals, but ignore the bad ones.
This phenomenon, called confirmation bias, was first described in the 5th century BC. Confirmation bias is pervasive in daily life, affecting human judgment in numerous cases of critical significance, for example, medical diagnostics, job interviews, court cases, scientific hypothesis testing - or political choices. Even so, why our brains tend towards confirmation has long remained elusive.
We set out to illuminate this question, starting from recent insights in neuroscience. Forming a decision requires the coordinated action of thousands of neurons that are distributed across many different regions of the brain. As people commit to a categorical decision ("I will vote for this candidate") this network of neurons settles into a stable state, called "attractor". What's more, brain regions involved in decision formation send strong feedback signals to regions processing the incoming information. We reasoned that this feedback might account for confirmation bias: once the brain's "decision regions" enter an attractor state and an initial decision is made, the feedback will amplify the processing of new information that is consistent with the initial decision while suppressing processing of information that contradicts the initial decision. Intriguingly, this is similar to the way our brains filter information when we selectively process a particular aspect of the sensory input to accomplish our goals: say, focusing on the announcement in a train station amidst the distracting noise in order to gather information about the train we are waiting for. So, we hypothesized that confirmation bias results from the selective filtering of incoming information, just like goal-directed attention.
We developed a novel behavioural laboratory task that enabled us to test this idea. Volunteers watched a cloud of flickering dots on a computer screen. When a beep played, some of the dots moved in unison, like the wind blowing some snowflakes into one direction. We then asked them to categorize the motion of the dots (to the left or to the right of a reference?). This judgment was hard because only a few dots moved coherently. We then showed them a second stream of moving dots, which might move in a slightly different direction. Finally, volunteers estimated the average direction across the two streams as precisely as possible.
Indeed, the volunteers selectively filtered new information (directions of the second stream of dots) so as to confirm their initial categorical judgment: New motion directions consistent with the initially chosen category (say, left from the reference) contributed strongly to the final estimation of the total direction; directions inconsistent with that category had little influence. The commitment to the initial categorical choice seems to have induced a change in brain state that made people interpret new information in a self-confirmatory way.
In psychology, confirmation bias had previously been studied in more complex decisions, as entailed in, for example, abstract reasoning or economic choices. It was striking to observe signatures of confirmation bias in decisions about basic sensory stimuli (snowflakes dart over a computer screen) that did not carry any particular meaning to participants. This suggests that confirmation bias is deeply rooted in the brain's machinery for decision-making.
To test whether our insights generalize to confirmation biases in complex decisions, we asked a different group of volunteers to do another version of the task: rapidly average a stream of numbers flashed on the screen. Halfway through the stream, they had to categorize the mean, by saying whether it was lower or higher than 50; at the end of the whole number stream, they typed in what they thought was the average across both. Again, they tended to up-weight numbers consistent with their chosen category (say, > 50) and down-weight numbers inconsistent with their choice.
Unravelling the sources of confirmation will help understand, and potentially expand, the bounds of human rationality. We have shown that the brain actively generates confirmation biases through a process similar to the filtering of input through selective attention. Our insights imply an intricate, hitherto unknown link between decision-making and selective attention, and they set the stage for pinpointing the mechanism at the level of the underlying brain circuits.
Original Article:
B. C. Talluri, A. E. Urai, K. Tsetsos, M. Usher, T. H. Donner, Confirmation Bias through Selective Overweighting of Choice-Consistent Evidence. Curr Biol 28, 3128-3135 e3128 (2018)Edited by:
Massimo Caine , Founder and Director
We thought you might like
The life-span of SARS-CoV-2 in pediatric patients
Jun 9, 2020 in Health & Physiology | 3.5 min read by Sarah Gibson , Charlotte Zhang , Oulan Li , Yi XuSnake uses its split jaws as a knife and fork
Dec 10, 2021 in Evolution & Behaviour | 4 min read by Yosuke KojimaMillions of years of evolution at stake on Madagascar
Aug 14, 2023 in Evolution & Behaviour | 4 min read by Nathan MichielsenPlagued for millennia: The complex transmission and ecology of prehistoric Yersinia pestis
Jul 31, 2023 in Microbiology | 3 min read by Aida Andrades Valtueña , Gunnar U. Neumann , Alexander HerbigMore from Psychology
Sharing a political ideology predicts more similar brain activity
Sep 1, 2023 in Psychology | 3.5 min read by Daantje de Bruin , Oriel FeldmanHallHow do people see, simplify, and solve problems?
Aug 25, 2023 in Psychology | 3.5 min read by Mark HoWomen are human too, but not according to the internet
Jul 21, 2023 in Psychology | 4 min read by April BaileyDeaf to arguments? How conspiracy beliefs shape opposition to wind farms
Jun 12, 2023 in Psychology | 3 min read by Kevin WinterCan Training Make Policing Fairer and More Effective?
May 22, 2023 in Psychology | 3 min read by David Weisburd , Cody TelepEditor's picks
Trending now
Popular topics