The Horizons Tracker

Research explores why we tend to employ wishful thinking.

wishful thinking research definition

“People aren’t purely truth-seekers—many beliefs are influenced by emotions and driven by what is pleasant or comforting. Like belief in an afterlife or optimism about health outcomes,” the researchers explain.

Self-deception

Understanding the intricacies of self-deception and its underlying causes proves challenging in real-world scenarios. To shed light on this phenomenon, the study employed a series of experiments involving over 1,700 participants, both in laboratory settings and online.

During these experiments, participants were briefly presented with various patterns, such as sets of differently oriented stripes or colored dots, and were tasked with identifying the patterns they perceived. Some of these patterns were associated with negative outcomes, intended to induce anxiety, such as a mild electric shock (conducted in the lab) or monetary loss (online).

Consistently, the study revealed that participants were less inclined to accurately identify patterns linked to potential shocks or financial losses. Instead, they tended to perceive patterns that aligned with more desirable outcomes. While previous research primarily focused on wishful thinking concerning positive outcomes, yielding mixed results, this study distinctly highlights how the negative emotion of anxiety regarding an outcome can drive wishful thinking.

Increasing realism

Furthermore, the researchers explored interventions aimed at fostering a more realistic outlook among participants. One approach involved simplifying the patterns to enhance their recognizability. The findings indicated that reducing uncertainty did, indeed, mitigate wishful thinking tendencies.

Another intervention entailed offering greater potential rewards for accurate pattern recognition. Surprisingly, this intervention had minimal impact, except in cases where participants could gather additional evidence about the specific pattern presented to them. This underscores the nuanced relationship between incentives and cognitive biases, suggesting that increased clarity may be more effective in curbing wishful thinking than mere incentive enhancement.

“When people had more time to collect evidence and were better rewarded for a correct answer, they became more realistic,” the researchers explain.

In the final phase of the experiments, where positive outcomes replaced negative ones, participants displayed no signs of wishful thinking. The absence of such tendencies suggests that mitigating negative emotions can indeed diminish over-optimism.

This observation underscores the significant role emotions play in shaping cognitive biases, highlighting the potential effectiveness of strategies aimed at alleviating negative effects to promote more realistic perceptions and judgments.

“Wishful thinking is important for humans in coping with anxiety about possible future events,” the researchers concede. The concern is situations in which too much optimism stops people from getting the information they need or from acting in a way that would benefit them.

“People can get too hopeful when things are uncertain. We observe this happening with climate change, when financial markets fluctuate, and even in personal health situations when people avoid medical help because they think everything will be fine. We need to know more about when wishful thinking helps and when it hurts.”

Facebook

The science of hope: More than wishful thinking

ASU researchers share tips for unleashing the power of hope in your own life

An illustration of rivers, mountains and forests, reading "the Roadmap to Hope."

Hoping to win the lottery? That would actually be a wish you have. Hoping to have a good day today? That’s being optimistic. 

According to social science researchers in Arizona State University’s  Center for the Advanced Study and Practice of Hope , having hope is harder work than we may think. They are unpacking the ways in which different populations conceive of hope and exploring how having higher hopes can improve our health and well-being.

What hope means

In psychology, hope is a cognitive practice that involves the intentional act of setting goals and working toward them with purpose. 

“Hopeful people are able to set goals, identify ways to reach their goals and feel as though they can do the work to achieve those goals,” said  Crystal Bryce , associate director of research in the Hope Center and clinical assistant professor in the  T. Denny Sanford School of Social and Family Dynamics.

John Parsi , executive director of the Hope Center, thinks about using hope like driving from ASU’s Tempe campus to the Downtown Phoenix campus.

“If there’s an accident on the I-10 or construction on the Broadway curve, I have to work around that to get to my destination. I can’t just stop in the middle of the highway and give up,” said Parsi. “I have to navigate through the traffic using hope theory and ultimately take agency to literally steer my actions.”

Parsi emphasizes that it is important to know the difference between hope and blind optimism. 

“Hope is an active process. Dreams and optimism are just belief structures,” he explained. “When you’re an optimistic person, you believe things in the world will turn out just fine, no matter what happens.”

Parsi says that because optimism doesn’t require a person to do anything, it can be a form of toxic positivity. But from the working scientific definition of hope, hope can only do good for a person.

“Hopeful people cannot just wish things into existence,” he said. “Hope requires a person to take responsibility for their wants and desires and take action in working towards them. Optimistic people see the glass as half full, but hopeful people ask how they can fill the glass full.”

Hope in higher education

One of the center’s ongoing projects is exploring what hope looks like in populations of ASU first-year students. Research has shown that more hopeful students are better off in terms of their health and ability to tackle future challenges.

“Students that have more hope do better in school, have more developed social and emotional skills, and foster stronger connections with their peers,” said  Ashley Fraser , a PhD student in the Sanford School and member of the Hope Center research team.

“The strategy of hope is fairly robust in its real-world impact. This isn’t a squishy science,” said Parsi. “Increased hope has been shown to be more effective than many mental health interventions in youth populations. Even as experts in this field, we get surprised time and time again by the impact and results these hope interventions have on the populations they are implemented in.”

The team places special emphasis on hope research in college students because positive psychology in the college experience is generally understudied.

“If you look at most studies around college students, historically it is really negative,” said Fraser. “Researchers often look at drinking, drug use, suicide and pregnancy, all of these negatively associated outcomes. Part of our center’s objective is to show that this is a great time in young people’s lives. This is really a time where people can change the course of their lives if they want to.”

The team’s data shows that first-year students with high hopes in their first semester of college were more likely to enroll in a second year than students with low hope. Increased hope in these students is also associated with prosocial behavior, better civic attitudes and increased community engagement in college. 

In the next stage of this study, the team wants to see how students’ levels of hope affect their well-being and social-emotional skills.

“The data has shown that the first year of college is a really great intervention point,” said Bryce. “We could really target first-semester students and teach them how to cultivate their hope. Some of them might not even know that hope is teachable. We are trying to develop exercises and practices for first-year students in fall 2021 to instill hope skills into their lives and see what impact this support has.”

Building hopeful communities

Many of the Hope Center’s other research projects revolve around the Kids at Hope , an affiliated nonprofit committed to working with youth-serving organizations to cultivate a culture of hope in community settings. The partnership allows researchers in the center to work with practitioners in the community and translate research into real-world impact.

“The hope research field is already fairly small,” said Parsi, who also serves as president of Kids at Hope. “But even rarer is this combination of a robust research center at a leading university and a nonprofit organization outlet to exercise our research and collect data from. It truly is a special partnership.”

In Michigan's Macomb County, the ASU team has been working with the state’s juvenile justice system to investigate how probation officers, educators and administrators within the incarceration community view and promote hope in youth on probation. Kids at Hope and the Hope Center team have also partnered with the Saddle Mountain Unified School District in Tonopah, Arizona, to see how hope practice curriculum can be integrated into schooling and how it impacts student relationships, engagement and achievement.

“Our lives are holistic. We are more than our work and more than our families, so we must set goals in all areas of our lives,” said Parsi. “This often starts with small goal setting and breaking down big goals into smaller, more reachable interim goals.”

In their hope curriculum, Kids at Hope teaches practitioners three universal truths: 

  • Every child — no exceptions — is capable of success.
  • Connection between adults and youth is meaningful and impactful.
  • Goal setting happens in many areas, including education and career, home and family, community and service, and hobbies and recreation.

“Everyone in these communities sees the glimmers of hope and understands how important it is to the success of their youth, but they’re not quite sure how to increase hope or support that kind of culture,” said Bryce. “That’s where we come in with our research on socialization to further their goals as organizations.”

How to have hope

Though they are experts on the theory of hope, Bryce, Fraser and Parsi are always working to become more hopeful people themselves. They offer some tips to help you get your hopes up in ways that benefit you and your community.

1. Take small steps for long-term success.

“Think about your goals for just today and make a list of actions that you need to do to accomplish those goals,” said Fraser, a self-proclaimed “hopeful realist.” “Then, envision your larger, long-term goals and see how your actions today will help you achieve those bigger, broader goals you have for yourself.”

2. Check in with yourself.

If you aren’t excited and enthusiastic about taking the next step in reaching your goals, odds are your goal needs to be adjusted. 

“Ask yourself if you’re excited about the future ahead of you and if your actions leading you there have energy behind them,” said Fraser. If you can answer “yes” to these questions, then it’s a good sign you are developing hope, according to hope theory.

3. Consider context.

The team encourages people to consider the cultural context they live in when evaluating how hopeful they are. Some of the center’s research revolves around how different ethnic populations, specifically Hispanic and Latino communities, view and practice hope.

“We do see something similar to hope in every culture. Part of that is because the human condition is designed to look forward,” explained Fraser. “The conception of hope looks different in religious communities, though the outcomes are still overwhelmingly positive. Hope is also a big positive impact in medical patient communities.”

In Fraser’s previous research, she investigated hope among post-war Colombian youth. She found that young people exposed to violence and trauma experienced a decrease in hope. Systemic oppression that is out of a person’s control can lead to uncontrollable feelings of hopelessness. 

“It’s important to understand that in some populations, we cannot just say ‘be more hopeful’ because there are contextual and cultural factors that can really hinder a child’s hope,” said Fraser. “Though hope does cross borders, there is more to discover about hope in the face of adversity.”

4. Remember that hope takes practice.

Even the researchers studying hope admit that they don’t always practice what they preach to others. 

“Hope is really hard. There are times where I don’t have as much hope as I want to,” said Bryce. “When I’m teaching a class or applying for grants, I constantly find myself trying to find those small steps to get to the larger goals I want to fulfill at the moment. We have to constantly remind ourselves of this, though we are entrenched in it all the time.”

In 2018, Parsi’s wife suffered a series of strokes. The couple had to pick up everything they had in Alaska to move to Arizona so that she could receive treatment at Barrow Neurological Institute.

“We had to be hopeful to get through that, but I definitely have my ebbs and flows,” said Parsi.  

“Life-altering events, like the ones that happened in my life, really force a person to take a step back and reevaluate the toolkit that we approach life’s challenges with. For me, that meant prioritizing different things in my life and setting new goals for myself.”

5. Foster a culture of hope.

The researchers at ASU’s Center for the Advanced Study and Practice of Hope emphasize that hope thrives within communities that support and uplift one another.

Parsi had trouble fostering hope in educational and social contexts of adversity. As the son of immigrants, a non-native Engish speaker and a first-generation college graduate, he often faced racist and xenophobic individuals and systems throughout his professional career.

“There would be a lot more people who overcame adversities in the world if we created a culture that values hope as a priority and invests in children and their ability to create,” said Parsi. 

As a hope researcher and community member, Parsi feels responsible to be a part of the success and failures of all the people in the communities he belongs to. 

“It is my responsibility not to pull the ladder up behind me, but instead to let the hardships I went through inform the idea that the success of our children belongs to all of us,” he said.  

Written by Maya Shrikant 

Top illustration by Ashley Quay

More Health and medicine

Four people standing around an emergency room bed.

Entrepreneurs, health care professionals come together to create medical solutions

Improving the emergency room experience. Revolutionizing blood draws. A more flexible SurgiSlinky that would replace bones.Those are just some of the ideas that came out of the eight-day Arizona…

Portrait of David Engelthaler.

Arizona public health leader to direct Health Observatory at ASU

Think about any health scare in recent Arizona history — from hantavirus and plague to anthrax and COVID-19 — and you can be sure David Engelthaler was working behind the scenes to help the state…

Blue gloved hand holding a plastic strip flu test

ASU researchers: Over-the-counter flu tests can 'revolutionize public health'

Influenza, more commonly known as the flu, causes 50,000-plus of deaths annually, but over-the-counter (OTC) tests for the disease are just beginning to emerge and are difficult to access.That’s a…

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Soc Cogn Affect Neurosci
  • v.7(8); 2012 Nov

Neural correlates of wishful thinking

Tatjana aue.

1 Swiss Center for Affective Sciences, University of Geneva, Geneva, Switzerland and 2 Center for Cognitive and Social Neuroscience, University of Chicago, Chicago, IL, USA

Howard C. Nusbaum

John t. cacioppo.

Wishful thinking (WT) implies the overestimation of the likelihood of desirable events. It occurs for outcomes of personal interest, but also for events of interest to others we like. We investigated whether WT is grounded on low-level selective attention or on higher level cognitive processes including differential weighting of evidence or response formation. Participants in our MRI study predicted the likelihood that their favorite or least favorite team would win a football game. Consistent with expectations, favorite team trials were characterized by higher winning odds. Our data demonstrated activity in a cluster comprising parts of the left inferior occipital and fusiform gyri to distinguish between favorite and least favorite team trials. More importantly, functional connectivities of this cluster with the human reward system were specifically involved in the type of WT investigated in our study, thus supporting the idea of an attention bias generating WT. Prefrontal cortex activity also distinguished between the two teams. However, activity in this region and its functional connectivities with the human reward system were altogether unrelated to the degree of WT reflected in the participants’ behavior and may rather be related to social identification, ensuring the affective context necessary for WT to arise.

For what a man more likes to be true, he more readily believes (Francis Bacon, 1561–1626).

INTRODUCTION

Human reasoning and decision-making are subject to distortions, which may originate in a variety of different reasons, ranging from the application of simple heuristics ( Goldstein and Gigerenzer, 2002 ) to more complex groupthink ( Kerschreiter et al ., 2008 ). These distortions are not necessarily irrational. Under some circumstances, simple heuristics can lead to more accurate inferences than strategies that arise from considerable knowledge about a situation ( Goldstein and Gigerenzer, 2002 ). Also, Zábojník (2004) developed a model that explains why most individuals legitimately believe that they are above average in various skills and abilities. According to this model, individuals test and improve their abilities until they are reasonably favorable, and thus, above average. Moreover, skills and abilities are not always normally distributed, leading occasionally to the phenomenon that >50% of the individuals in a population score indeed above average. These considerations show that some biases are quite rational—but others are not.

Francis Bacon's observation presaged work on a particularly pernicious systematic irrational bias, the tendency to overestimate the likelihood of desirable events and underestimate the likelihood of undesirable events. This cognitive bias has been called wishful thinking ( WT, McGuire, 1960 ). 1 In contrast to heuristics and the type of rational estimations of own capacities described before, WT originates in the application of individual preferences that are linked to affective experiences.

WT biased beliefs and decision-making have reliably been observed in many different contexts, especially when events are of personal relevance and people thus feel personally invested in the outcomes of their predictions ( Babad, 1987 ; Granberg and Holmberg, 1988 ). Biased predictions may work to maximize the anticipated reward or the hedonic valence of an expected outcome. WT has been documented whether the outcomes apply directly to the decision maker or they apply to someone with whom the decision maker identifies ( McGuire, 1960 ; Babad and Katz, 1991 ; Price, 2000 ). The particular case of WT for others is in line with research on social identity theory, which has shown that individuals generally prefer ingroup members over outgroup members ( Tajfel et al ., 1971 ) and that they display a need for a positive social identity ( Turner and Onorato, 1999 ). Overestimating the likelihood of positive events for other people or groups with whom we identify therefore may increase the experienced positivity of our own social identity, and, consequently, the hedonic valence of a predicted outcome, thus functioning as an affective reward. Evidently, WT is not a rare but a rather common phenomenon.

The concrete mechanisms underlying WT are still unknown. Krizan and Windschitl (2007) put forth three different stages at which WT could arise. First, at an early processing stage individuals could selectively attend to existing situational cues. Attentional deployment in WT may work in the sense of a confirmation bias, making us attend to supporting evidence for what we desire and neglect contradictory evidence. Such a mechanism would consequently serve the WT effect. Alternatively, WT could be generated by selective interpretation of available evidence, thus changing the sense of but not the access to the existing information in a situation. Attributing greater importance to the desirable than to the undesirable evidence in a situation would work in this service. Finally, a third alternative assumes WT to arise at a comparably late stage of information processing, namely at response formation. Consequently, the mechanism underlying WT could be either an attention bias, an interpretation bias, or a response bias.

Investigations of the neural correlates of wishful thinking are absent but could help identifying the mechanisms generating WT by supporting or contradicting either of these alternatives. For example, if early selective attention were at the basis of the WT effect, individuals should display WT-related activation in low-level attention-associated brain regions. Since, in the current study, the background information was presented visually, selective attention should be particularly visible in the visual cortex (e.g. Sabatinelli et al ., 2007 ). If, on the contrary, an interpretation bias or a response bias rather than an attention bias were responsible for the genesis of WT, we would expect the implication of higher level cognitive processing areas within the prefrontal cortex (PFC) because these mechanisms would imply either evidence reevaluation (reappraisal) or attempts at cognitive regulation (e.g. Ochsner et al ., 2002 ; Sharot et al ., 2007 ; see also Miller and Cohen, 2001 , for an extensive discussion on the importance of the prefrontal cortex for cognitive control). Additional justification for such a hypothesis comes from research on pathological gambling, which has been associated with biased cognitions leading to over-optimism and, importantly, with dysfunctions of the ventromedial PFC ( Cavendini et al ., 2002 ). Consequently, fMRI may help to reveal whether low-level attentive or higher level cognitive processes are at the origin of WT.

Moreover, because WT has been postulated to be a hedonic bias, activity in the areas supporting either of these mechanisms being at the origin of the WT phenomenon should be functionally connected to activity in the human reward system. This is because during WT people adopt expectations that they believe will yield affective rewards. Therefore, WT should recruit structures of the human reward system (including the striatum, the posterior cingulate and the amygdala; Heekeren et al ., 2007 ; Delgado et al ., 2008 ; Xue et al ., 2009 ). For instance, anticipated reward, such as pleasure experienced by favorite team soccer goals ( McLean et al ., 2009 ) has been reported to go along with activation in the dorsal striatum.

We investigated the neural correlates of WT in a social identification context based on National Football League (NFL) teams with participants who were recruited based on their reports that they cared about and followed closely a particular team in the NFL. Objective background information in this task was given in the same manner for all experimental scenarios, prohibiting the possibility that observed differences between scenarios would result from a differential access to available information. Participants in the current study estimated the winning odds for different NFL teams [favorite, neutral (filler items) and least favorite] on the basis of this background information. According to identification theory, outcomes bearing on the participant's favorite team are higher in self-relevance than outcomes bearing on one's least favorite team. Brain activity was recorded in an MRI scanner while the participants gave the estimations. In line with earlier research and behavioral pilot testing, we expected that participants would overestimate the success of the favorite team with respect to the least favorite team, that is, that they would exhibit WT.

The neural correlates of WT were investigated by (i) comparing favorite to least favorite team trials, because, during WT, individuals should well distinguish between the two teams, (ii) identifying functional connectivities of these areas with other brain regions (e.g. the human reward system) and (iii) by correlating these activation and connectivity data with the extent of WT demonstrated in the participants’ behavior. The latter point is particularly important because activation and connectivity patterns differing for favorite vs least favorite team trials could otherwise be simply a sign of differential preference for or identification with the two teams without being related to WT itself.

Participants

The study was approved by a local ethics committee (institutional review board for biological sciences at the University of Chicago). Written informed consent was obtained in accordance with the Helsinki Declaration of Human Rights (1999). Thirteen (two female) healthy University of Chicago students were recruited via ads posted on the Internet and in university buildings. They were aged between 19 and 34 years ( M  = 21.8, s.d. = 3.82) and indicated high interest in the NFL [ M  = 5.8, s.d. = 1.63, on a scale ranging from 1 (no interest) to 7 (intense interest)]. Five participants actively practiced American football ( M  = 8 h/week, s.d. = 6.08 h/week) at the time of the study. All but one watched football on a weekly basis (TV or local; M  = 4.5 h/week, s.d. = 3.40 h/week). Participants had normal or corrected-to-normal vision, and were not currently seeking treatment for affective disorders. Handedness was assessed with the Edinburgh Handedness Inventory ( Oldfield, 1971 ). Only individuals indicating ‘right’ for all items were included in the study. Participants were paid $15/h.

Setting and apparatus

MRI data were acquired on a 3T GE Signa scanner (GE Medical Systems, Milwaukee, WI, USA) with a standard quadrature GE head coil. Stimulus presentation and collection of participants’ button presses were controlled by e-prime 1.1 (Psychology Software Tools, Inc., Pittsburgh, PA, USA) running on a PC. Participants viewed the stimuli by the use of binocular goggles mounted on the head coil ∼2 in above the participants’ eyes. Button press responses were made on two MRI-compatible response boxes. Every button corresponded to a specific number (e.g. leftmost button left hand: 1; rightmost button right hand: 0).

High-resolution volumetric anatomical images [T1-weighted spoiled gradient-recalled (SPGR) images] were collected for every participant in 124 1.5-mm sagittal slices with 6° flip angle and 24 cm field of view (FOV). Functional data were acquired using a forward/reverse spiral acquisition with 40 contiguous 4.2-mm coronal slices with 0.5-mm gaps, for an effective slice thickness of 4.7 mm, in an interleaved order spanning the whole brain. Volumes were collected continuously with a repetition time (TR) of 3 s, an echo time (TE) of 28 ms and a FOV of 24 cm (flip angle = 84°, 64 × 64 matrix size, fat suppressed).

Upon participants’ arrival at the laboratory the nature of the experiment was explained and informed consent was obtained. Then, participants specified their favorite, neutral (team they neither liked nor disliked) and least favorite NFL teams.

Brain activity was recorded in an MRI scanner while the participants did 60 estimations (20 for each team). In each trial, participants were given background information about a single team (favorite, neutral, or least favorite). Trials for the neutral team acted as irrelevant filler trials in the current study. The background information comprised three pieces: (i) The probability of winning for the team in question if a given player played , (ii) the probability of winning if this specific player did not play and (iii) the probability that this player would actually play . Bayesian statistics permit the combination of these probabilities to determine the objective probability that the team will win the game ( Petty and Cacioppo, 1981 ). However, in the current study, participants were not given enough time to perform exact statistics.

The particular probabilities used for the favorite, neutral and least favorite teams were counterbalanced across participants. In order to make the teams more salient, in every trial, the team logos appeared in each of the four corners of the computer screen with the background information presented in between ( Figure 1 ). The participants’ task then was to specify what they believed to be the probability that the presented team would win its first playoff game given this information.

An external file that holds a picture, illustration, etc.
Object name is nsr081f1.jpg

Trial sequence. In every trial, the three statements were first presented for a random time of 3, 6 or 9 s, without any indication of the corresponding probabilities and the concerned team. Next, the club banner of the concerned team appeared in all four corners of the screen and the statement-related probabilities were projected to the right of the statements. From this moment, participants had 9 s to estimate the supposed winning odds. After this, a question prompted them to indicate their probability estimate by the use of two button boxes. Participants were given 6 s to do so. Immediately after, the next trial began with the presentation of the three statements.

Before beginning the study, participants were trained on the use of two button boxes, permitting them to indicate the probability estimates, and got the occasion to go through 10 practice trials to get familiar with the task. They were informed that they might ask questions about anything that was not perfectly clear to them. After the study, the participants indicated their sympathy with the different teams on three different Likert scales. These scales asked how much they liked the teams [ranging from 1 (extremely dislike) to 7 (extremely like)], as well as how pleasant [ranging from 1 (extremely unpleasant) to 7 (extremely pleasant)] and appealing [ranging from 1 (extremely unappealing) to 7 (extremely appealing)] the teams were to them.

Data analysis

Behavioral data.

Participants’ sympathy ratings for the favorite and least favorite team were contrasted with a paired samples t -test (α level of 0.05, one-tailed). In order to determine the extent of WT, for each trial, the objectively correct winning odds were subtracted from the participants’ estimates, thus yielding 60 difference scores per participant. Missing responses and outliers (estimates referring to difference scores deviating >3 s.d.'s from the mean difference score of a participant) constituted ∼5% of the data and were eliminated. A paired-samples t -test with an α level of 0.05 comparing participants’ estimates for the favorite team with the ones for the least favorite team was calculated. Given the directional nature of the hypothesis, a one-tailed test was performed.

Functional scans were realigned, normalized, time and motion corrected, temporally smoothed by a low-pass filter consisting of a 3-point Hamming window, and spatially smoothed by a 5-mm full-width at half-maximum (FWHM) Gaussian filter, using AFNI software. Next, a percent-signal change data transformation was applied to the functional scans. Individual deconvolution analyses incorporated the generation of impulse response functions (IRFs) of the BOLD signal on a voxel-wise basis ( Ward, 2001 ), permitting the estimation of the hemodynamic response for each condition relative to a baseline state (baseline, in our case, referred to the mean individual signal across the whole experimental session) without a priori assumptions about the specific form of an IRF. Deconvolution analyses comprised separate regressors for each time point (six included TRs, starting with the apparition of the club banners and the statement-related probabilities, i.e. the calculation process) of each experimental condition, and fitted these regressors by the use of a linear least squares model. Estimated individual percent-signal change for TRs two to six was averaged for each voxel in each condition for use in later statistical analyses. Output from the individual deconvolution analyses was converted to Talairach stereotaxic coordinate space ( Talairach and Tournoux, 1988 ) and interpolated to volumes with 3 mm 3 voxels.

Since WT should rely on a distinction between favorite and least favorite team trials, for each participant, three contrasts of interest were conducted for use in group-level whole-brain analyses: Percent signal change of each condition with respect to baseline as well as the differences in-percent-signal change for favorite vs least favorite team. Results were then entered into a cluster analysis with an individual voxel threshold of P  < 0.01, minimum cluster connection radius of 5.2, and a cluster volume of 702 µl (corresponding to 26 active, contiguous voxels). Minimum cluster volume was calculated by a Monte Carlo simulation with 10 000 iterations, assuming some interdependence between voxels (5 mm FWHM), resulting in a corrected whole-brain P -value of 0.05. Consequently, this method corrects for whole brain volume and thus, multiple testing at the voxel level.

To get insight into the coactivations of the brain areas differing between the favorite and least favorite teams with other brain areas (including the human reward system), we entered the regions found to significantly differ between the two teams as input seeds for context-dependent functional connectivity (also called psychophysiological interaction, PPI) analyses. Analyses for each cluster were conducted separately for each team as well as for the differential activity between the two teams. These analyses tested, within a whole brain analysis, for synchronized activities with the input seed regions. The exact method adopted can be found at: http://afni.nimh.nih.gov/sscc/gangc/CD-CorrAna.html . Results were entered into a cluster analysis with an individual voxel threshold of P  < 0.01, minimum cluster connection radius of 5.2 and a cluster volume of 702 µl (corresponding to 26 active, contiguous voxels).

Finally, and most importantly, we tested whether clusters displaying differential activation for the favorite and the least favorite team as revealed in the whole brain analysis were indeed related to WT and not just displaying affective preferences or differential identification with the two teams. We calculated between-subjects Spearman correlations between the differential activations in these clusters, on the one hand, and the difference in behavioral estimates for the two teams, on the other hand. We also extended this analysis to the functional connectivities: The differences (favorite vs least favorite trials) in Fisher's z correlations between seed and identified functionally connected clusters, on the one hand, were correlated with the difference in behavioral estimates for the two teams, on the other hand.

The favorite team was more liked than the least favorite team, t (12) = 10.31, P  < 0.000001 ( M 's = 6.5 and 2.0), and it was judged as more pleasant, t (12) = 7.66, P  < 0.00001 ( M 's = 6.2 and 1.8), and more appealing, t (12) = 7.96, P  < 0.000005 ( M 's = 6.2 and 2.2). More importantly, participants systematically overestimated the winning prospects for their favorite team compared with their least favorite team, t (12) = 2.13, P  < 0.05 (deviations from objectively correct estimates: M 's = 1.0, and −1.3, respectively), thus displaying WT. The 13 participants specified 9 different favorite teams and 8 different least favorite teams. Therefore, our results are not restricted to specific NFL teams.

Functional MRI data

Changes from baseline for both the favorite and the least favorite team included a decrease of default network activity (medial frontal and various temporal areas), but an increase in activity of the dorsolateral prefrontal cortex (dlPFC) and occipital areas ( Table 1 ). Activations and deactivations for the two teams were largely overlapping, with the slight difference of larger cluster sizes and, consequently, fewer clusters for the favorite team trials.

Whole-brain analyses

Contrast/AreasCluster volume (mm )Talairach coordinates (12)
Favorite team baseline
    B fusiform gurus, B inferior parietal lobule, B superior parietal lobule, B precuneus, B thalamus, B cingulate gyrus, B medial frontal gyrus, B superior frontal gyrus, B cerebellum154 332−1−32309.87<0.0000005
    B superior frontal gyrus, B medial frontal gyrus32 77814217−8.16<0.000004
    R supramarginal gyrus, R superior temporal gyrus10 28756−5023−9.41<0.0000007
    L inferior temporal gyrus, L superior temporal gyrus, L inferior frontal gyrus, L insula8208−436−9−7.08<0.00002
    L middle temporal gyrus, L superior temporal gyrus6426−52−5718−6.67<0.00003
    L cuneus, L lingual gyrus5184−11−862−7.36<0.000009
    B posterior cingulate, B cingulate gyrus3105−3−5128−4.69<0.0006
    R middle temporal gyrus, R superior temporal gyrus256548−9−10−8.36<0.000003
    R middle frontal gyrus23764228306.06<0.00006
    R precentral gyrus, R inferior frontal gyrus2349533275.55<0.0002
    R inferior frontal gyrus, R insula22144628−3−6.54<0.00003
    Lingual gyrus, R cuneus105319−89−3−6.23<0.00004
    L cingulate gyrus810−25−3127−4.68<0.0006
    L posterior cingulate, L lingual gyrus783−24−6554.46<0.0008
    R interior frontal gyrus7292711−13−4.88<0.0004
Least favorite team baseline
    B cingulate gyrus, L middle frontal gyrus, B medial frontal gyrus, B precuneus, B inferior parietal lobule, R superior occipital gyrus, L precentral gyrus75 060−6−31399.93<0.0000004
    B medial frontal gyrus, B superior frontal gyrus, B anterior cingulate41 71514122−8.42<0.000003
    B thalamus, L insula, L putamen, L caudate, L lentiform nucleus22 896−8−1306.29<0.00004
    R middle temporal gyrus, R superior temporal gyrus, R inferior frontal gyrus, R supramarginal gyrus, R inferior parietal lobule, R insula18 19850−2610−7.70<0.000006
    L superior temporal gyrus, L middle temporal gyrus, L angular gyrus, L supramarginal gyrus, L inferior parietal lobule11 691−51−4615−6.54<0.00003
    L lingual gyrus, L cuneus6993−11−850−6.68<0.00003
    B posterior cingulate, B cingulate gyrus, B precuneus5562−3−5128−5.82<0.00009
    R insula, R lentiform nucleus, R claustrum, R putamen, R caudate36992913115.19<0.0003
    L inferior frontal gyrus3591−3827−4−7.21<0.00002
    B cingulate gyrus35372−2037−7.05<0.00002
    R inferior frontal gyrus, R precentral gyrus2754533275.58<0.0001
    B cerebellum25382−50−264.44<0.0009
    R middle occipital gyrus, R inferior occipital gyrus, R fusiform gyrus237640−68−116.97<0.00002
    R cerebellum140431−49−294.78<0.0005
    R middle frontal gyrus12153829346.87<0.00002
    L parahippocampal gyrus945−23−36−10−4.93<0.0004
    R lingual gyrus94518−86−7−5.65<0.0002
    L middle frontal gyrus918−33490−5.96<0.00007
    B cerebellum7830−66−254.42<0.0009
    L middle frontal gyrus783−3919304.82<0.0005
    L posterior cingulate, L lingual gyrus729−23−6554.21<0.002
Favorite team least favorite team
    R superior parietal lobule, R precuneus140426−60434.87<0.0005
    L inferior occipital gyrus, L fusiform gyrus783−34−62−66.73<0.0001
    L medial frontal gyrus, L superior frontal gyrus702−254226.29<0.0001

All P s are based on two-tailed testing.

B = bilateral; L = left; R = right.

The BOLD contrast for favorite minus least favorite team revealed greater activity in (i) left medial and superior frontal gyri within the dorsomedial PFC [dmPFC; x , y , z  = −2, 54, 22; t (12) = 6.29, P  < 0.0001; cluster volume: 702 mm 3 ; Figure 2 a], (ii) right precuneus and right superior parietal lobule [ x , y , z  = 26, −60, 43; t (12) = 4.87, P  < 0.0005; cluster volume: 1404 mm 3 ; Figure 2 b] and (iii) left inferior occipital and left fusiform gyri [ x , y , z  = −34, −62, −6; t (12) = 6.73, P  < 0.0001; cluster volume: 783 mm 3 ; Figure 2 b] for the favorite team.

An external file that holds a picture, illustration, etc.
Object name is nsr081f2.jpg

Areas showing differential activation for the two teams. Significantly greater activation for the favorite team as compared with the least favorite team was observed in bilateral medial and superior frontal gyri within the medial prefrontal cortex [ a ; Talairach coordinates ( x , y , z ): −2, −54, 22], right precuneus and right superior parietal lobule ( b ; 26, −60, 43) and left interior occipital and left fusiform gyri ( b ; −34, −62, −6).

Context-dependent functional connectivity analyses

During the favorite team trials, but not during the least favorite team trials, the parietal seed cluster was functionally connected with an area in the left dlPFC ( Table 2 and Figure 3 ). However, the parietal cluster was not functionally connected to any other brain region in our study when the differential activity between the favorite and least favorite teams was examined, thus showing that the supposed difference in coactivations of the parietal cortex and the dlPFC between the favorite and the least favorite team did not reach significance.

An external file that holds a picture, illustration, etc.
Object name is nsr081f3.jpg

Results of the conducted connectivity analyses. Favorite: connectivity during favorite team trials; Least favorite: connectivity during least favorite team trials; Favorite-least favorite: differential connectivity between favorite and least favorite team trials.

Cluster/ContrastAreasCluster volume (mm )Talairach coordinates (12)
Parietal cluster
    Favorite teamL middle frontal gyrus783−3431265.61<0.0002
    Least favorite team–No significant connectivities
    Favorite least favorite team–No significant connectivities
Occipital cluster
    Favorite teamR claustrum, R insula, R putamen, R caudate1728303188.21<0.000003
R medial frontal gyrus12691339267.58<0.000007
L claustrum, L insula, L putamen837−2811176.74<0.00003
    Least favorite team–No significant connectivities
    Favorite least favorite teamR putamen, R caudate, R claustrum, R insula, B thalamus477918−221510.20<0.0000003
L claustrum, L insula, L putamen1485−2812157.78<0.000006
R posterior cingulate89120−5344.27<0.002
dmPFC cluster
    Favorite teamL caudate, L cingulate gyrus1593−117217.63<0.000007
R caudate, R cingulate gyrus1026204319.13<0.000001
R medial frontal gyrus9451333356.04<0.00006
L insula, L claustrum837−315174.94<0.0004
R superior frontal gyrus8372345227.26<0.00002
    Least favorite teamB medial frontal gyrus, B superior frontal gyrus, B anterior cingulate202513839−6.98<0.00002
R caudate, R cingulate gyrus99910−1630−6.59<0.00003
B medial frontal gyrus7020−961−4.48<0.0008
    Favorite least favorite teamR caudate, B cingulate gyrus364522−6227.97<0.000004
L caudate, L cingulate gyrus1485−106226.41<0.00004
L superior frontal gyrus972−734484.98<0.0004
R medial frontal gyrus8371235364.79<0.0005
B anterior cingulate783432234.71<0.0006

The occipital seed cluster displayed, for the favorite team, positive coupling with parts of right and left claustrum, insula and putamen. It further revealed positive connectivity with an area in the dmPFC. No functional connections with the occipital seed were observed for the least favorite team trials. The favorite vs least favorite PPI demonstrated differential connectivity for two clusters comprising left and right claustrum, insula and putamen, and a cluster in the posterior cingulate cortex. In all cases stronger positive connectivity was observed for the favorite as compared with the least favorite team.

Finally, activity in the dmPFC seed, during the favorite team trials, was positively coupled with activity in left and right caudate and posterior cingulate gyri, in two other areas in the dmPFC, as well as in insula and claustrum. Interestingly, negative coupling with right caudate and cingulate gyrus as well as two areas in the dmPFC were observed for the least favorite team; one of these dmPFC clusters extended into the anterior cingulate cortex (ACC). The favorite vs least favorite PPI revealed differential connectivity of the dmPFC seed with five clusters, including left and right caudate and cingulate gyri, bilateral ACC and two clusters in the dmPFC.

Between-subjects Spearman correlations

Table 3 displays the between-subjects Spearman correlations between the areas displaying differential activity or connectivities for the favorite vs least favorite team trials, on the one hand, and the difference in behavioral estimates for favorite vs least favorite team trials, on the other hand. Participants who displayed a greater extent of differential connectivities with the occipital cluster were characterized by stronger WT in the behavioral estimates, thus supporting the idea of an attention bias being the central mechanism underlying WT in our study.

Correlations with extent of WT displayed in behavior

MRI dataCorrelation with difference in behavioral estimate (favorite–least favorite)
Spearman's
Differential activation favorite least favorite team
    Parietal cluster0.50=0.08
    Occipital cluster0.46>0.11
    dmPFC cluster−0.02>0.95
Differential connectivities between the occipital cluster and
    R putamen, R caudate, R claustrum, R insula, R thalamus0.61<0.05
    L claustrum, L insula, L putamen0.66<0.05
    R posterior cingulate0.76<0.005
Differential connectivities between the dmPFC cluster and
    R caudate, R cingulate gyrus0.22>0.47
    L caudate, L cingulate gyrus−0.19>0.54
    L superior frontal gyrus0.54=0.06
    R medial frontal gyrus0.25>0.40
    R anterior cingulate−0.19>0.53

All P 's are based on two-tailed testing.

B = bilateral; L = left; R = right; N = 13.

Our participants displayed WT and treated their preferred sports team advantageously. In accordance with expectations, we observed systematic differences in participants’ winning estimates for the favorite as compared with the least favorite team. Additional statistical analyses revealed no difference between the neutral and the least favorite team but significantly higher estimates for the favorite than for the neutral team. Thus, the bias observed in our participants’ behavior is largely attributable to a preferential treatment of the favorite team.

On the neural level, we found differential activity between the favorite and the least favorite team in three key areas: The dmPFC, the precuneus/superior parietal lobule and the fusiform gyrus/occipital gyrus. However, activity in none of these areas was per se associated with the degree of WT displayed in our participants’ behavior. Earlier research has related activity in a network comprising the dmPFC and the right precuneus to self-relevance (here defined by social identity), self-reflection and social cognition ( Ochsner et al ., 2004 ; Amodio and Frith, 2006 ). Therefore, activity in these areas may have been related to processes accompanying the WT phenomenon while not being part of the definitive mechanism generating it. For instance, identification with the teams may have created a context of affective preferences without which WT would not have arisen. That we found differential activation in occipital (inferior occipital and fusiform gyri) and parietal areas (superior parietal lobule and precuneus) could further be a sign of selective attention to the information presented (cf Gerlach et al ., 1999 ; Sturm et al ., 2006 ) in the service of these affective preferences (e.g. McDermott et al ., 1999 ).

Limbic (i.e. areas in the anterior and posterior cingulate cortex) and dorsal striatal regions—both structures of the human reward system ( Heekeren et al ., 2007 ; Delgado et al ., 2008 ; Xue et al ., 2009 )—as well as closely related structures typically implicated in emotion processes such as the insula ( Craig, 2003 ; Critchley et al ., 2004 )—were not per se differentially active for the two teams. However, they were found to be differentially coupled with the occipital and the dmPFC clusters identified to differ between the favorite and the least favorite team trials in the activation analyses. Importantly, only the degree of differential connectivities with the occipital cluster was related to the extent of WT revealed in the participants’ behavior. Therefore, our results are supportive of an attention bias rather than interpretation or response bias being responsible for WT.

Specifically, the occipital cluster differing between the favorite team and the least favorite team was positively coupled with dorsal striatum (putamen and caudate), claustrum and insula during the favorite team trials. Furthermore, the posterior cingulate cortex displayed differential (favorite vs least favorite team) connectivity with this same cluster. Differential connectivity of these areas with the fusiform cortex is in line with the idea that the anticipated reward from seeing the favorite team winning guided the participants’ visual perception and attention in the favorite team trials, and thus supportive of the hypothesis that early selective attention in the service of hedonic needs is an important mechanism in the generation of WT. Interestingly, no coupling of any brain area with the occipital cluster was observed for the least favorite team. This result suggests that the synchronization of activities in the occipital cortex and the reward system we observed for the favorite team was related to positive rather than negative distinction processes in this study (up-regulation of expectancies for the favorite team by the application of selective visual attention). 2

Also during favorite team trials, positive coupling of activity in the dorsal striatum (caudate) was observed with activity in the area in the dmPFC that distinguished favorite and least favorite team trials. Conversely, negative coupling between these areas was demonstrated for the least favorite team. Thus, this connectivity pattern is consistent with the idea of a close link between identification and reward processing in our study. That differential activity and differential connectivities for the dmPFC cluster were altogether unrelated to the extent of WT displayed by our participants, on the contrary, is clearly inconsistent with the hypothesis that WT is generated by higher level biased cognitive evidence evaluation or response formation.

Opposed to this latter finding, alternative research on optimism bias suggests higher level cognitive processes to be of particular importance, instead. According to results reported by Sharot et al . (2007) , a mechanism of neural activity within limbic and control regions plays a pivotal role. The authors demonstrated reduced activity in the amygdala and the ACC when participants imagined negative future outcomes compared with positive future outcomes as well as positive and negative past outcomes. They suggested that the ACC applies less emotional salience to anticipated negative events (in the sense of an interpretation bias), which then leads to reduced amygdala activity.

The results of Sharot et al . (2007) may appear to diverge importantly from our own data. However, it has to be taken in consideration as well, that, in our experiment in contrast to Sharot's work on optimism, participants faced a favorite and least favorite team. Thus, they did not only anticipate positive and negative events for themselves. In some cases, in our study, negative outcomes may even have been specifically desired (i.e. anticipating low chances for the least favorite team to win the game) and the emotional salience of these outcomes did not need to be reduced. Thus, we would not expect a generally increased activity in the ACC for low vs high chances of winning (irrespective of consideration of team). For the same reason, we would not expect differential ACC activity for the two teams, irrespective of anticipated chances of winning. Moreover, the concrete aims of the study may have been more transparent and consciously perceivable in the Sharot et al . (2007) study than in our own study, thus leading to the implication of higher level cognitive processes.

In sum, our data propose that the occipital cluster and its functional connectivities are specifically involved in the type of WT investigated in our study, whereas activity in the dmPFC—possibly related to social identification—may just ensure the affective context necessary for WT to arise. We outlined that the occipital cluster has been related to visual attention, its functional connections (insula, posterior cingulate, caudate, putamen) to reward processing in earlier research. In our specific case, self-relevance may have activated structures of the human reward system, which in turn synchronized with the occipital cortex to guide visual attention in order to serve individual hedonic needs.

It is important to note, however, that our interpretations concerning the direction of information transmission in the brain during WT are premature since the PPI does not allow for causal inferences. Future research should investigate the flow of information, for instance, whether supposedly attention-related areas are recruited by the limbic and striatal areas, whether the flow of information goes the other way around, or whether influences are bi-directional.

Besides, the neural correlates we identified in our study do not need to be specific to WT. One can very well imagine highly similar neural response patterns in studies that confront individuals with alternative affective-laden situations without any need to take a decision or to make future predictions (e.g. in an affective priming context). Here, selective attention may take place nonetheless and activity in the visual cortex may display similar patterns of connectivities as in the current study.

Finally, our behavioral results demonstrate that WT appears even in tasks with demand for the application of rational Bayesian statistics. Greater effects of WT might be expected when no concrete and explicit objective background information is available, which is the case in most everyday life situations. These real-life situations should give even greater space for the employment of selective attention, for instance, by the activation of memory episodes that serve the goal of a positive identity. In such cases, the reward and regulation systems may be additionally functionally connected to memory-related neural networks.

Conflict of Interest

None declared.

Acknowledgments

This research was supported by National Institute of Mental Health Grant No. P50 MH72850 and the John Templeton Foundation.

1 , Weinstein (1980) defined this same phenomenon as optimism . We use the term wishful thinking because (i) this is the literature upon which we draw in the present study, (ii) McGuire's concept of wishful thinking predated Weinstein's use of the term optimism to refer to this phenomenon and (iii) contemporary researchers often use the term optimism not to mean Weinstein's definition but to mean the dictionary definition of optimism which implies that individuals display hopefulness and confidence about the future or the successful outcome of something without explicitly referring to individual preferences or affective experiences.

2 Remember that, consistent with this observation, our behavioral data did not reveal any systematic difference between the neutral and the least favorite team either and thus suggested a preferential treatment of the favorite team without specific punishment of the least favorite team.

  • Amodio DM, Frith CD. Meeting of minds: the medial frontal cortex and social cognition. Nature Reviews Neuroscience. 2006; 7 :268–77. [ PubMed ] [ Google Scholar ]
  • Babad E. Wishful thinking and objectivity among sports fans. Social Behaviour. 1987; 2 :231–40. [ Google Scholar ]
  • Babad E, Katz Y. Wishful thinking – against all odds. Journal of Applied Social Psychology. 1991; 21 :1921–38. [ Google Scholar ]
  • Cavendini P, Riboldi G, Keller R, D'Annucci A, Bellodi L. Frontal lobe dysfunction in pathological gambling patients. Biological Psychiatry. 2002; 51 :334–41. [ PubMed ] [ Google Scholar ]
  • Craig AD. A new view of pain as a homeostatic emotion. Trends in Neuroscience. 2003; 26 :303–7. [ PubMed ] [ Google Scholar ]
  • Critchley HD, Wiens S, Rotshtein T, Öhman A, Dolan RJ. Neural systems supporting interoceptive awareness. Nature Neuroscience. 2004; 7 :189–95. [ PubMed ] [ Google Scholar ]
  • Delgado MR, Gillis MM, Phelps EA. Regulating the expectation of reward via cognitive strategies. Nature Neuroscience. 2008; 11 :880–1. [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Gerlach C, Law I, Gade A, Paulson OB. Perceptual differentiation and category effects in normal object recognition. Brain. 1999; 122 :2159–70. [ PubMed ] [ Google Scholar ]
  • Goldstein DG, Gigerenzer G. Models of ecological rationality: The recognition heuristic. Psychological Review. 2002; 109 :75–90. [ PubMed ] [ Google Scholar ]
  • Granberg D, Holmberg S, editors. The Political System Matters: Social Psychology and Voting Behavior in Sweden and the United States. Cambridge: Cambridge University Press; 1988. [ Google Scholar ]
  • Heekeren HR, Wartenburger I, Marschner A, Mell T, Villringer A, Reischies FM. Role of the ventral striatum in reward-based decision making. Neuroreport. 2007; 18 :951–5. [ PubMed ] [ Google Scholar ]
  • Kerschreiter R, Schulz-Hardt S, Mojzisch A, Frey D. Biased information search in homogeneous groups: confidence as a moderator for the effect of anticipated task requirements. Personality and Social Psychology Bulletin. 2008; 34 :679–91. [ PubMed ] [ Google Scholar ]
  • Krizan Z, Windschitl PD. The influence of outcome desirability on optimism. Psychological Bulletin. 2007; 133 :95–121. [ PubMed ] [ Google Scholar ]
  • McDermott KB, Ojemann JG, Petersen SE, Ollinger JM, Snyder AZ, Akbudak E, et al. Direct comparison of episodic encoding and retrieval of words: an event-related fMRI study. Memory. 1999; 7 :661–78. [ PubMed ] [ Google Scholar ]
  • McGuire WJ. A syllogistic analysis of cognitive relationships. In: Rosenberg MJ, Hovland CI, McGuire WJ, Abelson RP, Brehm JW, editors. Attitude Organization and Change. New Haven, CT: Yale University Press; 1960. pp. 65–111. [ Google Scholar ]
  • McLean J, Brennan D, Wyper D, Condon B, Hadley D, Cavanagh J. Localization of regions of intense pleasure response evoked by soccer goals. Psychiatry Research – Neuroimaging. 2009; 171 :33–43. [ PubMed ] [ Google Scholar ]
  • Miller EK, Cohen JD. An integrative theory of prefrontal cortex function. Annual Review of Neuroscience. 2001; 24 :167–202. [ PubMed ] [ Google Scholar ]
  • Ochsner KN, Bunge SA, Gross JJ, Gabrieli JDE. Rethinking feelings: an fMRI study of the cognitive regulation of emotion. Journal of Cognitive Neuroscience. 2002; 14 :1215–29. [ PubMed ] [ Google Scholar ]
  • Ochsner KN, Knierim K, Ludlow DH, Hanelin J, Ramachandran T, Glover G, et al. Reflecting upon feelings: an fMRI study of neural systems supporting the attribution of emotion to self and other. Journal of Cognitive Neuroscience. 2004; 16 :1746–72. [ PubMed ] [ Google Scholar ]
  • Oldfield RC. The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia. 1971; 9 :97–113. [ PubMed ] [ Google Scholar ]
  • Petty RE, Cacioppo JT, editors. Attitudes and persuasion: classic and contemporary approaches. Dubuque, IA: Wm. C. Brown; 1981. [ Google Scholar ]
  • Price PC. Wishful thinking in the prediction of competitive outcomes. Thinking and Reasoning. 2000; 6 :161–72. [ Google Scholar ]
  • Sabatinelli D, Lang PJ, Keil A, Bradley MM. Emotional perception: correlation of functional MRI and event-related potentials. Cerebral Cortex. 2007; 17 :1085–91. [ PubMed ] [ Google Scholar ]
  • Sharot T, Riccardi AM, Raio CM, Phelps EA. Neural mechanisms mediating optimism bias. Nature. 2007; 450 :102–5. [ PubMed ] [ Google Scholar ]
  • Sturm W, Schmenk B, Fimm B, Specht K, Weis S, Thron A, et al. Spatial attention: more than intrinsic alerting? Experimental Brain Research. 2006; 171 :16–25. [ PubMed ] [ Google Scholar ]
  • Tajfel H, Flament C, Billig MG, Bundy RF. Social categorization and intergroup behaviour. European Journal of Social Psychology. 1971; 1 :149–77. [ Google Scholar ]
  • Talairach J, Tournoux P, editors. Co-Planar Stereotaxic Atlas of the Human Brain: 3D Proportional System: An Approach to Cerebral Imaging. New York, NY: Georg Thieme Verlag; 1988. [ Google Scholar ]
  • Turner JC, Onorato RS. Social identity, personality, and the self concept: a self categorization perspective. In: Tyler TR, Kramer RM, John OP, editors. The Psychology of the Social Self. Mahwah, NJ: Erlbaum; 1999. pp. 11–46. [ Google Scholar ]
  • Ward BD, editor. Deconvolution Analysis of fMRI Time Series Data (Technical Report) Milwaukee, WI: Biophysics Research Institute, Medical College of Wisconsin; 2001. [ Google Scholar ]
  • Weinstein ND. Unrealistic optimism about future life events. Journal of Personality and Social Psychology. 1980; 39 :806–820. [ Google Scholar ]
  • World Medical Association. Proposed revision of the Declaration of Helsinki. Bulletin of Medical Ethics. 1999; 147 :18–22. [ PubMed ] [ Google Scholar ]
  • Xue G, Lu ZL, Levin IP, Weller JA, Li XR, Bechara A. Functional dissociations of risk and reward processing in the medial prefrontal cortex. Cerebral Cortex. 2009; 19 :1019–27. [ PMC free article ] [ PubMed ] [ Google Scholar ]
  • Zábojník J. A model of rational bias in self-assessments. Economic Theory. 2004; 23 :259–82. [ Google Scholar ]

Logo

Heuristic Encyclopedia

Find everything you wanted to know about a heuristic

Our brain is a wondrous thing, but it's certainly not perfect. It uses mental shortcuts or heuristics to make 95% of decisions which can sometimes lead you to make erroneous choices. Heuristic hacks will help you avoid these common decision pitfalls, both at home and at work.

wishful thinking research definition

Table of Content

Learn about more heuristics and hacks:, accountability bias.

How Accountability Bias influences decision-making when held accountable for decisions

wishful thinking research definition

Heuristic Science Workshop

Explore the endless possibilities of decision heuristic science for your organization.

wishful thinking research definition

Please enter your details below to get a free heuristic workshop

Thank you for your interest in our free message optimization workshop. we'll be in touch with you soon..

Psychology For

Wishful Thinking: What It Is And How ‘wishful Thinking’ Affects Us

wishful thinking

Most of the time, desires are subordinated to reality. It is impossible to think that it is sunny – no matter how much we long for it – when we are walking in the rain. However, the theory of wishful thinking suggests that when we defend an idea we believe in, the boundaries of reality begin to disappear.

People cannot believe the most implausible things simply because they want to. Wishful thinking in English refers to the type of thinking that uses emotions and decides based on what would be more pleasant to imagine instead of using evidence or rationality as a basis to determine actions or opinions.

In this article we explain What does illusory thinking consist of? its foundations, as well as its main consequences when making decisions.

Table of Contents

What is wishful thinking?

It is said that human beings are rational beings; However, when deciding or forming opinions, a multitude of cognitive biases operate.

Wishful thinking refers, in psychology, to the process of forming opinions and making decisions. using desires and emotions, instead of relying on facts or rationality Illusionary thinking is based on emotions and is a product of the conflict between desire and reality.

Research suggests that if the situation remains static, people tend to imagine a positive resolution to events; This is known as optimism bias. Otherwise, for example, if some type of threat or negative event occurs, the reverse situation occurs and people tend to imagine a negative resolution to the situation.

In the case of illusory thinking, the subject only contemplates the arguments and events that support his opinion, he stops observing contrary evidence. Consequently, this type of decision-making process is considered to be based mainly on emotions.

Wishful thinking

However, even if it lacks rationality, sometimes favorable consequences can be derived from this cognitive bias. The so-called “Pygmalion effect” describes how illusory thinking can positively influence decision making, and help get better results

Forming accurate opinions based on evidence or rationality is a fundamental process when deciding, whether we are talking about politics, business strategy or any type of interaction that involves negotiating.

As Melnikoff asks in his article on motivation bias : When a lawyer defends his client or when a manager defends a new business strategy, are they limited by reality? Or does the objective make reality more easily distorted?

A concrete example of wishful thinking could be Tarot predictions. A person encouraged by a positive reading of cards, in which he is told that he is about to meet the love of his life, could suggest himself in different ways to make this happen, go out more than normal, be kinder to people. people you know etc. Although this does not mean that the relationship will turn out well, we can observe the effect on behavior that delusional thinking can produce

The wishful thinking process suggests that, when deciding, people will only value options that produce a positive outcome. At the same time, those possibilities in which the consequences are negative will be denied or not processed.

For this reason, wishful thinking is considered a form of bias and an inadequate process when guiding our behavior. This also implies a logical fallacy, in which one believes that something will be true simply by desiring it

As we see, numerous studies suggest that beliefs can be distorted if we have pre-established objectives, although current theories affirm that what is known as motivated bias, that is, the beliefs imposed by our desires, disappear when reality is firm.

Fundamentals of wishful thinking

Behind illusory thinking is imagination and desire, where evidence and reality are ignored to meet pre-established objectives or obtain the desired result.

Imagination is a human capacity that allows us to create or draw situations that are not real, that is, to make a representation in the absence of existing objects or situations. Although this capacity is not negative per se, since it is not only the foundation of artistic creation, it can also help us in the context of decision-making, for example, allowing us to imagine possible scenarios and anticipate results in different contexts. However, in wishful thinking, Not only is the imagination used, but the force of objective desire causes reality to be ignored and the different evidence, which can lead to a series of unnecessary risks.

It is in this process of denial where illusion intervenes above all. This manifests itself as a wrong interpretation of an event, but it does not cease to have a relationship with reality. An example that shows this correlation are mirages: the person dying of thirst in the desert will see an oasis of water. The desire, or in this case the need for water, will make us see the oasis.

Illusions, in some cases, are difficult to recognize , since they can take the form of reality. If we analyze it, when we buy the lottery we have a ticket. Therefore, it is important when deciding and making decisions or setting objectives, to be as realistic as possible and to have the greatest amount of evidence. Too much wishful thinking can lead to frustration and discouragement; On the contrary, a series of realistic goals will allow us to maintain motivation and, in the long run, better self-esteem.

Consequences of wishful thinking

The illusory thinking process involves denying evidence and rationality, therefore, it lacks objectivity. This most of the time leads to making wrong decisions with negative consequences for the subject.

There are different pseudosciences that are based on illusory thinking and the power of the mind as a method to achieve our goals. The placebo effect could also be a kind of wishful thinking, the positive effect of which has been demonstrated. However, the negative side of thinking this way could cause us not to go to the doctor or undergo necessary treatment, because we think that we can cure ourselves only by wanting to do so.

Another case of wishful thinking would have to do with our true capabilities Let’s imagine a person who wants to be a great concert artist and spends hours and hours imagining his performances and how he receives a standing ovation from the audience at the end of the performance. However, when push comes to shove he dedicates very few hours to what would really make him a better pianist, rehearsing.

As we see, this type of thinking can have negative consequences for self-esteem in the long run; since only with imagination and desire the person will never achieve their objectives and goals. Realistic thinking would involve analyzing what real abilities you have and how many hours and years of study are needed to become a great piano professional.

But, Why does our mind deceive us? Why is it due to irrational thoughts and cognitive biases? Apparently, the explanation would be in evolution; When making decisions, our brain cannot take into account all the necessary parameters. This way of acting was very beneficial for our ancestors, since they lived surrounded by external threats and it was important to quickly activate escape mechanisms.

However, we no longer live in an environment to which we have to continually adapt. And this way of thinking, rather than helping us, seems to limit us when it comes to making certain informed decisions; opting for our desires, instead of reality.

In conclusion, wishful thinking is based on imagination and desire , and although it may be positive in some specific case, as in the case of the “Pygmalion effect”, in the long run this decision process and belief formation affects our self-esteem and self-concept. Making decisions based on emotion and imagination, without taking reality into account, can cause frustration and other unpleasant feelings.

Related Posts:

The 9 Types of Thinking and Their Characteristics

  • Reference work entry
  • First Online: 01 January 2023
  • pp 1763–1772
  • Cite this reference work entry

wishful thinking research definition

  • Vlad P. Glăveanu 2 , 3  

169 Accesses

This entry proposes a sociocultural approach to wishes and wish-making, distinct from existing psychoanalytical and cognitive discussions of this phenomenon. It considers wish-making as part and parcel of everyday life and one of the main ways we have to engage with the future and attempt to construct it according to our needs and desires. Wishes are not signs of intrinsic irrationality, nor markers of pathological functioning; instead, they are embedded in human action and connect us to the possible, to others, and to the world. The entry starts by reviewing research into different types of wishes before considering in more detail theories and studies of wish-fulfilment and wishful thinking. It concludes with a reinterpretation of wishes as having their own scope, rationality, and as being intrinsically related to the ethical.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save.

  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
  • Available as EPUB and PDF
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

wishful thinking research definition

Husserl’s Phenomenology of Wishing

Aue, T., Nusbaum, H. C., & Cacioppo, J. T. (2012). Neural correlates of wishful thinking. Social Cognitive and Affective Neuroscience, 7 (8), 991–1000.

Article   Google Scholar  

Babad, E. (1987). Wishful thinking and objectivity among sports fans. Social Behavior, 4 , 231–240.

Google Scholar  

Babad, E. (1997). Wishful thinking among voters: Motivational and cognitive influences. International Journal of Public Opinion Research, 9 (2), 105–125.

Babad, E., & Katz, Y. (1991). Wishful thinking – Against all odds. Journal of Applied Social Psychology, 21 (23), 1921–1938.

Babad, E., & Yacobos, E. (1993). Wish and reality in voters’ predictions of election outcomes. Political Psychology, 14 (1), 37–54.

Babad, E., Hills, M., & O’Driscoll, M. (1992). Factors influencing wishful thinking and predictions of election outcomes. Basic and Applied Social Psychology, 13 (4), 461–476.

Bar-Hillel, M., & Budescu, D. (1995). The elusive wishful thinking effect. Thinking & Reasoning, 1 (1), 71–103.

Barnow, S., & Linden, M. (2002). Psychosocial risk factors of the wish to be dead in the elderly. Fortschritte der Neurologie–Psychiatrie, 70 (4), 185–192.

Brown, G. K., Steer, R. A., Henriques, G. R., & Beck, A. T. (2005). The internal struggle between the wish to die and the wish to live: A risk factor for suicide. American Journal of Psychiatry, 162 (10), 1977–1979.

Colace, C. (2013). Are the wish-fulfillment dreams of children the royal road for looking at the functions of dreams? Neuropsychoanalysis, 15 (2), 161–175.

Coulter, R. (2013). Wishful thinking and the abandonment of moving desires over the life course. Environment and Planning A, 45 (8), 1944–1962.

Darlington, A. S., Heule, F., & Passchier, J. (2013). Granting wishes: Parents’ perception of a wish fulfilment for a child with a life-threatening illness. Acta Paediatrica, 102 (10), 852–828.

Eagle, M. (1990). The concepts of need and wish in self psychology. Psychoanalytic Psychology, 7 (S), 71–88.

Ewing, B. (2009). Wish fulfillment: Palliative care and end-of-life intervention. Pediatric Nursing, 35 (2), 81–85.

Freud, S. (1900). The interpretation of dreams . Anthological translation by A. A. Brill, 1913. New York: Macmillan.

Galinsky, J. (2014). What do families, staff, and stakeholders think about children’s dream fulfilment? Issues in Comprehensive Pediatric Nursing, 37 (4), 203–211.

Gerson, M. J. (1984). Feminism and the wish for a child. Sex Roles, 11 (5–6), 389–399.

Ghent, E. (2002). Wish, need, drive: Motive in the light of dynamic systems theory and Edelman’s selectionist theory. Psychoanalytic Dialogues, 12 (5), 763–808.

Gordon, R., Franklin, N., & Beck, J. (2005). Wishful thinking and source monitoring. Memory & Cognition, 33 (3), 418–429.

Harpaz, I. (2002). Expressing a wish to continue or stop working as related to the meaning of work. European Journal of Work and Organizational Psychology, 11 (2), 177–198.

Heckhausen, H., & Kuhl, J. (1985). From wishes to action: The dead ends and short cuts on the long way to action. In M. Frese & J. Sabini (Eds.), Goal directed behavior: The concept of action in psychology (pp. 134–159). Hillsdale: Erlbaum.

Jeker, L., Micioni, G., Ruspa, M., Zeeb, M., & Campana, A. (1988). Wish for a child and infertility: Study on 116 couples. I. Interview and psychodynamic hypotheses. International Journal of Fertility, 33 (6), 411–420.

Johannessen, K. B., Oettingen, G., & Mayer, D. (2012). Mental contrasting of a dieting wish improves self-reported health behaviour. Psychology & Health, 27 (Suppl 2), 43–58.

Konijn, E. A., Nije Bijvank, M., & Bushman, B. J. (2007). I wish I were a warrior: The role of wishful identification in the effects of violent video games on aggression in adolescent boys. Developmental Psychology, 43 (4), 1038–1044.

Kovacs, M., & Beck, A. T. (1977). The wish to die and the wish to live in attempted suicides. Journal of Clinical Psychology, 33 (2), 361–365.

Krizan, Z., & Windschitl, P. D. (2009). Wishful thinking about the future: Does desire impact optimism? Social and Personality Psychology Compass, 3 (3), 227–243.

Krizan, Z., Miller, J. C., & Johar, O. (2010). Wishful thinking in the 2008 US presidential election. Psychological Science, 21 (1), 140–146.

Kruglanski, A. W., Jasko, K., & Friston, K. (2020). All thinking is “wishful” thinking. Trends in Cognitive Sciences . https://doi.org/10.1016/j.tics.2020.03.004 .

Lang, P. J., & Bradley, M. M. (2010). Emotion and the motivational brain. Biological Psychology, 84 (3), 437–450.

Leshem, O. A. (2017). What you wish for is not what you expect: Measuring hope for peace during intractable conflicts. International Journal of Intercultural Relations, 60 , 60–66.

Mele, A. R. (1984). Aristotle’s wish. Journal of the History of Philosophy, 22 (2), 139–156.

Miceli, M., & Castelfranchi, C. (2010). Hope: The power of wish and possibility. Theory & Psychology, 20 (2), 251–276.

Neuman, R., Rafferty, A., & Griffiths, T. (2014). A bounded rationality account of wishful thinking. In Proceedings of the annual meeting of the Cognitive Science Society (Vol. 36, No. 36), University of California, San Diego.

Orwell, G. (1945). London letter. Partisan Review, 12 , 467–472.

Parker, P. D., Thoemmes, F., Duineveld, J. J., & Salmela-Aro, K. (2015). I wish I had (not) taken a gap-year? The psychological and attainment outcomes of different post-school pathways. Developmental Psychology, 51 (3), 323–333.

Price, P. C. (2000). Wishful thinking in the prediction of competitive outcomes. Thinking & Reasoning, 6 (2), 161–172.

Rikxin, F., & White, W. A. (1913). Wishfulfillment and symbolism in fairy tales. Psychoanalytic Review, 1 (1), 94–107.

Schneider, W. (1998). Performance prediction in young children: Effects of skill, metacognition and wishful thinking. Developmental Science, 1 (2), 291–297.

Seeman, W., & Buck, R. (1952). On a behavioristic approach to the concept of wish-fulfillment. Journal of Abnormal and Social Psychology, 47 (1), 17–24.

Seybert, N., & Bloomfield, R. (2009). Contagion of wishful thinking in markets. Management Science, 55 (5), 738–751.

Sigall, H., Kruglanski, A., & Fyock, J. (2000). Wishful thinking and procrastination. Journal of Social Behavior and Personality, 15 (5), 283–296.

Stein, N., & Sanfilipo, M. (1985). Depression and the wish to be held. Journal of Clinical Psychology, 41 (1), 3–9.

Szabados, B. (1973). Wishful thinking and self-deception. Analysis, 33 (6), 201–205.

Tiggemann, M., Gardiner, M., & Slater, A. (2000). “I would rather be size 10 than have straight A’s”: A focus group study of adolescent girls’ wish to be thinner. Journal of Adolescence, 23 (6), 645–660.

Trope, Y., Gervey, B., & Liberman, N. (1997). Wishful thinking from a pragmatic hypothesis-testing perspective. In M. S. Myslobodsky (Ed.), The mythomanias: The nature of deception and self-deception (pp. 105–131). Mahwah: Lawrence Erlbaum.

Uhlaner, C. J., & Grofman, B. (1986). The race may be close but my horse is going to win: Wish fulfillment in the 1980 presidential election. Political Behavior, 8 (2), 101–129.

Vikan, A., & Clausen, S. E. (1993). Freud, Piaget, or neither? Beliefs in controlling others by wishful thinking and magical behavior in young children. The Journal of Genetic Psychology, 154 (3), 297–314.

Vosgerau, J. (2010). How prevalent is wishful thinking? Misattribution of arousal causes optimism and pessimism in subjective probabilities. Journal of Experimental Psychology: General, 139 (1), 32–48.

Watson, J. B. (1916). The psychology of wish fulfilment. The Scientific Monthly, 3 (5), 479–487.

Wellman, H. M., & Woolley, J. D. (1990). From simple desires to ordinary beliefs: The early development of everyday psychology. Cognition, 35 (3), 245–275.

Download references

Author information

Authors and affiliations.

School of Psychology, Dublin City University, Dublin, Ireland

Vlad P. Glăveanu

Centre for the Science of Learning and Technology, University of Bergen, Bergen, Norway

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Vlad P. Glăveanu .

Editor information

Editors and affiliations.

Dublin City University, Dublin, Ireland

Vlad Petre Glăveanu

Section Editor information

Marconi Institute for Creativity (MIC) Lab, University of Bologna, Bologna, Italy

Serena Mastria

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this entry

Cite this entry.

Glăveanu, V.P. (2022). Wish. In: Glăveanu, V.P. (eds) The Palgrave Encyclopedia of the Possible. Palgrave Macmillan, Cham. https://doi.org/10.1007/978-3-030-90913-0_179

Download citation

DOI : https://doi.org/10.1007/978-3-030-90913-0_179

Published : 26 January 2023

Publisher Name : Palgrave Macmillan, Cham

Print ISBN : 978-3-030-90912-3

Online ISBN : 978-3-030-90913-0

eBook Packages : Behavioral Science and Psychology Reference Module Humanities and Social Sciences Reference Module Business, Economics and Social Sciences

Share this entry

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Publish with us

Policies and ethics

  • Find a journal
  • Track your research

SEP home page

  • Table of Contents
  • Random Entry
  • Chronological
  • Editorial Information
  • About the SEP
  • Editorial Board
  • How to Cite the SEP
  • Special Characters
  • Advanced Tools
  • Support the SEP
  • PDFs for SEP Friends
  • Make a Donation
  • SEPIA for Libraries
  • Entry Contents

Bibliography

Academic tools.

  • Friends PDF Preview
  • Author and Citation Info
  • Back to Top

Self-Deception

Virtually every aspect of self-deception, including its definition and paradigmatic cases, is a matter of controversy among philosophers. Minimally, self-deception involves a person who (a)  as a consequence of some motivation or emotion, seems to acquire and maintain some false belief despite evidence to the contrary and (b) who may display behavior suggesting some awareness of the truth. Beyond this, philosophers divide over whether self-deception is intentional, whether it involves belief or some other sub- or non-doxastic attitude, whether self-deceivers are morally responsible for their self-deception, whether self-deception is morally problematic (and if it is, in what ways and under what circumstances), whether self-deception is beneficial or harmful, whether and in what sense collectives can be self-deceived (and how, if they can be self-deceived, this might affect individuals within such collectives), and whether our penchant for self-deception might be socially, psychologically or biologically adaptive or merely an accidental byproduct of our evolutionary history.

The discussion of self-deception and its associated puzzles sheds light on the ways motivation affects belief acquisition and retention and other belief-like cognitive attitudes; it also prompts us to scrutinize the notions of belief, intention, and the limits of such folk psychological concepts to adequately explain phenomena of this sort. Self-deception also requires careful consideration of the cognitive architecture that might accommodate this apparent irrationality regarding our beliefs.

Self-deception isn’t merely a philosophically interesting puzzle but a problem of existential concern. It raises the distinct possibility that we live with distorted views that may make us strangers to ourselves and blind to the nature of our morally significant engagements.

1. Paradoxes, Puzzles, and Problems of Self-Deception

2.1 temporal partitioning, 2.2 psychological partitioning, 3.1 revision of intention: non-intentionalist and deflationary approaches, 3.2 revision of belief: adjustment of attitude or content, 4. twisted or negative self-deception, 5.1 moral responsibility for self-deception, 5.2 the morality of self-deception, 6. origin of self-deception: adaptation or spandrel, 7.1 summative collective self-deception: self-deception across a collective, 7.2 non-summative collective self-deception: self-deception of a collective entity, other internet resources, related entries.

“What is self-deception?” sounds like a straightforward question, but the more philosophers have sought to answer it, the more puzzling it has become. Traditionally, self-deception has been modeled on interpersonal deception, where A intentionally gets B to believe some proposition p , all the while knowing or believing truly that ~ p . Such deception is intentional and requires the deceiver to know or believe that ~ p and the deceived to believe that p . One reason for thinking self-deception is analogous to interpersonal deception of this sort is that it helps us to distinguish self-deception from mere error since the acquisition and maintenance of the false belief are intentional, not accidental. It also helps to explain why we think self-deceivers are responsible for and open to the evaluation of their self-deception. If self-deception is properly modeled on interpersonal deception, self-deceivers intentionally get themselves to believe that p , all the while knowing or believing truly that ~ p . On this traditional model, then, self-deceivers apparently must (1) hold contradictory beliefs—the dual-belief requirement—and (2) intentionally get themselves to hold a belief they know or believe truly to be false.

The traditional model of self-deception, however, has been thought to raise two paradoxes: One concerns the self-deceiver’s state of mind—the so-called static paradox. How can a person simultaneously hold contradictory beliefs? The other concerns the process or dynamics of self-deception—the so-called dynamic or strategic paradox. How can a person intend to deceive herself without rendering her intentions ineffective? (Mele 1987a; 2001)

The dual-belief requirement raises the static paradox since it seems to pose an impossible state of mind, namely, consciously and simultaneously believing that p and ~ p . As deceiver, she believes that ~ p , and, as deceived, she believes that p . Accordingly, the self-deceiver consciously believes that p and ~ p . But if believing both a proposition and its negation in full awareness is an impossible state of mind, self-deception, as it has traditionally been understood, seems impossible as well. Static paradoxes also arise regarding motivation, intention, emotion, and the like insofar as self-deceivers seem to harbor psychological states of these kinds that seem deeply incompatible (Funkhouser 2019).

The requirement that the self-deceiver intentionally gets herself to hold a belief she knows to be false raises the dynamic or strategic paradox since it seems to involve the self-deceiver in an impossible project, namely, both deploying and being duped by some deceitful strategy. As deceiver, she must be aware that she’s deploying a deceitful strategy; but, as deceived, she must be unaware of this strategy for it to be effective. And yet it’s difficult to see how the self-deceiver could fail to be aware of her intention to deceive. A strategy known to be deceitful seems bound to fail. How could I be taken in by your efforts to get me to believe something false if I know what you’re up to? But if it’s impossible to be taken in by a strategy one knows is deceitful, then, again, self-deception, as it has traditionally been understood, seems to be impossible as well.

These paradoxes have led a minority of philosophers to be skeptical that self-deception is conceptually possible or even coherent (Paluch 1967; Haight 1980; Kipp 1980). Borge (2003) contends that accounts of self-deception inevitably give up central elements of our folk-psychological notions of “self” or “deception” to avoid paradox, leaving us to wonder whether this framework itself is what gets in the way of explaining the phenomenon. Such skepticism toward the concept may seem warranted, given the obvious paradoxes involved. Most philosophers, however, have sought some resolution to these paradoxes instead of giving up on the notion itself, not only because empirical evidence suggests that self-deception is not only possible but pervasive (Sahdra and Thagard 2003) but also because the concept does seem to pick out a distinct kind of motivated irrationality.

Philosophical accounts of self-deception roughly fall into two main groups: those that maintain that the paradigmatic cases of self-deception are intentional and those that deny this. Call these approaches intentionalist and revisionist , respectively. Intentionalists find the model of intentional interpersonal deception apt since it helps to explain the selectivity of self-deception and the apparent responsibility of self-deceivers, as well as provide a clear way of distinguishing self-deception from other sorts of motivated belief, such as wishful thinking. To avoid paradox, these approaches introduce a variety of divisions that shield the deceiving from the deceived mind. Revisionists are skeptical of these divisions and the ‘psychological exotica’ (Mele 2001) apparently needed to avoid the static and dynamic paradoxes. Instead, they argue that revision of the intention requirement, the belief requirement, or both offers a simpler account of self-deception that avoids the paradoxes raised by modeling it on intentional interpersonal deception.

2. Intentionalist Approaches

The chief problem facing intentional models of self-deception is the dynamic paradox, namely, that it seems impossible to form an intention to get oneself to believe what one currently disbelieves or believes is false. For one to carry out an intention to deceive oneself, one must know what one is doing; to succeed, one must be ignorant of this same fact. Intentionalists agree that self-deception is intentional or, at least, robustly purposive, but divide over whether it requires holding contradictory beliefs and, thus, over the specific content of the alleged intention involved (see §3.2 Revision of Belief ). Insofar as even the bare intention to acquire the belief that p for reasons having nothing to do with one’s evidence for p seems unlikely to succeed if directly known, most intentionalists introduce some sort of temporal or psychological partition to insulate self-deceivers from their deceptive stratagems. When self-deceivers are not consciously aware of what they truly believe or intend, it’s easier to see how they can play the role of the deceiver and the deceived. By dividing the mind into parts, temporally or psychologically, these approaches seek to show that self-deception does not involve paradox.

Some intentionalists argue that self-deception is a complex, temporally extended process during which a self-deceiver can consciously set out to deceive herself that p , knowing or believing that ~ p , and along the way lose her belief that ~ p , either forgetting her original deceptive intention entirely or regarding it as having, albeit accidentally, brought about the true belief she would have arrived at anyway (Sorensen 1985; Bermúdez 2000). So, for instance, an official involved in some illegal behavior might destroy any records of this behavior and create evidence that would cover it up (diary entries, emails, and the like), knowing that she will likely forget having done these things over the next few months. When her activities are investigated a year later, she has forgotten her tampering efforts and, based upon her falsified evidence, comes to believe falsely that she was not involved in the illegal activities of which she is accused. Here, the self-deceiver need never simultaneously hold contradictory beliefs even though she intends to bring it about that she believes that p , which she regards as false at the outset of the process of deceiving herself and true at its completion.

The self-deceiver need not even forget her original intention to deceive. Take an atheist who sets out to get herself to believe in God because it seems the best bet if God turns out to exist. She might well remember such an intention at the end of the process and deem that by God’s grace even this misguided path led her to the truth. What enables the intention to succeed in such cases is the operation of what Johnston (1988) terms ‘autonomous means’ (e.g., the normal degradation of memory, the tendency to believe what one practices, etc.), not the continued awareness of the intention, hinting that the process may be subintentional (See §3.1 Revision of Intention ).

While such temporal partitioning accounts appear to avoid the static and dynamic paradoxes, many, if not most, cases of self-deception aren’t of this temporally extended type. Regularly, self-deception seems to occur instantaneously (Jordan 2022), as when a philosopher self-deceives that her article is high quality even while reading the substantive and accurate criticisms in the rejection letter from the prestigious peer-reviewed journal she submitted it to (example due to Mele 2001). Additionally, many of these temporally extended cases lack the distinctive opacity, indirection, and tension associated with garden-variety cases of self-deception (Levy 2004).

Another strategy employed by intentionalists is the division of the self into psychological parts that play the role of the deceiver and deceived, respectively. These strategies range from positing strong division in the self, where the deceiving part is a relatively autonomous subagency capable of belief, desire, and intention (Rorty 1988), to more moderate division, where the deceiving part still constitutes a separate center of agency (Pears 1984, 1986; 1991), to the relatively modest division of Davidson (1982, 1986), where there need only be a boundary between conflicting attitudes and intentions.

Such divisions are prompted in large part by the acceptance of the dual-belief requirement. It isn’t simply that self-deceivers hold contradictory beliefs, which though strange, isn’t impossible since one can believe that p and believe that ~ p without believing that p & ~ p . The problem such theorists face stems from the appearance that the belief that ~ p motivates and thus forms a part of the intention to bring it about that one acquires and maintains the false belief that p (Davidson 1986). So, for example, the Nazi official’s recognition that his actions implicate him in serious evil motivates him to implement a strategy to deceive himself into believing he is not so involved; he can’t intend to bring it about that he holds such a false belief if he doesn’t recognize it is false, and he wouldn’t want to bring such a belief about if he didn’t recognize the evidence to the contrary. So long as this is the case, the deceiving self, whether it constitutes a separate center of agency or something less robust, must be hidden from the conscious self being deceived if the self-deceptive intention is to succeed.

While these psychological partitioning approaches seem to resolve the static and dynamic puzzles, they do so by introducing a picture of the mind that raises puzzles of its own. On this point, there appears to be consensus even among intentionalists that self-deception can and should be accounted for without invoking speculative or stipulative divisions not already used to explain non-self-deceptive behavior, what Talbott (1995) calls ‘innocent’ divisions. That said, recent, if controversial, research (e.g., Bargh and Morsella 2008; Hassin, Bargh, and Zimmerman 2009; Huang and Bargh 2014) seems to support the possibility of the sort of robust unconscious but flexible goal pursuit that could explain the way self-deceivers are able to pursue their deceptive goal while retaining the beliefs necessary for navigating the shifting evidential terrain (see Funkhouser and Barrett 2016, 2017 and Doody 2016 for skepticism about the applicability of this research to self-deception). If this kind of research shows that no ‘psychological exotica’ are necessary to explain self-deception, there is less pressure to deflate the phenomenon in ways that minimize the active, strategic role self-deceivers play in the process.

3. Revisionist Approaches

A number of philosophers have moved away from modeling self-deception directly on intentional interpersonal deception, opting instead to revise either the intention or the belief requirement traditional intentionalist models assume. Those revising the intention requirement typically treat self-deception as a species of motivationally biased belief, thus avoiding the problems involved with intentionally deceiving oneself. Call these non-intentionalist and deflationary approaches .

Those revising the belief requirement do so in a variety of ways. Some posit other, non-doxastic or quasi-doxastic attitudes toward the proposition involved (‘misrepresentation’ Jordan 2020, 2022; ‘hope,’ ‘suspicion,’ ‘doubt,’ ‘anxiety’ Archer 2013; ‘besires’ Egan 2009; ‘pretense’ Gendler 2007; ‘imagination’ Lazar 1999). Others alter the content of the proposition believed (Holton 2001; Funkhouser 2005; Fernández 2013), while others suggest the doxastic attitudes involved are indeterminate, somehow ‘in-between believing’ (Schwitzgebel 2001; Funkhouser 2009) or subject to shifting degrees of credulity throughout the process of self-deception (Chan and Rowbottom 2019). Call these revision of belief approaches.

Deflationary approaches focus primarily on the process of self-deception, while the revision of belief approaches focus on the product. A revision of either of these aspects, of course, has ramifications for the other. For example, if self-deception doesn’t involve belief, but some other non-doxastic attitude (product), then one may well be able to intentionally enter that state without paradox (process). This section considers non-intentional and deflationary approaches and the worries such approaches raise ( §3.1 ). It also considers revision of belief approaches ( §3.2 ).

Non-intentionalists argue that most ‘garden-variety’ cases of self-deception can be explained without adverting to subagents, or unconscious beliefs and intentions, which, even if they resolve the static and dynamic puzzles of self-deception, raise puzzles of their own. If such non-exotic explanations are available, intentionalist explanations seem unwarranted and unnecessary.

Since the central paradoxes of self-deception arise from modeling self-deception on intentional interpersonal deception, non-intentionalists suggest this model be jettisoned in favor of one that takes ‘to be deceived’ to be nothing more than believing falsely or being mistaken in believing something (Johnston 1988; Mele 2001). For instance, Sam mishears that it will be a sunny day and relays this misinformation to Joan with the result that she believes it will be a sunny day. Joan is deceived into believing it will be sunny, and Sam has deceived her, albeit unintentionally. Initially, such a model may not appear promising for self-deception since simply being mistaken about p or accidentally causing oneself to be mistaken about p doesn’t seem to be self-deception at all but some sort of innocent error. Non-intentionalists, however, argue that in cases of self-deception, the false belief is not accidental but, rather, motivated by desire (Mele 2001), emotion (Lazar 1999), anxiety (Johnston 1988; Barnes 1997), or some other attitude regarding p or related to p . So, for instance, when Allison believes, against the preponderance of evidence available to her, that her daughter is not having learning difficulties, non-intentionalists will explain the various ways she misreads the evidence by pointing to such things as her desire that her daughter not have learning difficulties, her fear that she has such difficulties, or anxiety over this possibility. In such cases, Allison’s self-deceptive belief that her daughter is not having learning difficulties fulfills her desire, quells her fear, or reduces her anxiety, and it’s this function—not an intention—that explains why her belief formation process is biased. Allison’s false belief is not an innocent mistake but a consequence of her motivational states.

Non-intentionalists divide over the dual-belief requirement. Some accept the requirement, seeing the persistent efforts to resist the conscious recognition of the unwelcome truth or to reduce the anxiety generated by this recognition as characteristic of self-deception (Bach 1981; Johnston 1988). So, in Allison’s case, her belief that her daughter is having learning difficulties, along with her desire that this not be the case, motivates her to employ means to avoid this thought and to believe the opposite.

Others, however, argue the needed motivation can as easily be supplied by uncertainty or ignorance whether p , or suspicion that ~ p (Mele 2001; Barnes 1997). Thus, Allison need not hold any opinion regarding her daughter’s having learning difficulties for her false belief to count as self-deception since it’s her regarding evidence in a motivationally biased way in the face of evidence to the contrary, not her recognition of this evidence, that makes her belief self-deceptive. Accordingly, Allison needn’t intend to deceive herself nor believe at any point that her daughter, in fact, has learning difficulties. If we think someone like Allison is self-deceived, then self-deception requires neither contradictory beliefs nor intentions regarding the acquisition or retention of the self-deceptive belief. Such approaches are ‘deflationary’ in the sense that they take self-deception to be explicable without reaching for exotic cognitive architecture since neither intentions nor dual-beliefs are required to account for self-deception. (For more on the dual-belief requirement, see §3.2 Revision of Belief .)

Mele (2001, 2012) has offered the most fully articulated deflationary account, and his view has been the target of the most scrutiny, so it is worth stating what he takes to be the jointly sufficient conditions for entering self-deception:

  • The belief that p , which S acquires, is false
  • S treats data relevant, or at least seemingly relevant, to the truth value of p in a motivationally biased way
  • This biased treatment is a nondeviant cause of S’ s acquiring the belief that p
  • The body of data possessed by S at the time provides greater warrant for ~ p than for p
  • S consciously believes at the time that there is a significant chance that ~ p
  • S ’s acquiring the belief that p is a product of “reflective, critical reasoning,” and S is wrong in regarding that reasoning as properly directed

Mele (2012) added the last two conditions to clarify his account in view of some of the criticisms of this kind of approach addressed in the next section.

To support non-intentionalism, some have looked to the purposive mechanisms of deception operating in non-human organisms as a model (Smith 2014), while others have focused on neurobiological mechanisms triggered by affect to explain the peculiarly purposive responses to evidence involved in self-deception (Lauria et al . 2016).

Deflationary Worries and Modifications

Critics contend these deflationary accounts do not adequately distinguish self-deception from other sorts of motivated believing (such as wishful thinking), nor can they explain the peculiar selectivity associated with self-deception, its characteristic ‘tension,’ the way it involves a failure of self-knowledge, or the agency of the self-deceiver.

Self-Deception and Wishful Thinking : What distinguishes wishful thinking from self-deception, according to intentionalists, just is that the latter is intentional while the former is not (Bermúdez 2000). Specifically, wishful thinking does not seem ‘deceptive’ in the requisite sense. Non-intentionalists respond that what distinguishes wishful thinking from self-deception is that self-deceivers recognize evidence against their self-deceptive belief whereas wishful thinkers do not (Bach 1981; Johnston 1988) or that they merely possess, without recognizing it, greater counterevidence than wishful thinkers (Mele 2001). In either case, self-deceivers exert more agency than wishful thinkers over their belief formation. In wishful thinking, motivation triggers a belief formation process in which the person does not play an active, conscious role, “while in self-deception the subject is a willing participant in directing cognition towards the doxastic embrace of the favored proposition” (Scott-Kakures 2002; see also Szabados 1973). While the precise relationship between wishful thinking and self-deception is clearly a matter of debate, non-intentionalists offer plausible ways of distinguishing the two that do not invoke the intention to deceive.

Self-Deception and Selectivity : Another worry—termed the ‘selectivity problem’—originally raised by Bermúdez (1997, 2000) is that deflationary accounts don’t seem to be able to explain the selective nature of self-deception (i.e., why motivation seems only selectively to produce bias). Why is it, such intentionalists ask, that we are not rendered biased in favor of the belief that p in many cases where we have a very strong desire that p (or anxiety or some other motivation related to p )? Intentionalists argue that an intention to get oneself to acquire the belief that p offers the most straightforward answer to this question.

Others, following Mele (2001, 2012, 2020), contend that selectivity may be explained in terms of the agent’s assessment of the relative costs of erroneously believing that p or ~ p (see Friedrich (1993) and Trope and Lieberman (1996) on lay hypothesis testing). Essentially, this approach suggests that the minimization of costly errors is the central principle guiding hypothesis testing. So, for example, Josh would be happier believing falsely that the gourmet chocolate he finds so delicious isn’t produced by exploited farmers than falsely believing that it is since he desires that it not be so produced. Because Josh considers the cost of erroneously believing his favorite chocolate is tainted by exploitation to be very high—no other chocolate gives him the same pleasure—it takes a great deal more evidence to convince him that his chocolate is so tainted than it does to convince him otherwise. It’s the low subjective cost of falsely believing the chocolate is not tainted that facilitates Josh’s self-deception. But we can imagine Josh having the same strong desire that his chocolate not be tainted by exploitation and yet having a different assessment of the cost of falsely believing it’s not tainted. Say, for example, he works for an organization promoting fair trade and non-exploitive labor practices among chocolate producers and believes he has an obligation to accurately represent the labor practices of the producer of his favorite chocolate and would, furthermore, lose credibility if the chocolate he himself consumes is tainted by exploitation. In these circumstances, Josh is more sensitive to evidence that his favorite chocolate is tainted—despite his desire that it not be—since the subjective cost of being wrong is higher for him than it was before. It is the relative, subjective costs of falsely believing p and ~ p that explain why desire or other motivation biases belief in some circumstances and not others.

While error-cost accounts offer some explanation of selectivity, one might still complain that even when these conditions are met, self-deception needn’t follow (Bermúdez 2000, 2017). Some non-intentionalist respond that these conditions are as complete an explanation of self-deception as is possible. Given the complexity of the factors affecting belief formation and the lack of a complete account of its etiology, we shouldn’t expect a complete explanation in each case (Funkhouser 2019; Mele 2020). Others argue that attention to the role of emotion in assessing and filtering evidence sheds further light on the process. According to such approaches, affect plays a key role in triggering the conditions under which motivation leads to self-deception (Galeotti 2016a; Lauria et al. 2016; Lauria and Preissmann 2018). Our emotionally-loaded appraisal of evidence, in combination with evidential ambiguity and our potential to cope with the threatening reality, helps to explain why motivation tips us toward self-deception. Research on the role of dopamine regulation and negative somatic markers provides some empirical support for this sort of affective model (Lauria et al. 2016; Lauria and Preissmann 2018).

Non-intentionalists also point out that intentionalists have selectivity problems of their own since it isn’t clear why intentions are formed in some cases rather than others (Jurjako 2013) or why some intentions to acquire a self-deceptive belief succeed while others do not (Mele 2001, 2020) See Bermúdez (2017) for a response to these worries.

Self-Deception and Tension : A number of philosophers have complained that deflationary accounts fail to explain certain ‘tensions’ or conflicts supposed to be present in cases of genuine self-deception (Audi 1997; Bach 1997; Nelkin 2002; Funkhouser 2005; Fernández 2013; Jordan 2022). Take Ellen, who says she is doing well in her biology class but systematically avoids looking at the results on her quizzes and tests. She says she doesn’t need to look; she knows she didn’t miss anything. When her teacher tries to catch her after class to discuss her poor grades, she rushes off. Similarly, when she sees an email from her teacher with the subject line “your class performance,” she ignores it. The prospect of looking at the test results, talking with her teacher, or reading her email sends a flash of dread through her and a pit to her stomach, even as she projects calm and confidence. Ellen’s behavior and affective responses suggest to these critics that she knows she isn’t doing well in her biology class, despite her avowals to the contrary. Ellen’s case highlights a variety of tensions that may arise in the self-deceived. Philosophers have focused on tensions that arise with respect to evidence (since there is a mismatch between what the evidence warrants and what is believed or avowed); to unconscious doxastic attitudes (since they are at variance with those consciously held); to self-knowledge (since one has a false second-order belief about what one believes); to the various components of belief—behavioral, verbal, emotional, physical—(since what one does, says, feels, or experiences can come apart in a number of ways) (Funkhouser 2019); and to authorship (since one may be aware of authoring her self-deception and presenting her self-deception to herself as not having been authored, i.e., as the truth) (Jordan 2022). These tensions seem to be rooted in the self-deceiver’s awareness of the truth. But, since deflationary accounts deny the dual-belief requirement, specifically, that self-deceivers must hold the true belief that ~p , it’s not clear why self-deceivers would experience tension or display behaviors in conflict with their self-deceptive belief p .

Deflationary theorists counter that suspicion that ~ p , thinking there’s a significant possibility that ~ p, may suffice to explain these kinds of tensions (Mele 2001, 2009, 2010, 2012). Clearly, a person who self-deceptively believes that p and suspects that ~p may experience tension; moreover, such attitudes combined with a desire that p might account for the behaviors highlighted by critics.

While these attitudes may explain some of the tension in self-deception, a number of critics think they are inadequate to explain deep-conflict cases in which what self-deceivers say about p is seriously at odds with non-verbal behavior that justifies the attribution of the belief that ~p to them (Audi, 1997; Patten 2003; Funkhouser 2005; Gendler 2007; Fernández 2013). While some propose these cases are just a type of self-deception that deflationary approaches cannot explain (Funkhouser 2009, 2019; Fernández 2013), others go further, suggesting these cases show that deflationary approaches aren’t accounts of self-deception at all but of self-delusion since deflationary self-deceivers seem single-mindedly to hold the false belief (Audi 2007; Funkhouser 2005; Gendler 2007).

Some defenders of deflation acknowledge that the significant difference between what deflationary accounts have in view (namely, people who do not believe the unwelcome truth that ~p , having a motivation-driven, unwarranted skepticism toward it) and what deep-conflict theorists do (namely, people who know the unwelcome truth that ~p and avoid reflecting on it or encountering evidence for it) warrants questioning whether these phenomena belong to the same psychological kind, but argue that it’s the deep-conflict cases that represent something other than self-deception. Those who unconsciously hold the warranted belief and merely say or pretend they hold the unwarranted one (Audi 1997; Gendler 2007) hardly seem deceived (Lynch 2012). They seem more to resemble what Longeway (1990) calls escapism , the avoidance of thinking about what we believe to escape reality (See Lynch 2012).

Funkhouser (2019) suggests that this dispute over the depth of conflict and intensity of tension involved in self-deception is, in part, a dispute over which cases are central, typical, and most interesting. Whether deep-conflict cases constitute a distinct psychological kind or whether they reflect people’s pre-theoretical understanding of self-deception remains unclear, but deflationary approaches seem to be capable of explaining at least some of the behavior such theorists insist justifies attributing an unconscious belief that ~p . Deep-conflict theorists need to explain why we should think when one avows that p , one does not also believe it to some degree, and why the behavior in question cannot be explained by nearby proxies like suspicion or doubt that p (Mele 2010, 2012). Some deflationary theorists contend that a degree of belief model might render deep-conflict cases more sensible (see Shifting Degrees of Belief ).

Self-Deception and Self-Knowledge : Several theorists have argued that deflationary approaches miss certain failures of self-knowledge involved in cases of self-deception. Self-deceivers, these critics argue, must hold false beliefs about their own belief formation process (Holton 2001; Scott-Kakures 2002), about what beliefs they actually hold (Funkhouser 2005; Fernández 2013), or both. Holton (2001), for instance, argues that Mele’s conditions for being self-deceived are not sufficient because they do not require self-deceivers to hold false beliefs about themselves. It seems possible for a person to acquire a false belief that p as a consequence of treating data relevant to p in a motivationally biased way when the data available to her provides greater warrant for p and still retain accurate self-knowledge. Such a person would readily admit to ignoring certain data because it would undermine a belief she cherishes. She makes no mistakes about herself, her beliefs, or her belief formation process. Such a person, Holton argues, would be willfully ignorant but not self-deceived. If, however, her strategy was sufficiently opaque to her, she would be apt to deny she was ignoring relevant evidence and even affirm her belief was the result of what Scott-Kakures (2002) calls “reflective, critical reasoning.” These erroneous beliefs represent a failure of self-knowledge that seems, according to these critics, essential to self-deception, and they distinguish it from wishful thinking (see above), willful blindness, and other nearby phenomena.

In response to such criticisms, Mele (2009, 2012) has offered the following sufficient condition: S ’s acquiring the belief that p is a product of “reflective, critical reasoning,” and S is wrong in regarding that reasoning as properly directed. Some worry that meeting this condition requires a degree of awareness about one’s reasons for believing that would rule out those who do not engage in reflection on their reasons for belief (Fernández 2013) and fails to capture errors about what one believes that seem essential for dealing with deep-conflict cases (Fernández 2013; Funkhouser 2005). Whether Mele’s (2009) proposed condition requires too much sophistication from self-deceivers is debatable but suggests a way of accounting for the intuition that self-deceivers fail to know themselves without requiring them to harbor hidden beliefs or intentions.

Self-Deception and Agency: Some worry that deflationary explanations render self-deceivers victims of their own motivations; they don’t seem to be agents with respect to their self-deception but unwitting patients. But, in denying self-deceivers engage in intentional activities for the purpose of deceiving themselves, non-intentionalists needn’t deny self-deceivers engage in any intentional actions. It’s open to them to accept what Lynch terms agentism : “Self-deceivers end up with their unwarranted belief as a result of their own actions motivated by the desire that p ” (Lynch 2017). According to agentism, motivation affects belief by means of intentional actions, not simply by triggering biasing mechanisms. Self-deceivers can act with intentions like “find any problem with unwelcome evidence” or “find any p -supporting evidence” with a view of determining whether p is true. These kinds of intentions may explain the agency of self-deceivers and how they could be responsible for self-deception and not merely victims of their own motivations. It isn’t perfectly clear whether Mele-style deflationists are committed to agentism, but even if they are, questions remain about whether such unwittingly deceptive intentional actions are enough to render self-deceivers true agents of their deception since they think that they are engaging in actions to determine the truth, not to deceive themselves.

Approaches that focus on revising the notion that self-deception requires holding that p and ~p , the dual-belief requirement implied by traditional intentionalism, either introduce some “doxastic proxy” (Baghramian and Nicholson 2013) to replace one or both beliefs or alter the content of the self-deceiver’s belief in a way that preserves tension without involving outright conflict. These approaches resolve the doxastic paradox either by denying that self-deceivers hold the unwelcome but warranted belief ~p (Talbott 1995; Barnes 1997; Burmúdez 2000; Mele 2001), denying they hold the welcome but unwarranted belief p (Audi 1982, 1988; Funkhouser 2005; Gendler 2007; Fernández 2013; Jordan 2020, 2022), denying they hold either belief p or ~p (Archer 2013; Porcher 2012), or contending they have shifting degrees of belief regarding p (Chan and Rowbottom 2019). Lauria et al. (2016) argue for an integrative approach that accommodates all these products of self-deception on the basis of empirical research on the role affect plays in assessments of evidence.

Denying the Unwelcome Belief : Both intentionalists and non-intentionalists may question whether self-deceivers must hold the unwelcome but warranted belief. For intentionalists, what’s necessary is some intention to form the target belief p , and this is compatible with having no views at all regarding p (lacking any evidence for or against p) or believing p is merely possible (possessing evidence too weak to warrant belief that p or ~p ) (Bermúdez 2000). Moreover, rejecting this requirement relieves the pressure to introduce deep divisions (Talbott 1995). For non-intentionalists, the focus is on how the false belief is acquired, not whether a person believes it’s contradictory. For them, it suffices that self-deceivers acquire the unwarranted false belief that p in a motivated way (Mele 2001). The selectivity and tension typical of self-deception can be explained without attributing ~p since nearby proxies like suspicion that ~ p can do the same work. Citing Rorty’s (1988) case of Dr. Androvna, a cancer specialist who believes she does not have cancer but who draws up a detailed will and writes uncharacteristically effusive letters suggesting her impending departure, Mele (2009) points out that Androvna’s behavior might easily be explained by her holding that there’s a significant chance I have cancer. And this belief is compatible with Androvna’s self-deceptive belief that she does not, in fact, have cancer.

Denying the Welcome Belief : Another strand of revision of belief approaches focuses on the welcome belief that p , proposing alternatives to this belief that function in ways that explain what self-deceivers typically say and do. Self-deceivers display ambiguous behavior that not only falls short of what one would expect from a person who believes that p but seems to justify the attribution of the belief that ~p . For instance, Androvna’s letter-writing and will-preparation might be taken as reasons for attributing to her the belief that she won’t recover, despite her verbal assertions to the contrary. To explain the special pattern of behavior displayed by self-deceivers, some of these theorists propose proxies for full belief, such as sincere avowal (Audi 1982, 1988); pretense (Gendler 2007); an intermediate state between belief and desire, or ‘besire’ (Egan 2009); some other less-than-full belief state akin to imaginations or fantasies (Lazar 1999); or simply ‘misrepresentation’ (Jordan 2020, 2022). Such states may guide and motivate action in many, though not all, circumstances while being relatively less sensitive to evidence than beliefs.

Others substitute a higher-order belief to explain the behavior of self-deceivers as another kind of proxy for the belief that p (Funkhouser 2005; Fernández 2013). On such approaches, self-deceivers don’t believe p ; they believe that they believe that p , and this false second-order belief—“I think that I believe that p ”—underlies and underwrites their sincere avowal that p as well as their ability to entertain p as true. Self-deception, then, is a kind of failure of self-knowledge, a misapprehension or misattribution of one’s own beliefs. By shifting the content of the self-deceptive belief to the second-order, this approach avoids the doxastic paradox and explains the characteristic ‘tension’ or ‘conflict’ attributed to self-deceivers in terms of the disharmony between the first-order and second-order beliefs, the latter explaining their avowed belief and the former their behavior that goes against that avowed belief (Funkhouser 2005; Fernández 2013).

Denying both the Welcome Belief and the Unwelcome Belief : Given the variety of proxies that have been offered for both the welcome and the unwelcome belief, it should not be surprising that some argue that self-deception can be explained without attributing either belief to self-deceivers, a position Archer (2013) refers to as ‘nondoxasticism.’ Porcher (2012) recommends against attributing beliefs to self-deceivers on the grounds that what they believe is indeterminate since they are, as Schwitzgebel (2001, 2010) contends, “in-between-believing,” neither fully believing that p nor fully not believing that p . For Porcher (2012), self-deceivers show the limits of the folk psychological concepts of belief and suggest the need to develop a dispositional account of self-deception that focuses on the ways that self-deceivers’ dispositions deviate from those of stereotypical full belief. Funkhouser (2009) also points to the limits of folk psychological concepts and suggests that in cases involving deep conflict between behavior and avowal, “the self-deceived produce a confused belief-like condition so that it is genuinely indeterminate what they believe regarding p .” Archer (2013), however, rejects the claim that the belief is indeterminate or that folk psychological concepts are inadequate, arguing that folk psychology offers a wide variety of non-doxastic attitudes such as ‘hope,’ ‘suspicion,’ ‘anxiety,’ and the like that are more than sufficient to explain paradigm cases of self-deception without adverting to belief.

Shifting Degrees of Belief : Some contend that attention to shifting degrees of belief offers a better explanation of paradigm cases of self-deception—especially the behavioral tensions—and avoids the static paradox (Chan and Rowbottom 2019). In their view, many so-called non-doxastic attitudes entail some degree of belief regarding p. Shifts in these beliefs are triggered by and track shifts in information and non-doxastic propositional attitudes such as desire, fear, anxiety, and anger. For instance, a husband might initially have a high degree of belief in his spouse’s fidelity that plummets when he encounters threatening evidence. His low confidence reveals afresh how much he wants her fidelity and prompts him to despair. These non-doxastic attitudes trigger another shift by focusing his attention on evidence of his spouse’s love and fidelity, leaving him with a higher degree of confidence than his available evidence warrants. On this shifting belief account, the self-deceiver holds both p and ~p at varying levels of confidence that are always greater than zero (example due to Chan and Rowbottom 2019).

While revision of belief approaches suggest a number of non-paradoxical ways of thinking about self-deception, some worry that those approaches denying that self-deceivers hold the welcome but unwarranted belief that p eliminate what is central to the notion of self-deception, namely, deception (see, e.g., Lynch 2012; Mele 2010). Whatever the verdict, these revision of belief approaches suggest that our way of characterizing belief may not be fine-grained enough to account for the subtle attitudes or meta-attitudes that self-deceivers bear on the proposition in question. Taken together, these approaches make it clear that the question regarding what self-deceivers believe is by no means resolved.

‘Twisted’ or negative self-deception differs from ‘straight’ or positive self-deception because it involves the acquisition of an unwelcome as opposed to a welcome belief (Mele 1999, 2001; Funkhouser 2019). Roughly, the negatively self-deceived have a belief that is neither warranted nor wanted in consequence of some desire, emotion, or combination of both. For instance, a jealous husband, uncertain about his wife’s fidelity, comes to believe she’s having an affair on scant and ambiguous evidence, something he certainly doesn’t want to be the case. Intentionalists may see little problem here, at least in terms of offering a unified account, since both positive and negative self-deceivers intend to produce the belief in question, and nothing about intentional deception precludes getting the victim to believe something unpleasant or undesirable. That said, intentionalists typically see the intention to believe p as serving a desire to believe p (Davidson 1985; Talbott 1995; Bermúdez 2000, 2017), so they still face the difficult task of explaining why negative self-deceivers intend to acquire a belief they don’t want (Lazar 1999; Echano 2017). Non-intentionalists have a steeper hill to climb since it’s difficult to see how someone like the anxious husband could be motivated to form a belief that he doesn’t at all desire. The challenge for the non-intentionalist, then, is to supply a motivational story that explains the acquisition of such an unwelcome belief. Ideally, the aim is to provide a unified explanation for both the positive and negative varieties with a view to theoretical simplicity. Attempts to provide such an explanation now constitute a significant and growing body of literature that centers on the nature of the motivations involved and the precise role affect plays in the process.

Since the desire for the welcome belief cannot serve as the motive for acquiring the unwelcome one, non-intentionalists have sought some ulterior motive (Pears 1984), such as the reduction of anxiety (Barnes 1997), the avoidance of costly errors (Mele 1999, 2001), or denial that the motivation is oriented toward the state of the world at all (Nelkin 2002).

The jealous husband might be motivated to believe his wife is unfaithful because it supports the vigilance needed to eliminate all rival lovers and preserve the relationship—both of which he desires (Pears 1984). Similarly, I might be anxious about my house burning down and come to hold the unwelcome belief that I’ve left the burner on. Ultimately, acquiring the unwelcome belief reduces my anxiety because it prompts me to double-check the stove (Barnes 1997). Some are skeptical that identifying such ulterior desires or anxieties is always possible or necessary (Lazar 1999; Mele 2001). Many, following Mele (2001, 2003), see a simpler explanation of negative cases in terms of the way motivation, broadly speaking, affects the agent’s assessment of the relative costs of error. The jealous husband—not wanting to be made a fool—sees the cost of falsely believing his spouse is faithful as high, while the person anxious about their house burning down sees the cost of falsely believing the burner is on as low. Factors such as what the agent cares about and what she can do about the situation affect these error cost assessments and may explain, in part, the conditions under which negative self-deception occurs.

Since negative self-deception often involves emotions—fear, anxiety, jealousy, rage—a good deal of attention has been given to how this component is connected to the motivation driving negative self-deception. Some, like Mele (2001, 2003), acknowledge the possibility that emotion alone or in combination with desire is fundamental to what motivates bias in these cases but remain reluctant to say such affective motives are essential or entirely distinguishable from the desires involved. Others worry that leaving motivation so ambiguous threatens the claim to provide a unified explanation of self-deception (Galeotti 2016a). Consequently, some have sought a more central role for affect, seeing emotion as triggering or priming motivationally biased cognition (Scott-Kakures 2000, 200; Echano 2017) or as operating as a kind of evidential filter in a pre-attentive—non-epistemic—appraisal of threatening evidence (Galeotti 2016a). On this latter affective-filter view, our emotions may lead us to see evidence regarding a situation we consider significant to our wellbeing as ambiguous and therefore potentially distressing, especially when we deem our ability to deal with the unwelcome situation as limited. Depending on how strong our affective response to the distressing evidence is, we may end up discounting evidence for the situation we want, listening instead to our negative emotions (anxiety, fear, sorrow, etc.), with the result that we become negatively self-deceived (see Lauria and Preissmann 2018). Research on the role of dopamine regulation and negative somatic markers provides some neurobiological evidence in support of this sort of affective-filter model and its potential to offer a unified account of positive and negative self-deception (Lauria et al. 2016; Lauria and Preissmann 2018).

While the philosophers considered so far take the relevant motives to be about the state of the world, some hold that the relevant motives have to do with self-deceivers’ states of mind. If this latter desire-to-believe approach is taken, then there may be just one general motive for both kinds of self-deception. Nelkin (2002, 2012), for instance, argues that the motivation for self-deceptive belief formation should be restricted to a desire to believe that p and that this is compatible with not wanting p to be true. I might want to hold the belief that I have left the stove burner on but not want it to be the case that I have actually left it on. The belief is desirable in this instance because holding it ensures it won’t be true. What unifies cases of self-deception—both twisted and straight—is that the self-deceptive belief is motivated by a desire to believe that p ; what distinguishes them is that twisted self-deceivers do not want p to be the case, while straight self-deceivers do. Some, like Mele (2009), argue that such an approach is unnecessarily restrictive since a variety of other motives oriented toward the state of the world might lead one to acquire the unwelcome belief; for example, even just wanting to not be wrong about the welcome belief (see Nelkin 2012 for a response). Others, like Galeotti (2016a), worry that this desire-to-believe account renders self-deceivers’ epistemic confusion into something bordering on incoherence since it seems to imply self-deceivers want to believe p regardless of the state of the world, and such a desire seems absurd even at an unconscious level.

Whether the motive for self-deception aims at the state of the world or the state of the self-deceiver’s mind, the role of affect in the process remains a significant question that further research in neurobiology may shed light upon. The role of affect has been underappreciated but seems to be gathering support and will no doubt guide future theorizing, especially on negative self-deception.

5. Morality and Self-Deception

Even though much of the contemporary philosophical discussion of self-deception has focused on epistemology, philosophical psychology, and philosophy of mind, historically, the morality of self-deception has been the central focus of discussion. Self-deception has been thought to be morally wrong or, at least, morally dangerous insofar as it represents a threat to moral self-knowledge, a cover for immoral activity, or a violation of authenticity. Some thinkers, what Martin (1986) calls ‘the vital lie tradition,’ however, have held that self-deception can, in some instances, be salutary, protecting us from truths that would make life unlivable (e.g., Rorty 1972, 1994). There are two major questions regarding the morality of self-deception: First, can a person be held morally responsible for self-deception, and if so, under what conditions? Second, is there anything morally problematic with self-deception, and if so, what and under what circumstances? The answers to these questions are clearly intertwined. If self-deceivers cannot be held responsible for self-deception, then their responsibility for whatever morally objectionable consequences self-deception might have will be mitigated if not eliminated. Nevertheless, self-deception might be morally significant even if one cannot be taxed for entering it. To be ignorant of one’s moral self, as Socrates saw, may represent a great obstacle to a life well lived, whether or not one is at fault for such ignorance.

Whether self-deceivers can be held responsible for their self-deception is largely a question of whether they have the requisite control over the acquisition and maintenance of their self-deceptive belief. In general, intentionalists hold that self-deceivers are responsible since they intend to acquire the self-deceptive belief, usually recognizing the evidence to the contrary. Even when the intention is indirect, such as when one intentionally seeks evidence in favor of p or avoids collecting or examining evidence to the contrary, self-deceivers seem to intentionally flout their own normal standards for gathering and evaluating evidence. So, minimally, they are responsible for such actions and omissions.

Initially, non-intentionalist approaches may seem to remove the agent from responsibility by rendering the process by which she is self-deceived subintentional. If my anxiety, fear, or desire triggers a process that ineluctably leads me to hold the self-deceptive belief, it seems I cannot be held responsible for holding that belief. How could I be held responsible for processes that operate without my knowledge and which are set in motion without my intention? Most non-intentionalist accounts, however, do allow for the possibility that self-deceivers are responsible for individual episodes of self-deception or for the vices of cowardice and lack of self-control from which they spring. To be morally responsible in the sense of being an appropriate target for praise or blame requires, at least, that agents have control over the actions in question. Mele (2001), for example, argues that many sources of bias are controllable and that self-deceivers can recognize and resist the influence of emotion and desire on their belief acquisition and retention, particularly in matters they deem to be important—morally or otherwise. The extent of this control, however, is an empirical question. Nelkin (2012) argues that since Mele’s account leaves the content of motivation driving the bias unrestricted, the mechanism in question is so complex that “it seems unreasonable to expect the self-deceiver to guard against” its operation.

Other non-intentionalists take self-deceivers to be responsible for certain epistemic vices, such as cowardice in the face of fear or anxiety and lack of self-control with respect to the biasing influences of desire and emotion. Thus, Barnes (1997) argues that self-deceivers “can, with effort, in some circumstances, resist their biases” (83) and “can be criticized for failing to take steps to prevent themselves from being biased; they can be criticized for lacking courage in situations where having courage is neither superhumanly difficult nor costly” (175). Whether self-deception is due to a character defect or not, ascriptions of responsibility depend upon whether the self-deceiver has control over the biasing effects of her desires and emotions.

Some question whether self-deceivers do have such control. For instance, Levy (2004) has argued that deflationary accounts of self-deception that deny the contradictory belief requirement should not suppose that self-deceivers are typically responsible since it is rarely the case that self-deceivers possess the requisite awareness of the biasing mechanisms operating to produce their self-deceptive belief. Lacking such awareness, self-deceivers do not appear to know when or on which beliefs such mechanisms operate, rendering them unable to curb the effects of these mechanisms, even when they operate to form false beliefs about morally significant matters. Lacking the control necessary for moral responsibility in individual episodes of self-deception, self-deceivers seem also to lack control over being the sort of person disposed to self-deception.

Non-intentionalists may respond by claiming that self-deceivers often are aware of the potentially biasing effects their desires and emotions might have and can exercise control over them (DeWeese-Boyd 2007). They might also challenge the idea that self-deceivers must be aware in the ways Levy suggests. One well-known account of control, employed by Levy, holds that a person is responsible just in case she acts on a mechanism that is moderately responsive to reasons (including moral reasons), such that were she to possess such reasons, this same mechanism would act upon those reasons in at least one possible world (Fischer and Ravizza 1999). Guidance control, in this sense, requires that the mechanism in question be capable of recognizing and responding to moral and non-moral reasons sufficient for acting otherwise. Nelkin (2011, 2012), however, argues that reasons-responsiveness should be seen as applying primarily to the agent , not the mechanism, requiring only that the agent has the capacity to exercise reason in the situation under scrutiny. The question isn’t whether the biasing mechanism itself is reasons-responsive but whether the agent governing its operation is—that is, whether self-deceivers typically could recognize and respond to moral and non-moral reasons to resist the influence of their desires and emotions and instead exercise special scrutiny of the belief in question. According to Nelkin (2012), expecting self-deceivers to have such a capacity is more likely if we understand the desire driving their bias as a desire to believe that p , since awareness of this sort of desire would make it easier to guard against its influence on the process of determining whether p . Van Loon (2018) points out that discussion of moral responsibility and reasons-responsiveness have focused on actions and omissions that indirectly affect belief formation when it is more appropriate to focus on epistemic reasons-responsiveness. Attitudinal control, not action control, is what is at issue in self-deception. Drawing on McHugh’s (2013, 2014, 2017) account of attitudinal control, Van Loon argues that self-deceivers on Mele-style deflationary accounts are responsible for their self-deceptive beliefs because they recognize and react to evidence against their self-deceptive belief across a wide-range of counter-factual scenarios even though their recognition of this evidence does not alter their belief and their reaction to such evidence leads them viciously to hold the self-deceptive belief.

Galeotti (2016b) rejects the idea that control is the best way to think about responsibility in cases of self-deception since self-deceivers on deflationary approaches seem both confused and relatively powerless over the process. Instead, following a modified version of Sher’s (2009) account of responsibility, she contends that self-deceivers typically have, but fail to recognize, evidence that their acts related to belief formation are wrong or foolish and so fall below some applicable standard (epistemic, moral, etc.). The self-deceiver is under motivational pressure, not incapacitated, and on exiting self-deception, recognizes “the faultiness of her condition and feels regret and shame at having fooled herself” (Galeotti 2016b). This ex-post reasons-responsiveness suggests self-deceivers are responsible in Sher’s sense even if their self-deception is not intentional.

In view of these various ways of cashing out responsibility, it’s plausible that self-deceivers can be morally responsible for their self-deception on deflationary approaches and certainly not obvious that they couldn’t be.

Insofar as it seems plausible that, in some cases, self-deceivers are apt targets for censure, what prompts this attitude? Take the case of a mother who deceives herself into believing her husband is not abusing their daughter because she can’t bear the thought that he is a moral monster (Barnes 1997). Why do we blame her? Here we confront the nexus between moral responsibility for self-deception and the morality of self-deception. Understanding what obligations may be involved in cases of this sort will help to clarify the circumstances in which ascriptions of moral responsibility are appropriate. While some instances of self-deception seem morally innocuous, and others may even be thought salutary in various ways (Rorty 1994), most theorists have thought there to be something morally objectionable about self-deception or its consequences in many cases. Self-deception has been considered objectionable because it facilitates harm to others (Linehan 1982; Gibson 2020; Clifford 1877) and to oneself, undermines autonomy (Darwall 1988; Baron 1988), corrupts conscience (Butler 1726), violates authenticity (Sartre 1943), manifests a vicious lack of courage and self-control that undermines the capacity for compassionate action (Jenni 2003), violates an epistemic duty to properly ground self-ascriptions (Fernández 2013), violates a general duty to form beliefs that “conform to the available evidence” (Nelkin 2012), or violates a general duty to respect our own values (MacKenzie 2022).

Linehan (1982) argues that we have an obligation to scrutinize the beliefs that guide our actions that is proportionate to the harm to others such actions might involve. When self-deceivers induce ignorance of moral obligations in connection with the particular circumstances, likely consequences of actions, or nature of their own engagements, by means of their self-deceptive beliefs, they may be culpable for negligence with respect to their obligation to know the nature, circumstances, likely consequences, and so forth of their actions (Jenni 2003; Nelkin 2012). Self-deception, accordingly, undermines or erodes agency by reducing our capacity for self-scrutiny and change (Baron 1988). If I am self-deceived about actions or practices that harm others or myself, my abilities to take responsibility and change are also severely restricted.

Joseph Butler (1726), in his well-known sermon “On Self-Deceit,” emphasizes the ways in which self-deception about one’s moral character and conduct—‘self-ignorance’ driven by inordinate ‘self-love’—not only facilitates vicious actions but hinders the agent’s ability to change. Such ignorance, claims Butler, “undermines the whole principle of good … and corrupts conscience, which is the guide of life” (1726). Holton (2022) explores the way our motivation to see ourselves as morally good may play a role in this lack of moral self-knowledge. Existentialist philosophers such as Kierkegaard and Sartre, in very different ways, view self-deception as a threat to ‘authenticity’ insofar as self-deceivers fail to take responsibility for themselves and their engagements in the past, present, and future. By alienating us from our own principles, self-deception may also threaten moral integrity (Jenni 2003). MacKenzie (2022) might be seen as capturing precisely what’s wrong with this sort of inauthenticity when she contends that we have a duty to properly respect our values, even non-moral ones. Since self-deception is always about something we value in some way, it represents a failure to properly respect ourselves as valuers. Others note that self-deception also manifests a certain weakness of character that disposes us to react to fear, anxiety, or the desire for pleasure in ways that bias our belief acquisition and retention in ways that serve these emotions and desires rather than accuracy (Butler 1726; Clifford 1877). Such epistemic cowardice (Barnes 1997; Jenni 2003) and lack of self-control may inhibit the ability of self-deceivers to stand by or apply moral principles they hold by biasing their beliefs regarding particular circumstances, consequences, or engagements or by obscuring the principles themselves. Gibson (2020), following Clifford (1877), contends that self-deception increases the risk of harm to the self and others and the cultivation of epistemic vices like credulity that may have devastating social ramifications. In all these ways and a myriad of others, philosophers have found some self-deception objectionable in itself or for the consequences it has, not only on our ability to shape our lives but also for the potential harm it may cause to ourselves and others.

Evaluating self-deception and its consequences for ourselves and others is a difficult task. It requires, among other things: determining the degree of control self-deceivers have; what the self-deception is about (Is it important morally or otherwise?); what ends the self-deception serves (Does it serve mental health or as a cover for moral wrongdoing?); how entrenched it is (Is it episodic or habitual?); and, whether it is escapable (What means of correction are available to the self-deceiver?). As Nelkin (2012) contends, determining whether and to what degree self-deceivers are culpably negligent will ultimately need to be determined on a case-by-case basis in light of answers to such questions about the stakes at play and the difficulty involved. Others like Mackenzie (2022) hold that every case of self-deception is a violation of a general duty to respect our own values, though some cases are more egregious than others.

If self-deception is morally objectionable for any of these reasons, we ought to avoid it. But, one might reasonably ask how that is possible given the subterranean ways self-deception seems to operate. Answering this question is tricky since strategies will vary with the analyses of self-deception and our responsibility for it. Nevertheless, two broad approaches seem to apply to most accounts, namely, the cultivation of one’s epistemic virtues and the cultivation of one’s epistemic community (Galeotti 2016b). One might avoid the self-deceptive effects of motivation by cultivating virtues like impartiality, vigilance, conscientiousness, and resistance to the influence of emotion, desire, and the like. Additionally, one might cultivate an epistemic community that holds one accountable and guides one away from self-deception (Rorty 1994; Galeotti 2016b, 2018). By binding ourselves to communities we have authorized to referee our belief formation in this way, we protect ourselves from potential lapses in epistemic virtue. These kinds of strategies might indirectly affect our susceptibility to self-deception and offer some hope of avoiding it.

Quite aside from the doxastic, strategic, and moral puzzles self-deception raises, there is the evolutionary puzzle of its origin. Why do human beings have this capacity in the first place? Why would natural selection allow a capacity to survive that undermines the accurate representation of reality, especially when inaccuracies about individual ability or likely risk can lead to catastrophic errors?

Many argue that self-deceptively inflated views of ourselves, our abilities, our prospects, or our control—so-called ‘positive illusions’—confer direct benefits in terms of psychological well-being, physical health, and social advancement that serve fitness (Taylor and Brown 1994; Taylor 1989; McKay and Dennett 2009). Just because ‘positive illusions’ make us ‘feel good,’ of course, it does not follow that they are adaptive. From an evolutionary perspective, whether an organism ‘feels good’ or is ‘happy’ is not significant unless it enhances survival and reproduction. McKay and Dennett (2009) argue that positive illusions are not only tolerable; evolutionarily speaking, they contribute to fitness directly. Overly positive beliefs about our abilities or chances for success appear to make us more apt to exceed our abilities and achieve success than more accurate beliefs would (Taylor and Brown 1994; Bandura 1989). According to Johnston and Fowler (2011), overconfidence is “advantageous, because it encourages individuals to claim resources they could not otherwise win if it came to a conflict (stronger but cautious rivals will sometime fail to make a claim, and it keeps them from walking away from conflicts they would surely win).” Inflated attitudes regarding the personal qualities and capacities of one’s partners and children would also seem to enhance fitness by facilitating the thriving of offspring (McKay and Dennett 2009).

Alternatively, some argue that self-deception evolved to facilitate interpersonal deception by eliminating the cues and cognitive load that consciously lying produces and by mitigating retaliation should the deceit become evident (von Hippel and Trivers 2011; Trivers 2011, 2000, 1991). On this view, the real gains associated with ‘positive illusions’ and other self-deceptions are byproducts that serve this greater evolutionary end by enhancing self-deceivers’ ability to deceive. Von Hippel and Trivers (2011) contend that “by deceiving themselves about their own positive qualities and the negative qualities of others, people are able to display greater confidence than they might otherwise feel, thereby enabling them to advance socially and materially.” Critics have pointed to data suggesting high self-deceivers are deemed less trustworthy than low self-deceivers (McKay and Dennett 2011). Others have complained that there is little data to support this hypothesis (Dunning 2011; Van Leeuwen 2007a, 2013a), and what data there is shows us to be poor lie-detectors (Funkhouser 2019; Vrij 2011). Some challenge this theory by noting that a simple disregard for the truth would serve as well as self-deception and have the advantage of retaining true representations (McKay and Prelec 2011) or that often self-deceivers are the only ones deceived (Van Leeuwen 2007a; Kahlil 2011). Van Leeuwen (2013a) raises the concern that the wide variety of phenomena identified by this theory as self-deception render the category so broad that it is difficult to tell whether it is a unified phenomenon traceable to particular mechanisms that could plausibly be sensitive to selection pressures. Funkhouser (2019) worries that the unconscious retention of the truth that von Hippel and Trivers (2011) propose would generate tells of its own and that the psychological complexity of this explanation is unnecessary if the goal is to deceive others (which is itself contentious) since that goal would be easier to achieve through self-delusion. So, von Hippel and Trivers’ (2011) theory may explain self-delusion but not cases of self-deception marked by deep conflict (Funkhouser 2017b).

In view of these shortcomings, Van Leeuwen (2007a) argues the capacity for self-deception is a spandrel —a byproduct of other aspects of our cognitive architecture—not an adaptation in the strong sense of being positively selected. While Funkhouser (2017b) agrees that the basic cognitive architecture that allows motivation to influence belief formation—as well as the specific tools used to form or maintain biased belief—were not selected for the sake of self-deception, it nevertheless makes sense to say, for at least some contents, that self-deception is adaptive (2017b).

Whether it is an adaptation or a spandrel, it’s possible this capacity has nevertheless been retained as a consequence of its fitness value. Lopez and Fuxjager (2011) argue that the broad research on the so-called “winner effect”—the increased probability of achieving victory in social or physical conflicts following prior victories—lends support to the idea that self-deception is at least weakly adaptive since self-deception in the form of positive illusions, like past wins, confers a fitness advantage. Lamba and Nityananda (2014) test the theory that self-deceived are better at deceiving others—specifically, whether overconfident individuals are overrated by others and underconfident individuals underrated. In their study, students in tutorials were asked to predict their own performance on the next assignment as well as that of each of their peers in the tutorial in terms of absolute grade and relative rank. Comparing these predictions and the actual grades given on the assignment suggests a strong positive relationship between self-deception and deception. Those who self-deceptively rated themselves higher were rated higher by their peers as well. These findings lend suggestive support to the claim self-deception facilitates other deception. While these studies certainly do not supply all the data necessary to support the theory that the propensity to self-deception should be viewed as adaptation, they do suggest ways to test these evolutionary hypotheses by focusing on specific phenomena.

Whether or not the psychological and social benefits identified by these theories explain the evolutionary origins of the capacity for self-deceit, they may well shed light on its prevalence and persistence, as well as point to ways to identify contexts in which this tendency presents high collective risk (Lamba and Nityananda 2014).

7. Collective Self-Deception

Collective self-deception has received scant direct philosophical attention as compared with its individual counterpart. Collective self-deception might refer simply to a group of similarly self-deceived individuals or to a group-entity (such as a corporation, committee, jury, or the like) that is self-deceived. These alternatives reflect two basic perspectives that social epistemologists have taken on ascriptions of propositional attitudes to collectives. On the one hand, such attributions might be taken summatively as simply an indirect way of attributing those states to members of the collective (Quinton 1975/1976). This summative understanding, then, considers attitudes attributed to groups to be nothing more than metaphors expressing the sum of the attitudes held by their members. To say that students think tuition is too high is just a way of saying that most students think so. On the other hand, such attributions might be understood non-summatively as applying to collective entities, themselves ontologically distinct from the members upon which they depend. These so-called ‘plural subjects’ (Gilbert 1989, 1994, 2005) or ‘social integrates’ (Pettit 2003), while supervening upon the individuals comprising them, may well express attitudes that diverge from individual members. For instance, saying NASA believed the O-rings on the space shuttle’s booster rockets to be safe need not imply that most or all the members of this organization personally held this belief, only that the institution itself did. The non-summative understanding, then, considers collectives to be, like persons, apt targets for attributions of propositional attitudes and potentially of moral and epistemic censure as well. Following this distinction, collective self-deception may be understood in either a summative or non-summative sense.

In the summative sense, collective self-deception refers to a self-deceptive belief shared by a group of individuals whom each come to hold the self-deceptive belief for similar reasons and by similar means, varying according to the account of self-deception followed. We might call this self-deception across a collective. In the non-summative sense, the subject of collective self-deception is the collective itself, not simply the individuals comprising it. The following sections offer an overview of these forms of collective self-deception, noting the significant challenges posed by each.

Understood summatively, we might define collective self-deception as the holding of a false belief in the face of evidence to the contrary by a group of people as a result of shared desires, emotions, or intentions (depending upon the account of self-deception) favoring that belief. Collective self-deception is distinct from other forms of collective false belief—such as might result from deception or lack of evidence—insofar as the false belief issues from the agents’ own self-deceptive mechanisms (however these are construed), not the absence of evidence to the contrary or presence of misinformation. Accordingly, the individuals constituting the group would not hold the false belief if their vision weren’t distorted by their attitudes (desire, anxiety, fear, or the like) toward the belief. What distinguishes collective self-deception from solitary self-deception is just its social context; namely, that it occurs within a group that shares both the attitudes bringing about the false belief and the false belief itself.

Merely sharing desires, emotions, or intentions favoring a belief with a group does not entail that the self-deception is properly social since these individuals may well self-deceive regardless of the fact that their motivations are shared with others (Dings 2017; Funkhouser 2019); they are just individually self-deceiving in parallel. What makes collective self-deception social , according to Dings (2017), is that others are a means used in each individual’s self-deception. So, when a person situates herself in a group of like-minded people in response to an encounter with new and threatening evidence, her self-deception becomes social. Self-deception also becomes social in Dings’ (2017) view when a person influences others to make them like-minded with regard to her preferred belief, using their behavior to reinforce her self-deception . Within highly homogeneous social groups, however, it may be difficult to tell who is using the group instrumentally in these ways, especially when that use is unwitting. Moreover, one may not need to seek out such a group of like-minded people if they already comprise one’s community. In this case, those people may become instrumental to one’s self-deception simply by dint of being there to provide insulation from threatening evidence and support for one’s preferred belief. In any case, this sort of self-deception is both easier to foster and more difficult to escape, being abetted by the self-deceptive efforts of others within the group.

Virtually all self-deception has a social component, being wittingly or unwittingly supported by one’s associates (see Ruddick 1988). In the case of collective self-deception, however, the social dimension comes to the fore since each member of the collective unwittingly helps to sustain the self-deceptive belief of the others in the group. For example, my cancer-stricken friend might self-deceptively believe her prognosis to be quite good. Faced with the fearful prospect of death, she does not form accurate beliefs regarding the probability of her full recovery, attending only to evidence supporting full recovery and discounting or ignoring altogether the ample evidence to the contrary. Caring for her as I do, I share many of the anxieties, fears, and desires that sustain my friend’s self-deceptive belief, and as a consequence, I form the same self-deceptive belief via the same mechanisms. In such a case, I unwittingly support my friend’s self-deceptive belief, and she mine—our self-deceptions are mutually reinforcing. We are collectively or mutually self-deceived, albeit on a very small scale. Ruddick (1988) calls this ‘joint self-deception,’ and it is properly social just in case each person is instrumental in the formation of the self-deceptive belief in the other (Dings 2017).

On a larger scale, sharing common attitudes, large segments of a society might deceive themselves together. For example, we share a number of self-deceptive beliefs regarding our consumption patterns. Many of the goods we consume are produced by people enduring labor conditions we do not find acceptable and in ways that we recognize are environmentally destructive and likely unsustainable. Despite our being at least generally aware of these social and environmental ramifications of our consumptive practices, we hold the overly optimistic beliefs that the world will be fine, that its peril is overstated, that the suffering caused by the exploitive and ecologically degrading practices are overblown, that our own consumption habits are unconnected to these sufferings, and even that our minimal efforts at conscientious consumption are an adequate remedy (see Goleman 1989). When self-deceptive beliefs such as these are held collectively, they become entrenched, and their consequences, good or bad, are magnified (Surbey 2004).

The collective entrenches self-deceptive beliefs by providing positive reinforcement through others sharing the same false belief as well as by protecting its members from evidence that would destabilize the target belief. There are, however, limits to how entrenched such beliefs can become and remain. Social support cannot be the sole or primary cause of the self-deceptive belief, for then the belief would simply be the result of unwitting interpersonal deception and not the deviant belief formation process that characterizes self-deception. If the environment becomes so epistemically contaminated as to make counter-evidence inaccessible to the agent, then we have a case of simple false belief, not self-deception. Thus, even within a collective, a person is self-deceived just in case her own motivations skew the belief formation process that results in her holding the false belief. But, to bar this from being a simple case of solitary self-deception, others must be instrumental to her belief formation process such that if they were not part of that process, she would not be self-deceived (Dings 2017). For instance, I might be motivated to believe that climate change is not a serious problem and form that false belief as a consequence. In such a case, I’m not socially self-deceived, even if virtually everyone I know shares a similar motivation and belief. But, say I encounter distressing evidence in my environmental science class that I can’t shake on my own. I may seek to surround myself with like-minded people, thereby protecting myself from further distressing evidence and providing myself with reassuring evidence. Now, my self-deception is social, and this social component drives and reinforces my own motivations to self-deceive.

Relative to solitary self-deception, the collective variety presents greater external obstacles to avoiding or escaping self-deception and is, for this reason, more entrenched. If the various proposed psychological mechanisms of self-deception pose an internal challenge to the self-deceiver’s power to control her belief formation, then these social factors pose an external challenge to the self-deceiver’s control. Determining how superable this challenge is will affect our assessment of individual responsibility for self-deception as well as the prospects of unassisted escape from it.

Collective self-deception can also be understood from the perspective of the collective itself in a non-summative sense. Though there are varying accounts of group belief, generally speaking, a group can be said to believe, desire, value, or the like just in case its members “jointly commit” to these things as a body (Gilbert 2005). A corporate board, for instance, might be jointly committed as a body to believe, value, and strive for whatever the CEO recommends. Such commitment need not entail that each individual board member personally endorses such beliefs, values, or goals, only that they do so as members of the board (Gilbert 2005). While philosophically precise accounts of non-summative self-deception remain largely unarticulated—an exception is Galeotti’s (2018) detailed analysis of how collective self-deception occurs in the context of politics—the possibilities mirror those of individual self-deception. When collectively held attitudes motivate a group to espouse a false belief despite the group’s possession of evidence to the contrary, we can say that the group is collectively self-deceived in a non-summative sense.

For example, Robert Trivers (2000) suggests that ‘organizational self-deception’ led to NASA’s failure to represent accurately the risks posed by the space shuttle’s O-ring design, a failure that eventually led to the Challenger disaster. The organization as a whole, he argues, had strong incentives to represent such risks as small. As a consequence, NASA’s Safety Unit mishandled and misrepresented data it possessed that suggested that under certain temperature conditions, the shuttle’s O-rings were not safe. NASA, as an organization, then, self-deceptively believed the risks posed by O-ring damage were minimal. Within the institution, however, there were a number of individuals who did not share this belief, but both they and the evidence supporting their belief were treated in a biased manner by the decision-makers within the organization. As Trivers (2000) puts it, this information was relegated “to portions of … the organization that [were] inaccessible to consciousness (we can think of the people running NASA as the conscious part of the organization).” In this case, collectively held values created a climate within NASA that clouded its vision of the data and led to its endorsement of a fatally false belief.

Collective self-deceit may also play a significant role in facilitating unethical practices by corporate entities. For example, a collective commitment by members of a corporation to maximizing profits might lead members to form false beliefs about the ethical propriety of the corporation’s practices. Gilbert (2005) suggests that such a commitment might lead executives and other members to “simply lose sight of moral constraints and values they previously held.” Similarly, Tenbrunsel and Messick (2004) argue that self-deceptive mechanisms play a pervasive role in what they call ‘ethical fading,’ acting as a kind of ‘bleach’ that renders organizations blind to the ethical dimensions of their decisions. They argue that such self-deceptive mechanisms must be recognized and actively resisted at the organizational level if unethical behavior is to be avoided. More specifically, Gilbert (2005) contends that collectively accepting that “certain moral constraints must rein in the pursuit of corporate profits” might shift corporate culture in such a way that efforts to respect these constraints are recognized as part of being a good corporate citizen. In view of the ramifications this sort of collective self-deception has for the way we understand corporate misconduct and responsibility, understanding its specific nature in greater detail remains an important task.

Collective self-deception, understood in either the summative or non-summative sense, raises significant questions, such as whether individuals within collectives bear responsibility for their self-deception or the part they play in the collective’s self-deception and whether collective entities can be held responsible for their epistemic failures (see Galeotti 2016b, 2018 on these questions). Finally, collective self-deception prompts us to ask what means are available to collectives and their members to resist, avoid, and escape self-deception. Galeotti (2016b, 2018) argues for a variety of institutional constraints and precommitments to keep groups from falling prey to self-deception.

Given the capacity of collective self-deception to entrench false beliefs and to magnify their consequences—sometimes with disastrous results—collective self-deception is not just a philosophical puzzle; it is a problem that demands attention.

  • Ames, R.T., and W. Dissanayake (eds.), 1996, Self and Deception , New York: State University of New York Press.
  • Archer, Sophie, 2013, “Nondoxasticism about Self‐Deception,” Dialectica , 67(3): 265–282.
  • –––, 2018, “Why ‘believes’ is not a vague predicate,” Philosophical Studies ,175 (12): 3029–3048.
  • Audi, R., 2007, “Belief, Intention, and Reasons for Action,” in Rationality and the Good , J. Greco, A. Mele, and M. Timmons (eds.), New York: Oxford University Press.
  • –––, 1989, “Self-Deception and Practical Reasoning,” Canadian Journal of Philosophy , 19: 247–266.
  • –––, 1982, “Self-Deception, Action, and Will,” Erkenntnis , 18: 133–158.
  • –––, 1976, “Epistemic Disavowals and Self-Deception,” The Personalist , 57: 378–385.
  • Bach, K., 1997, “Thinking and Believing in Self-Deception,” Behavioral and Brain Sciences , 20: 105.
  • –––, 1981, “An Analysis of Self-Deception,” Philosophy and Phenomenological Research , 41: 351–370.
  • Baghramian, M., and A. Nicholson, 2013, “The Puzzle of Self-Deception,” Philosophy Compass , 8(11): 1018–1029.
  • Bagnoli, C., 2012, “Self-deception and agential authority,” Humana Mente , 20: 99–116
  • Bargh, John and Morsella, Ezequiel, 2008, “The Unconscious Mind,” Perspectives on Psychological Science , 3: 73–79.
  • Barnes, A., 1997, Seeing through Self-Deception , New York: Cambridge University Press.
  • Baron, M., 1988, “What is Wrong with Self-Deception,” in Perspectives on Self-Deception , B. McLaughlin and A. O. Rorty (eds.), Berkeley: University of California Press.
  • Bayne, T. & J. Fernandez Eds., 2009, Delusion and Self-Deception: Affective and Motivational Influences on Belief Formation , New York: Psychology Press.
  • Bok, S., 1989, “Secrecy and Self-Deception,” in Secrets: On the Ethics of Concealment and Revelation , New York: Vintage.
  • –––, 1980, “The Self Deceived,” Social Science Information , 19: 923–935.
  • Bermúdez, José Luis, 2017,“Self-deception and Selectivity: Reply to Jurjako,” Croatian Journal of Philosophy , 17(1): 91–95.
  • –––, 2000, “Self-Deception, Intentions, and Contradictory Beliefs,” Analysis , 60(4): 309–319.
  • –––, 1997, “Defending Intentionalist Accounts of Self-Deception,” Behavioral and Brain Sciences , 20: 107–8.
  • Bird, A., 1994, “Rationality and the Structure of Self-Deception,” in S. Gianfranco (ed.), European Review of Philosophy (Volume 1: Philosophy of Mind), Stanford: CSLI Publications.
  • Borge, S., 2003, “The Myth of Self-Deception,” The Southern Journal of Philosophy , 41: 1–28.
  • Borman, David, 2022, “Self-Deception and Moral Interests,” European Journal of Philosophy , first online 17 January 2022. doi:10.1111/ejop.12756
  • Brown, R., 2003, “The Emplotted Self: Self-Deception and Self-Knowledge.,” Philosophical Papers , 32: 279–300.
  • Butler, J., 1726, “Upon Self-Deceit,” in D.E. White (ed.), 2006, The Works of Bishop Butler , Rochester: Rochester University Press. [ Available online ]
  • Cerovac, I., 2015, “Intentionalism as a theory of self-deception,” Balkan Journal of Philosophy , 7: 145–150.
  • Chisholm, R. M., and Feehan, T., 1977, “The Intent to Deceive,” Journal of Philosophy , 74: 143–159.
  • Clifford, W. K., 1877, “The Ethics of Belief,” in The Ethics of Belief and Other Essays , Amherst, NY: Prometheus Books, 70–97.
  • Cook, J. T., 1987, “Deciding to Belief without Self-deception,” Journal of Philosophy , 84: 441–446.
  • Correia, Vasco, 2014, “From Self-deception to self-control: Emotional biases and the virtues of precommitment,” Croatian Journal of Philosophy , 14(3): 309–323.
  • Dalton, P., 2002, “Three Levels of Self-Deception (Critical Commentary on Alfred Mele’s Self-Deception Unmasked),” Florida Philosophical Review , 2(1): 72–76.
  • Darwall, S., 1988, “Self-Deception, Autonomy, and Moral Constitution,” in Perspectives on Self-Deception , B. McLaughlin and A. O. Rorty (eds.), Berkeley: University of California Press.
  • Davidson, D., 1986, “Deception and Division,” in J. Elster (ed.) 1986, 79–92; reprinted in D. Davidson, Problems of Rationality , with introduction by Marcia Cavell and interview with Ernest Lepore, Oxford: Clarendon Press, 2004, 199–212.
  • –––, 1982, “Paradoxes of Irrationality,” in Philosophical Essays on Freud , R. Wollheim and J. Hopkins (eds.), Cambridge: Cambridge University Press.
  • Demos, R., 1960, “Lying to Oneself,” Journal of Philosophy , 57: 588–95.
  • Dennett, D., 1992, “The Self as a Center of Narrative Gravity,” in Consciousness and Self: Multiple Perspectives , F. Kessel, P. Cole, and D. Johnson (eds.), Hillsdale, NJ: L. Erlbaum.
  • de Sosa, R., 1978, “Self-Deceptive Emotions,” Journal of Philosophy , 75: 684–697.
  • –––, 1970, “Self-Deception,” Inquiry , 13: 308–321.
  • DeWeese-Boyd, I., 2014, “Self-Deceptive Religion and the Prophetic Voice,” Journal for Religionsphilosophie , 3: 26–37.
  • –––, 2007, “Taking Care: Self-Deception, Culpability and Control,” teorema , 26(3): 161–176.
  • Dings, Roy, 2017, “Social Strategies in Self-deception,” New Ideas in Psychology , 47: 16–23.
  • Doody, P., 2017, “Is There Evidence of Robust, Unconscious Self- Deception? A Reply to Funkhouser and Barrett,” Philosophical Psychology , 30(5): 657–676.
  • Dunn, R., 1995, “Motivated Irrationality and Divided Attention,” Australasian Journal of Philosophy , 73: 325–336.
  • Dunning, D., 2011, “Get Thee to a Laboratory,” Commentary on target article, “The Evolution and Psychology of Self-Deception,” by W. von Hippel and R. Trivers, Behavioral and Brain Sciences , 34(1): 18–19.
  • –––, 1995, “Attitudes, Agency and First-Personality,” Philosophia , 24: 295–319.
  • –––, 1994, “Two Theories of Mental Division,” Australasian Journal of Philosophy , 72: 302–316.
  • Dupuy, J-P. (ed.), 1998, Self-Deception and Paradoxes of Rationality (Lecture Notes 69), Stanford: CSLI Publications.
  • Dyke, D., 1633, The Mystery of Selfe-Deceiving , London: William Standby.
  • Echano, M., 2017, “The Motivating Influence of Emotion on Twisted Self-Deception,” Kritike , 11(2): 104–120.
  • Edwards, S., 2013, “Nondoxasticism about Self-Deception,” Dialectica , 67(3): 265–282.
  • Egan, Andy, 2009, “Imagination, Delusion, and Self-Deception,” in Delusions, Self-Deception, and Affective Influences on Belief-formation , T. Bayne and J. Fernandez (eds.), New York: Psychology Press.
  • Elster, J. (ed.), 1986, The Multiple Self , Cambridge: Cambridge University Press.
  • Fairbanks, R., 1995, “Knowing More Than We Can Tell,” The Southern Journal of Philosophy , 33: 431–459.
  • Fernández, Jordi, 2013, “Self-deception and self-knowledge,” Philosophical Studies 162(2): 379–400.
  • Fingarette, H., 1998, “Self-Deception Needs No Explaining,” The Philosophical Quarterly , 48: 289–301.
  • –––, 1969, Self-Deception , Berkeley: University of California Press; reprinted, 2000.
  • Friedrich, James, 1993, “Primary error detection and minimization PEDMIN strategies in social cognition: A reinterpretation of confirmation bias phenomena,” Psychological Review , 100, 298–319.
  • Fischer, J. and Ravizza, M., 1998, Responsibility and Control . Cambridge: Cambridge University Press.
  • Funkhouser, Eric, 2019, Self-Deception , New York: Routledge.
  • –––, 2017a, “Beliefs as Signals: A New Function for Belief,” Philosophical Psychology , 30(6): 809–831.
  • –––, 2017b, “Is Self-Deception an Effective Non-Cooperative Strategy?,” Biology and Philosophy , 32: 221–242.
  • –––, 2009, “Self-Deception and the Limits of Folk Psychology,” Social Theory and Practice , 35(1): 1–13.
  • –––, 2005, “Do the Self-Deceived Get What They Want?,” Pacific Philosophical Quarterly , 86(3): 295–312.
  • Funkhouser, Eric, and David Barrett, 2017, “Reply to Doody,” Philosophical Psychology , 30(5): 677–681.
  • –––, 2016, “Robust, unconscious self-deception: Strategic and flexible,” Philosophical Psychology , 29(5): 1–15.
  • Funkhouser, Eric, and Kyle Hallam, 2022, “Self-Handicapping and Self-Deception: A Two-Way Street,” Philosophical Psychology ,
  • Galeotti, Anna Elisabetta, 2018, Political Self-Deception , Cambridge: Cambridge University Press.
  • –––, 2016a, Straight and Twisted Self-Deception,” Phenomenology and Mind , 11: 90–99.
  • –––, 2016b, “The Attribution of Responsibility to Self-Deceivers,” Journal of Social Philosophy , 47(4): 420–438.
  • –––, 2012, “Self-Deception: Intentional Plan or Mental Event?” Humana Mente , 20: 41–66.
  • Gendler, T. S., 2007, “Self-Deception as Pretense,” Philosophical Perspectives , 21: 231–258.
  • Gibson, Quinn Hiroshi, 2020, “Self-Deception as Omission,” Philosophical Psychology , 33(5): 657–678.
  • Gilbert, Margaret, 2005, “Corporate Misbehavior and Collective Values,” Brooklyn Law Review , 70(4): 1369–80.
  • –––, 1994, “Remarks on Collective Belief,” in Socializing Epistemology , F. Schmitt (ed.), Lanham, MD: Rowman and Littlefield.
  • –––, 1989, On Social Facts , London: Routledge.
  • Goleman, Daniel, 1989, “What is negative about positive illusions?: When benefits for the individual harm the collective,” Journal of Social and Clinical Psychology , 8: 190–197.
  • Graham, G., 1986. “Russell’s Deceptive Desires,” The Philosophical Quarterly , 36: 223–229.
  • Haight, R. M., 1980, A Study of Self-Deception , Sussex: Harvester Wheatsheaf.
  • Hales, S. D., 1994, “Self-Deception and Belief Attribution,” Synthese , 101: 273–289.
  • Hassin, Ran, John Bargh, & Shira Cohen-Zimerman, 2009, “Automatic and Flexible: The Case of Nonconscious Goal Pursuit,” Social Cognition , 27: 20–36.
  • Hernes, C., 2007, “Cognitive Peers and Self-Deception,” teorema , 26(3): 123–130.
  • Hauerwas, S. and Burrell, D., 1977, “Self-Deception and Autobiography: Reflections on Speer’s Inside the Third Reich,” in Truthfulness and Tragedy , S. Hauerwas with R. Bondi and D. Burrell, Notre Dame: University of Notre Dame Press.
  • Holton, Richard, 2022, “Self-Deception and the Moral Self,” in Manuel Vargas, and John M. Doris (eds.), The Oxford Handbook of Moral Psychology , Oxford University Press, 262–284.
  • –––, 2001, “What is the Role of the Self in Self-Deception?,” Proceedings of the Aristotelian Society , 101(1): 53–69.
  • Jenni, K., 2003, “Vices of Inattention,” Journal of Applied Philosophy , 20(3): 279–95.
  • Johnson, D., and Fowler, J. 201, “The Evolution of Overconfidence,” Nature , 477: 317–320.
  • Johnston, M., 1988, “Self-Deception and the Nature of Mind,” in Perspectives on Self-Deception , B. McLaughlin and A. O. Rorty (eds.), Berkeley: University of California Press.
  • Jordan, Maiya, 2022, “Instantaneous Self-Deception,” Inquiry , 65(2): 176–201.
  • –––, 2020, “Literal Self-Deception,” Analysis , 80(2): 248–256.
  • –––, 2019, “Secondary Self-Deception,” Ratio , 32(2): 122–130.
  • Jurjako, Mark, 2013, “Self-deception and the selectivity problem,” Balkan Journal of Philosophy , 5: 151–162.
  • Kahlil, E., 2011, “The Weightless Hat,” Commentary on target article, “The Evolution and Psychology of Self-Deception,” by W. von Hippel and R. Trivers, Behavioral and Brain Sciences , 34(1): 30–31.
  • Kirsch, J., 2005, “What’s So Great about Reality?,” Canadian Journal of Philosophy , 35(3): 407–428.
  • Lauria, Federico and Delphine Preissmann, 2018, “What Does Emotion Teach Us about Self-Deception? Affective Neuroscience in Support of Non-Intentionalism,” Les Ateliers de l’Éthique / The Ethics Forum , 13(2): 70–94.
  • Lauria, F., Preissmann, D., and Cle menta, F., 2016, "Self-deception as Affective Coping: An Empirical Perspective on Philosophical Issues." Consciousness and Cognition , 41: 119–134.
  • Lazar, A., 1999, “Deceiving Oneself Or Self-Deceived?,” Mind , 108: 263–290.
  • –––, 1997, “Self-Deception and the Desire to Believe,” Behavioral and Brain Sciences , 20: 119–120.
  • Lamba, S and Nityananda, V., 2014, “Self-Deceived Individuals are Better at Deceiving Others,” PLOS One , 9/8: 1–6.
  • Levy, N., 2004, “Self-Deception and Moral Responsibility,” Ratio (new series) , 17: 294–311.
  • Linehan, E. A. 1982, “Ignorance, Self-deception, and Moral Accountability,” Journal of Value Inquiry , 16: 101–115.
  • Lockhard, J. and Paulhus, D. (eds.), 1988, Self-Deception: An Adaptive Mechanism? , Englewood Cliffs: Prentice-Hall.
  • Longeway, J., 1990, “The Rationality of Self-deception and Escapism,” Behavior and Philosophy , 18: 1–19.
  • Lopez, J., and M. Fuxjager, 2012, “Self-deception’s adaptive value: Effects of positive thinking and the winner effect,” Consciousness and Cognition , 21(1): 315–324.
  • Lynch, Kevin, 2022, “Being Self-Deceived about One’s Own Mental State,” Philosophical Quarterly , 72(3): 652–672.
  • –––, 2017, “An Agentive Non-Intentionalist Theory of Self-Deception,” Canadian Journal of Philosophy , 47(6): 779–798.
  • –––, 2014, “Self-deception and shifts of attention,” Philosophical Explorations , 17(1): 63–75.
  • –––, 2013, “Self-Deception and Stubborn Belief,” Erkenntnis , 78(6): 1337–1345.
  • –––, 2012, “On the ‘tension’ inherent in Self-Deception,” Philosophical Psychology , 25(3): 433–450.
  • –––, 2010, “Self-deception, religious belief, and the false belief condition,” Heythrop Journal , 51(6): 1073–1074.
  • –––, 2009, “Prospects for an Intentionalist Theory of Self-Deception,” Abstracta , 5(2): 126–138.
  • MacKenzie, J., 2022, “Self-Deception as a Moral Failure,” The Philosophical Quarterly , 72(2): 402-421.
  • Martin, M., 1986, Self-Deception and Morality , Lawrence: University Press of Kansas.
  • ––– (ed.), 1985, Self-Deception and Self-Understanding , Lawrence: University Press of Kansas.
  • Martínez Manrique, F., 2007, “Attributions of Self-Deception,” teorema , 26(3): 131–143.
  • McHugh, Conor, 2017, “Attitudinal Control,” Synthese , 194(8): 2745–2762.
  • –––, 2014, “Exercising Doxastic Freedom,” Philosophy and Phenomenological Research , 88(1): 1–37.
  • –––, 2013, “Epistemic Responsibility and Doxastic Agency,” Philosophical Issues , 23(1): 132–157.
  • McLaughlin, B. and Rorty, A. O. (eds.), 1988, Perspectives on Self-Deception , Berkeley: University of California Press.
  • McKay, R. and Dennett, D., 2009, “The Evolution of Misbelief,” Behavioral and Brain Sciences , 32(6): 493–561.
  • McKay, R., Prelec, D., “Protesting Too Much: Self-Deception and Self-Signaling,” Commentary on target article, “The Evolution and Psychology of Self-Deception,” by W. von Hippel and R. Trivers, Behavioral and Brain Sciences , 34(1): 34–35.
  • Mele, Alfred, 2020, “Self-Deception and Selectivity,” Philosophical Studies , 177: 2697–2711.
  • –––, 2012, “When Are We Self-Deceived?” Humana Mente Journal of Philosophical Studies , 20: 1–15.
  • –––, 2010, “Approaching Self-Deception: How Robert Audi and I Part Company,” Consciousness and Cognition , 19: 745–750.
  • –––, 2009, “Delusional Confabulations and Self-Deception,” in W. Hirstein (ed.), Confabulation: Views from Neuroscience, Psychiatry, Psychology, and Philosophy , Oxford: Oxford University Press, pp. 139–157.
  • –––, 2009, “Have I Unmasked Self-Deception or Am I Self-Deceived?” in C. Martin (ed.), The Philosophy of Deception , Oxford: Oxford University Press, pp. 260–276.
  • –––, 2007, “Self-Deception and Three Psychiatric Delusions: On Robert Audi’s Transition from Self-Deception to Delusion,” in M. Timmons, J. Greco, and A. Mele (eds.), Rationality and the Good , Oxford: Oxford University Press, pp. 163–175.
  • –––, 2007, “Self-Deception and Hypothesis Testing,” in M. Marraffa, M. De Caro, and F. Feretti (eds.), Cartographies of the Mind , Dordrecht: Kluwer, pp. 159–167.
  • –––, 2006, “Self-Deception and Delusions,” European Journal of Analytic Philosophy , 2: 109–124.
  • –––, 2003, “Emotion and Desire in Self-Deception,” in A. Hatzimoysis (ed.), Philosophy and the Emotions , Cambridge: Cambridge University Press, pp. 163–179.
  • –––, 2001, Self-Deception Unmasked , Princeton: Princeton University Press.
  • –––, 2000, “Self-Deception and Emotion,” Consciousness and Emotion , 1: 115–137.
  • –––, 2000, “Self-Deception and Emotion,” Consciousness and Emotion , 1: 115–139.
  • –––, 1999, “Twisted Self-Deception,” Philosophical Psychology , 12: 117–137.
  • –––, 1997, “Real Self-Deception,” Behavioral and Brain Sciences , 20: 91–102.
  • –––, 1987a, Irrationality: An Essay on Akrasia, Self-Deception, Self-Control , Oxford: Oxford University Press.
  • –––, 1987b, “Recent Work on Self-deception,” American Philosophical Quarterly , 24: 1–17.
  • –––, 1983, “Self-Deception,” Philosophical Quarterly , 33: 365–377.
  • Mijović-Prelec, D., and Prelec, D., 2010, “Self-deception as Self-Signaling: A Model and Experimental Evidence,” Philosophical Transactions of the Royal Society B , 365: 227–240.
  • Moran, R., 1988, “Making Up Your Mind: Self-Interpretation and Self-constitution,” Ratio (new series) , 1: 135–151.
  • Nelkin, D., 2012, “Responsibility and Self-Deception: A Framework,” Humana Mente Journal of Philosophical Studies , 20: 117–139.
  • –––, 2002, “Self-Deception, Motivation, and the Desire to Believe,” Pacific Philosophical Quarterly , 83: 384–406.
  • Nicholson, A., 2007.“Cognitive Bias, Intentionality and Self-Deception,” teorema , 26(3): 45–58.
  • Noordhof, P., 2009, “The Essential Instability of Self-Deception,” Social Theory and Practice , 35(1): 45–71.
  • –––, 2003, “Self-Deception, Interpretation and Consciousness,” Philosophy and Phenomenological Research , 67: 75–100.
  • Paluch, S., 1967, “Self-Deception,” Inquiry , 10: 268–78.
  • Patten, D., 2003, “How do we deceive ourselves?,” Philosophical Psychology , 16(2): 229–46.
  • Pears, D., 1991, “Self-Deceptive Belief Formation,” Synthese , 89: 393–405.
  • –––, 1984, Motivated Irrationality , New York: Oxford University Press.
  • Pettit, Philip, 2006, “When to Defer to Majority Testimony — and When Not,” Analysis , 66(3): 179–187.
  • –––, 2003, “Groups with Minds of Their Own,” in Socializing Metaphysics , F. Schmitt (ed.), Lanham, MD: Rowman and Littlefield.
  • Philström, S., 2007, “Transcendental Self-Deception,” teorema , 26(3): 177–189.
  • Porcher, J., 2012, “Against the Deflationary Account of Self-Deception,” Humana Mente , 20: 67–84.
  • Quinton, Anthony, 1975/1976, “Social Objects,” Proceedings of the Aristotelian Society , 75: 1–27.
  • Räikkä, J. 2007, “Self-Deception and Religious Beliefs,” Heythrop Journal , 48: 513–526.
  • Rorty, A. O., 1994, “User-Friendly Self-Deception,” Philosophy , 69: 211–228.
  • –––, 1983, “Akratic Believers,” American Philosophical Quarterly , 20: 175–183.
  • –––, 1980, “Self-Deception, Akrasia and Irrationality,” Social Science Information , 19: 905–922.
  • –––, 1972, “Belief and Self-Deception,” Inquiry , 15: 387–410.
  • Rowbottom, D. & Chan, C., 2019, “Self-Deception and Shifting Degrees of Belief,” Philosophical Psychology , 32: 1204–1220.
  • Sahdra, B. and Thagard, P., 2003, “Self-Deception and Emotional Coherence,” Minds and Machines , 13: 213–231.
  • Sanford, D,1988, “Self-Deception as Rationalization,” in Perspectives on Self-Deception , B. McLaughlin and A. O. Rorty (eds.), Berkeley: University of California Press.
  • Sartre, J-P., 1946, L’etre et le néant , Paris: Gallimard; trans. H. E. Barnes, 1956, Being and Nothingness , New York, Washington Square Press.
  • Scott-Kakures, Dion, 2021, “Self-Deceptive Inquiry: Disorientation, Doubt, Dissonance,” Midwest Studies in Philosophy , 45: 457–482.
  • –––, 2002, “At Permanent Risk: Reasoning and Self-Knowledge in Self-Deception,” Philosophy and Phenomenological Research , 65: 576–603.
  • –––, 2001, “High anxiety: Barnes on What Moves the Unwelcome Believer,” Philosophical Psychology , 14: 348–375.
  • –––, 2000, “Motivated Believing: Wishful and Unwelcome,” Noûs , 34: 348–375.
  • Sher, George, 2009, Who Knew? Responsibility without Awareness , Oxford: Oxford University Press.
  • Smith, D. L., 2014, "Self-Deception: A Teleofunctional Approach," Philosophia, 42: 181–199.
  • –––, 2013, "The Form and Function of Self-Deception: A Biological Model," Sistemi intelligenti , 3: 565–580.
  • Sorensen, R., 1985, “Self-Deception and Scattered Events,” Mind , 94: 64–69.
  • Surbey, M., 2004, “Self-deception: Helping and hindering personal and public decision making,” in Evolutionary Psychology, Public Policy and Personal Decisions , C. Crawford and C. Salmon (eds.), Mahwah, NJ: Lawrence Earlbaum Associates.
  • Schwitzgebel, E., 2001, “In-Between Believing,” Philosophical Quarterly , 51: 76–82.
  • Szabados, B., 1973, “Wishful Thinking and Self-Deception,” Analysis , 33(6): 201–205.
  • Talbott, W. J., 1997, “Does Self-Deception Involve Intentional Biasing,” Behavior and Brain Sciences , 20: 127.
  • –––, 1995, “Intentional Self-Deception in a Single Coherent Self,” Philosophy and Phenomenological Research , 55: 27–74.
  • Taylor, S. and Brown, J., 1994, “Positive Illusion and Well-Being Revisited: Separating Fact from Fiction,” Psychological Bulletin , 116: 21–27.
  • Taylor, S. and Brown, J., 1988, “Illusion and Well-Being: A Social Psychological Perspective on Mental Health,” Psychological Bulletin , 103(2): 193–210.
  • Tenbrusel, A.E. and D. M Messick, 2004, “Ethical Fading: The Role of Self-Deception in Unethical Behavior,” Social Justice Research , 7(2): 223–236.
  • Trivers, R., 2011, The Folly of Fools: The Logic of Deceit and Self-Deception in Human life , New York: Basic Books.
  • Trivers, R., 2000, “The Elements of a Scientific Theory of Self-Deception,” in Evolutionary Perspectives on Human Reproductive Behavior , Dori LeCroy and Peter Moller (eds.), Annals of the New York Academy of Sciences , 907: 114–131.
  • Trivers, R., 1991, “Deceit and Self-Deception: The relationship between Communication and Consciousness,” Man and Beast Revisited , 907: 175–191.
  • Van Fraassen, B., 1995, “Belief and the Problem of Ulysses and the Sirens,” Philosophical Studies , 77: 7–37.
  • –––, 1984, “Belief and Will,” Journal of Philosophy , 81: 235–256.
  • Van Leeuwen, N., 2013a, “Review of Robert Trivers’ The Folly of Fools: The Logic of Deceit and Self-Deception in Human Life,” Cognitive Neuropsychiatry , 18(1-2): 146–151.
  • –––, 2013b, “Self-Deception,” in International Encyclopedia of Ethics , H. LaFollette (ed.), New York: Wiley-Blackwell.
  • –––, 2009, “Self-Deception Won’t Make You Happy,” Social Theory and Practice , 35(1): 107–132.
  • –––, 2007a, “The Spandrels of Self-deception: Prospects for a biological theory of a mental phenomenon,” Philosophical Psychology , 20(3): 329–348.
  • –––, 2007b, “The Product of Self-Deception,” Erkenntnis , 67(3): 419–437.
  • Van Loon, Marie, 2018, “Responsibility for self-deception,” Les Ateliers de l’Éthique / the Ethics Forum , 13(2): 119–134.
  • Von Hippel, W. & Trivers, R., 2011, “The Evolution and Psychology of Self-Deception,” Behavioral and Brain Sciences , 34(1): 1–56.
  • Vrij, A., 2011, “Self-deception, lying, and the ability to deceive,” Behavioral and Brain Sciences , 34(1): 40–41.
  • Wei, Xintong, 2020, “The role of pretense in the process of self-deception,” Philosophical Explorations , 23(1): 1–14.
  • Whisner, W., 1993, “Self-Deception and Other-Person Deception,” Philosophia , 22: 223–240.
  • –––, 1989, “Self-Deception, Human Emotion, and Moral Responsibility: Toward a Pluralistic Conceptual Scheme,” Journal for the Theory of Social Behaviour , 19: 389–410.
How to cite this entry . Preview the PDF version of this entry at the Friends of the SEP Society . Look up topics and thinkers related to this entry at the Internet Philosophy Ontology Project (InPhO). Enhanced bibliography for this entry at PhilPapers , with links to its database.

[Please contact the author with suggestions.]

action | belief | Davidson, Donald | delusion | lying and deception: definition of | moral responsibility | Sartre, Jean-Paul | self-knowledge | weakness of will

Acknowledgments

The author would like to thank Margaret DeWeese-Boyd, Douglas Young, and the editors for their help in constructing and revising this entry.

Copyright © 2023 by Ian Deweese-Boyd < ian . deweese-boyd @ gordon . edu >

  • Accessibility

Support SEP

Mirror sites.

View this site from another server:

  • Info about mirror sites

The Stanford Encyclopedia of Philosophy is copyright © 2023 by The Metaphysics Research Lab , Department of Philosophy, Stanford University

Library of Congress Catalog Data: ISSN 1095-5054

  • More from M-W
  • To save this word, you'll need to log in. Log In

wishful thinking

Definition of wishful thinking

Examples of wishful thinking in a sentence.

These examples are programmatically compiled from various online sources to illustrate current usage of the word 'wishful thinking.' Any opinions expressed in the examples do not represent those of Merriam-Webster or its editors. Send us feedback about these examples.

Word History

1932, in the meaning defined above

Dictionary Entries Near wishful thinking

wishful thinker

Cite this Entry

“Wishful thinking.” Merriam-Webster.com Dictionary , Merriam-Webster, https://www.merriam-webster.com/dictionary/wishful%20thinking. Accessed 27 Aug. 2024.

More from Merriam-Webster on wishful thinking

Nglish: Translation of wishful thinking for Spanish Speakers

Subscribe to America's largest dictionary and get thousands more definitions and advanced search—ad free!

Play Quordle: Guess all four words in a limited number of tries.  Each of your guesses must be a real 5-letter word.

Can you solve 4 words at once?

Word of the day.

See Definitions and Examples »

Get Word of the Day daily email!

Popular in Grammar & Usage

Plural and possessive names: a guide, 31 useful rhetorical devices, more commonly misspelled words, why does english have so many silent letters, your vs. you're: how to use them correctly, popular in wordplay, 8 words for lesser-known musical instruments, it's a scorcher words for the summer heat, 7 shakespearean insults to make life more interesting, birds say the darndest things, 10 words from taylor swift songs (merriam's version), games & quizzes.

Play Blossom: Solve today's spelling word game by finding as many words as you can using just 7 letters. Longer words score more points.

Hope is more than just wishful thinking

wishful thinking research definition

We hear parents, teachers, and students use the word  hope  every day. But what exactly does it mean? When we read or hear the word, we might think of a positive outlook or desire, yet its true definition is nebulous. It implies that something will automatically or magically occur without effort. Even though having an optimistic outlook is important to overall well-being, a person with an unstructured and immeasurable concept of hope is prey to vague expectations—as if he is a passive bystander waiting for an outcome to come about.

What happens when it is instead viewed as an active construct? Studies have found that hope, when used in a proactive manner, can become a useful framework to help achieve goals and contribute to personal and psychological health.

Positive Psychology: A Strengths Approach

Historically, psychologists have approached the study of psychological well-being from a deficit perspective, focusing on treating and alleviating pathologies. Over time, they have taken an increasingly proactive and positive approach to the study and development of individuals and their happiness. Positive psychologists focus on developing personal strengths, fostering the growth of positive responses to adversity, and strengthening social and emotional foundations in patient’s lives. They study well-being, contentment, and satisfaction with the past, flow and happiness in the present, and optimism for the future.

Hope is one of many positive psychological constructs that contribute to overall health. Karl Menninger was one of the first to study hope as a psychological construct by defining hope as the positive expectation for attaining goals. He presented the idea that hope, while a basic part of daily human operation, was ill-defined and obscure, and theorized that some mental illness reflected a lack of hope. He also believed that successful treatment included reestablishing hope for those who were suffering.

Contemporary researchers have continued to build upon a “strengths approach” to examining the role of hope in psychological well-being. Hope is the link between goals dreamed today and the attainment of those goals in the future. Hopeful individuals view themselves as being able to create paths (“pathways thinking”) to achieve their goals; they initiate steps (“agency thinking”) toward achieving goals and sustain their course along a route to success.

What Hope Research Tells Us

C. R. Snyder and his team of contemporary researchers have written about hopefulness. Snyder developed the Adult Dispositional Hope Scale and the Children’s Hope Scale to offer researchers an active, measureable framework for the study of hope. Using Snyder’s hope scales, research has found that high-hope individuals:

  • Are able to set goals, make flexible plans to achieve the goals they set, and take action toward goal attainment. When comparing high-hope and low hope subjects, high-hopers think more positively about themselves, set higher goals, and select more goals.
  • Have a stronger belief in the likelihood that they will achieve their goals. They focus on success.
  • Possess self-referential beliefs in situations of adversity. Those who are hopeful have an undercurrent of internal self-statements: “I can,” “I’ll make it,” and “I won’t give up.”
  • Trust in themselves to be able to adjust to prospective trouble and losses.

On the other hand, individuals lacking hope:

  • Believe that pathways to their goals are unavailable to them; they set low goals and have a sense of uncertainty and failure about being able to achieve their goals.
  • Have a tendency to experience negative emotions when working toward their goals.

Hope and the Gifted

Research on the contribution of hope to the overall well-being of gifted learners and the long-term importance on school outcomes is just beginning. So far, researchers have found that gifted students with high hope achieve successful school-related outcomes, demonstrate higher success on standardized achievement tests, and set higher global academic goals and expectancies of success. Students with low hope experience greater anxiety and self-doubt.

High-hope students don’t belittle themselves when they’re not successful. They don’t let their failures affect their ultimate sense of worth, but rather attribute their failure to a lack of effort or strategies for success. High-hope students have superior academic and interpersonal satisfactions when compared to those with low hope.

Researchers have also examined levels of hopefulness and the contribution hope makes to personal well-being in college freshman honors participants and early entrants, with a subsequent check-in during sophomore year. Findings include:

  • Being hopeful was strongly associated with positive personal well-being.
  • Gifted students who enter college with a higher sense of hope experience greater positive feelings of overall personal well-being initially—and after one year in college.
  • Honors college students who enter college with a strong ability to identify multiple paths toward goals appear to have moved toward achieving those goals after one year of college.

Increasing Student Hope

Although children in America generally feel hopeful, all children do not have the same level of hopefulness. At times, gifted children can find themselves in settings that are socially and intellectually stagnant. Gifted students may have a hard time finding intellectual peers and stimulating cognitive challenges commensurate with their abilities in a regular classroom. School can seem hopeless without goal-setting skills, without identifying paths to attain goals, and with low confidence toward accomplishing goals. Ideally, parents, teachers, and counselors should collaborate to help children develop goal-setting skills.

Setting goals :  Teaching goal setting is foundational to assisting gifted children in the development of hopefulness. Gifted children need to be active participants in setting their own personal, social, and academic goals. In early grades, gifted students should be encouraged to set simple and specific goals that move them toward getting something accomplished (“approach goals”) rather than avoiding something (“avoidance goals”). As a child ages, goal setting can become more complex. Listing and ranking goals can help students learn the skill of prioritizing. Setting multiple goals should be encouraged in middle and high school, as multiple goals offer a fallback position if students encounter an intransigent obstacle. Students must also be taught how to identify and set markers of success.

Developing “pathways thinking” :  Parents and educators can help gifted students develop pathways thinking by breaking down larger goals into smaller components that are approached, and eventually completed, in a sequenced and logical way. Students should identify multiple routes to both small and large goals and practice overcoming obstacles. In this way, they learn to become flexible thinkers who see obstructions as pathways that don’t work. If a pathway is identified as unfeasible, students must learn the skills to switch lanes and find alternate routes.

Developing personal meaning or “agency thinking” :  It’s essential that children set goals that are their own—with personal meaning—rather than assuming the goals of their peers, teachers, or parents. Motivation, persistence, and performance are undermined if goals are not personally determined. Gifted children should also set “stretch” goals, keep journals of internal dialogues, and engage in team-related activities to foster personal meaning. In addition, developing memories of positive experiences, either through personal successes or by the example of others, helps keep gifted children resilient when they face difficulty reaching their goals.

The aim of those who parent and educate gifted kids is to foster academic success and happiness. Hope, when framed as a goal-directed and active process, helps students thrive academically and personally. It is therefore much more than wishful thinking about positive outcomes. When the gifted are able to set goals, see multiple routes to goals, and move toward the attainment of those goals, they have hope for fulfilling their potential. So the next time you use or hear the word hope, think twice about both what it means and your role in helping those who are hopeful.

This blog post, an excerpt from  Parenting for High Potential  (March 2014), is by Janette Boazman, the chair of education at the University of Dallas.

Editor's note: This is part of a series of blog posts that is collaboratively published every week by the Thomas B. Fordham Institute and National Association for Gifted Children. Each post in the series exists both here on  Flypaper  and on  the NAGC Blog .

  • Skip to main content
  • Skip to primary sidebar
  • Skip to footer

Psychology Spot

All About Psychology

Wishful thinking: When we see only what we want to see

wishful thinking

A psychoanalyst believed that dreaming of fish was the cause of all psychological disorders. When patients came to him and began to talk about their problems, the analyst interrupted them and asked:

Psychoanalyst: Excuse me, but did you dream last night?

Patient: I do not know… Yes, I think so.

Psychoanalyst: You have dreamed of fish, didn’t you?

Patient: Ah… no, no.

Psychoanalyst: What did you dream then?

Patient: Well, I dreamed I was walking along a road.

Psychoanalyst: Was there a puddle on the road?

Patient: Well … I do not know.

Psychoanalyst: But, could there have been one?

Patient: I think so.

Psychoanalyst: There could have been some fish in that puddle?

Patient: No, no.

Psychoanalyst: On the road of your dream there was any restaurant?

Patient: No.

Psychoanalyst: But there may have been one…

Patient: Well, I think maybe there could be a restaurant.

Psychoanalyst: Do they served any fish at that restaurant?

Patient: Well, I think as a restaurant perhaps they served some fish.

Psychoanalyst: Aha! I knew it! You dreamed fish.

Neither rational nor logical

When we want something, we try to find ways to get it, and the thought is the main tool we have to decide which directions take. Each of these decisions will gradually approaching or moving us away from what we want. As the psychoanalyst of the story, we trust those decisions because we believe that our thinking is logical and rational, we do not consider that it is profoundly influenced by our desires, expectations and dreams.

In fact, the desire tells us that we are missing something, and  wishful thinking  is the mental response to that need. The problem is that sometimes this desire is so great that thought becomes its slave.

Then we are not able to see the signs that we’re going off the road, we see only what we want to see and draw the conclusions we want, ignoring the reality. And this can cause us serious problems.

The pitfalls of wishful thinking

Wishful thinking comes from desire, not from reality, so often become a source of many problems in our daily lives.

– Ensures that we focus too much into results.  Wishful thinking is mostly concrete, is aimed at achieving a goal in the shortest time possible. That sense of urgency caused by desire is dangerous, it prevents us from planning properly and perceive the signals we are on the wrong road. Basically, the wishful thinking gives us the logical justifications we need to make mistakes and take the path we want, even if it is not the best one or convenient.

– Prevents us from seeing evidence and obstacles.  Wishful thinking makes sure that we draw conclusions and make decisions based on what we like, rather than analyzing the details taking reality into account. This kind of thinking is based on emotions rather than on evidence. The basic problem is that we want so much for something to happen or be true to reject evidence of the contrary. In fact, a study conducted at the University of New York showed that wishful thinking feeds fantasies about our future that we focus more on benefits than disadvantages, and at the end, we take worst decisions.

– Cause great frustrations.  One of the main problems caused by wishful thinking is that we end up living in the fantasy world that we built in our mind, based on erroneous conclusions that we have taken. In practice, it is as if we lived in a house of cards. Obviously, as soon as the wind blows that castle falls. Then we are overwhelmed by frustration, disappointment and sadness.

Why do we allow wishful thinking take the reins?

A study at the University of Illinois showed that our thinking goes something like memory. That is, sometimes distorts some memories forgetting some details and adding new ones because it appears us more convenient and to avoid painful memories , or to consolidate the image we have of ourselves, avoiding dissonances.

These psychologists asked a group of people to read a set of predictions, some were very convenient for them while others absolutely negative. The task of each participant was determining the reliability of the source. Interestingly, people tended to indicate that the most convenient statements came from reliable sources, while negative predictions were attributed to unreliable sources. In practice, our desire to ensure that things are working well makes us lose the perspective.

We can also let wishful thinking take the reins to avoid having to deal with a complex reality, that we are not willing to accept. When things do not work as we would like, instead of accepting reality we decide to look only at what we like and confirms our view of the world. That way we feel better.

Of course, this mechanism works normally at a subconscious level.

How to use the wishful thinking in our favor?

Wishful thinking in itself is not negative, it is only necessary to learn how to use it for our advantage. In fact, the desire that lies at its base has an enormous motivating power which we can exploit through the WOOP technique (Wish, Outcome, Obstacle and Plan):

  • Wish. Think of something you want intensely. That desire becomes an achievable goal. Keep in mind that while desire can be vague and general, the goal should be concrete and quantifiable.
  • Outcome. Imagine the best possible outcome, let yourself be filled with a feeling of accomplishment.
  • Obstacle. Come back to reality, concentrate in the obstacles that keep you from achieving this goal and choose the greatest.
  • Plan. Think of possible solutions and select a concrete and effective action that will allow you to overcome this obstacle.

Kappes, H. B. & Oettingen, G. (2012) Wishful Information Preference. Positive Fantasies Mimic the Effects of Intentions. Pers Soc Psychol Bull; 38(7): 870-881.

Gordon, R. et. Al. (2005) Wishful thinking and source monitoring. Memory & Cognition; 33(3): 418-429.

Jennifer Delgado

Psychologist Jennifer Delgado

I am a psychologist and I spent several years writing articles for scientific journals specialized in Health and Psychology. I want to help you create great experiences. Learn more about me .

wishful thinking research definition

3 good reasons NOT to go to the psychologist

26/08/2024 By Jennifer Delgado

wishful thinking research definition

5 habits that destroy your self-esteem

wishful thinking research definition

How to stop punishing yourself for a past mistake?

24/08/2024 By Jennifer Delgado

Cambridge Dictionary

  • Cambridge Dictionary +Plus

Meaning of wishful thinking in English

Your browser doesn't support HTML5 audio

  • appealingly
  • beyond someone's expectations
  • bright spot
  • hopefulness
  • it's an ill wind (that blows nobody any good) idiom
  • over-optimistic
  • over-optimistically
  • pin your hopes on something/someone idiom
  • please God idiom

wishful thinking | American Dictionary

Examples of wishful thinking, translations of wishful thinking.

Get a quick, free translation!

{{randomImageQuizHook.quizId}}

Word of the Day

a very short holiday

Trial, judge, and jury: talking about what happens when a criminal is caught

Trial, judge, and jury: talking about what happens when a criminal is caught

wishful thinking research definition

Learn more with +Plus

  • Recent and Recommended {{#preferredDictionaries}} {{name}} {{/preferredDictionaries}}
  • Definitions Clear explanations of natural written and spoken English English Learner’s Dictionary Essential British English Essential American English
  • Grammar and thesaurus Usage explanations of natural written and spoken English Grammar Thesaurus
  • Pronunciation British and American pronunciations with audio English Pronunciation
  • English–Chinese (Simplified) Chinese (Simplified)–English
  • English–Chinese (Traditional) Chinese (Traditional)–English
  • English–Dutch Dutch–English
  • English–French French–English
  • English–German German–English
  • English–Indonesian Indonesian–English
  • English–Italian Italian–English
  • English–Japanese Japanese–English
  • English–Norwegian Norwegian–English
  • English–Polish Polish–English
  • English–Portuguese Portuguese–English
  • English–Spanish Spanish–English
  • English–Swedish Swedish–English
  • Dictionary +Plus Word Lists
  • English    Noun
  • American    Noun
  • Translations
  • All translations

To add wishful thinking to a word list please sign up or log in.

Add wishful thinking to one of your lists below, or create a new one.

{{message}}

Something went wrong.

There was a problem sending your report.

  • Dictionaries home
  • American English
  • Collocations
  • German-English
  • Grammar home
  • Practical English Usage
  • Learn & Practise Grammar (Beta)
  • Word Lists home
  • My Word Lists
  • Recent additions
  • Resources home
  • Text Checker

Definition of wishful thinking noun from the Oxford Advanced American Dictionary

wishful thinking

Definitions on the go

Look up any word in the dictionary offline, anytime, anywhere with the Oxford Advanced Learner’s Dictionary app.

wishful thinking research definition

IMAGES

  1. PPT

    wishful thinking research definition

  2. What Is Wishful Thinking and 5 Types of People Who Are Most Prone to It

    wishful thinking research definition

  3. Wishful Thinking

    wishful thinking research definition

  4. PPT

    wishful thinking research definition

  5. PPT

    wishful thinking research definition

  6. Wishful Thinking as a Proof Technique

    wishful thinking research definition

COMMENTS

  1. Wishful thinking

    Wishful thinking. Wishful thinking is the formation of beliefs based on what might be pleasing to imagine, rather than on evidence, rationality, or reality. It is a product of resolving conflicts between belief and desire. [ 1] Methodologies to examine wishful thinking are diverse.

  2. Research Explores Why We Tend To Employ Wishful Thinking

    Instead, they tended to perceive patterns that aligned with more desirable outcomes. While previous research primarily focused on wishful thinking concerning positive outcomes, yielding mixed results, this study distinctly highlights how the negative emotion of anxiety regarding an outcome can drive wishful thinking. Increasing realism

  3. The science of hope: More than wishful thinking

    According to social science researchers in Arizona State University's Center for the Advanced Study and Practice of Hope, having hope is harder work than we may think. They are unpacking the ways in which different populations conceive of hope and exploring how having higher hopes can improve our health and well-being.

  4. Neural correlates of wishful thinking

    Wishful thinking (WT) implies the overestimation of the likelihood of desirable events. It occurs for outcomes of personal interest, but also for events of interest to others we like. We investigated whether WT is grounded on low-level selective attention or on higher level cognitive processes including differential weighting of evidence or ...

  5. All Thinking is 'Wishful' Thinking: Trends in Cognitive Sciences

    Whereas the term 'wishful thinking' is typically reserved for the motivation for specific certainty, the motivations to avoid it, or to have (or avoid) nonspecific certainties can be no less motivating or 'wished for', supporting our claim that 'all thinking is wishful thinking'; in other words, all thinking is motivated.

  6. Wishful Thinking: Definition, Example & How Wishful Thinking Works

    Wishful Thinking is a near-universal tendency to expect positive outcomes and consequences from events, even when the likelihood of a positive outcome is just 50-50 or worse. Wishful Thinking, as a heuristic, describes how wishing for pleasant outcomes/effects influences decisions. It also applies to making decisions based on underestimating ...

  7. Wishful Thinking: What It Is And How 'wishful Thinking' Affects Us

    Wishful thinking refers, in psychology, to the process of forming opinions and making decisions. using desires and emotions, instead of relying on facts or rationality Illusionary thinking is based on emotions and is a product of the conflict between desire and reality. Research suggests that if the situation remains static, people tend to ...

  8. PDF Wishful Thinking final

    wishful thinking can be fruitfully characterized in terms of the mechanisms that generate them and the problems associated with them. We highlight the array of mechanisms associated with wishful thinking, as well as the fact that it can be evaluated both from epistemic and ethical perspectives.

  9. Wish

    Wishful thinking is addressed next, an area of research informed much more by cognitive and neuroscience theories and methods and, therefore, "biased" towards discovering errors and biases in our (wishful) engagement with the world.

  10. Self-Deception

    In wishful thinking, motivation triggers a belief formation process in which the person does not play an active, conscious role, "while in self-deception the subject is a willing participant in directing cognition towards the doxastic embrace of the favored proposition" (Scott-Kakures 2002; see also Szabados 1973).

  11. Wishful Thinking and Procrastination

    Wishful thinking is the extent to which cognitions (e.g. expectations, judgments) are affected by motivations (e.g., desire for an outcome). Participants (N = 64) expected to have to complete a ...

  12. Wishful seeing: Motivational influences on visual perception of the

    We can also call such motivationally biased reasoning by its more everyday name, wishful thinking. Our mission in the research reported herein has been to ask how "deep" wishful thinking might go. Does wishful thinking have an accomplice—wishful seeing—that resides in perception? We begin by reflecting on the history of these questions.

  13. Hope Psychology: What Are The Benefits of Hope?

    This definition of hope is based on "Hope Theory," a positive psychology concept developed by American psychologist Charles Snyder. Positive psychology is the scientific study of what makes ...

  14. Wishful thinking Definition & Meaning

    The meaning of WISHFUL THINKING is the attribution of reality to what one wishes to be true or the tenuous justification of what one wants to believe. How to use wishful thinking in a sentence.

  15. Hope is more than just wishful thinking

    It is therefore much more than wishful thinking about positive outcomes. When the gifted are able to set goals, see multiple routes to goals, and move toward the attainment of those goals, they have hope for fulfilling their potential. So the next time you use or hear the word hope, think twice about both what it means and your role in helping ...

  16. Wishful thinking: When we see only what we want to see

    Wishful thinking makes sure that we draw conclusions and make decisions based on what we like, rather than analyzing the details taking reality into account. This kind of thinking is based on emotions rather than on evidence. The basic problem is that we want so much for something to happen or be true to reject evidence of the contrary.

  17. WISHFUL THINKING definition

    WISHFUL THINKING meaning: 1. the imagining or discussion of a very unlikely future event or situation as if it were possible…. Learn more.

  18. WISHFUL THINKING

    WISHFUL THINKING definition: 1. the imagining or discussion of a very unlikely future event or situation as if it were possible…. Learn more.

  19. Wishful Seeing: How Preferences Shape Visual Perception

    We discuss classic and current research demonstrating wishful seeing across two perceptual tasks, showing that people categorize ambiguous visual information and represent their environments in ways that align with their desires. Further, we outline when and how wishful seeing occurs.

  20. WISHFUL THINKING Definition & Meaning

    Wishful thinking definition: interpretation of facts, actions, words, etc., as one would like them to be rather than as they really are; imagining as actual what is not.. See examples of WISHFUL THINKING used in a sentence.

  21. wishful thinking noun

    Definition of wishful thinking noun in Oxford Advanced American Dictionary. Meaning, pronunciation, picture, example sentences, grammar, usage notes, synonyms and more.

  22. Wishful thinking

    wishful thinking: 1 n the illusion that what you wish for is actually true Type of: fancy , fantasy , illusion , phantasy something many people believe that is false