Despite robust evidence that misinformation continues to influence event-related reasoning after a clear retraction, evidence for the continued influence of misinformation on person impressions is mixed. Across four experiments, we investigated the impact of person-related misinformation and its correction on dynamic (moment-to-moment) impression formation. Participants formed an impression of a protagonist, “John”, based on a series of behaviour descriptions, including misinformation that was later retracted. Person impressions were recorded after the presentation of each behaviour description. As predicted, we found a strong effect of information valence on person impressions: negative misinformation had a greater impact on person impressions than positive misinformation (Experiments 1 and 2). Furthermore, in each experiment participants fully discounted the misinformation once retracted, regardless of whether the misinformation was negative or positive. This was true even when the other behaviour descriptions were congruent with (Experiment 2) or causally related to the retracted misinformation (Experiments 3 and 4). To test for generalisation, Experiment 4 used a different misinformation statement; it again showed no evidence for the continued influence of retracted misinformation on person impressions. Our findings indicate that person-related misinformation can be effectively discounted following a clear retraction.
In the digital information age, people are exposed to vast amounts of factual and false information. Consider the false allegation that Bill Gates was the architect of the COVID-19 pandemic, manufacturing the virus to profit from a future vaccine (Reuters, 2020), or the false claim that Hillary Clinton was involved in a child paedophile ring at a pizza parlour (O’Rourke, 2016). Of course, false allegations are also often levelled against less-known people both on social media and in offline contexts. This raises the question: Even when discredited, do false allegations continue to influence person impressions? Or, as the saying goes, ‘does mud stick’? Despite the threat misinformation poses to a person’s reputation, there is a limited understanding of the factors that moderate misinformation’s influence on impression formation. The present study examined this issue, testing the extent to which misinformation and subsequent corrections dynamically shape person impressions when these are first formed.
Continued Influence Effect
Misinformation penetrates deep into social networks, spreading further and faster than true information (Juul & Ugander, 2021; Vosoughi et al., 2018). This pernicious impact has inspired a wealth of psychological research into misinformation and its effects. Perhaps most worrying is the finding that misinformation often continues to influence people’s judgements, reasoning, and decision making even after it has been retracted or corrected. This phenomenon is known as the continued influence effect (CIE; Johnson & Seifert, 1994; Wilkes & Leatherbarrow, 1988) and it has been demonstrated across a large body of experimental work (for meta-analyses, see Chan et al., 2017; Walter & Murphy, 2018; Walter & Tukachinsky, 2020; for a review, see Ecker et al., 2022).
Research investigating the CIE has focused on event misinformation. In a common CIE paradigm, participants read a narrative text that describes a fictitious event, which contains a critical piece of (typically causal) information (e.g., that negligent storage of flammable materials started a warehouse fire); this misinformation is later retracted (e.g., it is clarified that no flammable materials were present) or not retracted. Although a retraction reduces the number of references to the misinformation participants make in response to inferential-reasoning questions about the event (e.g., “Why was there so much black smoke?”), participants often continue to rely on the retracted misinformation (e.g., Ecker et al., 2010, 2015). Similar effects are observed using variations to this paradigm with shorter statements (see Brydges et al., 2020; Gordon et al., 2017), implied rather than explicitly stated misinformation (e.g., Rich & Zaragoza, 2016; Tay et al., 2022), and real-world misinformation (e.g., the unfounded link between vaccination and autism; see Swire et al., 2021), adding to the robustness of the CIE for event misinformation. These findings align with the broader belief-perseverance literature, which shows that original beliefs are often retained despite disconfirming evidence (C. A. Anderson et al., 1980; Asch, 1946; Downey & Christensen, 2006; Green & Donahue, 2011).
One theoretical account of the CIE relates to the way people process event information. When learning about an unfolding event, people build mental models, which are characterised by their temporal sequence and causal structure (Bower & Morrow, 1990; van Oostendorp & Bonebakker, 1999). When misinformation is retracted, this can present a threat to model coherence because the retraction leaves a “gap” in the model. Under these circumstances, the correction may not be integrated into the model, and during reasoning preference may be given to an incorrect but complete mental model rather than a correct but incomplete one (Brydges et al., 2018; Ecker, Lewandowsky, Chang, et al., 2014; Johnson-Laird, 2012; Kendeou et al., 2019; Schul & Mayo, 2014). In support of this theoretical account, it has been found that misinformation is more resistant to correction if it is central to an event narrative, providing a causal explanation for, or otherwise enhancing comprehension of, the event (Hamby et al., 2020; Johnson & Seifert, 1994; Rich & Zaragoza, 2020; van Oostendorp & Bonebakker, 1999). These findings suggest that the CIE may arise as a consequence of the way people represent event information in memory.
CIE in Impression Formation
In contrast to event-related reasoning, the empirical evidence for a CIE in person-related reasoning (i.e., impression formation) is mixed. On the one hand, there is substantial evidence that discredited misinformation can continue to inform person impressions. Examples include inadmissible evidence affecting jury decisions (see Steblay et al., 2006, for a review), continued stigma towards victims of miscarriages of justice (Brooks & Greenberg, 2021; Clow & Leach, 2015), and discounted rumours swaying voter preferences (Jardina & Traugott, 2019; Weeks & Garrett, 2014). In an experiment investigating misinformation and political figures, Thorson (2016) found that misinformation about a (fictional) candidate accepting donations from a convicted felon led participants to evaluate the candidate more negatively compared to a no-misinformation control condition, even when the misinformation had been corrected (although primarily if the political candidate was affiliated with the participant’s non-preferred political party; see also Bullock, 2007). These findings point to the presence of a CIE in impression formation, although this may depend on specific task features (i.e., if an explicit judgement is required) and whether there is motivation to keep the original information active (see Mensink & Rapp, 2011; Rapp & Kendeou, 2007, 2009).
However, other studies have failed to observe a CIE in impression formation (e.g., Cobb et al., 2013; De keersmaecker & Roets, 2017; Ecker & Rodricks, 2020). When investigating the impact of discredited positive misinformation, Cobb et al. (2013) presented participants with a mock news story which favoured a political candidate. When the story was later corrected, participants reported more negative impressions of the candidate compared to those who had not received the positive misinformation; in other words, this study not only found no CIE, but even an overcorrection effect. This demonstrates that corrections can be fully effective in impression formation. Ecker and Rodricks (2020) investigated the effects of discredited negative misinformation on person impressions. Their study featured observations about a fictitious student named “John”. Participants received descriptions of behaviours that John had engaged in; these were mainly behaviours of neutral valence (e.g., “John gave a presentation to his class”), but in some conditions a behaviour of negative valence was included (“John slapped his girlfriend”) that was subsequently retracted or not retracted. Results showed participants in the retraction condition fully discounted the discredited misinformation in their person impressions. Taken together, this suggests people can fully discredit retracted person-related misinformation.1
The mixed findings for a CIE in impression formation may be due, in part, to the different methodological approaches used. In some studies participants were presented with misinformation targeting political candidates (Cobb et al., 2013; Thorson, 2016) while in other studies the misinformation related to people with no political affiliation (De keersmaecker & Roets, 2017; Ecker & Rodricks, 2020). Studies investigating the CIE in impression formation have also differed in the type of allegation and valence of the misinformation (positive and negative; Cobb et al., 2013; Ecker & Rodricks, 2020). These factors may moderate the CIE in impression formation, particularly when the target type (politician vs. private citizen) and type of misinformation (positive vs. negative) coheres with pre-existing knowledge (e.g., see Cappella et al., 2015; Walter & Tukachinsky, 2020). As a result, task characteristics may influence the robustness of the CIE in impression formation, and may help explain the mixed results. Thus, it is important to systematically investigate these factors and their influence on the CIE in impression formation.
To better understand the mixed findings, our attention turns to the robustness of the CIE as it relates to person impressions, and the conditions under which it can be detected. Given that impression formation is a dynamic process where people continually update their impressions to integrate new information (Asch, 1946; Kashima & Kerekes, 1994), a CIE might be present at the time of correction (i.e., a primacy effect; see Park, 1986) but then fade as more and more information is received and the impressions continue to be updated. Demonstration of a short-lived CIE could contribute to the mixed evidence if the CIE is generally tested too late to enable detection in an impression-formation task (i.e., at the end of the task; see Cappella et al., 2018; Connor Desai & Reimers, 2019). This possibility is tested here using a dynamic (moment-to-moment) person-impression rating measure that is taken across the entire task.
Another factor that has not been systematically investigated is the valence of the misinformation, and the extent to which it might moderate a CIE in impression formation. While some studies have investigated positively-valenced misinformation and its impact on person impressions (e.g., Cobb et al., 2013), and others have considered the effect of negatively-valenced misinformation on person impressions (e.g., Ecker & Rodricks, 2020; Thorson, 2016), no studies (to the best of our knowledge) have compared both. As a result, we do not know the role that misinformation valence plays in this context.2 Valence has been found to impact impression formation in general, with negative information found to be more influential than positive information (i.e., negativity bias; Fiske, 1980; Skowronski & Carlston, 1989; see Rusconi et al., 2020 for a review). It is possible that this negativity bias extends to the misinformation context. To determine the role of misinformation valence, the effect of positively- and negatively-valenced misinformation on person impressions was tested in Experiments 1 and 2.
Another explanation for the inconsistent findings may relate to misinformation coherence, specifically, the extent to which the misinformation coheres with prior information. When information about a person takes the form of unrelated traits or behaviours presented within a randomly ordered list, as is typical of impression-formation research (e.g., Carlston & Smith, 1996; Srull & Wyer, 1989), model coherence may be enhanced by retracting an “outlying” trait description or behaviour that is inconsistent with the initial person model (Fiske & Linville, 1980). For example, in the study by Ecker and Rodricks (2020), John slapping his girlfriend was the only negative behaviour mentioned; when this was retracted, updating of the person model may have been readily and effectively executed because it enhanced overall model coherence (also see Mende-Siedlecki et al., 2013; Srull & Wyer, 1989). In contrast, when the retracted misinformation coheres with other information, removing it may threaten model coherence, thereby promoting continued reliance on the retracted misinformation (e.g., Thorson, 2016). This possibility is consistent with the role that misinformation coherence has been theorised to take in event-related reasoning; here, retracted event-related misinformation which coheres with other information continues to be relied upon because it helps to maintain the causal structure of the mental model of the event (Ecker et al., 2011; Johnson & Seifert, 1994). To test this explanation, coherence-building elements were added in Experiments 2–4.
The Present Study
To investigate the impact of person-related misinformation and its correction on impression formation, we combined the CIE paradigm—which provides participants with misinformation that is or is not subsequently retracted—with a dynamic person-impression rating task. This task records trial-to-trial impression updates, thus measuring how the behaviour descriptions provided to participants—including a misinformation item and its retraction—influence person impressions over time. To manipulate participants’ impressions of a target person, we selected behaviour statements from a pool of pre-rated statements (Mickelberg et al., 2022) based on their morality (Brambilla et al., 2021) and believability (Cone et al., 2019), two dimensions known to be important in impression formation. We expected that misinformation would affect person impressions, and that a retraction would reverse this effect, at least partially; by measuring impressions after each piece of information was received, we anticipated that we would be able to detect any CIE even if this effect was short-lived. In addition, if lack of model coherence affected the CIE (Cobb et al., 2013; Ecker & Rodricks, 2020), then adding coherence-building elements to the materials should promote a CIE. Finally, to clarify the role of valence, we compared both positive and negative misinformation.
Evidence for a CIE in impression formation was examined over four experiments. In each, participants were presented with a sequence of 27 behaviour statements that related to a fictitious person named “John”. The statements described behaviours John had engaged in and included a target misinformation statement that was later retracted or not retracted. Participants were told the information came from several of John’s acquaintances. Person impressions were recorded after the presentation of each behaviour statement. In Experiment 1, behaviour statements included a series of neutral filler statements (e.g., “John had ordered his favourite dish from a Chinese restaurant”) and a target statement whose valence was either negative (“John had an affair with his best friend’s wife”), positive (“John jumped off a boat to save a drowning friend even though this put John’s own life at risk”) or neutral (“John went to a fancy restaurant but couldn’t pronounce the items on the menu”). Statements were presented as stand-alone items in the manner typical of impression-formation research (Cone & Ferguson, 2015; Gregg et al., 2006; Kerpelman & Himmelfarb, 1971). Experiments 2–4 extended Experiment 1 by replacing some of the neutral filler statements with coherence-building elements that we hypothesised would make misinformation more resistant to correction: Experiment 2 included behaviour statements that were congruent with the target behaviour, and Experiments 3 and 4 included behaviour statements that were causally related to the target behaviour. Experiments 1–3 used the same target statements, whereas Experiment 4 included new target materials. To foreshadow, no evidence for a CIE in impression formation was obtained in any of the experiments.
Experiment 1
Experiment 1 presented information about the protagonist “John” in five conditions. The first four conditions varied according to the valence of the target statement (negative or positive) and whether the target statement was later retracted or not retracted. The fifth condition used a neutral target statement and served as a baseline. The experiment thus used a 2 × 2 between-subjects design (negative retraction; positive retraction; negative no-retraction; positive no-retraction), plus a neutral control.
Method
All experiments were approved by the University of Western Australia’s Human Research Ethics Office. Participants received an ethics-approved information sheet and provided informed consent (by clicking ‘next’ to proceed to the study). The experiments were conducted in accordance with Australia’s National Statement on Ethical Conduct in Human Research (NHMRC, 2018).
Participants
A sample of 500 U.S.-based participants (282 male, 213 female, 4 other, 1 preferred not to say) were recruited through the online crowd-sourcing platform Prolific (https://prolific.com); age range was 18–84 years (M = 32.42, SD = 11.31). Participants were randomly assigned to one of the five conditions (n = 100 per condition).
Materials
Impression-Formation Task
Participants viewed 27 behaviour statements concerning a person named John, presented one-by-one. After reading each statement, participants rated their impression of John using a dynamic impression-rating measure ranging from −50 (extremely negative) to 50 (extremely positive), with 0 indicating neutral. A marker on the scale indicated the participant’s current rating (see Figure 1).
Note. (A) shows a neutral filler statement; (B) shows the retraction of the negative target statement.
Note. (A) shows a neutral filler statement; (B) shows the retraction of the negative target statement.
The impression-formation task had two phases: (1) presentation of initial information, and (2) presentation of updated information. The first phase encompassed trials 1–17. In this phase, participants were told that the study investigated how people form impressions of others. Participants were then informed they would be given information about a person named “John”. They were told this information had been collected from four of his acquaintances, who provided examples of his behaviour that they had observed over the past few months. A 20 s delay was imposed to allow time to read the instructions before participants could proceed to the first behaviour statement. Over the 17 trials, participants were presented with 16 neutral filler statements, presented in a random order, and a target statement presented at trial 13. The target statement depended on condition, and was either negative, positive, or neutral (see preceding section for the statements used). At the start of the second phase, participants were told that John’s acquaintances were interviewed a second time, to verify the initial information they provided and add new information about John. During the second phase (trials 18–27), ten update statements were provided, in which six new neutral filler statements were presented, two neutral filler statements from the first phase were confirmed (e.g., “CONFIRMED: John was running late, so he drove to work rather than taking the bus”), and two statements from the first phase were retracted. In the retraction and neutral conditions, this included retraction of a neutral filler statement and the target statement (e.g., “RETRACTION: John did NOT have an affair with his best friend’s wife”); in the no-retraction conditions, two filler statements were retracted instead. The presentation order of the updated statements was randomised, except that the target-statement retraction always occurred at trial 19.
The behaviour statements were selected from a corpus of pre-tested stimuli (Mickelberg et al., 2022). Each statement was associated with a morality score ranging from −4 (very morally bad) to 4 (very morally good), and a believability score ranging from 0 (not believable) to 8 (very believable). We selected statements with high believability scores (> 6) and morality scores appropriate for neutral filler statements or the neutral target (i.e., close to 0) or strongly valenced (close to ±4) for the positive and negative target statements. To illustrate, the neutral target statement “John went to a fancy restaurant but couldn’t pronounce the items on the menu” was associated with morality = 0.06, believability = 6.75. By contrast, the negative target statement was strongly negative (“John had an affair with his best friend’s wife”; morality = −3.40, believability = 6.84) and the positive target statement was strongly positive (“John jumped off a boat to save a drowning friend even though this put John’s own life at risk”; morality = 3.55, believability = 6.79). See Supplementary Materials for the full list of behaviour statements selected across experiments.
Reysen Likability Scale
The Reysen likability scale (Reysen, 2005) measured the likability of the protagonist John. The scale consists of 11 items (e.g., “This person is likeable”) rated on a Likert scale ranging from 1 (very strongly disagree) to 7 (very strongly agree; see Supplementary Materials for the full scale). The scale was previously used by Ecker and Rodricks (2020) to test for a CIE in impression formation, and has been associated with strong internal consistency reliability (Cronbach’s α = .88; Reysen, 2005). In the present experiments, reliability ranged from α = .88 to α = .95. Across experiments, likability scores were moderately positively correlated with impression ratings at trial 27 (r = .53 to r = .69) but not strongly enough to be considered collinear.
Recognition Test
A recognition test was included to measure the extent to which participants attended to and encoded the behaviour statements. Eleven multiple-choice questions, each with three to four response options, tested recognition of the initial and updated behaviour statements (e.g., “What instrument did John learn to play as a child?”—“The piano”, “The drums”, “The violin”; see Supplementary Materials for the full list of questions).
Procedure
The experiment was performed online. After reading information about the task and providing informed consent, participants completed the impression-formation task, followed by the likability scale and recognition test, and were then debriefed and thanked for their participation. The median completion time was 8 minutes. Participants were paid £1.00 (approximately US$1.25) upon completion of the study.
Results
The data were analysed using linear regression models. All analyses were performed and all figures were created in R (v4.0.3; R Core Team, 2021). Regression models were estimated using the lm function of base R. Figures were created using ggplot2 (v3.3.5; Wickham, 2016). The data, R script, and supplementary information associated with Experiments 1–4 are available at https://osf.io/ymuap.
Performance on the recognition test was high (M = 88.61%, SD = 13.13%). Only 12 participants scored below 50%; excluding them did not qualitatively affect the results, so analyses of the full-sample data are reported. A manipulation check was conducted to test that target statements at trial 13 affected impressions. Dependent variables for investigating continued influence were impression ratings immediately post-retraction (trial 19) and after a delay (trial 27). The likability scores gave another measure of participants’ impressions after a delay. Secondary analyses (for trials 1–12 and 14–18) can be found in Supplementary Materials.
Manipulation Check
As illustrated in Figure 2, participants in all conditions reported a somewhat positive impression of John from the outset, and this gradually improved over trials 1–12 (see Supplementary Materials). The target statement was then presented at trial 13. To ensure the target-statement manipulation was successful (i.e., target statements affected impressions), we tested the impact of the positive and negative target statements on person impressions at trial 13. First, however, we ascertained that at trial 13, there was no evidence of a statistical difference (based on random noise) between the negative-retraction and negative no-retraction conditions (p = .328), or between the positive-retraction and positive no-retraction conditions (p = .626). For the purpose of comparing to the neutral condition, the conditions were thus collapsed into a combined negative or positive condition according to target-statement valence.
Note. The mean impression ratings for the neutral condition are duplicated in panels A and B. The shaded areas are 95% confidence intervals. The vertical line at trial 13 indicates target-statement presentation; the vertical line at trial 19 indicates target-statement retraction (in the retraction conditions).
Note. The mean impression ratings for the neutral condition are duplicated in panels A and B. The shaded areas are 95% confidence intervals. The vertical line at trial 13 indicates target-statement presentation; the vertical line at trial 19 indicates target-statement retraction (in the retraction conditions).
Impression ratings at trial 13 were entered into a linear regression, with target-statement valence as the predictor (negative; positive; neutral). This showed that the negative conditions returned lower impression ratings than the neutral control (i.e., impressions were more negative when participants had received the negative target statement; β = −42.43, SE = 2.58, t = −16.43, p < .001), and the positive conditions returned higher impression ratings than the neutral control (i.e., impressions were more positive when participants had received the positive target statement; β = 26.90, SE = 2.58, t = 10.42, p < .001). The difference in beta-weight magnitude (β = −42.43 compared to β = 26.90) suggests that participants’ impressions of John were more strongly affected by the negative target statement than the positive target statement. This was confirmed by comparing the absolute difference in impression ratings from trial 12 to trial 13 (negative: 50.05 > positive: 26.09; t(198) = 7.10, p < .001).
Between trial 13 (the target statement) and trial 19 (the target-statement retraction, in the retraction conditions), the effect of the negative target statement weakened over trials but impression ratings remained lower than in the neutral condition. The effect of the positive target statement also weakened between trials 13 and 19 but impression ratings remained higher than in the neutral condition. The differences to neutral were statistically significant (see Supplementary Materials).
Tests for Continued Influence
Evidence for a CIE would be observed if the target statements continued to influence person impressions after being retracted. We tested for a CIE at two time-points: immediately following the target-statement retraction (trial 19) and after a delay (trial 27).
Immediate Test
The trial-19 impression ratings were entered into linear regressions with condition as a predictor (retraction; no-retraction; neutral), with separate regression models for the negative and positive conditions. For negative misinformation, the model was significant, F(2, 297) = 66.37, p < .001, R2 = .31. In the negative no-retraction condition, impression ratings were lower than in the neutral control condition, β = −26.78, SE = 3.60, t = −7.43, p < .001, indicating that the effect of the negative target statement persisted at trial 19. When the negative target statement was retracted in the negative-retraction condition, impression ratings were higher than in the negative no-retraction condition, β = 38.13, SE = 3.69, t = 10.34, p < .001, and also higher than in the neutral control condition, β = 11.35, SE = 2.84, t = 4.00, p < .001. This indicates the retraction of the negative target statement was fully effective; in fact, participants overcorrected at immediate test.
For positive misinformation, the model was also significant, F(2, 297) = 18.68, p < .001, R2 = .11. In the positive no-retraction condition, impression ratings were higher than in the neutral control condition, β = 8.57, SE = 2.95, t = 2.91, p = .004, indicating that the effect of the positive target statement persisted at trial 19. Note that the effect of the positive information on person impressions was not as strong as the negative information on this trial (β = 8.57 compared to β = −26.78). This was confirmed by comparing the absolute difference in impression ratings between trials 12 and 19 for the negative and positive no-retraction conditions (negative: 38.48 > positive: 19.27; t(198) = 5.89, p < .001). When the positive target statement was retracted in the positive-retraction condition, impression ratings were lower than in the positive no-retraction condition, β = −20.30, SE = 3.60, t = −5.64, p < .001, and also lower than in the neutral control condition, β = −11.73, SE = 3.42, t = −3.43, p < .001. This indicates the retraction of the positive target statement was fully effective, and that again participants overcorrected at immediate test.
Delayed Test
Next, the same set of analyses was conducted on the trial-27 impression ratings. For negative misinformation, the model was significant, F(2, 297) = 30.82, p < .001, R2 = .17. In the negative no-retraction condition, impression ratings were lower than in the neutral control condition, β = −24.76, SE = 3.67, t = −6.74, p < .001, indicating that the effect of the negative target statement persisted at trial 27. For the negative-retraction condition, impression ratings were higher than those in the negative no-retraction condition, β = 23.30, SE = 3.81, t = 6.13, p < .001, but were comparable to control, β = −1.43, SE = 3.10, t = −0.46, p = .646. This indicates that while the retraction of the negative target statement was still effective after a delay, there was no longer evidence of an overcorrection.
For positive misinformation, the model was not significant, F(2, 297) = 1.41, p = .245, R2 = .01, indicating there was no statistical evidence of a difference between conditions. Thus, there was no evidence for a CIE (the positive-retraction and neutral conditions did not differ); however, no strong conclusions can be drawn from this because the effect of the positive misinformation appeared to have “worn off” regardless of the retraction (i.e., the positive no-retraction and neutral conditions did not differ).
Likability Scores
Scores on the likability scale were consistent with the results of the impression-formation task (see Figure 3). Likability scores were entered into a linear regression with condition as the predictor (retraction; no-retraction; neutral), and with separate models for the negative and positive conditions. For negative misinformation, the model was significant, F(2, 297) = 18.15, p < .001, R2 = .11. Likability was lower in the negative no-retraction condition compared to the neutral control condition, β = −0.71, SE = 0.13, t = −5.31, p < .001, and the negative-retraction condition, β = −0.60, SE = 0.13, t = −4.60, p < .001. There was no evidence of a statistical difference between the negative-retraction and neutral conditions, β = −0.10, SE = 0.12, t = −0.91, p = .364. This again indicates the retraction of negative misinformation was fully effective (i.e., no CIE).
Note. Bars indicate condition means; error bars are 95% bootstrapped confidence intervals; data points indicate participant mean scores; violins provide distributional information.
Note. Bars indicate condition means; error bars are 95% bootstrapped confidence intervals; data points indicate participant mean scores; violins provide distributional information.
For positive misinformation, the model was not significant, F(2, 297) = 2.30, p = .102, R2 = .02, indicating there was no statistical evidence of a difference between conditions. This indicates there was neither an impact of the misinformation, nor a CIE.
Discussion
Experiment 1 tested for a CIE in impression formation and whether it was moderated by misinformation valence (i.e., negative vs. positive). Results showed that, first of all, negative misinformation had a strong and immediate negative impact on person impressions, and (when not retracted) the impact endured after a delay. Positive misinformation had a weaker immediate impact, and the impact did not endure after a delay. Once retracted, negative and positive misinformation was fully discounted, leaving no evidence of a continued impact on person impressions. For negative misinformation, an overcorrection was observed immediately following the retraction, such that impression ratings became more positive than those in the neutral control condition. Person impressions then returned to the neutral baseline after a delay. A temporary overcorrection was also observed immediately following the retraction of the positive misinformation (as in Cobb et al., 2013). The immediate overcorrection of person impressions demonstrates the effectiveness of a retraction in correcting person-related misinformation.
Prior to a retraction, our findings show that negative misinformation is given greater weight than positive misinformation. This suggests that person impressions can be influenced more by a single negative statement than by a single positive statement (despite the statements being close in extremity according to pre-ratings; see Mickelberg et al., 2022). This demonstrates an information valence effect, adding to the robustness of previous findings of a negativity bias in impression formation (Rozin & Royzman, 2001; Rusconi et al., 2020; Skowronski & Carlston, 1989; see also Mensink & Rapp, 2011).
Our finding that people can fully discount retracted misinformation when making person impression judgements supports some previous studies (Cobb et al., 2013; Ecker & Rodricks, 2020) but contrasts with CIE research in general, which typically finds that there is still continued reliance on corrected misinformation (Chan et al., 2017; Ecker et al., 2022; Walter & Tukachinsky, 2020). A possible explanation for this is that the misinformation used in Experiment 1 may not have adequately cohered with the other person information provided, and this may have made it easier for participants to fully discount the misinformation. In other words, the misinformation used in Experiment 1 did not provide a causal explanation for an event (as is typical in CIE studies), nor did it serve to strengthen mental-model coherence due to it lacking consistency with the other behavioural information presented. Thus it could be that the retraction did not present a threat to model coherence, and may not have created a “gap” in the participants’ mental model of the protagonist (Ecker et al., 2011). Consequently, there was no reason to rely on the misinformation post-retraction to maintain model coherence, explaining the absence of a CIE.
To test this explanation, Experiment 2 introduced coherence-building elements in the form of behaviour statements that were congruent with (i.e., conceptually related to) the target statements, to promote the integration of the misinformation with the other protagonist information. We hypothesised that improved congruency between the general information and misinformation about John would give rise to a CIE.
Experiment 2
Experiment 2 included behaviour statements that are congruent with the target statement. For instance, in the conditions with the negative target statement “John had an affair with his best friend’s wife”, the congruent statements included “John cheated in a card game” (also suggesting a poor moral character) and “John and his best friend had a fight” (consistent with John having an affair with the friend’s wife). We reasoned that including the congruent statements would promote the integration of the target statement within participants’ mental model of John.
Method
Experiment 2 used the same design as Experiment 1, but included six behaviour statements that were congruent with the target statement. There were four experimental conditions as in Experiment 1 (negative retraction; positive retraction; negative no-retraction; positive no-retraction), but Experiment 2 had two neutral conditions. One (the negative-congruent neutral condition) had a neutral target statement and the same congruent information as the negative conditions, to serve as the baseline for the negative-retraction and negative no-retraction conditions. The other (the positive-congruent neutral condition) had a neutral target statement and the same congruent information as the positive conditions, to serve as the baseline for the positive-retraction and positive no-retraction conditions.
Participants
A convenience sample of 600 first-year undergraduate students from the University of Western Australia (438 females, 155 males, 3 other, 4 preferred not to say) aged 16–62 years (M = 20.13, SD = 4.88) were recruited. Participants were randomly assigned to one of the six conditions (n = 100 per condition).
Materials
Impression-Formation Task
The task was identical to that used in Experiment 1, but six of the neutral filler statements presented in trials 8, 9, 10, 11, 15, and 17 were replaced with behaviour statements that were either valence-congruent or meaning-congruent with the target statement.
Valence-Congruent Statements. Two behaviour statements that were valence-congruent with the target statement were presented prior to the target statement (at trials 8 and 10). Valence-congruent statements were selected from the same corpus of behaviour stimuli (Mickelberg et al., 2022). The valence-congruent statements were matched according to the valence of the target statement (positive or negative) but were selected to be milder in valence (i.e., morality close to ±2) and similarly high in believability (> 6). Two negative behaviour statements were selected for the negative conditions (“John cheated in a card game while playing with a group of his friends”; morality = −1.91, believability = 6.71; and “John found a wallet with $50 in it, took the money out and left the wallet on the floor”; morality = −2.35; believability = 6.81). Two positive behaviour statements were used for the positive conditions (“John shaved his head when he found out his partner had cancer and required radiation therapy;” morality = 2.46, believability = 7.23; and “John volunteers at a dog refuge, walking the dogs and cleaning their kennels once a week”; morality = 2.67, believability = 7.04).
Meaning-Congruent Statements. Four meaning-congruent statements were generated by the authors, designed to align with the (negative and positive) target statements but to be neutral in morality. The two meaning-congruent statements that aligned with the negative target statement (that John had an affair with his best friend’s wife) were “John and his best friend had a fight” and “John went out for a coffee with his best friend’s wife”. The two meaning-congruent statements that aligned with the positive target statement (that John saved a friend from drowning) were “John attended first aid training at the local sailing club” and “John went to a Surf Life Saving awards dinner”. Two statements (one positive, one negative) were introduced prior to the presentation of the target statement (at trials 9 and 11) and two (one positive, one negative) were presented after the target statement (at trials 15 and 17). It was anticipated that these statements would be interpreted as neutral fillers by participants who did not receive the related target statements, so for simplicity we gave the same four meaning-congruent statements in every condition (expecting only two to be actually meaning-congruent in a given condition, or none in the neutral conditions). The presentation order of the four statements was counterbalanced across participants.
Procedure
The procedure was identical to Experiment 1. Median completion time was 10 minutes. Participants received partial course credit upon completion of the study.
Results
Performance on the recognition test was high (M = 93.73%, SD = 8.99%). Only 4 participants scored below 50%; excluding them did not qualitatively affect the results so analyses of the full-sample data are reported.
Manipulation Check
As illustrated in Figure 4, prior to the presentation of valence-congruent information at trial 8, participants in all conditions reported a positive impression of John from the outset, and this gradually improved over trials 1 to 7 (see Supplementary Materials). In the negative conditions the valence-congruent behaviour statements at trials 8 and 10 caused the impression ratings to become more negative over trials 8 to 12. By contrast, in the positive conditions the valence-congruent statements at trials 8 and 10 caused the impression ratings to become more positive over trials 8 to 12 (see Supplementary Materials).
Note. The shaded areas are 95% confidence intervals. The vertical lines at trials 8 and 10 indicate valence-congruent-statement presentations. The vertical line at trial 13 indicates target-statement presentation and the vertical line at trial 19 indicates target-statement retraction (in the retraction conditions). The meaning-congruent statements were presented at trials 9, 11, 15, and 17.
Note. The shaded areas are 95% confidence intervals. The vertical lines at trials 8 and 10 indicate valence-congruent-statement presentations. The vertical line at trial 13 indicates target-statement presentation and the vertical line at trial 19 indicates target-statement retraction (in the retraction conditions). The meaning-congruent statements were presented at trials 9, 11, 15, and 17.
The target statement was then presented at trial 13. To ensure the target-statement manipulation was successful, we tested the impact of the positive and negative target statements on person impressions at trial 13. Again, we first checked that at trial 13, there was no evidence of a statistical difference between the negative-retraction and negative no-retraction conditions (p = .406) or the positive-retraction and positive no-retraction conditions (p = .052); thus, the conditions were collapsed into a combined negative or positive condition according to target-statement valence, for the purpose of comparing to the (negative-congruent and positive-congruent) neutral conditions.
Impression ratings at trial 13 were entered into a linear regression, with target-statement valence as a predictor and separate models for the negative conditions (negative; negative-congruent neutral) and positive conditions (positive; positive-congruent neutral). This showed that the negative conditions returned lower impression ratings than the (negative-congruent) neutral condition (i.e., person impressions were more negative when participants had received the negative target statement; β = −35.72, SE = 2.20, t = −16.24, p < .001), and the positive conditions returned higher impression ratings than the (positive-congruent) neutral condition (i.e., ratings became more positive when participants had received the positive target statement; β = 15.90, SE = 2.10, t = 7.58, p < .001). The difference in beta-weight magnitude (β = −35.72 vs. β = 15.90), suggests that participants’ impressions of John were more strongly affected by the negative than the positive target statement. This was confirmed by comparing the absolute difference in impression ratings from trial 12 to trial 13 (negative: 38.77 > positive: 19.89; t(198) = 6.82, p < .001).
Between trial 13 (the target statement) and trial 19 (the target-statement retraction, in the retraction conditions), the effect of the negative target statement weakened over trials, but ratings remained lower than in the (negative-congruent) neutral condition (the difference to the negative-congruent neutral condition was statistically significant). There was no effect of the positive target statement over trials 13 and 19 (the difference to the positive-congruent neutral condition was not significant; see Supplementary Materials).
Tests for Continued Influence
The CIE was again tested at two time-points: immediately following the target-statement retraction (trial 19) and after a delay (trial 27).
Immediate Test
The trial-19 impression ratings were entered into linear regressions with condition as a predictor (retraction, no-retraction, neutral), with separate models run for the negative and positive conditions. For negative misinformation, the model was significant, F(2, 297) = 29.30, p < .001, R2 = .16. In the negative no-retraction condition, impression ratings were lower than in the (negative-congruent) neutral condition, β = −15.82, SE = 3.33, t = −4.76, p < .001, indicating that the effect of the negative target statement persisted at trial 19. When the negative target statement was retracted in the negative-retraction condition, impression ratings were higher than in the negative no-retraction condition, β = 24.93, SE = 3.44, t = 7.25, p < .001, and also higher than in the (negative-congruent) neutral condition, β = 9.11, SE = 3.12, t = 2.92, p = .004. This indicates the retraction of the negative target statement was fully effective; in fact, participants overcorrected at the immediate test.
For positive misinformation, the model was significant, F(2, 297) = 8.47, p < .001, R2 = .05. However, there was no evidence of a statistical difference between the positive no-retraction and (positive-congruent) neutral conditions, β = −4.06, SE = 3.03, t = −1.34, p = .182, indicating that the effect of the positive target statement was not observable at trial 19. Note that the effect of the positive information again was not as strong as the negative information by this trial (β = −15.82 vs. β = −4.06), indicating that the effect of the negative target statement was more persistent than that of the positive target statement. This was confirmed by comparing the absolute difference in impression ratings between trials 12 and 19 for the negative and positive no-retraction conditions (negative: 23.21 > positive: 12.59; t(198) = 4.68, p < .001). When the positive target statement was retracted in the positive-retraction condition, impression ratings were lower than in the positive no-retraction condition, β = −8.95, SE = 3.39, t = −2.64, p = .009, and also lower than in the (positive-congruent) neutral condition, β = −13.01, SE = 3.27, t = −3.98, p < .001. This means there was no evidence of a CIE for positive misinformation; however, no strong conclusions can be drawn from this because the effect of the positive misinformation had “worn off” regardless of retraction (i.e., the positive no-retraction and [positive-congruent] neutral conditions did not differ).
Delayed Test
Next, the same set of analyses was conducted on the trial-27 impression ratings. For negative misinformation, the model was significant, F(2, 297) = 19.93, p < .001, R2 = .12. In the negative no-retraction condition, impression ratings were lower than in the (negative-congruent) neutral condition, β = −12.98, SE = 3.35, t = −3.88, p < .001, indicating that the effect of the negative target statement persisted at trial 27. For the negative-retraction condition, impression ratings were higher than those in the negative no-retraction condition, β = 19.95, SE = 3.23, t = 6.18, p < .001, and also those in the (negative-congruent) neutral condition, β = 6.97, SE = 3.03, t = 2.30, p = .023. This indicates the retraction of the negative misinformation was effective, and there was still evidence of an overcorrection after a delay.
For positive misinformation, the model was not significant, F(2, 297) = 0.12, p = .889, R2 < .01, indicating there was no statistical evidence of a difference between conditions. Thus, there was no evidence for a CIE (the positive-retraction and [positive-congruent] neutral conditions did not differ); however, as with the immediate test, no strong conclusions can be drawn from this because the effect of the positive misinformation had “worn off” regardless of retraction.
Likability Scores
Likability scores were entered into a linear regression with condition as a predictor (retraction, no-retraction, neutral), with separate models for the negative and positive conditions. For negative misinformation, the model was significant, F(2, 297) = 21.27, p < .001, R2 = .13 (see Figure 5). Likability was lower in the negative no-retraction condition compared to the (negative-congruent) neutral condition, β = −0.51, SE = 0.11, t = −4.84, p < .001, and the negative-retraction condition, β = −0.59, SE = 0.09, t = −6.46, p < .001. There was no evidence of a statistical difference between the negative-retraction and (negative-congruent) neutral conditions, β = 0.08, SE = 0.10, t = 0.82, p = .415. This again indicates the retraction of negative misinformation was fully effective (i.e., no CIE).
Note. Bars indicate condition means; error bars are 95% bootstrapped confidence intervals; data points indicate participant mean scores; violins provide distributional information.
Note. Bars indicate condition means; error bars are 95% bootstrapped confidence intervals; data points indicate participant mean scores; violins provide distributional information.
For positive misinformation, there was no statistical evidence of a difference between the conditions, F(2, 297) = 0.25, p = .780, R2 < .01. This indicates there was neither an impact of the misinformation, nor a CIE.
Discussion
The Experiment 2 finding of a stronger and more enduring impact of negative misinformation compared to positive misinformation replicated the negativity bias shown in Experiment 1. Also in line with Experiment 1, Experiment 2 again demonstrated that people can fully discount discredited negative or positive misinformation. For negative misinformation, the lack of CIE was attributable to a fully effective retraction; in fact, there was again an immediate overcorrection such that impression ratings became more positive compared to the neutral-condition baseline. Unlike Experiment 1, this overcorrection was maintained after a delay. By contrast, the effect of positive misinformation wore off over time even without a retraction. This replicates a pattern shown in impression ratings and likability scores at the end of Experiment 1, but extended also to impression ratings at the time of the retraction.
Thus, contrary to our prediction, the results of Experiment 2 showed that adding information that is congruent with the misinformation did not promote a CIE in impression formation (immediately or after a delay). We predicted that including congruent information would foster integration of the misinformation into participants’ mental model of the protagonist, making the misinformation more resistant to correction (Ecker et al., 2010; Ecker, Lewandowsky, Fenton, et al., 2014). This prediction was not supported: we found no evidence that people continued to rely on misinformation after it was retracted.
It is possible that the congruent information (e.g., John cheated in a card game) was not sufficiently related to the misinformation (e.g., John having an affair) to affect its integration into the mental model of the protagonist. Another possibility is that the relation between the misinformation and the supporting information needs to be more causal, rather than simply congruent (Connor Desai & Reimers, 2022; Hamby et al., 2020; Johnson & Seifert, 1994). Experiment 3 therefore included causally-related information, with a view to making the misinformation more resistant to correction. If the misinformation provides a causal explanation for another piece of information provided, people may be motivated to hold onto the misinformation despite a retraction due to its functional role in their mental model; thus, a CIE should be more likely to emerge with the addition of causally-related information.
Finally, as mentioned earlier, the CIE is typically detected using inferential-reasoning questions rather than impression ratings (Brydges et al., 2018; Johnson & Seifert, 1994; Lewandowsky et al., 2012; Wilkes & Leatherbarrow, 1988). It could be that the impression-rating questions used in Experiments 1 and 2 are not ideally suited to measuring continued influence in the person-impression paradigm. In light of this, an additional inferential-reasoning measure was introduced in Experiment 3 to determine whether a CIE may be observed in inferential judgements even in the absence of a CIE in impression ratings.
Experiment 3
Experiment 3 tested whether inclusion of a statement that is causally related to the target misinformation statement would promote a CIE in impression formation. Given the effects of positive misinformation tended to dissipate in Experiments 1 and 2 even without a retraction, Experiment 3 focused exclusively on negative misinformation. Experiment 3 included a statement (“John was spotted in a hotel lobby with his best friend’s wife”) that had a pre-tested causal relation to the target statement (“John had an affair with his best friend’s wife”). We reasoned that this statement (the “causally-related statement” from here on) would promote stronger integration of the target statement within participants’ mental model of the protagonist (given it would be unlikely for the event in the causally-related statement to occur without the target statement being true), making the misinformation more resistant to retraction. In other words, retracting the target statement would leave the information in the causally-related statement unexplained, thus creating a motivation to dismiss the retraction and retain the misinformation in the mental model. Experiment 3 used the same impression-formation task as Experiment 1 (i.e., without the congruent information from Experiment 2), but with the addition of the causally-related statement. In addition, Experiment 3 included an indirect measure of misinformation reliance, using inferential-reasoning questions (for a similar approach, see Brydges et al., 2018; Ecker et al., 2017), allowing us to investigate whether inferential reasoning about the target person is subject to a CIE even if impression formation per se is not.
Method
Participants
To pilot-test candidate causally-related statements, we recruited 100 U.S.-based participants via Prolific. Two participants were excluded due to uniform responding, leaving a sample of N = 98 (67 female, 29 male, 1 other, 1 preferred not to say) aged 18–57 (M = 27.74, SD = 8.14). For Experiment 3, a separate sample of 300 U.S.-based participants (206 females, 81 males, 12 other, 1 preferred not to say) aged 18–69 (M = 29.91, SD = 10.56) was recruited via Prolific. Participants were randomly assigned to one of three conditions (n = 100 per condition). Prolific users who had participated in Experiment 1 were not invited to participate.
Materials
Causally-Related Statement Selection
Pilot testing was conducted to select a statement that people would interpret as being causally related to the target statement. Participants were told to assume—in line with the misinformation retraction—that John was NOT having an affair, and given this information were asked to rate the conditional likelihood of ten scenarios (e.g., “John went to the drycleaner to get a lipstick stain removed from his collar”), using a 5-point Likert scale from 1 (very unlikely) to 5 (very likely). The statement “John was spotted with his best friend’s wife in a hotel lobby” (M = 1.82, SD = 0.98; see Supplementary Materials) was selected on the basis that it was rated to be unlikely (under the assumption there was no affair) but without being overly indicative of an affair (compared to, e.g., the “wife announced she was pregnant with John’s baby”). The pilot test took approximately 2 minutes, and participants were paid £0.30 (approximately US$0.35) for their participation.
Inference Questions
In Experiment 3, seven inference questions (two open-ended questions and five Likert rating scales) were designed to measure reliance on the misinformation during inferential reasoning. The open-ended questions (presented as the first and last items) asked participants for information about John (“If you could tell someone about one specific thing John has done, what would it be?”; “Describe briefly in one sentence what kind of relationship John has with his best friend’s wife”). The rating scales asked participants to rate their endorsement of John across five statements (e.g., “John is an honorable man”; see Supplementary Materials) using an 11-point Likert scale from 0 (strongly disagree) to 10 (strongly agree).
Inference-Score Calculation. Responses to the open-ended inference questions were coded by two scorers blind to the experimental conditions, following a standardised guide (see Supplementary Materials). All scoring discrepancies were resolved through discussion. Any unambiguous reference to the target statement was scored 1 (e.g., “John was having an affair with his best friend’s wife”). References to the misinformation suggesting an ambiguous level of endorsement were scored 0.5 (e.g., “John may have been having an affair?”). Responses were scored 0 where the misinformation was not mentioned or was specifically discredited (e.g., “It was suggested that John was having an affair but that was not the case”). To put the Likert scale inference questions on the same scale as the open-ended questions, responses were divided by 10 (e.g., Brydges et al., 2018; Ecker et al., 2010); scales that were positively worded were reverse-scored. All responses were then summed to create an overall inference score, which ranged from 0 to 7, with higher scores indicating stronger endorsement of the misinformation.
Procedure
The impression-formation task in Experiment 3 matched that in Experiment 1, but with the causally-related statement replacing a neutral filler statement at trial 15 (shortly after the target statement was presented). Following the impression-formation task, participants completed the likability scale and inference questions, and the recognition test.3 Participants were then debriefed and thanked for their participation. Median completion time was 10 minutes, and participants were paid £1.50 (approximately US$2.05).
Results
Performance on the recognition test was high (M = 92.67%, SD = 7.11%). All participants scored at least 50%.
Manipulation Checks
As illustrated in Figure 6, participants in all conditions reported a positive impression of John from the outset, and this gradually improved over trials 1–12 (see Supplementary Materials). The target statement was then presented at trial 13. To ensure the target-statement manipulation was successful, we tested the impact of the negative target statement on person impressions at trial 13. Again, at trial 13, there was no evidence of a statistical difference between the negative-retraction and negative no-retraction conditions (p = .604), and thus for the purpose of comparing to the neutral condition, conditions were collapsed into a combined negative condition. Impression ratings at trial 13 were entered into a linear regression with target-statement valence as the predictor (negative; neutral). This showed that the negative conditions returned lower impression ratings than the neutral control (i.e., impressions were more negative when participants had received the negative target statement; β = −51.04, SE = 2.77, t = −18.42, p < .001).
Note. The shaded areas are 95% confidence intervals. The vertical line at trial 13 indicates target-statement presentation, the vertical line at trial 15 indicates causally-related statement presentation, and the vertical line at trial 19 indicates target-statement retraction (in the retraction condition).
Note. The shaded areas are 95% confidence intervals. The vertical line at trial 13 indicates target-statement presentation, the vertical line at trial 15 indicates causally-related statement presentation, and the vertical line at trial 19 indicates target-statement retraction (in the retraction condition).
Between trial 13 (the target statement) and trial 19 (the target-statement retraction, in the retraction condition), the effect of the negative target statement weakened but ratings remained lower than the neutral-condition baseline (the difference to neutral was statistically significant; see Supplementary Materials).
Tests for Continued Influence
The CIE was tested at two time-points: immediately following the target-statement retraction (trial 19) and after a delay (trial 27).
Immediate Test
The trial-19 impression ratings were entered into a linear regression with condition as a predictor (negative retraction; negative no-retraction; neutral). The model was significant, F(2, 297) = 54.69, p < .001, R2 = .27. In the negative no-retraction condition, impression ratings were lower than in the neutral control condition, β = −24.32, SE = 3.56, t = −6.83, p < .001, indicating that the effect of the negative target statement persisted at trial 19. When the negative target statement was retracted in the negative-retraction condition, impression ratings were higher than in the negative no-retraction condition, β = 34.59, SE = 3.61, t = 9.58, p < .001, and also higher than in the neutral control condition, β = 10.27, SE = 2.98, t = 3.45, p < .001. This indicates the retraction of the negative target statement was fully effective; in fact, participants again overcorrected at immediate test.
Delayed Test
Next, the same set of analyses was conducted on the trial-27 impression ratings. The model was significant, F(2, 297) = 53.26, p < .001, R2 = .26. In the negative no-retraction condition, impression ratings were lower than in the neutral control condition, β = −26.79, SE = 3.59, t = −7.47, p < .001, indicating that the effect of the negative target statement persisted at trial 27. For the negative-retraction condition, impression ratings were higher than those in the negative no-retraction condition, β = 33.39, SE = 3.37, t = 9.90, p < .001, and also higher than in the neutral control condition, β = 6.60, SE = 3.31, t = 1.99, p = .048. This indicates that the retraction of the negative target statement was still fully effective after a delay, and again there was evidence of an overcorrection.
Likability Scores
Likability scores were entered into a linear regression with condition as a predictor (negative retraction; negative no-retraction; neutral). The model was significant, F(2, 297) = 20.40, p < .001, R2 = .12. Likability was lower in the negative no-retraction condition compared to the neutral control condition, β = −0.62, SE = 0.12, t = −5.16, p < .001, and the negative-retraction condition, β = −0.66, SE = 0.12, t = −5.62, p < .001. There was no evidence of a statistical difference between the negative-retraction and neutral conditions, β = 0.05, SE = 0.11, t = 0.42, p = .673, indicating the retraction was fully effective.
Inference Scores
Inference scores supported the observations made based on the impression ratings and likability scores (see Figure 7). Inference scores were entered into a linear regression with condition as a predictor (negative retraction; negative no-retraction; neutral). The model was significant, F(2, 297) = 210.10, p < .001, R2 = .59. Inference scores were higher (i.e., participants made more references to the negative misinformation) for the negative no-retraction condition compared to the neutral condition, β = 2.62, SE = 0.20, t = 13.05, p < .001, and the negative-retraction condition, β = 3.79, SE = 0.17, t = 22.46, p < .001. In the negative-retraction condition, inference scores were lower than in the neutral condition, β = −1.23, SE = 0.20, t = −5.96, p < .001. This indicated that the retraction was fully effective and there was no evidence of a CIE in inferential reasoning; in fact, participants overcorrected.
Note. Bars indicate condition means; error bars are 95% bootstrapped confidence intervals; data points indicate participant mean scores; violins provide distributional information.
Note. Bars indicate condition means; error bars are 95% bootstrapped confidence intervals; data points indicate participant mean scores; violins provide distributional information.
Discussion
In line with Experiments 1 and 2, Experiment 3 showed that non-retracted negative information has an enduring impact on person impressions. Experiment 3 further corroborated the finding that people can fully discount discredited misinformation in impression formation. Results again showed that there was an overcorrection immediately following the retraction of the negative misinformation and after a delay, such that impressions were more positive than in the neutral condition. In addition, Experiment 3 extended these findings to inferential reasoning, suggesting that retracted misinformation is not relied upon to make subsequent inferences in the context of person impressions.
Thus, despite our predictions, the introduction of an inferential-reasoning measure did not allow us to detect a CIE in person-related reasoning. This is in contrast to the robust findings for the CIE in the context of events, where retracted misinformation reliably continues to influence inferential reasoning (Chan et al., 2017; Ecker et al., 2022; Walter & Tukachinsky, 2020). These findings were obtained despite the inclusion of a causally-related statement designed to strengthen person-model coherence and make the misinformation more resistant to retraction.
Experiment 3 also showed that participants who received the causally-related statement but not the negative target statement (i.e., those in the neutral condition) formed a more negative impression of John than those who received both but subsequently had the negative target statement retracted (negative retraction condition). The inference questions showed a similar pattern, with more references to the misinformation in the neutral compared to the negative-retraction condition, despite the fact that the misinformation was not presented in the neutral condition. This suggests that participants inferred the misinformation even if it was not explicitly mentioned, based on the innuendo provided by the causally-related statement (i.e., participants learning only that John had been spotted in a hotel lobby with his friend’s wife inferred they were having an affair); it also demonstrates that participants perceived a relation between the causally-related statement and the target statement, in line with pre-testing. In the retraction condition, the effect of the causally-related statement on participants’ impressions and inferences was counteracted by the target statement’s retraction, leading to a more neutral response. Other CIE literature has shown that misinformation can still affect people’s judgements when it is not explicitly mentioned but merely hinted at (Ecker, Lewandowsky, Chang, et al., 2014; Rich & Zaragoza, 2016; but see Tay et al., 2022).
The experiments reported so far used the same misinformation statement, which may limit the generalisability of our results. While single statements are typical in continued-influence and impression-formation research (Brydges et al., 2020; Cone & Ferguson, 2015; Gordon et al., 2017; Rydell et al., 2007; Skowronski, 2002; Skowronski & Carlston, 1989; Srull & Wyer, 1989), it is possible that varying the misinformation statement could return different results (i.e., presence of a CIE). To test this, we ran a final experiment that used a different target statement and a different causally-related statement.
Experiment 4
Experiment 4 tested whether the inclusion of a different target statement and a different causally-related statement would promote a CIE in impression formation. This was done to guard against the possibility that the null effects observed in Experiments 1–3 might be restricted to the materials used. Experiment 4 used the same impression-formation task as Experiment 3.
Method
Participants
For Experiment 4, a sample of 300 U.S.-based participants (183 females, 110 males, 5 other, 2 preferred not to say) aged 18–74 (M = 37.78, SD = 13.86) was recruited via Prolific. Participants were randomly assigned to one of three conditions (n = 100 per condition). Prolific users who had participated in Experiment 1 or Experiment 3 were not invited to participate.
Materials
Impression-Formation Task
The task was identical to that used in Experiment 3, but with a new target statement and a new causally-related statement.
Target statement. The new target statement was selected from the corpus of pre-tested behavioural stimuli (Mickelberg et al., 2022) and met similar criteria to the original negative target statement (morality value close to −4, believability > 5). The target statement selected was “John kicked his pet dog hard in the head when it didn’t come when called” (morality = −3.31, believability = 5.82).
Causally-related statement. The causally-related statement was generated by the authors. The causally-related statement needed to (i) be causally related to the target statement, and (ii) be perceived as neutral and not imply the misinformation when presented in the absence of the target statement (i.e., in the neutral condition). The causally-related statement selected was “John accompanied his wife to the vet to have his dog treated for a head injury”.
Inference Questions
As per Experiment 3, seven inference questions (two open-ended questions and five Likert rating scales) were designed to measure reliance on the misinformation during inferential reasoning. The questions were updated to reflect the new misinformation statement: open-ended questions (presented as the first and last items) asked participants for information about John (“If you could tell someone about one specific thing John has done, what would it be?”; “Describe briefly in one sentence how you think John’s pet dog could have received a head injury”). The rating scales asked participants to rate their endorsement of John across five statements (e.g., “John is a good person”; see Supplementary Materials) using an 11-point Likert scale from 0 (strongly disagree) to 10 (strongly agree).
Inference-Score Calculation. The inference-score calculation matched Experiment 3, with an updated scoring guide to reflect the new target statement (see Supplementary Materials).
Procedure
The procedure was identical to Experiment 3. Median completion time was 10 minutes, and participants were paid £1.50 (approximately US$2.05).
Results
Performance on the recognition test was high (M = 91.30%, SD = 8.59%). One participant was excluded due to technical difficulties preventing them from completing the task, leaving a final sample of N = 299. One participant scored below 50% in the recognition test; excluding them did not qualitatively affect the results so they were included in the analysis.
Manipulation Checks
As illustrated in Figure 8, participants in all conditions reported a positive impression of John from the outset, and this gradually improved over trials 1–12 (see Supplementary Materials). The target statement was then presented at trial 13. To ensure the target-statement manipulation was successful, we tested the impact of the negative target statement on person impressions at trial 13. At trial 13, there was no evidence of a statistical difference between the negative-retraction and negative no-retraction conditions (p = .076), and thus for the purpose of comparing to the neutral condition, these were collapsed into a combined negative condition. Impression ratings at trial 13 were entered into a linear regression with target-statement valence as the predictor (negative, neutral). This showed that the negative conditions returned lower impression ratings than the neutral control (i.e., impressions were more negative when participants had received the negative target statement; β = −58.38, SE = 2.44, t = −23.84, p < .001).
Note. The shaded areas are 95% confidence intervals. The vertical line at trial 13 indicates target-statement presentation, the vertical line at trial 15 indicates causally-related statement presentation, and the vertical line at trial 19 indicates target-statement retraction (in the retraction condition).
Note. The shaded areas are 95% confidence intervals. The vertical line at trial 13 indicates target-statement presentation, the vertical line at trial 15 indicates causally-related statement presentation, and the vertical line at trial 19 indicates target-statement retraction (in the retraction condition).
Between trial 13 (the target statement) and trial 19 (the target-statement retraction, in the retraction condition), the effect of the negative target statement weakened but ratings remained lower than the neutral-condition baseline (the difference to neutral was statistically significant; see Supplementary Materials).
Tests for Continued Influence
The CIE was tested at two time-points: immediately following the target-statement retraction (trial 19) and after a delay (trial 27).
Immediate Test
The trial-19 impression ratings were entered into a linear regression with condition as a predictor (negative retraction; negative no-retraction; neutral). The model was significant, F(2, 296) = 96.77, p < .001, R2 = .39. In the negative no-retraction condition, impression ratings were lower than in the neutral control condition, β = −38.38, SE = 3.46, t = −11.09, p < .001, indicating that the effect of the negative target statement persisted at trial 19. When the negative target statement was retracted in the negative-retraction condition, impression ratings were higher than in the negative no-retraction condition, β = 42.15, SE = 3.62, t = 11.64, p < .001, and did not differ from the neutral control condition, β = 3.77, SE = 2.94, t = 1.28, p = .202. This indicates the retraction of the negative target statement was fully effective.
Delayed Test
Next, the same set of analyses was conducted on the trial-27 impression ratings. The model was significant, F(2, 296) = 59.88, p < .001, R2 = .29. In the negative no-retraction condition, impression ratings were lower than in the neutral control condition, β = −34.71, SE = 3.67, t = −9.46, p < .001, indicating that the effect of the negative target statement persisted at trial 27. For the negative-retraction condition, impression ratings were higher than those in the negative no-retraction condition, β = 31.88, SE = 3.73, t = 8.55, p < .001, and did not differ from the neutral control condition, β = −2.83, SE = 3.15, t = −0.90, p = .370. This indicates that the retraction of the negative target statement was still fully effective after a delay.
Likability Scores
Likability scores were entered into a linear regression with condition as a predictor (negative retraction; negative no-retraction; neutral). The model was significant, F(2, 296) = 81.88, p < .001, R2 = .36. Likability was lower in the negative no-retraction condition compared to the neutral control condition, β = −1.62, SE = 0.15, t = −10.98, p < .001, and the negative-retraction condition, β = −1.51, SE = 0.15, t = −10.14, p < .001. There was no evidence of a statistical difference between the negative-retraction and neutral conditions, β = −0.12, SE = 0.13, t = −0.91, p = .362, indicating that the retraction was fully effective.
Inference Scores
Inference scores supported the observations made based on the impression ratings and likability scores (see Figure 9). Inference scores were entered into a linear regression with condition as a predictor (negative retraction, negative no-retraction, neutral). The model was significant, F(2, 296) = 410.50, p < .001, R2 = .74. Inference scores were higher for the negative no-retraction condition compared to the neutral condition, β = 4.19, SE = 0.17, t = 25.40, p < .001, and the negative-retraction condition, β = 3.93, SE = 0.18, t = 21.53, p < .001. There was no evidence of a statistical difference between negative-retraction and neutral conditions, β = 0.26, SE = 0.14, t = 1.82, p = .071. This indicates that the retraction was fully effective and there was no evidence of a CIE in inferential reasoning.
Note. Bars indicate condition means; error bars are 95% bootstrapped confidence intervals; data points indicate participant mean scores; violins provide distributional information.
Note. Bars indicate condition means; error bars are 95% bootstrapped confidence intervals; data points indicate participant mean scores; violins provide distributional information.
Equivalence Analysis
In line with Experiments 1–3, Experiment 4 showed that people can fully discount discredited misinformation in impression formation (the only difference across experiments was that there was some evidence for overcorrection in the immediate test of Experiments 1–3 and the delayed test in Experiments 2 and 3). The same pattern of results was observed for the likability scores in Experiments 1–4. However, this conclusion regarding the absence of evidence for a CIE in impression formation is based on null results, and it therefore needs to be made with caution.
To corroborate the absence of an effect, an equivalence analysis (Counsell & Cribbie, 2015; Lakens, 2017) was conducted on the non-significant linear regression results between the negative-retraction condition and neutral condition across Experiments 1–4, using impression ratings and likability scores. Equivalence analysis allows for direct inferences to be made about the absence of an effect or the presence of a negligibly small effect. We used the Anderson-Hauck procedure (S. Anderson & Hauck, 1983) in the reg.equiv function in R (Alter & Counsell, 2021). The Anderson-Hauck procedure has been shown to have greater statistical power than the two one-sided tests procedure (TOST; Schuirmann, 1987) when comparing regression coefficients at smaller sample sizes (e.g., Alter & Counsell, 2021; Counsell & Cribbie, 2015). In line with Campbell (2023), we chose standardised regression coefficients of ±0.1 (i.e., a small effect size according to Cohen, 1988), and anything that falls within the +0.1 to −0.1 range was determined to be a negligible effect. Across all experiments, the results indicated that there was insufficient evidence for negligible effects (i.e., no consistent evidence in favour of the null) across all the measures (impression ratings and likability scores), meaning that the true population effect could be larger/smaller than ±0.1 (see Supplementary Materials).4
Discussion
Experiment 4 used a new set of materials and showed much the same pattern of results as Experiments 1–3. Non-retracted negative misinformation had an enduring impact on person impressions, and a clear retraction eliminated the influence of misinformation on impression formation (see also De keersmaecker & Roets, 2017; Ecker & Rodricks, 2020). Impression ratings did not differ from the neutral control condition immediately following the retraction of the negative misinformation, or after a delay. Contrary to Experiments 1–3, there was no evidence of an overcorrection in Experiment 4. As per Experiment 3, no CIE was detected in person-related inferential reasoning.
General Discussion
The present study investigated the impact of person-related misinformation and its correction on person impressions. As predicted, negative misinformation had a greater impact on person impressions compared to positive misinformation, and the impact was longer-lived (when not retracted), representing a negativity bias in impression formation (Rusconi et al., 2020; Ybarra, 2001). Critically, across four experiments, we found no statistical evidence for the continued influence of retracted misinformation on person impressions. This was true for negative misinformation and for positive misinformation.
The study set out to clarify mixed findings in the existing literature regarding the CIE in impression formation, by testing three explanations. Firstly, we suggested that a short-lived CIE may have gone undetected in previous studies, as they only tested for it at the end of the task. We therefore introduced a dynamic impression-rating measure, but did not detect a CIE, neither at the time the misinformation was retracted, nor at the end of the task. Secondly, as previous studies had used a mixture of negative and positive misinformation, we suggested that misinformation valence may moderate the CIE in impression formation. However, we found no CIE with either negative or positive misinformation. Finally, we suggested there may have been differences in misinformation coherence in previous studies, and this could have affected whether or not a CIE was observed. However, we found that the introduction of coherence-building elements did not promote a CIE in impression formation.
Our findings are consistent with Ecker and Rodricks (2020) and Cobb (2013), suggesting that people can fully discount retracted person-related misinformation. Moreover, the present results are in line with the broader impression-formation literature whereby person judgements can be rapidly updated in light of new information (Rydell et al., 2007; Skowronski & Carlston, 1992)—this was shown by the success of the retractions, but also by the fact that a single piece of misinformation impacted person impressions.
Our results contrast with the robust evidence for the CIE in event-related reasoning (Chan et al., 2017; Walter & Tukachinsky, 2020). One way to interpret this result is that corrections of event-related and person-related misinformation may be processed differently. While the retraction of causal event-related misinformation can threaten mental-model coherence, promoting a CIE, mental-model coherence may be less important (or more robust) in impression formation. As a result, mental models may be easier to update when person-related misinformation is retracted, and therefore misinformation can be corrected more effectively. If there are fundamental differences between how events and people are represented in mental models (e.g., more emphasis on temporal and causal relations for events; Bower & Morrow, 1990; van Oostendorp & Bonebakker, 1999), this may explain why even the addition of coherence-building elements did not lead to a CIE in impression ratings (Experiments 2–4) and inferential reasoning (Experiments 3 and 4). To confidently establish if there are fundamental differences between events and people, it would be ideal to compare event and person-related misinformation within a single study, with target type (event or person) as an independent variable. However, designing a study in which events can be equivalently compared to people is a non-trivial challenge.
We found no evidence of a CIE across three different pieces of misinformation (one positive and two negative), indicating that the result is somewhat generalisable. However, a CIE may occur for other pieces of misinformation. Going forward, it will be important to test new statements that vary in content (e.g., morality vs. competence), while being careful not to introduce confounds (e.g., evaluative extremity) that have been shown to moderate person judgements (Rusconi et al., 2017; see Brambilla et al., 2021, for a review).
It should also be noted that equivalence tests did not return evidence in favour of the null hypothesis. Our null findings therefore highlight the need for additional research, rather than serving as strong evidence against the existence of continued influence in the context of person impressions. Specifically, we hope that the present paper will contribute to the bigger picture by drawing attention to our null findings (thus avoiding the ‘file-drawer problem,’ which has contributed to limited reproducibility of, and eroded credibility in, psychological findings; Ferguson & Heene, 2012; Open Science Collaboration, 2015) and allowing interested researchers to establish better-calibrated estimates of the ‘true’ effect size of the CIE in the context of person impressions, thus ultimately advancing CIE theory.
Limitations and Future Research Directions
Despite the consistent results returned across four experiments, there are several limitations of the current research. First, the present set of experiments used only a single fictional person (“John”). Although a previous study on continued influence in person impressions found equivalent (null) effects regardless of the fictional protagonist’s name and implied cultural background (i.e., “John” vs. “Vladimir”; Ecker & Rodricks, 2020), future research could use a range of different names, including female or gender-nonspecific names. Moreover, more research is needed on how misinformation influences impressions of familiar versus unfamiliar people. For example, a CIE may be more likely to emerge if participants already hold a negative attitude towards a person based on previous encounters, existing knowledge, or incompatible worldviews (Thorson, 2016; see also Bullock, 2007). Given the documentation of clear cases of misinformation impacting person judgements in other contexts (Brooks & Greenberg, 2021; Clow et al., 2012; Jardina & Traugott, 2019; Steblay et al., 2006; Weeks & Garrett, 2014), a long-term research goal will be to identify the boundary conditions under which person misinformation does versus does not continue to influence person impressions after a retraction.
Second, when the misinformation was retracted in the present study, it was done using an unambiguous retraction (i.e., “John did NOT have an affair with his best friend’s wife”) and there was no reason to question the retraction’s credibility. As believability of new evidence is a predictor of impression revision (Cone et al., 2019), the unambiguous nature of the retraction may have contributed to the elimination of the retracted misinformation’s influence. That said, previous studies that found a CIE with event misinformation also used unambiguous retractions (e.g., Ecker et al., 2010; Johnson & Seifert, 1994), and apart from the clear wording, no specific measures were taken in the present study to boost the credibility of the retraction or its source; thus, this line of reasoning may not provide a satisfactory explanation for the absence of an effect. However, it is possible that a CIE in impression formation might arise if the correction is more tentative (MacFarlane et al., 2021) or the correction source is less credible (e.g., if a retraction is provided by the accused person themselves; also see Connor Desai et al., 2020; Ecker & Antonio, 2021), which is something that future studies might examine.
Finally, the use of a dynamic impression-rating measure may have led to greater retraction efficacy than would otherwise be expected. Research has shown that memory for event information is a predictor of susceptibility to the CIE (Sanderson et al., 2021). It could be that eliciting trial-to-trial impression updates improved participants’ memory of the behaviour statements, allowing a higher-fidelity mental model to be created. As a result, it may be easier to integrate the retraction and update the mental model, reducing participants’ reliance on the retracted misinformation (Ecker et al., 2017; Kendeou et al., 2014, 2019). A complementary explanation is based on evidence that eliciting repeated judgements (similar to the dynamic impression-rating measure) can lead people to give greater weight to more recent information when forming person impressions (i.e., recency effect, see Kashima & Kerekes, 1994, for a review). This could lead participants to prioritise the retraction over the initial misinformation when making their post-retraction impression ratings, reducing reliance on the misinformation. While a dynamic measure was necessary to assess the time course of information updating, future research should consider whether paradigm selection (i.e., continuous vs. single judgements) affects results. The use of dynamic and static impression measures could also provide insight into how misinformation shapes person impressions across different real-world information environments (e.g., social media vs. traditional media), where the frequency of updating may vary.
Conclusion
The continued reliance on person-related misinformation can have serious implications for those who are the target of false allegations, as well as for broader society. The present study found that when people form an impression of a fictional person, negative misinformation has a greater impact than positive misinformation. However, and regardless of valence, when misinformation was unequivocally retracted, participants fully discounted the retracted misinformation, even when other behaviour descriptions were congruent or causally related to the retracted misinformation. Thus, for the scenarios considered, once person-related misinformation was retracted, person impressions could be fully updated, suggesting that mud does not always stick.
Author Contributions
AM, BW, NF and UE contributed to conception and design. BW developed the experiment. BW and AM collected and stored the data. AM analysed the data. AM wrote the manuscript, with input from all authors. All authors revised the article and contributed to interpretation of data. All authors approved the final submitted version of the manuscript.
Competing Interests
The authors declared that they had no conflicts of interest with respect to authorship or the publication of this article.
Data Accessibility Statement
All the participant data can be found on the project page at OSF: https://osf.io/ymuap
Funding
Research reported in this publication was supported by a Postgraduate Research Scholarship from the Defence Science and Technology Group of the Department of Defence and an Australian Government Research Training Program Scholarship to the first author (AM), an Australian Research Council grant FT190100708 to the third author (UE), and an Office of National Intelligence and Australian Research Council grant NI210100224 to the last author (NF).
Footnotes
In providing evidence for and against the CIE in impression formation, we refer to studies that have measured person impressions explicitly (i.e., direct person judgements). It should be noted that in the impression updating literature, implicit, indirect judgements tend to be more resistant to updating (see Rydell et al., 2007), although even implicit impressions can be influenced by newly provided information to the extent that it is diagnostic (Cone & Ferguson, 2015), believable (Cone et al., 2019) and subject to reinterpretation (Mann & Ferguson, 2015; Okten et al., 2019).
We note that in the traditional event-related CIE literature, there have been suggestions that positive misinformation may be easier to correct than negative misinformation, but hitherto this notion has not been substantiated (see Chang et al., 2019; Walter & Tukachinsky, 2020).
An additional impression rating was added at the very end of the procedure to determine if there were any carry-over effects following the inference questions. This was not the case. For the interested reader, the results of the final impression rating can be found in the Supplementary Materials.
Because the equivalence bounds were conducted following the main analysis, the equivalence analysis was rerun using a medium effect size (±0.3). Equivalence tests at the ±0.3 level indicated a negligible effect (evidence in favour of the null) for the likability scores (Experiments 1–3) but not for the dynamic impression ratings (Experiments 1 and 4) or likability scores (Experiment 4).