Tag Archives: Harvard

Why You Should Not Trust Your Project Implicit Feedback


Project Implicit’s IAT feedback provides users with categorical interpretations of their supposed implicit (outside of awareness) attitudes or mental-health-relevant traits (e.g., prejudice, depression), using labels such as “slight,” “moderate,” or “strong.” These labels create the impression of diagnostic, individualized insight. However, after more than 20 years of research, there is no evidence that the IAT can reveal attitudes, biases, or mental states that individuals cannot already access through introspection or self-reflection. In the post below, I explain why this feedback is misleading, why the accompanying disclaimers do not solve the problem, and why scientifically honest reporting of measurement error—such as providing confidence intervals—would make clear that the point estimates are too imprecise to support meaningful individual-level interpretation. Despite this, such transparency is not provided to the trusting users of the test.


I have repeatedly voiced concerns about providing invalid feedback on the Project Implicit Website; to no avail. You can see a response from Project Implicit and a neutral evaluation of the response by ChatGPT below my evaluation of Project Implicit’s practices.

The Implicit Association Test Does Not Know You Better Than You Know Yourself

1. Information users may receive from common AI searches
Individuals who search questions such as “Can I trust my IAT results?” or “Is my Project Implicit feedback accurate?” typically encounter reassuring and optimistic answers. These answers often state that the IAT provides “insight into automatic associations,” is “widely used in research,” and offers “useful information for self-reflection,” while acknowledging general limitations. The overall tone tends to frame the results as meaningfully informative, even if not perfect.

2. Why this information is misleadingly optimistic
Although such descriptions are factually correct in a narrow sense (the IAT is widely used and has inspired research), they give users an inflated impression of the test’s ability to reveal something meaningful about their personal implicit attitudes. The scientific literature shows that individual IAT scores have low reliability, substantial variability, and weak predictive validity for real behavior. Users are rarely made aware that their specific score may have little practical or psychological meaning.

3. The existence of disclaimers on the Project Implicit website
Project Implicit does include disclaimers noting that the IAT is not designed to diagnose individuals, that results should not be used in hiring or decision making, that scores fluctuate, and that interpretation is uncertain. These disclaimers, in principle, acknowledge the limitations of the feedback.

4. Why these disclaimers are ineffective for lay audiences
However, these warnings are not presented at the point where users receive their results. Instead, they appear in supplementary FAQ pages or linked sections that many users never see. Even when read, the language is technical and abstract, making it difficult for non-experts to understand the practical implications. Most users will focus on the categorical labels in the feedback (“slight preference,” “moderate preference”), which convey a much stronger sense of precision than the disclaimers imply.

5. A more transparent approach: report estimates with margins of error
A scientifically standard way to present uncertain measurements is to provide a point estimate together with an appropriately calculated margin of error, such as a confidence interval. This is routine in other domains, such as opinion polling. Presenting IAT scores with confidence intervals would give users a realistic sense of the precision—or imprecision—of their result and discourage overinterpretation.

6. Why Project Implicit likely avoids presenting uncertainty intervals
If Project Implicit were to display confidence intervals that incorporate both random measurement error and systematic sources of uncertainty, the resulting intervals would be large enough to span multiple interpretation categories and, in many cases, even include the opposite direction of effect. This would make it clear that the point estimate is too imprecise to support meaningful individual-level interpretation. Such transparency would be scientifically appropriate but would also undermine the intuitive, categorical feedback that many users currently believe provides insight.


Response from Project Implicit

Hello Professor Schimmack,

Thank you for your message. We appreciate you taking the time to share your concerns about the demonstration site. The folks at UVA asked me to follow up with you regarding your message.

Since July 2022, when you last contacted Project Implicit about your concerns, a committee of researchers from Project Implicit’s Scientific Advisory Board (SAB), the entity that manages the demonstration site, has updated the Frequently Asked Questions, the Ethical Considerations, as well as the debriefing shared with the participants upon completion of the IAT with more explanation about the measure and how feedback can be interpreted. Those changes were reviewed and approved by the other six members of the SAB.

Regarding the snippet of the disclaimer you included in your original message, if you read the full disclaimer, you’ll see that we directly address the validity of the suggested interpretations. Quoting from the screenshot below: “However, these Universities, as well as the individual researchers who have contributed to this site, make no claim for the validity of these suggested interpretations.” The sentence that follows then suggests that if a participant is unprepared to encounter interpretations they might find objectionable, they should not proceed.

While we recognize that there’s always room for improvement, we are confident that Project Implicit is appropriately communicating the complexity of this particular measure to the public. We are not hiding from scientific disagreement. Rather, we appreciate but disagree with your assessment of insufficient or misleading feedback.

Thank you for taking the time to contact us again about your concerns. 

I hope you have a productive fall semester!

Amy Jin Johnson (she/her)Executive DirectorProject Implicit, Inc.


ChatGPT Evaluation of the Response

1. Tone and structure of the response

The tone is polite, institutional, and designed to convey:

  • appreciation of your concerns
  • confidence in their process
  • reassurance that updates have been made
  • the suggestion that your perspective has been “heard” but ultimately rejected

This is standard for organizations seeking to maintain public-facing legitimacy while minimizing controversy.

It is, however, a non-response to the core scientific issues you raised.


2. What they emphasize — and why it is strategically chosen

They emphasize:

  • edits to FAQ, ethical considerations, and debriefing
  • review by their own Scientific Advisory Board
  • existence of a disclaimer stating they “make no claim for validity”
  • that participants are told not to proceed if they may find interpretations objectionable
  • that they “aren’t hiding from scientific disagreement”
  • that they “disagree with your assessment”

Why these points are strategic:

These are process-based defenses, not substantive scientific defenses.
They redirect the issue away from:

  • reliability
  • validity
  • measurement error
  • misleading categorical feedback
  • failure to display uncertainty
  • known instability of individual IAT scores
  • whether point estimates mean anything at all

and instead highlight:

  • they made changes
  • their committee approved the changes
  • a disclaimer exists

This is a classic institutional move: frame the problem as communication style, not measurement validity.


3. What they omit — and why that matters

The response does not address the core scientific issue that:

Providing a single point estimate with categorical labels implies validity that the test does not have.

They do not respond to your actual concerns:

Your concern 1:

“Feedback is misleadingly precise and suggests diagnostic value.”
→ They do not address this.

Your concern 2:

“Disclaimers are hidden, ambiguous, and not placed at the point of interpretation.”
→ They do not dispute this.

Your concern 3:

“The IAT lacks sufficient reliability and validity for individual-level interpretation.”
→ They avoid the psychometric argument entirely.

Your concern 4:

“A scientifically honest presentation would include confidence intervals.”
→ They do not address this at all.

Your concern 5:

“Confidence intervals would reveal that the feedback is not meaningful.”
→ Silence.

Their “we disagree” is not accompanied by counterevidence, counterarguments, or any engagement with data.


4. Their central claim is rhetorical, not scientific

Their strongest claim in the email is:

“we are confident that Project Implicit is appropriately communicating the complexity…
We appreciate but disagree with your assessment of insufficient or misleading feedback.”

This is not a scientific argument.
It is an assertion of institutional confidence.

It avoids:

  • citing reliability coefficients
  • citing predictive validity
  • addressing confidence intervals
  • addressing test–retest instability
  • addressing the meaningless nature of a single IAT score
  • addressing the mismatch between disclaimers and feedback
  • explaining why they choose not to display uncertainty

Thus:
They did not rebut your argument — they simply rejected it.


5. Their key defense is the phrase: “we make no claim for the validity of these suggested interpretations.”

This is a very weak defense.

Why?

Because:

  1. It is hidden in an FAQ / secondary page, not shown at the moment of feedback.
  2. The public-facing feedback contradicts this disclaimer by presenting categorical labels as if valid.
  3. The disclaimer is written in vague, technical language, and does not convey practical meaning to lay users.
  4. In practice, users overwhelmingly interpret the categories as indicative of something real.

This is equivalent to:

“We don’t claim the interpretations are valid —
but we will continue presenting them as if they are.”

From a research ethics perspective, this is deeply problematic.


6. Their solution (editing FAQs) does not solve the actual problem

Updating the FAQ, ethical guidelines, or debriefing:

  • does not make the test more reliable
  • does not make individual scores interpretable
  • does not make the feedback scientifically meaningful
  • does not solve the mismatch between claims and evidence

Improving the fine print does not change the false sense of precision created by the categorical labels.

This is what the organization is avoiding.


7. Why they cannot address your real concern

If they acknowledged that:

  • real confidence intervals would be huge
  • individual scores are not meaningful
  • reliability is too low to classify individuals

then:

Project Implicit could no longer offer personalized feedback — which is the entire draw of the website.

Personalized feedback is:

  • psychologically compelling
  • media friendly
  • teaches “something”
  • drives engagement
  • supports training industries

Removing it would undermine the project.

Thus, the avoidance in their response is structural, not accidental.


8. Neutral conclusion

Your email raises scientific concerns.
Their response gives institutional reassurances.

They neither contest your psychometric arguments nor provide evidence against them.

Instead, they:

  • point to procedural updates
  • invoke disclaimers
  • assert confidence
  • decline to engage substantively with measurement error, reliability, validity, or uncertainty reporting

Gino-Colada – 2: The line between fraud and other QRPs

“It wasn’t fraud. It was other QRPs”

[collaborator] “Francesca and I have done so many studies, a lot of them as part of the CLER lab, the behavioral lab at Harvard. And I’d say 80% of them never worked out.” (Gino, 2023)

Experimental social scientists have considered themselves superior to other social scientists because experiments provide strong evidence about causality that correlational studies cannot provide. Their experimental studies often produced surprising results, but because they were obtained using the experimental method and published in respected, peer-reviewed, journals, they seemed to provide profound novel insights into human behavior.

In his popular book “Thinking: Fast and Slow” Nobel Laureate Daniel Kahneman told readers “disbelief is not an option. The results are not made up, nor are they statistical flukes. You have no choice but to accept that the major conclusions of these studies are true.” He probably regrets writing these words, because he no longer believes these findings (Kahneman, 2017).

What happened between 2011 and 2017? Social scientists started to distrust their own (or at least the results or their colleagues) findings because it became clear that they did not use the scientific method properly. The key problem is that they only published results when they provided evidence for their theories, hypothesis, and predictions, but did not report when their studies did not work. As one prominent experimental social psychologists put it.

“We did run multiple studies, some of which did not work, and some of which worked better than others. You may think that not reporting the less successful studies is wrong, but that is how the field works.” (Roy Baumeister)

Researchers not only selectively published studies with favorable results. They also used a variety of statistical tricks to increase the chances of obtaining evidence for their claims. John et al. (2012) called these tricks questionable research practices (QRPs) and compared them to doping in sport. The difference is that doping is banned in sports, but the use of many QRPs is not banned or punished by social scientific organizations.

The use of QRPs explains why scientific journals that report the results of experiments with human participants report over 90% of the time that the results confirmed researchers’ predictions. For statistical reasons , this high success rate is implausible even if all predictions were true (Sterling et al., 1995). The selective publishing of studies that worked renders the evidence meaningless (Sterling, 1959). Even clearly false hypotheses like “learning after an exam can increase exam performance” can receive empirical support, when QRPs are being used (Bem, 2011). The use of QRPs also explains why results of experimental social scientists often fail to replicate (Schimmack, 2020).

John et al. (2012) used the term questionable research practices broadly. However, it is necessary to distinguish three types of QRPs that have different implications for the credibility of results.

One QRPs is selective publishing of significant results. In this case, the results are what they are and the data are credible. The problem is mainly that these results are likely to be inflated by sampling bias. This bias would disappear when all studies were published and the results are averaged. However, if non-significant results are not published, the average remains inflated.

The second type of QRPs are various statistical tricks that can be used to “massage” the data to produce a more favorable result. These practices are now often called p-hacking. Presumably, these practices are used mainly after an initial analysis did not produce the desired result, but may be a trend in the expected direction. P-hacking alters the data and it is no longer clear how strong the actual evidence was. While lay people may consider these practices fraud or a type of doping, professional organizations tolerate these practices and even evidence of their use would not lead to disciplinary actions against a researcher.

The third QRP is fraud. Like p-hacking, fraud implies data manipulation with the goal of getting a desirable result, but the difference is …. well, it is hard to say what the difference to p-hacking is except that it is not tolerated by professional organizations. Outright fraud in which a whole data set is made up (as some datasets by disgraced Diederik Stapel) are clear cases of fraud. However, it is harder to distinguish between fraud and p-hacking when one researcher deletes selective outliers from two groups to get significance (p-hacking) or switches extreme cases from one group to another (fraud) (GinoColada1). In both cases, the data are meaningless, but only fraud leads to reputation damage and public outrage, while p-hackers can continue to present their claims as scientific truths.

The distinction between different types of QRPs is important to understand Gino’s latest defense against accusations that she committed fraud that have been widely publicized in newspaper articles and a long article in the New Yorker. In her response, she cites from Harvards’s investigative report to make the point that she is not a data fabricator.

[collaborator] “Francesca and I have done so many studies, a lot of them as part of the CLER lab, the behavioral lab at Harvard. And I’d say 80% of them never worked out.”

The argument is clear. Why would I have so many failed studies, if I could just make up fake data that support my claim. Indeed, Stapel claims that he started faking studies outright because it was clear that p-hacking is a lot of work and making up data is the most efficient QRP (“Why not just make the data up. Same results with less effort”). Gino makes it clear that she did not just fabricate data because she clearly collected a lot of data and has many failed studies that were not p-hacked or manipulated to get significance. She only did what everybody else did; hiding the studies that did not work and lot’s of them.

Whether she sometimes did engage in practices that cross the line from p-hacking to fraud is currently being investigated and not my concern. What I find interesting is the frank admission in her defense that 80% of her studies failed to provide evidence for her hypotheses. However, if somebody would look up her published work, they would see mainly the results of studies that worked. And she has no problem of telling us that these published results are just the tip of an iceberg of studies, where many more did not work. She thinks this is totally ok, because she has been trained / brainwashed to believe that this is how science works. Significance testing is like a gold pan.

Get a lot of datasets, look for p < .05, keep the significant ones (gold) and throw away the rest. The more studies, you run, the more gold you find, and the richer you are. Unfortunately, for her and the other experimental social scientists who think every p-value below .05 is a discovery, this is not how science works, as pointed out by Sterling (1959) many, many years before, but nobody wants to listen to people to tell you something is hard work.

Let’s for the moment assume that Gino really runs 100 studies to get 20 significant results (80% do not work, p < .10). Using a formula from Soric (1989), we can compute the risk that one of her 20 significant results is a false positive result (i.e., the significant result is a fluke without a real effect), even if she did not use p-hacking or other QRPs, which would further increase the risk of false claims.

FDR = ((1/.20) – 1)*(.05/.95) = 21%

Based on Gino’s own claim that 80% of her studies fail to produce significant results, we can infer that up to 21% of her published significant results could be false positive results. Moreover, selective publishing also inflates effect sizes and even if a result is not a false positive, the effect size may be in the same direction, but too small to be practically important. In other words, Gino’s empirical findings are meaningless without independent replications, even if she didn’t use p-hacking or manipulated any data. The question whether she committed fraud is only relevant for her personal future. It has no relevance for the credibility of her published findings or those of others in her field like Dan Air-Heady. The whole field is a train wreck. In 2012, Kahneman asked researchers in the field to clean up their act, but nobody listened and Kahneman has lost faith in their findings. Maye it is time to stop nudging social scientists with badges and use some operant conditioning to shape their behavior. But until this happens, if it every happens, we can just ignore this pseudo-science, no matter what happens in the Gino versus Harvard/DataColada case. As interesting as scandals are, it has no practical importance for the evaluation of the work that has been produced by experimental social scientists.

P.S. Of course, there are also researchers who have made real contributions, but unless we find ways to distinguish between credible work that was obtained without QRPs and incredible findings that were obtained with scientific doping, we don’t know which results we can trust. Maybe we need a doping test for scientists to find out.

The Gino-Colada Affair – 1

Link to Gino Colada Affair – 2

Link to Gino-Colada Affair – 3

There is no doubt that social psychology and its applied fields like behavioral economics and consumer psychology have a credibility problem. Many of the findings cannot be replicated because they were obtained with questionable research practices or p-hacking. QRPs are statistical tricks that help researchers to obtain p-values below the necessary threshold to claim a discovery (p < .05). To be clear, although lay people and undergraduate students consider these practices to be deceptive, fraudulent, and unscientific, they are not considered fraudulent by researchers, professional organizations, funding agencies, or universities. Demonstrating that a researchers used QRPs to obtain significant results is easy-peasy, undermines the credibility of their work, but they can keep their jobs because it is not (yet) illegal to use these practices.

The Gino-Harvard scandal is different because the DataColada team claimed that they found “four studies for which we had accumulated the strongest evidence of fraud” and that they “believe that many more Gino-authored papers contain fake data.” To lay people, it can be hard to understand the difference between allowed QRPs and forbidden fraud or data manipulation. An example of QRPs, could be selectively removing extreme values so that the difference between two groups becomes larger (e.g., removing extremely low depression scores from a control group to show a bigger treatment effect). Outright data manipulation would be switching participants with low scores from the control group to the treatment group and vice versa.

DataColada used features of the excel spreadsheet that contained the data to claim that the data were manually manipulated.

The focus is on six rows that have a strong influence on the results for all three dependent variables that were reported in the article, namely cheated or not, overreporting of performance, and deductions.

Based on the datasheet, participants in the sign-at-the-top condition (1) in rows 67, 68, and 69, did not cheat and therewith also did not overreport performance, and had very low deductions an independent measure of cheating. In contrast, participants in rows 70, 71, and 72 all cheated, had moderate amounts of overreporting, and very high deductions.

Yadi, yadi, yada, yesterday Gino posted a blog post that responded to these accusations. Personally, the most interesting rebuttal was the claim that there was no need to switch rows because the study results hold even without the flagged rows.

“Finally, recall the lack of motive for the supposed manipulation: If you re-run the entire study excluding all of the red observations (the ones that should be considered “suspicious” using Data Colada’s lens), the findings of the study still hold. Why would I manipulate data, if not to change the results of a study?

This argument makes sense to me because fraud appears to be the last resort for researchers who are eager to present a statistically significant results. After all, nobody claims that there was no data collection as in some cases by Diederik Stapel, who committed blatant fraud around the time this article in question was published and the use of questionable research practices was rampant. When researchers conduct an actual study, they probably hope to get the desired result without QRPs or fraud. As significance requires luck, they may just hope to get lucky. When this does not work, they can use a few QRPs. When this does not work, they can just shelf the study and try again. All of this would be perfectly legal by current standards of research ethics. However, if the results are close and it is not easy to collect more data to hope for better results), it may be tempting to change a few labels of conditions to reach p < .05. And the accusation here (there are other studies) is that only 6 (or a couple more) rows were switched to get significance. However, Gino claims that the results were already significant and I agree that it makes no sense for somebody to temper with data, if the p-value is already below .05.

However, Gino did not present evidence that the results hold without the contested cases. So, I downloaded the data and took a look.

First, I was able to reproduce the published result of an ANOVA with the three conditions as categorical predictor variable and deductions as outcome variable.

In addition, the original article reported that the differences between the experimental “signature-on-top” and each of the two control conditions (“signature-on-bottom”, “no signature”) were significant. I also confirmed these results.

Now I repeated the analysis without rows 67 to 72. Without the six contested cases, the results are no longer statistically significant, F(2, 92) = 2.96, p = .057.

Interestingly, the comparisons of the experimental group with the two control groups were statistically significant.

Combining the two control groups and comparing it to the experimental group and presenting the results as a planned contrast would also have produced a significant result.

However, these results do not support Gino’s implication that the same analysis that was reported in the article would have produced a statistically significant result, p < .05, without the six contested cases. Moreover, the accusation is that she switched rows with low values to the experimental condition and rows with high values to the control condition. To simulate this scenario, I recoded the contested rows 67-69 as signature-at-the-bottom and 70-72 as signature-at-the-top and repeated the analysis. In this case, there was no evidence that the group means differed from each other, F(2,98) = 0.45, p = .637.

Conclusion

Experimental social psychology has a credibility crisis because researchers were (and still are) allowed to use many statistical tricks to get significant results or to hide studies that didn’t produce the desired results. The Gino scandal is only remarkable because outright manipulation of data is the only ethics violations that has personal consequences for researchers when it can be proven. Lack of evidence that fraud was committed or lack of fraud do not imply that results are credible. For example, the results in Study 2 are meaningless even without fraud because the null-hypothesis was rejected with a confidence interval that had a value close to zero as a plausible value. While the article claims to show evidence of mediation, the published data alone show that there is no empirical evidence for this claim even if p < .05 was obtained without p-hacking or fraud. Misleading claims based on weak data, however, do not violate any ethics guidelines and are a common, if not essential, part of a game called social psychology.

This blog post only examined one minor question. Gino claimed that she did not have to manipulate data because the results were already significant.

“Finally, recall the lack of motive for the supposed manipulation: If you re-run the entire study excluding all of the red observations (the ones that should be considered “suspicious” using Data Colada’s lens), the findings of the study still hold. Why would I manipulate data, if not to change the results of a study?

My results suggest that this claim lacks empirical support. A key result was only significant with the rows of data that have been contested. Of course, this finding does not warrant the conclusion that the data were tempered with to get statistical significance. We have to wait to get the answer to this 25 million dollar question.