First Approximation
A Washington State University study (Prof. Mesut Cicek) found that ChatGPT correctly classified scientific hypotheses as true or false ~80% of the time — but once performance is adjusted for random guessing (chance alone would yield ~50%), the effective score drops to roughly "D grade" territory, indicating only modest true reasoning ability. A further red flag: ChatGPT frequently gave *contradictory answers* to the same scientific question across repeated trials, undermining trust in its consistency as a scientific tool. The 80% headline number is real but misleading — it's the chance-adjusted performance and the inconsistency rate that matter most, and both are damning.
- The exact statistical method used for chance-adjustment (Cohen's kappa? Corrected accuracy?) — the primary peer-reviewed paper was not directly accessed
- Whether GPT-4 or other models were tested — results may not generalize beyond ChatGPT-3.5
- Domain breakdown — inconsistency may be far worse for cutting-edge vs. established science (logged as hypothesis)
- NASA SWOT claim — entirely unrelated; a pipeline clustering artifact, not a genuine part of this topic
The primary peer-reviewed paper (journal, DOI, full methodology) was never retrieved. All findings flow through secondary science journalism. The study's conclusions are credible and internally consistent, but independent replication and peer scrutiny of the chance-adjustment methodology remain essential before treating this as settled evidence.
ChatGPT should not be trusted as a standalone scientific reasoning tool. The 80% accuracy figure, often cited as impressive, is substantially undermined by the chance-adjustment analysis and — more practically — by the finding that it gives contradictory answers to identical questions. For researchers, this suggests LLMs in their current form are better used as literature-search assistants than as hypothesis evaluators.
Findings
fact
WSU Study: ChatGPT's 80% Accuracy Masks Weak Underlying Reasoning
A March 2026 Washington State University study by Prof. Mesut Cicek tested ChatGPT on hundreds of scientific hypotheses (true/false classification). Surface accuracy was ~80%, but once adjusted for random guessing, performance dropped to only "modest" levels. Additionally, ChatGPT frequently contradicted itself when asked the same question multiple times — sometimes flipping answers back and forth. This suggests the AI's apparent competence in scientific reasoning is partially an artifact of binary-choice probability rather than genuine understanding.
fact
ChatGPT ~80% Raw Accuracy on Scientific Hypotheses, but Modest After Guessing Adjustment
A Washington State University study (reported by Science Daily) tested ChatGPT on hundreds of scientific hypotheses across multiple disciplines. Raw classification accuracy was approximately 80%. However, when results were adjusted to account for the probability of getting answers right by random guessing, the adjusted performance was described as only "modest." The study also found that ChatGPT produced contradictory answers when asked the same question multiple times — a consistency failure that further undermines its reliability for scientific reasoning tasks.
pattern
Single-Source Concentration: All ChatGPT Reasoning Claims from Science Daily
All three substantive claims about ChatGPT's scientific reasoning limitations originate from a single source (Science Daily, credibility score 0.5) with very low independent corroboration (0.18). Science Daily is a news aggregator that summarizes academic press releases, not an independent scientific source. There is no corroborating coverage from independent scientific publications, AI research outlets, or other news organizations in the database. This makes these claims plausible but unverified beyond a single reporting pipeline.
connection
Misplaced NASA/SWOT Claim in This Topic Cluster
One claim in this topic cluster (claim_id: 23f85b66) — "Perhaps less than previously thought, according to new data from the SWOT mission" — belongs to NASA/JPL reporting about oceanography or Earth science (likely sea level or river discharge data from the SWOT satellite mission). It was likely mislabeled or misclassified into this ChatGPT topic. This claim has zero corroboration and is semantically unrelated to ChatGPT's scientific reasoning limitations. It should be investigated separately and removed from this cluster.
fact
ChatGPT Scores ~80% Raw But "Modest" After Guessing Adjustment
A study reported by Science Daily found ChatGPT correctly classified scientific hypotheses as true or false approximately 80% of the time across hundreds of hypotheses. However, when adjusted to account for the probability of random guessing, the effective score drops substantially — leading researchers to characterize ChatGPT's scientific reasoning as "only modest." The 80% raw figure is misleading because a coin flip would yield 50% correct on true/false items; the adjusted score factors out this baseline.
fact
ChatGPT Produces Contradictory Answers to Repeated Scientific Questions
The same Science Daily study found that ChatGPT frequently produces contradictory answers when asked the same scientific question multiple times. This inconsistency is a separate and compounding limitation — not only is performance modest overall, but the model's unreliability across repeated queries means users cannot trust it to give the same answer consistently. This raises concerns about using ChatGPT as a reliable scientific reasoning tool, particularly in research or educational contexts where reproducibility matters.
fact
WSU Study: ChatGPT Scores "D" on Scientific Hypothesis Classification
Washington State University professor Mesut Cicek and colleagues conducted a study feeding more than 700 scientific hypotheses into ChatGPT, repeating each query 10 times. ChatGPT achieved approximately 76-80% raw accuracy in classifying hypotheses as true or false. However, the researchers downgraded this to a "D" performance rating because when adjusted for chance (random guessing on a true/false binary has a 50% baseline), the effective net accuracy is only modestly above chance. Additionally, ChatGPT was inconsistent, giving contradictory answers to the same question across the 10 repetitions — about 27% of the time it would flip its answer. This was published/reported around March 17, 2026.
fact
The "80% accuracy" framing is misleading without baseline context
The claim that "ChatGPT correctly classified scientific hypotheses ~80% of the time" sounds impressive, but the researchers explicitly adjusted for the 50% random-guessing baseline inherent to true/false questions. After this adjustment (e.g., using Cohen's kappa or corrected accuracy), the net performance above chance is much more modest — this is why the study rated ChatGPT's performance as equivalent to a "D" grade. The 80% raw figure is not false, but without the baseline context it significantly overstates ChatGPT's actual reasoning ability. This is a key nuance that the ScienceDaily headline captures but which requires the full study to understand properly.
pattern
Contradictory answers reveal reliability problem beyond accuracy
The most practically significant finding from the WSU study may not be ChatGPT's accuracy rate, but its inconsistency: when the same scientific hypothesis was asked 10 times, ChatGPT gave contradictory answers in a substantial minority of cases. This is separate from the accuracy-versus-chance issue and represents a reliability/reproducibility problem. A tool that gives different answers to identical queries cannot be trusted as a scientific reasoning assistant, regardless of its average accuracy. This has implications for how AI tools are used in research, systematic reviews, and hypothesis testing.
gap_filled
NASA SWOT claim is an unrelated database artifact
One of the 4 claims in this topic — "Perhaps less than previously thought, according to new data from the SWOT mission" — is from NASA JPL and is clearly unrelated to ChatGPT scientific reasoning. It was incorrectly grouped into this topic, likely due to proximity in the ingestion pipeline. The SWOT mission is a US-French satellite that measures river water. This claim should be excluded from analysis of ChatGPT's scientific reasoning limitations.
What We Don't Know
study —
Primary peer-reviewed paper by Prof. Mesut Cicek (WSU) on ChatGPT scientific hypothesis classification
All claims in this topic trace back to a single Science Daily summary. Without the primary paper, we cannot verify the methodology (how "adjusted for random guessing" was calculated), the exact fields of science tested, sample size, or whether findings have been peer-reviewed and published in a journal. Independent replication is also unknown.
study —
Original peer-reviewed study on ChatGPT scientific reasoning from Washington State University
Science Daily covers the study's conclusions but the primary source — the actual journal article with methodology, sample size, disciplines tested, and the specific guessing-adjustment formula used — has not been retrieved or verified. Without this, we cannot independently assess whether the 80% figure is meaningful and how "modest" the adjusted score actually is.
document —
Primary peer-reviewed paper by Prof. Mesut Cicek (WSU) on ChatGPT scientific hypothesis classification — journal name, DOI, and exact methodology
The exact statistical method used to adjust for random guessing (e.g., Cohen's kappa vs. corrected accuracy) determines how severe the performance downgrade truly is. Without the full paper, the "D grade" rating cannot be independently verified or compared against other AI benchmarks.
Connected Investigations
unexpected
structural
AI Tool Unreliability and Supply Chain Attack Converge on Same Infrastructure Week
In the week of March 19-22, 2026, Trivy — a tool used by developer teams to assess software security — was compromised, potentially corrupting CI/CD secrets across thousands of organizations. Simultaneously, a WSU study confirmed ChatGPT gives contradictory answers to identical scientific questions. Both findings target the same underlying trust layer: automated tools that developers and researchers rely on to verify correctness are demonstrably unreliable, one through adversarial compromise, one through inherent stochasticity. Organizations using Trivy to 'verify' security may have been doing so with a compromised verifier while using ChatGPT to interpret results.
Connecting: Trivy Supply Chain Compromise + ChatGPT Scientific Reasoning Limitations
unexpected
cascading
Trivy Compromise and ChatGPT Inconsistency Both Undermine the Same Emerging Workflow
Security teams increasingly combine automated vulnerability scanners (like Trivy) with LLM-assisted analysis (like ChatGPT) to triage findings. The Trivy compromise means scanner outputs during the March 19 window cannot be trusted. The WSU study means ChatGPT-assisted interpretation of those outputs is itself unreliable — giving contradictory answers to identical questions. Any organization that used this combined workflow during the compromise window faces a double epistemic failure: the data was potentially poisoned AND the tool used to reason about it is demonstrably inconsistent. No coverage has noted this compounding vulnerability.
Connecting: Trivy Supply Chain Compromise + ChatGPT Scientific Reasoning Limitations
unexpected
structural
Both 'Scientific Truth' Stories Reveal the Same Peer Review Access Gap
The ChatGPT scientific reasoning study (WSU) and the origins-of-life asteroid findings both confirm striking results through secondary science journalism while the primary peer-reviewed papers remain unaccessed by investigators. In both cases, the methodology details that determine how strong the conclusions actually are — chance-adjustment method for ChatGPT; nucleobase delivery survival rates for asteroids — sit behind a journalism layer that strips exactly the caveats that matter. The same epistemic gap that makes ChatGPT unreliable for science (surface accuracy without depth) is reproduced in how the science about ChatGPT is itself being reported.
Connecting: ChatGPT Scientific Reasoning Limitations + Origins of Life Building Blocks