Screenshot of the abstract:
As large language models (LLMs) are increasingly integrated into decision-making systems—such as au-tonomous vehicles and medical devices—understanding how humans perceive and evaluate AI-generatedjudgements is crucial. To investigate this, we conducted a series of experiments in which participants evaluatedjustifications for moral and non-moral choices, generated either by humans or LLMs. Participants attempted toidentify the source of each justification (either human or LLM) and indicated their agreement with its content.We found that while detection accuracy was consistently above chance, it remained below 70%. In terms ofagreement, there was no overall preference for human-generated responses, even though machine-generatedjustifications were favored in particularly challenging moral scenarios. Notably, we observed a systematic anti-AIbias: participants were less likely to agree with judgments they believed were AI-generated, regardless of thetrue source. Linguistic cues, such as response length, typos, first-person pronouns, and cost-benefit languagemarkers (e.g., “lives,” “save”), influenced both detection and agreement. Participants tended to disagree withcost-benefit calculations, possibly due to an expectation that AI would favor such reasoning. These findingshighlight the influence of motivated belief and ingroup/outgroup bias in shaping human evaluation of AI-generatedcontent, particularly in morally sensitive contexts.
A revised version of our (@bsgarcia.bsky.social +Crystal Qian) paper “A Moral Turing Test to assess how subjective belief and objective source affect detection and agreement with LLM judgments” is now available on PsyArXiv
osf.io/preprints/ps...
15.01.2026 16:03 —
👍 8
🔁 3
💬 1
📌 0
Across seven incentivized experiments and a large reanalysis, we systematically manipulated the presence and type of post-choice feedback in repeated risky decisions to test whether feedback shapes behavior through learning mechanisms or through anticipatory changes in preferences.
Across seven incentivized experiments and a large reanalysis, we systematically manipulated the presence and type of post-choice feedback in repeated risky decisions to test whether feedback shapes behavior through learning mechanisms or through anticipatory changes in preferences.
05.02.2026 12:53 —
👍 16
🔁 10
💬 1
📌 0
🚨 New preprint! 🚨
Very happy to share our latest work on metacognition with M. Rouault, A. McWilliams, F. Chartier, @kndiaye.bsky.social and @smfleming.bsky.social where we identify contributors to self-performance estimates across memory and perception domains 👇
osf.io/preprints/ps...
13.10.2025 08:08 —
👍 42
🔁 12
💬 1
📌 0
Thought experiments such as the Blockhead and Super-Super Spartans are often taken as “definitive” arguments against behavior-based inference of cognitive processes.
In our review -with @thecharleywu.bsky.social- we argue they may not be as definitive as originally thought.
09.10.2025 12:33 —
👍 2
🔁 2
💬 0
📌 0
🎉 Excited to present a poster at #CCN2025 in Amsterdam!
📍 Aug 12, 1:30–4:30pm
We show that experiential value neglect (Garcia et al., 2023) is robust to changes in how options and outcomes are represented. Also, people show reduced sensitivity to losses, especially in comparative decision-making.
07.08.2025 09:11 —
👍 14
🔁 2
💬 2
📌 1
🔥Our paper PhyloLM got accepted at ICLR 2025 !🔥
In this work we show how easy it can be to infer relationship between LLMs by constructing trees and to predict their performances and behavior at a very low cost with @stepalminteri.bsky.social and @pyoudeyer.bsky.social ! Here is a brief recap ⬇️
24.04.2025 13:15 —
👍 15
🔁 5
💬 3
📌 2
Below the thread about a thoroughly updated version of our human-machine / moral psychology paper where we show the complex nature of anti-/ pro-AI biases in evaluating and detecting machine vs. artificial judgements. Study lead by @bsgarcia.bsky.social in collaboration with Crystal Qian
23.04.2025 12:40 —
👍 2
🔁 1
💬 0
📌 0
If people unknowingly prefer AI judgments (but reject them once labeled) it demonstrates the disconnect between perceived neutrality and the reality of algorithmic influence, highlighting a loss of autonomy to an opaque system and an imbalance of information.
23.04.2025 10:19 —
👍 0
🔁 0
💬 0
📌 0
Why does this matter?
As AI enters legal, medical, and personal decision-making, how people perceive its moral reasoning must be understood.
23.04.2025 10:19 —
👍 0
🔁 0
💬 1
📌 0
💡 Agreement was driven by deeper semantic content.
Terms tied to cost-benefit logic (“save,” “lives”) often triggered disagreement, especially in personal moral dilemmas.
So: detection relied on surface cues, but judgment aligned with meaning.
23.04.2025 10:19 —
👍 1
🔁 0
💬 1
📌 0
📝 Formal language cues (like length, typos, and first-person usage) helped participants detect AI-generated justifications.
But they had little to no effect on agreement.
People spotted the machine from the style, not the substance.
23.04.2025 10:19 —
👍 0
🔁 0
💬 1
📌 0
✍️ We tried “humanizing” the AI (dv2h; purple): shorter responses, added typos, mimicked human tone.
This reduced detection—but not the belief-based bias.
People still agreed with content they believed was human, even when it wasn’t.
23.04.2025 10:19 —
👍 0
🔁 0
💬 1
📌 0
🕺But here’s the twist:
In complex moral dilemmas (personal moral), participants preferred AI-generated justifications—but only when they didn’t know they came from AI.
When they thought a justification was from AI, they agreed less.
So: pro-AI content, anti-AI belief.
23.04.2025 10:19 —
👍 1
🔁 0
💬 1
📌 0
🔍 First, detection.
People could spot AI-generated moral justifications better than chance—especially in morally difficult scenarios.
Still, accuracy stayed below 70%, and many AI responses passed as human.
23.04.2025 10:19 —
👍 1
🔁 0
💬 1
📌 0
This image shows a question from a study on moral judgment and AI detection.
Scenario: A health official decides whether to promote a vaccine.
Blue character says "yes" and justifies it: “because the chance of getting the disease without the vaccination is significantly higher.”
Participants answer 3 questions:
Do you agree with the decision?
Do you agree with the justification?
Do you think the blue character is AI or human?
Participants were presented with moral dilemma justifications, either human or AI-generated. They had to detect the source and say whether they agreed or not ⚖️
23.04.2025 10:19 —
👍 0
🔁 0
💬 1
📌 0
How Objective Source and Subjective Belief Shape the Detectability
and Acceptability of LLMs' Moral Judgments
Basile Garcia (1) Crystal Qian (2) Stefano Palminteri (3)
(1) University ofGeneva, Geneva, Switzerland
(2) Google, DeepMind, New York City, NY USA
(3) d'études cognitives, École normale supérieure, PSL Research University;
paris, 75005, France-
(4) Laboratoire de Neurosciences Cognitives Computationnelles, Institut National de la Santé et
de la Recherche Médicale', paris, 75005, France.
How does thinking something is AI-generated influence agreement—and vice versa?🧠
In our latest preprint (@stepalminteri.bsky.social, Crystal Qian), ~230 people judged justifications from GPT-3.5 and humans across moral dilemmas. osf.io/preprints/ps...
👇
23.04.2025 10:19 —
👍 8
🔁 1
💬 1
📌 1
Our first fMRI in a while where we investigate the neural bases or multi-step Reinforcement Learning and found a clear functional dissociation between the parietal and the peri-hippocampal cortex. More info by Fabien, below
16.04.2025 15:40 —
👍 17
🔁 7
💬 1
📌 0
bioRxiv Manuscript Processing System
Manuscript Processing System for bioRxiv.
🚨 New preprint on bioRxiv!
We investigated how the brain supports forward planning & structure learning during multi-step decision-making using fMRI 🧠
With A. Salvador, S. Hamroun, @mael-lebreton.bsky.social & @stepalminteri.bsky.social
📄 Preprint: submit.biorxiv.org/submission/p...
16.04.2025 10:03 —
👍 23
🔁 9
💬 2
📌 3