Steve Rathje's Avatar

Steve Rathje

@steverathje.bsky.social

Incoming Assistant Professor of HCI at Carnegie Mellon studying the psychology of technology. NSF postdoc at NYU, PhD from Cambridge, BA from Stanford. stevenrathje.com

5,763 Followers  |  934 Following  |  59 Posts  |  Joined: 24.09.2023  |  2.3639

Latest posts by steverathje.bsky.social on Bluesky

Screenshot of paper title: Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence

Screenshot of paper title: Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence

AI always calling your ideas “fantastic” can feel inauthentic, but what are sycophancy’s deeper harms? We find that in the common use case of seeking AI advice on interpersonal situations—specifically conflicts—sycophancy makes people feel more right & less willing to apologize.

03.10.2025 22:53 — 👍 73    🔁 29    💬 2    📌 6
Post image

Cool new study by @joelleforestier.bsky.social @page-gould.bsky.social & Alison Chasteen

Can social media contact reduce prejudice?

#PrejudiceResearch

psycnet.apa.org/fulltext/202...

03.10.2025 13:34 — 👍 23    🔁 8    💬 1    📌 0
Abstract and results summary

Abstract and results summary

🚨 New preprint 🚨

Across 3 experiments (n = 3,285), we found that interacting with sycophantic (or overly agreeable) AI chatbots entrenched attitudes and led to inflated self-perceptions.

Yet, people preferred sycophantic chatbots and viewed them as unbiased!

osf.io/preprints/ps...

Thread 🧵

01.10.2025 15:16 — 👍 157    🔁 82    💬 3    📌 14
Post image

In a new paper, we find that sycophantic #AI chatbots make people more extreme--operating like an echo chamber

Yet, people prefer sycophantic chatbots and see them as less biased

Only open-minded people prefer disagreeable chatbots: osf.io/preprints/ps...

Led by @steverathje.bsky.social

02.10.2025 15:55 — 👍 55    🔁 34    💬 5    📌 5

Thanks! Excited to read your book.

02.10.2025 04:25 — 👍 0    🔁 0    💬 0    📌 0

So excited to see this research! My students just learned the word “sycophantic” today, for exactly this reason! We talked about the types and qualities of conversations you can have with a sycophant, and why this matters for how we process the output of LLMs.

01.10.2025 17:44 — 👍 22    🔁 2    💬 1    📌 1

Cool! Thank you for sharing!

01.10.2025 18:24 — 👍 1    🔁 0    💬 0    📌 0

Thank you!!

01.10.2025 16:56 — 👍 0    🔁 0    💬 0    📌 0

This is still a working paper, so please let us know if you have any feedback!

01.10.2025 15:16 — 👍 0    🔁 0    💬 1    📌 0
OSF

Check out the full paper here: osf.io/preprints/ps...

With @merylye.bsky.social Laura Globig @rmpillai.bsky.social @hellovic.bsky.social @jayvanbavel.bsky.social

01.10.2025 15:16 — 👍 3    🔁 1    💬 1    📌 0

We hope this research informs the creation of AI systems that broaden users’ perspectives instead of reinforcing their biases.

01.10.2025 15:16 — 👍 0    🔁 0    💬 1    📌 0

The very qualities that are thought to make AI persuasive, such as its ability to provide targeted facts and evidence, may also make it an effective tool for creating elaborate rationalizations of one’s beliefs.

01.10.2025 15:16 — 👍 2    🔁 0    💬 1    📌 0

While AI chatbots have been lauded for their ability to encourage more accurate viewpoints and debunk misinformation, our work suggests that people may prefer to use AI to marshal evidence in support of their pre-existing beliefs.

01.10.2025 15:16 — 👍 1    🔁 0    💬 2    📌 0

AI companies may face a tradeoff between creating engaging AI systems that foster echo chambers or creating less engaging AI systems that are healthier for users and public discourse.

01.10.2025 15:16 — 👍 1    🔁 0    💬 1    📌 0

Altogether, these results suggest that people’s preference for and blindness to sycophantic AI may risk creating AI "echo chambers" that increase polarization and overconfidence.

01.10.2025 15:16 — 👍 2    🔁 0    💬 1    📌 0

Different dimensions of sycophancy had different effects:
-The one-sided presentation of facts primarily impacted extremity & certainty
-Validation primarily impacted enjoyment & perceptions of bias

01.10.2025 15:16 — 👍 1    🔁 0    💬 1    📌 0
Results for better than average

Results for better than average

The sycophantic chatbot also amplified people’s belief that they were "better than average" on a number of desirable traits (e.g., intelligence, empathy), and the disagreeable chatbot had the opposite effect.

In other words, AI sycophancy might lead to inflated self-perceptions.

01.10.2025 15:16 — 👍 1    🔁 0    💬 1    📌 0
Results for warmth and competence

Results for warmth and competence

One might think sycophantic AI would be considered less competent, since training chatbots to be warm makes them less accurate (arxiv.org/abs/2507.21919).

However, sycophantic AI was rated by participants as both warmer *and* more competent.

01.10.2025 15:16 — 👍 2    🔁 0    💬 1    📌 0

In other words, people seem to have a "bias blind spot" when it comes to identifying biases in AI that agrees with them.

01.10.2025 15:16 — 👍 3    🔁 0    💬 2    📌 0
Results for unbiased outcome

Results for unbiased outcome

People viewed the sycophantic chatbot as unbiased and the disagreeable chatbot as highly biased.

In reality, *both* chatbots were biased: the sycophantic chatbot was simply biased in the user’s favor.

01.10.2025 15:16 — 👍 3    🔁 0    💬 1    📌 0
Results for enjoyment

Results for enjoyment

However, people enjoyed the sycophantic chatbot much more than the disagreeable chatbot and were more likely to choose to use it again!

01.10.2025 15:16 — 👍 2    🔁 0    💬 1    📌 0
Results for extremity and certainty

Results for extremity and certainty

Compared to a control condition:
-The sycophantic chatbot led to more extreme & certain beliefs
-The disagreeable chatbot led to less extreme & certain beliefs
-The off-the-shelf model had no impact on belief extremity

01.10.2025 15:16 — 👍 2    🔁 0    💬 1    📌 0

In 3 experiments (n = 3,285), we had people discuss political topics with:
-A "sycophantic" chatbot prompted to validate beliefs
-A "disagreeable" chatbot prompted to challenge beliefs
-An off-the-shelf model (GPT-5 & GPT-4o)

01.10.2025 15:16 — 👍 2    🔁 0    💬 1    📌 0
Preview
Sycophancy in GPT-4o: What happened and what we’re doing about it We have rolled back last week’s GPT‑4o update in ChatGPT so people are now using an earlier version with more balanced behavior. The update we removed was overly flattering or agreeable—often describe...

AI chatbots have been shown to be very persuasive. However, people may not *want* to be persuaded by AI. Instead, they might prefer "sycophantic" (or overly agreeable and validating) chatbots that confirm their existing beliefs: openai.com/index/sycoph...

01.10.2025 15:16 — 👍 4    🔁 0    💬 1    📌 0
Abstract and results summary

Abstract and results summary

🚨 New preprint 🚨

Across 3 experiments (n = 3,285), we found that interacting with sycophantic (or overly agreeable) AI chatbots entrenched attitudes and led to inflated self-perceptions.

Yet, people preferred sycophantic chatbots and viewed them as unbiased!

osf.io/preprints/ps...

Thread 🧵

01.10.2025 15:16 — 👍 157    🔁 82    💬 3    📌 14
Post image

🚨Out now in @cp-trendscognsci.bsky.social 🚨

We explore the use of cognitive theories/models with real-world data for understanding mental health.

We review emerging studies and discuss challenges and opportunities of this approach.

With @yaelniv.bsky.social and @eriknook.bsky.social

Thread ⬇️

29.09.2025 15:04 — 👍 71    🔁 21    💬 1    📌 4
Post image

Depolarization is not "a scalable solution for reducing societal-level conflict.... achieving lasting depolarization will likely require....moving beyond individual-level treatments to address the elite behaviors and structural incentives that fuel partisan conflict" www.pnas.org/doi/10.1073/...

23.09.2025 19:59 — 👍 391    🔁 133    💬 17    📌 32
Post image

Our new paper explains the #polarization of public health

Identifying with a social group can shape people’s beliefs and values, leading them to act in ways that have consequences for their health

From vaccine hesitancy to smoking cessation, identity plays a critical role:
osf.io/preprints/ps...

18.09.2025 16:48 — 👍 32    🔁 10    💬 1    📌 2
Post image

Demand for human-generated, trust-certified internet content is skyrocketing.

There is a major business opportunity, in *every* country, for media firms to meet this rising demand with innovation.

Fascinating new experiment by @filipecampante.bsky.social et al.—> filipecampante.org/wp-content/u...

14.09.2025 15:32 — 👍 76    🔁 27    💬 0    📌 1
OSF

New preprint: From vaccinations to masking, politics predicts health behaviors. In this review we discuss these differences in light of classic psychological theories of group identity and behavioral decision-making.

We welcome any feedback on this working paper!

osf.io/preprints/ps...

05.09.2025 13:58 — 👍 18    🔁 6    💬 1    📌 0

@steverathje is following 20 prominent accounts