Also sharing a beautiful illustration of these ideas by my lovely and talented π©βπ¨ friend, Adhara Martellini!
08.08.2025 14:57 β π 4 π 1 π¬ 0 π 0@setayeshradkani.bsky.social
PhD candidate in Brain and Cognitive Sciences @MIT studying legitimacy, punishment and social learning
Also sharing a beautiful illustration of these ideas by my lovely and talented π©βπ¨ friend, Adhara Martellini!
08.08.2025 14:57 β π 4 π 1 π¬ 0 π 0Bottom line: The same punishment can teach different lessons to different people depending on their prior beliefs, even when everyone is reasoning rationally. So, even well-meant punishment can widen divides or fuel polarization.
10/10
Finding 4οΈβ£: In a separate study, we found that repeated punishments can fail to close societal divides, and may even polarize initially shared beliefs. Our model predicts when punishment works in reducing polarization and when it backfires!
9/N
Finding 3οΈβ£: Our computational model simultaneously captures peopleβs belief updates about the act and the authority, even in novel prior conditions that the model has never seen before π; even better than control models that are fit to predict each belief separately!
8/N
Finding 2οΈβ£: Peopleβs prior beliefs shape their reasoning about punishment. The same punishment can lead to contrasting inferences, depending on the value and uncertainty of prior beliefs about both the act and the authority.
7/N
Finding 1οΈβ£: From observing punishment (or no punishment!), people simultaneously update their beliefs about the wrongness of the target act, and about the authorityβs motivations and values.
6/N
Across 3 studies, we used imaginary villages to experimentally control peopleβs pre-existing beliefs about the target act and the authority. We then measured how observing punishment with different severities moves their beliefs.
5/N
We used an inverse planning framework: people assume authorities plan punishment to achieve their desires based on their beliefs. By inverting this model, people infer the hidden beliefs and desires that most likely produced the observed punishment.
4/N
We solved this puzzle by building a computational model that characterizes βhowβ people interpret punishment.
Key insight: People use their pre-existing beliefs and opinions to simultaneously evaluate both the norm to be learned and the authority whoβs punishing.
But how?!
3/N
This is a puzzle in life and in punishment literature!
Even when nobody disagrees about the factsβeverybody knows what action happened, who punished it, and what they did to punish itβdifferent observers of the same punishment could come to drastically different conclusions.
2/N
π¨Out in PNASπ¨
with @joshtenenbaum.bsky.social & @rebeccasaxe.bsky.social
Punishment, even when intended to teach norms and change minds for the good, may backfire.
Our computational cognitive model explains why!
Paper: tinyurl.com/yc7fs4x7
News: tinyurl.com/3h3446wu
π§΅
1. This Friday I hosted a workshop on morality with some fabulous humans who took take time out of their lives to talk about ideas with each other. Here is a thread with brief summaries of what folks presented in case you too would like to hear about their science:
#PsychSciSky #SocialPsyc #DevPsyc
The Saxe Lab @ MIT is hiring! We seek one lab manager to start in summer 2025. Research in our lab focuses on social cognition (learn more on saxelab.mit.edu).
Please apply at: tinyurl.com/saxe2025 (Job ID 31993).
Review of applications starts on March 24, 2025.
Sharing appreciated. Thank you!
Does poverty lead to risk taking or risk avoidance? Turns out, to both. Our new paper (with D. Nettle & W. Frankenhuis) in @royalsocietypublishing.org explains why, and conducts preregistered tests of our βdesperation thresholdβ model.
royalsocietypublishing.org/doi/10.1098/...
A π§΅
Thanks for this! Could I please be added as well?
11.11.2024 21:09 β π 0 π 0 π¬ 1 π 0Huge thanks to @dgrand.bsky.social , @falklab.bsky.social and @anthlittle.bsky.social for their feedback on this work, and to our funding sources!
13/13
We need these authorities to succeed in being seen as independent and truthful, because in this space of uncertainty, those are the voices that can move people toward an accurate outcome
12/N
Given the current π state of things, this work offers insight into the challenges faced by independent election observers, public health professionals, and others seeking to cultivate credibility as debunkers
11/N
Finding 3: Differing beliefs about authorities can spread polarization
When beliefs about the authority diverge in the original domain (election fraud), their debunking can polarize the two groups in a new domain (e.g., public health) even when the groups initially share the same perspective
10/N
Finding 2: Credibility isnβt just about unbiasedness
Consistent with other work in political science, authorities seen as biased can still be effective debunkers β if they are seen as biased *in favor* of the perspective they are debunking, although this depends on the uncertainty of beliefs
9/N
if people are certain, debunking β even by a reasonably unbiased, committed authority β fails:
1) The groups' beliefs about the perspective remain polarized
2) The groups' initially shared beliefs about the authority additionally diverge!
8/N
Finding 1: Debunking can workβ¦sometimes
If initial beliefs in the perspective (βthe election was unfairβ) are held with some uncertainty, then groups can eventually converge on shared beliefs, if they also believe the authority is unbiased and committed to the truth. But β¦
7/N
Using simulations, we capture how three related beliefs evolve: (1) commitment to the initial perspective; (2) views on the bias of the authority; (3) views on the authorityβs commitment to the truth
6/N
We use this inverse-planning framework [https://pubmed.ncbi.nlm.nih.gov/19729154/] to develop a computational model of how observersβ beliefs evolve as each group witnesses 5 debunking acts that counter a single perspective on a topic (e.g., βthe election was unfairβ)
5/N
We make minimal assumptions that both groups:
- Are made up of equally rational Bayesians
- Differ in their belief on a topic (ground truth may not be known)
- Hold shared beliefs about the authorityβs motives with some uncertainty
(see the paper for a discussion of why these assumptions)
4/N
We focus instead on how observers interpret debunking acts (and update their beliefs) based on their intuitive theory of the authorityβs motives: degree of bias, commitment to the truth
3/N
Lots of great work on misinformation in politics, health, etc. focuses on the content and delivery method of corrections, and on the role of biases, identities, etc. in sustaining polarized beliefs
2/N
New work in PNAS Nexus with @mlandauwells.bsky.social and @rebeccasaxe.bsky.social
We ask: if two groups hold opposing beliefs, when does debunking by an authority lead to belief convergence and when does polarization persist?
Paper: tinyurl.com/mr2e4kvx
MIT News: tinyurl.com/3bud9z4f
π§΅1/N