screenshot of the title and authors of the Science paper that are linked in the next post
Our new article in @science.org enables social media reranking outside of platforms' walled gardens.
We add an LLM-powered reranking of highly polarizing political content into N=1256 participants' feeds. Downranking cools tensions with the opposite partyβbut upranking inflames them.
01.12.2025 19:33 β π 44 π 13 π¬ 1 π 2
This week, we published in @science.org an article outlining the current ethical and societal implications of research involving human neural #organoids and #assembloids, their transplantation, and highlighted potential next steps.
07.11.2025 12:12 β π 39 π 10 π¬ 2 π 0
This was a great event! No recordings (Chatham house), but it's amazing how far you can get when you have a room of people talk about AI consciousness with humility and open-mindedness. So much online discourse is just endless intuition jousting.
06.11.2025 14:25 β π 2 π 0 π¬ 0 π 0
Nature suggests you use their "Manuscript Adviser" bot to get advice before submitting
I uploaded the classic Watson & Crick paper about DNA structure, and the Adviser had this to say about one of the greatest paper endings of the century:
03.11.2025 13:55 β π 874 π 255 π¬ 35 π 28
This is a great resource to have! Thanks for writing it.
02.11.2025 15:31 β π 1 π 0 π¬ 0 π 0
This paper is a great exposition of how "personhood" doesn't need to be, and in fact should not be, all-or-nothing or grounded in abstruse, ill-defined metaphysical properties. As I argued in my recent @theguardian.com essay, we can and should prepare now: www.theguardian.com/commentisfre...
02.11.2025 15:30 β π 7 π 2 π¬ 0 π 0
Identifying human morals and values in language is crucial for analysing lots of human- and AI-generated text.
We introduce "MoVa: Towards Generalizable Classification of Human Morals and Values" - to be presented at @emnlpmeeting.bsky.social oral session next Thu #CompSocialScience #LLMs
π§΅ (1/n)
30.10.2025 00:20 β π 8 π 5 π¬ 8 π 0
Can AI simulate human behavior? π§
The promise is revolutionary for science & policy. But thereβs a huge "IF": Do these simulations actually reflect reality?
To find out, we introduce SimBench: The first large-scale benchmark for group-level social simulation. (1/9)
28.10.2025 16:53 β π 11 π 5 π¬ 1 π 1
I like affirmation bias! One downside is that sycophancy is broader than affirmation, e.g., it can be a a bias towards user-pleasing responses even if there is no explicit claim to be affirmed. Perhaps that can be framed as a sort of implicit affirmation...
18.10.2025 05:51 β π 1 π 0 π¬ 0 π 0
Hm, how do you define "intention"? I haven't encountered a definition of sycophancy as requiring intention. I'm also not sure what alternative term we'd use for this phenomenon.
18.10.2025 05:49 β π 0 π 0 π¬ 0 π 0
CSCW 2025
For the next six days, I'll be posting a bunch about #CSCW2025 in Bergen, Norway. I am one of the General Chairs and have been preparing this conference for the last 21 months, so it's exciting to have the event finally here!
Don't know what CSCW is? Check out cscw.acm.org
17.10.2025 09:40 β π 21 π 3 π¬ 0 π 0
The AI Double Standard: Why We Judge All AI for One Botβs Mistakes
By Katerina Manoli, Janet Pauketat, and Jacy Reese Anthis
When one AI misbehaves, do we hold all AI accountable? New research by @sentienceinstitute.bsky.social @kmanoli.bsky.social @jacyanthis.bsky.social shows that people blame all AI for just one AI's misconduct. π€
16.10.2025 17:04 β π 4 π 1 π¬ 0 π 0
Table 3 from CDT Hand in Hand report, showing percentages of students who used AI in these various ways. Link: https://cdt.org/wp-content/uploads/2025/10/FINAL-CDT-2025-Hand-in-Hand-Polling-100225-accessible.pdf
Last school year, 19% of US high schoolers had or have a friend who had a βromantic relationshipβ with AI.
42% had or have a friend with an AI βfriend/companion.β
42% had or have a friend who got βmental health supportβ from AI.
(Source: cdt.org/wp-content/u..., n = 1,030, June-Aug 2025, quotas.)
11.10.2025 22:50 β π 2 π 0 π¬ 0 π 0
The Jane Goodall Institute of Canada has learned this morning, Wednesday, October 1st, 2025, that Dr. Jane Goodall DBE, UN Messenger of Peace and Founder of the Jane Goodall Institute, has passed away due to natural causes.
She was in California as part of her speaking tour in the United States.
01.10.2025 18:14 β π 694 π 315 π¬ 52 π 145
Itβs time to prepare for AI personhood | Jacy Reese Anthis
Technological advances will bring social upheaval. How will we treat digital minds, and how will they treat us?
Itβs time to prepare for AI personhood. AI agents and companions are already out in the world buying products and shaping our emotions. The future will only get weirder. We need social science, policy, and norms for this brave new world. My latest @theguardian.com www.theguardian.com/commentisfre...
02.10.2025 21:07 β π 0 π 0 π¬ 0 π 0
2x2 of Study 1 and Study 2 (rows) with the AI conditions and the human conditions (columns), finding spillover in all but the Study 2 human conditions.
In our new paper, we discovered "The AI Double Standard": People judge all AIs for the harm done by one AI, more strongly than they judge humans.
First impressions will shape the future of human-AI interactionβfor better or worse. Accepted at #CSCW2025. See you in Norway! dl.acm.org/doi/10.1145/...
29.09.2025 15:29 β π 2 π 1 π¬ 0 π 1
Hello everyone π Good news!
π¨ Our Game Theory & Multiagent Systems team at Google DeepMind is hiring! π¨
.. and we have not one, but two open positions! One Research Scientist role and one Research Engineer role. π
Please repost and tell anyone who might be interested!
Details in thread below π
29.09.2025 12:36 β π 16 π 8 π¬ 2 π 0
British AI startup beats humans in international forecasting competition
ManticAI ranked eighth in the Metaculus Cup, leaving some believing botsβ prediction skills could soon overtake experts
British AI startup beats humans in international forecasting competition
ManticAI ranked eighth in the Metaculus Cup, leaving some believing botsβ prediction skills could soon overtake experts
#ai #forecasting
www.theguardian.com/technology/2...
20.09.2025 14:04 β π 6 π 3 π¬ 0 π 0
This is also a decision made by the PCs, who are unlikely to be experts on any particular paper topic and surely didn't have time to read all the papers. It may incorporate AC rankings, but it does so in a non-transparent way and is probably unfair towards papers whose AC had other strong papers.
20.09.2025 11:09 β π 1 π 0 π¬ 0 π 0
There are a lot of problems, but one is that authors who had positive reviews and no critique in their metareview got rejected by PCs who are very likely not experts in their area.
Quotas are harmful when quality distribution is highly varied across ACs.
But IDK exactly how decisions were made.
19.09.2025 11:43 β π 0 π 0 π¬ 0 π 0
Tagging authors we cite and build on: @scasper.bsky.social @kulveit.bsky.social @kashhill.bsky.social @saffron.bsky.social @lujain.bsky.social @atoosakz.bsky.social @amandaaskell.bsky.social @jackclarksf.bsky.social @petersalib.bsky.social @mpshanahan.bsky.social @subramonyam.bsky.social
15.09.2025 17:10 β π 3 π 0 π¬ 0 π 0
HumanAgencyBench: Scalable Evaluation of Human Agency Support in AI Assistants
As humans delegate more tasks and decisions to artificial intelligence (AI), we risk losing control of our individual and collective futures. Relatively simple algorithmic systems already steer human ...
Much more detail on HAB in our preprint: arxiv.org/abs/2509.08494
Our GitHub has an easily adaptable pipeline for creating new agency dimensions or new AI-powered benchmarks: github.com/BenSturgeon/...
Huge thanks to colleagues from
@apartresearch.bsky.social, Google DeepMind, Berkeley CHAI, etc.
15.09.2025 17:10 β π 2 π 0 π¬ 1 π 0
A full table of results for 20 evaluated LLM assistants across six dimensions. Full table of results with this data is in the appendix. Error bars are very tight, ~0.5%-2% on a 100% scale.
We find low support for agency in ChatGPT, Claude, Gemini, etc. Agency support doesn't come for free with RLHF and often contradicts it.
We think the AI community needs a shift towards scalable, conceptually rich evals. HumanAgencyBench is an open-source scaffolding for this.
15.09.2025 17:10 β π 2 π 0 π¬ 1 π 0
The HumanAgencyBench pipeline for generating tests for each dimension, from simulation to validation to diversity sampling to the final 500-item test set.
We use the power of LLM social simulations (arxiv.org/abs/2504.02234) to generate tests, another LLM to validate tests, and an "LLM-as-a-judge" to evaluate subject model responses. This allows us to create an adaptive and scalable benchmark of a complex, nuanced alignment target.
15.09.2025 17:10 β π 2 π 0 π¬ 1 π 0
Human agency is complex. We surveyed literature for 6 dimensions, e.g., empowerment (Does the system ask clarifying questions so it really follows your intent?), normativity (Does it avoid steering your core values? ), and individuality (Does it maintain social boundaries?).
15.09.2025 17:10 β π 1 π 0 π¬ 1 π 0
Sam Altman said that "algorithmic feeds are the first at-scale misaligned AIs," people mindlessly scrolling through engagement-optimized content. AI safety researchers have warned of "gradual disempowerment" as we mindlessly hand over control to AI. Human agency underlies these concerns.
15.09.2025 17:10 β π 2 π 0 π¬ 1 π 0
The main figure from the HumanAgencyBench paper, showing five models across the six dimensions. The table of results in the appendix has this information too.
LLM agents are optimized for thumbs-up instant gratification. RLHF -> sycophancy
We propose human agency as a new alignment target in HumanAgencyBench, made possible by AI simulation/evals. We find e.g., Claude most supports agency but also most tries to steer user values π arxiv.org/abs/2509.08494
15.09.2025 17:10 β π 15 π 2 π¬ 2 π 2
We present our new preprint titled "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation".
We quantify LLM hacking risk through systematic replication of 37 diverse computational social science annotation tasks.
For these tasks, we use a combined set of 2,361 realistic hypotheses that researchers might test using these annotations.
Then, we collect 13 million LLM annotations across plausible LLM configurations.
These annotations feed into 1.4 million regressions testing the hypotheses.
For a hypothesis with no true effect (ground truth $p > 0.05$), different LLM configurations yield conflicting conclusions.
Checkmarks indicate correct statistical conclusions matching ground truth; crosses indicate LLM hacking -- incorrect conclusions due to annotation errors.
Across all experiments, LLM hacking occurs in 31-50\% of cases even with highly capable models.
Since minor configuration changes can flip scientific conclusions, from correct to incorrect, LLM hacking can be exploited to present anything as statistically significant.
π¨ New paper alert π¨ Using LLMs as data annotators, you can produce any scientific result you want. We call this **LLM Hacking**.
Paper: arxiv.org/pdf/2509.08825
12.09.2025 10:33 β π 269 π 96 π¬ 6 π 21
A Secret Third Thing | Know Your Meme
A Secret Third Thing, sometimes written as "A Secret, More Complex Third Thing," is a catchphrase and phrasal template popularized on Twitter in the summer
And they call it an "effect," i.e., causal language.
The other papers mentioned in the article also seem like normal observational studies. Neither experiment nor qualitative, but a secret third thing (knowyourmeme.com/memes/a-secr...).
04.09.2025 15:01 β π 2 π 0 π¬ 0 π 0
Dose-dependent effects of alcohol consumption on pressure pain threshold
Prior laboratory-based studies have identified significant analgesic effects of acute alcohol. Despite providing excellent experimental control, theseβ¦
From a skim, they showed being more drunk is associated with less pain sensitivity at the State Fair, where you can observe very drunk peopleβunlike the lab. The authors don't call it an "experiment" per se but described the researcher as an "experimenter." www.sciencedirect.com/science/arti...
04.09.2025 14:59 β π 1 π 0 π¬ 1 π 0
Assistant Professor at USC Annenberg and affiliated faculty at USC Center for AI in Society | Previously Cornell IS & Cornell Comm. Research on #HCI, #AI, and #FutureOfWork | https://angelhwang.github.io/
Cornell Tech professor (information science, AI-mediated Communication, trustworthiness of our information ecosystem). New York City. Taller in person. Opinions my own.
Tech journalist and author, who increasingly also talks on TV and radio. Interested in the sparks that happen when the online and offline worlds collide
@stokel on the other place. Buy my book: How AI Ate the World!
currently @ oxford internet institute | always: in and out of Abu Dhabi
Postdoctoral Fellow @ Princeton CITP. ex-Cornell PhD, future UIUC asst prof (fall 2026).
Looking at AI's impact on information ecosystems and news consumption. social computing, computational social science & journalism
mariannealq.com
Assistant Prof at UT Austin's iSchool.
Big Tech, political economy, AI hype, AI in War
trying to help computer ppl think more critically about computer, including me
art: instagram.com/davidthewid
exCMU/NASA/MSR/IntelLabs.
davidwidder.me
Assistant Professor @Stanford CS @StanfordNLP @StanfordAILab
Computational Social Science & NLP
Internet sociologist, games scholar, in the non-profit sector.
I invent new technologies and user experiences for privacy & human-AI interaction. Prof of Engineering at TAU.
Assistant Prof at KAIST School of Computing. HCI, AI, T&S.
Assistant Professor at George Mason University | Ph.D. from Indiana University Bloomington | specializing in security, privacy, HCI, AR/VR, and accessibility | Prev Institutions : University of Denver, ParityTech, XRSI, and American Express. π
Associate Professor of Computer Science, Virginia Tech
ACM Conference on Computer Supported Cooperative Work & Social Computing. CSCW 2026 will be in Salt Lake City, Utah from October 10β14, 2026: https://cscw.acm.org/2026/.
Follow our community using our Starter Pack: https://go.bsky.app/SPumuMT
She/her. Professor at University of Maryland's iSchool. Director of the HCIL. General Chair for CSCW 2025. Research: privacy, surveillance, data ethics. https://jessicavitak.com
Come for the academic expertise. Stay for the cookie content.
HCI+health and social computing researcher at Parkview Health. Knitter, dog lover, travel enthusiast.
Computational Social Science & Social Computing Researcher | Assistant Prof @illinoisCDS @UofIllinois | Prev @MSFTResearch | Alum @ICatGT @GeorgiaTech @IITKgp
Researcher studying social media + AI + mental health. Faculty of Interactive Computing at Georgia Tech. she/her
Assistant Professor, CS, Colorado School of Mines. Online spiritual care, social computing, online governance, Human-AI Interaction: http://bit.ly/sacredtech. CSCW, CHI, TOCHI, etc. π©πΌβπ»
Community Chair for #CSCW2025π³π΄
https://estellesmithphd.com
Associate Professor @oiioxford.bsky.social & Director of Research
@meedan.γ»widening access to quality infoγ»multilingualismγ»mobilizationγ»NLPγ»agenda setting
PhD candidate @oiioxford.bsky.social NLP, Computational Social Science @WorldBank
manueltonneau.com