Finally we do test it empirically: finding some models where the embedding matrix of the LLM already provides decently interpretable nearest neighbors
But this was not the full story yet...
@mariusmosbach.bsky.social and @elinorpd.bsky.social nudged me to use contextual embeddings
11.02.2026 15:10 โ ๐ 1 ๐ 1 ๐ฌ 1 ๐ 0
Really cool new work with surprising results! Highly recommend checking out the demo ๐
11.02.2026 15:20 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0
Grok fact-checks our paper on Grok fact-checking - and it approves!
04.02.2026 13:49 โ ๐ 28 ๐ 7 ๐ฌ 1 ๐ 0
๐ญ How do LLMs (mis)represent culture?
๐งฎ How often?
๐ง Misrepresentations = missing knowledge? spoiler: NO!
At #CHI2026 we are bringing โจTALESโจ a participatory evaluation of cultural (mis)reps & knowledge in multilingual LLM-stories for India
๐ arxiv.org/abs/2511.21322
1/10
02.02.2026 21:38 โ ๐ 42 ๐ 21 ๐ฌ 1 ๐ 2
this is amazing! made quick NYC & boston posters
30.01.2026 21:05 โ ๐ 3 ๐ 0 ๐ฌ 0 ๐ 0
Potato is a great platform for researchers! Highly recommend (plus a great development team behind it)
30.01.2026 15:41 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
Microsoft Research NYC is hiringย a researcher in the space of AI and society!
29.01.2026 23:27 โ ๐ 62 ๐ 39 ๐ฌ 1 ๐ 2
Iโve had a similar experience except with knitting / crocheting!
29.01.2026 18:21 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0
Federal agents with weapons drawn, moments before murdering American citizens on the streets of Minneapolis at the dawn of 2026.
What should academics be doing right now?
I have been writing up some thoughts on what the research says about effective action, and what universities specifically can do.
davidbau.github.io/poetsandnurs...
It's on GitHub. Suggestions and pull requests welcome.
github.com/davidbau/poe...
26.01.2026 03:27 โ ๐ 37 ๐ 16 ๐ฌ 0 ๐ 4
Whoa! Thatโs a nice view! Orโฆ well, Iโm sure itโs nice on a clear day
26.01.2026 05:58 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
I'll be presenting this work on January 25th (Hall 2, poster 41) at #AAAI2026 in Singapore!
Please stop by and reach out if you'd like to chat ๐
23.01.2026 14:49 โ ๐ 3 ๐ 0 ๐ฌ 0 ๐ 0
๐https://arxiv.org/abs/2406.17737
Work done with Deb Roy and Jad Kabbara
@jad-kabbara.bsky.social
at @mit.edu @medialab.bsky.social
23.01.2026 14:42 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0
This pattern, which we refer to as targeted underperformance, suggests that models systematically lower information quality for some users.
As LLMs increasingly mediate access to knowledge ๐๐ง , these dynamics risk amplifying epistemic inequity at scale.
6/6
23.01.2026 14:42 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Hereโs one concrete example:
The same factual SciQ question posed to Claude
โ
Answered for a control user (no bio)
โ Refused for a less-educated Russian user
5/6
23.01.2026 14:42 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Across models, we observe systematic drops in accuracy and truthfulness for users who are:
โข Less educated
โข Non-native English speakers
โข From outside the U.S.
These effects compound and are largely invisible ๐ to standard evaluations.
4/6
23.01.2026 14:42 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 1
We evaluated GPT-4, Claude Opus, and Llama-3-8B in a Multiple Choice setup with questions taken from TruthfulQA and SciQ. Each question is conditioned on a user bio where we vary three user traits:
โข Education level ๐
โข Country of origin ๐
โข English proficiency ๐ฃ๏ธ
3/6
23.01.2026 14:42 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Spoiler alert: we find the answer is often no! โ ๏ธ
LLM accuracy and truthfulness systematically degrade for some users in ways that standard benchmarks, focused on best-case performance, fail to capture.
2/6
23.01.2026 14:42 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
๐ Excited to share our new paper which was accepted to #AAAI2026!
As LLMs become increasingly used as sources of factual knowledge, we ask:
Do they perform equitably across users of different backgrounds?
๐งตโฌ๏ธ
1/6
23.01.2026 14:42 โ ๐ 2 ๐ 1 ๐ฌ 1 ๐ 1
Yay!
Out of curiosity, what is the process of going from reviewer to AC and then to SAC? Do they just ask you out of the blue one day? Or do you apply?
20.01.2026 09:25 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0
Most LLM evals use API calls or offline inference, testing models in a memory-less silo. Our new Patterns paper shows this misses how LLMs actually behave in real user interfaces, where personalization and interaction history shape responses: arxiv.org/abs/2509.19364
12.12.2025 20:42 โ ๐ 38 ๐ 11 ๐ฌ 1 ๐ 1
Elinor Poole-Dayan, Jiayi Wu, Taylor Sorensen, Jiaxin Pei, Michiel A. Bakker: Benchmarking Overton Pluralism in LLMs https://arxiv.org/abs/2512.01351 https://arxiv.org/pdf/2512.01351 https://arxiv.org/html/2512.01351
02.12.2025 06:29 โ ๐ 2 ๐ 1 ๐ฌ 0 ๐ 0
20.11.2025 20:56 โ ๐ 22 ๐ 3 ๐ฌ 1 ๐ 0
Thoughtful (as always) blog post from Nicholas Carlini. "Are large language models worth it?" A nice read giving his perspective on risks of ML models.
Post: nicholas.carlini.com/writing/2025...
For people who prefer, this is the video of the talk from @colmweb.org www.youtube.com/watch?v=PngH...
19.11.2025 16:56 โ ๐ 34 ๐ 11 ๐ฌ 1 ๐ 1
Extremely thrilled to talk about our new paper: "Who Evaluates AIโs Social Impacts? Mapping Coverage And Gaps In First And Third Party Evaluations".
This is the first big project output from the
@eval-eval.bsky.social coalition! Thread below:
13.11.2025 14:34 โ ๐ 18 ๐ 7 ๐ฌ 1 ๐ 0
Congratulations @sivareddyg.bsky.social ! ๐ฅณ Incredibly well deserved!!
14.11.2025 17:11 โ ๐ 3 ๐ 0 ๐ฌ 0 ๐ 0
YouTube video by UVM Office of Research
IC2S2 2026 | Burlington, Vermont
We're excited to announce that the website and registration for IC2S2 2026 (July 28-31) will launch in early December! The Vermont Complex Systems Institute @vcsi.bsky.social at the University of Vermont will be hosting IC2S2 in 2026: youtube.com/watch?v=p412S4GnPkc&feature=youtu.be
13.11.2025 15:36 โ ๐ 34 ๐ 19 ๐ฌ 0 ๐ 1
A staircase in the new School of Computer, Data & Information Sciences building at Wisconsin Madison. Tan wood structures surround tapestry art and a small indoor garden.
A view from above of the staircases in the Wisconsin CDIS building
An shot from below of winding wooden staircases and a glass atrium rooftop. The new School of Computer, Data & Information Sciences building at Wisconsin Madison.
A bicolor white cat with seal-colored markings, looking upwards with big wide dark eyes.
It's the season for PhD apps!! ๐ฅง ๐ฆ โ๏ธ โ๏ธ
Apply to Wisconsin CS to research
- Societal impact of AI
- NLP โโ CSS and cultural analytics
- Computational sociolinguistics
- Human-AI interaction
- Culturally competent and inclusive NLP
with me!
lucy3.github.io/prospective-...
11.11.2025 22:32 โ ๐ 51 ๐ 16 ๐ฌ 1 ๐ 1
@bennokrojer.bsky.social didn't your lab have something like this happen
07.11.2025 21:02 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0
such a valuable resource! thanks for sharing
07.11.2025 13:51 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0
Itโs grad school application season, and I wanted to give some public advice.
Caveats:
-*-*-*-*
โจ> These are my opinions, based on my experiences, they are not secret tricks or guarantees
โจ> They are general guidelines, not meant to cover a host of idiosyncrasies and special cases
06.11.2025 14:55 โ ๐ 112 ๐ 58 ๐ฌ 4 ๐ 7
PhD supervised by Tim Rocktรคschel and Ed Grefenstette, part time at Cohere. Language and LLMs. Spent time at FAIR, Google, and NYU (with Brenden Lake). She/her.
Assistant Professor of Computational Social Science at Center for Critical Computational Studies (C3S), Goethe University Frankfurt | prev. postdoc at Center for Adaptive Rationality, MPIB Berlin | PhD political science Hertie School | psychologist
Assistant Prof of AI & Decision-Making @MIT EECS
I run the Algorithmic Alignment Group (https://algorithmicalignment.csail.mit.edu/) in CSAIL.
I work on value (mis)alignment in AI systems.
https://people.csail.mit.edu/dhm/
The Paul G. Allen School of Computer Science & Engineering at the University of Washington educates tomorrow's innovators while developing solutions to humanity's greatest challenges.
One of the world's top schools in information science. Offering degrees in Informatics, Library and Information Science (MLIS), Information Management (MSIM), and Museology, and doctorates in information science. Seattle, WA.
https://ischool.uw.edu/
Director, Center for Tech Responsibility@Brown. FAccT OG. AI Bill of Rights coauthor. Former tech advisor to President Biden @WHOSTP. He/him/his. Posts my own.
NLP, Linguistics, Cognitive Science, AI, ML, etc.
Job currently: Research Scientist (NYC)
Job formerly: NYU Linguistics, MSU Linguistics
Assistant Professor at UW and Staff Research Scientist at Google DeepMind. Social Reinforcement Learning in multi-agent and human-AI interactions. PhD from MIT. Check out https://socialrl.cs.washington.edu/ and https://natashajaques.ai/.
๐ PhD student @ischool.uw.edu (he/him)
๐ค Interested in pluralistic alignment & scalable oversight in the context of social media algorithms
Academic Website: https://sohamde.in/ | ๐Seattle, WA
MIT postdoc, incoming UIUC CS prof
katedonahue.me
phd @ cornell information science
NLP, AI, LM research director at Ai2; Professor at UW
Assistant Professor @ UW iSchool. Interested in computational social science, social networks & causal inference.
http://martinsaveski.com
MIT Researcher, he/him, Senior Visiting Researcher @ Ritsumeikan, Co-Founder of Humanyze, former Senior Researcher @ HBS, author of People Analytics. AI, management, law, corporate governance, psychology, anthropology, ethics, and similar topics
PhD candidate @nlphuji.bsky.social | Using NLP to help scientists ๐
https://noy-sternlicht.github.io
MSc Computer Science @mcgill.ca. Interested in NLP, narrative understanding, computational social science, NLP for social good, and AI for human development. He/him https://davidghobson1.github.io/