Thereβs a lot more detail in the full paper, and I would love to hear your thoughts and feedback on it!
Check out the preprint here: arxiv.org/pdf/2506.10150
17.06.2025 15:13 β π 1 π 0 π¬ 0 π 0
Huge thanks to my amazing collaborators: Fai Poungpeth, @diyiyang.bsky.social, Erina Farrell, @brucelambert.bsky.social, and @mattgroh.bsky.social π
17.06.2025 15:13 β π 1 π 0 π¬ 1 π 0
LLMs, when benchmarked against reliable expert judgments, can be reliable tools for overseeing emotionally sensitive AI applications.
Our results show we can use LLMs-as-judge to monitor LLMs-as-companion!
17.06.2025 15:13 β π 1 π 0 π¬ 1 π 0
For example, in one of the conversations in our dataset, a response that an expert saw as "dismissingβ the speakerβs emotions, a crowdworker interpreted as "validating" their emotions instead!
17.06.2025 15:13 β π 1 π 0 π¬ 1 π 0
These misjudgments from crowdworkers have huge implications for AI training and deploymentβ
If we use flawed evaluations to train and monitor "empathic" AI, we risk creating systems that propagate a broken standard of what good communication looks like.
17.06.2025 15:13 β π 1 π 0 π¬ 1 π 0
So why the gap between experts/LLMs and crowds?
Crowdworkers often
- have limited attention
- rely on heuristics like βitβs the thought that countsβ
- focusing on intentions rather than actual wording
show systematic rating inflation due to social desirability bias
17.06.2025 15:13 β π 1 π 0 π¬ 1 π 0
And when experts disagree, LLMs struggle to find a consistent signal too.
Hereβs how expert agreement (Krippendorff's alpha) varied across empathy sub-components:
17.06.2025 15:13 β π 1 π 0 π¬ 1 π 0
But hereβs the catch: LLMs are reliable when experts are reliable.
The reliability of expert judgments depends on the clarity of the construct. For nuanced, subjective components of empathic communication, experts often disagree.
17.06.2025 15:13 β π 1 π 0 π¬ 1 π 0
We analyzed thousands of annotations from LLMs, crowdworkers, and experts on 200 real-world conversations
And specifically looked at 21 sub-components of empathic communication from 4 evaluative frameworks
The result? LLMs consistently matched expert judgments better than crowdworkers did! π₯
17.06.2025 15:13 β π 1 π 0 π¬ 1 π 0
How do we reliably judge if AI companions are performing well on subjective, context-dependent, and deeply human tasks? π€
Excited to share the first paper from my postdoc (!!) investigating when LLMs are reliable judges - with empathic communication as a case study π§
π§΅π
17.06.2025 15:13 β π 4 π 0 π¬ 1 π 1
Super cool opportunity to work with brilliant scientists and fantastic mentors @mattgroh.bsky.social and Dashun Wang ππ
Feel free to reach out!
02.04.2025 14:35 β π 2 π 0 π¬ 0 π 0
Decision-Point Guided Safe Policy Improvement
Within batch reinforcement learning, safe policy improvement (SPI) seeks to ensure that the learnt policy performs at least as well as the behavior policy that generated the dataset. The core challeng...
Our paper: Decision-Point Guided Safe Policy Improvement
We show that a simple approach to learn safe RL policies can outperform most offline RL methods. (+theoretical guarantees!)
How? Just allow the state-actions that have been seen enough times! π€―
arxiv.org/abs/2410.09361
23.01.2025 18:23 β π 3 π 1 π¬ 0 π 0
CS PhD student @StanfordNLP
https://cs.stanford.edu/~shaoyj/
Cognitive scientist at Stanford. Open science advocate. Symbolic Systems Program director. Bluegrass picker, slow runner, dad. http://langcog.stanford.edu
Prof @ Wesleyan U; Co-Director of WesMediaProject (@wesmediaproject.bsky.social & mediaproject.wesleyan.edu); policomm, polisci, public opinion & health policy scholar; member of Collaborative on Media & Messaging for Health & Social Policy (commhsp.org)
Iβm a reasonable man, get off my case.
sanjaysrivastava.com
Director of Education for MediaSmarts, Canada's centre for digital media literacy. He/him. Open to correction.
Assistant professor of CS at UC Berkeley, core faculty in Computational Precision Health. Developing ML methods to study health and inequality. "On the whole, though, I take the side of amazement."
https://people.eecs.berkeley.edu/~emmapierson/
NLP, ML & society, healthcare.
PhD student at Berkeley, previously CS at MIT.
https://rajivmovva.com/
Professor, Department of Psychology and Center for Brain Science, Harvard University
https://gershmanlab.com/
assoc prof, uc irvine cogsci & LPS: perception+metacognition+subjective experience, fMRI+models+AI
meta-science, global education, phil sci
prez+co-founder, neuromatch.io
fellow, CIFAR brain mind & consciousness
meganakpeters.org
she/her πππ views mine
Professor of Computer Science, Stanford - HCI & Design, Co-Founder & Co-Director Stanford HAI @stanfordhai.bsky.social
https://pinchunc.github.io
Postdoc at Oxford | PhD in Cognitive Sciences & MSc in Statistics from UC Irvine | Sleep & Memory
Associate Professor focused on #collectiveintelligence and #negotiation at UCL School of Management // previously Kellogg School of Mgmt, Annenberg Penn School of Comm, Community Mediation, AmeriCorps // only following identifiable people
interested in digital ecosystems and beautiful data.
Cognitive scientist studying how we see + think @ Johns Hopkins University. π¨π¦
Lab: https://perception.jhu.edu/
Associate Professor, Department of Psychology, Harvard University. Computation, cognition, development.
Asst Prof, Dartmouth CS. AI/ML, HCI, Multimodality, Music/Audio. Prev: PhD @ MIT, Allen Institute for AI, Netflix Research, Berklee College of Music.
https://nsingh1.host.dartmouth.edu
The Northwestern Institute on Complex Systems (NICO) incubates innovative collaborations that leverage complexity, networks, and data science.
Asst Professor, IROM Dept, UT Austin | FAccT EC | PhD, Machine Learning & Public Policy, Carnegie Mellon | Algorithmic fairness, human-AI collab | π¨π΄ π she/her/ella.