Max Lamparth, Ph.D.'s Avatar

Max Lamparth, Ph.D.

@mlamparth.bsky.social

Postdoc at @Stanford, @StanfordCISAC, Stanford Center for AI Safety, and the SERI program | Focusing on interpretable, safe, and ethical AI decision-making.

519 Followers  |  642 Following  |  56 Posts  |  Joined: 22.11.2024  |  1.7206

Latest posts by mlamparth.bsky.social on Bluesky

Post image Post image Post image Post image

ICYMI: The 2025 SERI Symposium explored the risks that emerge from the intersection of complex global challenges & policies designed to mitigate them, bringing together leading experts & researchers from across the Bay Area who specialize in a range of global risks
www.youtube.com/watch?v=wF20...

17.04.2025 20:40 โ€” ๐Ÿ‘ 3    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
MENTAT: A Clinician-Annotated Benchmark for Complex Psychiatric Decision-Making The official Stanford AI Lab blog

In their latest blog post for Stanford AI Lab, CISAC Postdoc @mlamparth.bsky.social and colleague Declan Grabb dive into MENTAT, a clinician-annotated dataset tackling real-world ambiguities in psychiatric decision-making.
ai.stanford.edu/blog/mentat/

11.04.2025 17:34 โ€” ๐Ÿ‘ 2    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

That sounds familiar. Thank you for sharing :)

04.04.2025 23:05 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Did you add anything to that query or is this the output for just that prompt? ๐Ÿ˜…

04.04.2025 22:27 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
MENTAT: A Clinician-Annotated Benchmark for Complex Psychiatric Decision-Making The official Stanford AI Lab blog

Thank Stanford AI Lab for featuring our work in a new blog post!

We created a dataset that goes beyond medical exam-style questions and studies the impact of patient demographic on clinical decision-making in psychiatric care on fifteen language models

ai.stanford.edu/blog/mentat/

04.04.2025 22:15 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Position: We Need An Adaptive Interpretation of Helpful, Honest, and Harmless Principles The Helpful, Honest, and Harmless (HHH) principle is a foundational framework for aligning AI systems with human values. However, existing interpretations of the HHH principle often overlook contextua...

The Helpful, Honest, and Harmless (HHH) principle is key for AI alignment, but current interpretations miss contextual nuances. CISAC postdoc @mlamparth.bsky.social & colleagues propose an adaptive framework to prioritize values, balance trade-offs, and enhance AI ethics.
arxiv.org/abs/2502.06059

11.03.2025 22:58 โ€” ๐Ÿ‘ 4    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Thank you for your support! In the short term, we hope to provide an evaluation data set for the community, because there is no existing equivalent at the moment, and highlight some issues. In the long term, we want to motivate extensive studies to enable oversight tools for responsible deployment.

26.02.2025 18:21 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Supported through @stanfordmedicine.bsky.social, Stanford Center for AI Safety,
@stanfordhai.bsky.social, @fsi.stanford.edu , @stanfordcisac.bsky.social StanfordBrainstorm

#AISafety #ResponsibleAI #MentalHealth #Psychiatry #LLM

26.02.2025 17:46 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

9/ Great collaboration with
Declan Grabb, Amy Franks, Scott Gershan, Kaitlyn Kunstman, Aaron Lulla, Monika Drummond Roots, Manu Sharma, Aryan Shrivasta, Nina Vasan, Colleen Waickman

26.02.2025 17:46 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Moving Beyond Medical Exam Questions: A Clinician-Annotated Dataset of Real-World Tasks and Ambiguity in Mental Healthcare Current medical language model (LM) benchmarks often over-simplify the complexities of day-to-day clinical practice tasks and instead rely on evaluating LMs on multiple-choice board exam questions. Th...

8/ MENTAT is open-source.

Weโ€™re making it available to the community to push AI research beyond test-taking and toward real clinical reasoning with dedicated eval questions and 20 designed questions for few-shot prompting or similar approaches.
Paper arxiv.org/abs/2502.16051

26.02.2025 17:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

7/ High scores on multiple choice QA โ‰  Free-form decisions.

๐Ÿ“‰ High accuracy in multiple-choice tests does not necessarily translate to consistent open-ended responses (free-form inconsistency as measured in this paper: arxiv.org/abs/2410.13204).

26.02.2025 17:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

6/ Impact of demographic information on decision-making

๐Ÿ“‰ Bias alert: All models performed differently across categories based on patient age, gender coding, and ethnicity. (Full plots in the paper)

26.02.2025 17:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

5/ We put 15 LMs to the test. The results?

๐Ÿ“‰ LMs did great on more factual tasks (diagnosis, treatment).
๐Ÿ“‰ LMs struggled with complex decisions (triage, documentation).
๐Ÿ“‰ (Mental) health fine-tuned models (higher MedQA scores) dont outperform their off-the-shelf parent models.

26.02.2025 17:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

4/ The questions in the triage and documentation categories are designed to be ambiguous to reflect the challenges and nuances of these tasks, for which we collect annotations and create a preference dataset to enable more nuanced analysis with soft labels.

26.02.2025 17:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

3/ Each question has five answer options for which we remove all non-decision-relevant demographic information of patients to allow for detailed studies of how patient demographic information (age, gender, ethnicity, nationality, โ€ฆ) impacts model performance.

26.02.2025 17:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

2/ Introducing MENTAT ๐Ÿง  (MENtal health Tasks AssessmenT): A first-of-its-kind dataset designed and annotated by mental health experts with no LM involvement. It covers real clinical tasks in five categories:
โœ… Diagnosis
โœ… Treatment
โœ… Monitoring
โœ… Triage
โœ… Documentation

26.02.2025 17:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Itโ€™s Time to Bench the Medical Exam Benchmark Medical licensing examinations, such as the United States Medical Licensing Examination, have become the default benchmarks for evaluating large language models (LLMs) in health care. Performance o...

1/ Current clinical AI evaluations rely on medical board-style exams that favor factual recall. Real-world decision-making is complex, subjective, and with ambiguity even to human expert decision-makersโ€”spotlighting critical AI safety issues also in other domains. Also: ai.nejm.org/doi/full/10....

26.02.2025 17:07 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Post image

๐Ÿšจ New paper!

Medical AI benchmarks over-simplify real-world clinical practice and build on medical exam-style questionsโ€”especially in mental healthcare. We introduce MENTAT, a clinician-annotated dataset tackling real-world ambiguities in psychiatric decision-making.

๐Ÿงต Thread:

26.02.2025 17:07 โ€” ๐Ÿ‘ 9    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Now also on arxiv.org/abs/2502.14143 !

21.02.2025 20:03 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
https://www.cooperativeai.com/post/new-report-multi-agent-risks-from-advanced-ai

I'm very happy to have contributed to the report.

Read the full report or the executive summary here t.co/jsoa3y1bLm (also coming to arxiv)

20.02.2025 20:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

We analyze key failure modes (conflict, collusion, and miscommunication), and describe seven risk factors that can lead to these failures (information asymmetries, network effects, selection pressures, destabilizing dynamics, commitment and trust, emergent agency, and multi-agent security).

20.02.2025 20:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Check out our new report on multi-agent security led by Lewis Hammond and the Cooperative AI Foundation! With the deployment of increasingly agentic AI systems across domains, this research area becomes more crucial.

20.02.2025 20:30 โ€” ๐Ÿ‘ 4    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

Submitting a benchmark to
ICML? Check out our NeurIPS Spotlight paper BetterBench! We outline best practices for benchmark design, implementation & reporting to help shift community norms. Be part of the change! ๐Ÿ™Œ

+ Add your benchmark to our database for visibility: betterbench.stanford.edu

27.01.2025 22:02 โ€” ๐Ÿ‘ 11    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

It was fun to contribute to this new dataset evaluating at the frontier of human expert knowledge! Beyond accuracy, the results also demonstrate the necessity for novel uncertainty quantification methods for LMs attempting challenging tasks and decision-making.

Check out the paper at: lastexam.ai

24.01.2025 17:44 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Getting rejected with one 10/10 review score and the same reviewer arguing that the other reviewers have unrealistic expectations hits different.๐Ÿค”
Oh well, time to refine ๐Ÿ˜

23.01.2025 01:30 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
CS120: Introduction to AI Safety Introduction to AI Safety

Webpage: web.stanford.edu/class/cs120/...
I will also update the reading list once at some point. I would love to get feedback or paper recommendations!

06.01.2025 17:02 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Want to learn more about safe AI and the challenges of creating it?

Check out the public syllabus (slides and recordings) of my course: "CS120 Introduction to AI Safety". The course is designed for people with all backgrounds, including non-technical. #AISafety #ResponsibleAI

06.01.2025 17:02 โ€” ๐Ÿ‘ 7    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Second Draft of the General-Purpose AI Code of Practice published, written by independent experts Independent experts present the second draft of the General-Purpose AI Code of Practice, based on the feedback received on the first draft, published on 14 November 2024.

As one of the vice chairs of the EU GPAI Code of Practice process, I co-wrote the second draft which just went online โ€“ feedback is open until mid-January, please let me know your thoughts, especially on the internal governance section!

digital-strategy.ec.europa.eu/en/library/s...

19.12.2024 16:59 โ€” ๐Ÿ‘ 14    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 1

Great collaboration between Stanford's Center for AI Safety, Brainstorm Lab for Mental Health Innovation, @stanfordmedicine.bsky.social's Department of Psychiatry and Behavioral Sciences, @stanfordcisac.bsky.social, and @fsistanford.bsky.social.

#ResponsibleAI #AISafety

19.12.2024 16:59 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Risks from Language Models for Automated Mental Healthcare: Ethics... Amidst the growing interest in developing task-autonomous AI for automated mental health care, this paper addresses the ethical and practical challenges associated with the issue and proposes a...

This op-ed is also based on our CoLM publication "Risks from Language Models for Automated Mental Healthcare: Ethics and Structure for Implementation" which you can find here: openreview.net/forum?id=1pg...

19.12.2024 16:59 โ€” ๐Ÿ‘ 2    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

@mlamparth is following 20 prominent accounts