Natasha Jaques's Avatar

Natasha Jaques

@natashajaques.bsky.social

Assistant Professor at UW and Staff Research Scientist at Google DeepMind. Social Reinforcement Learning in multi-agent and human-AI interactions. PhD from MIT. Check out https://socialrl.cs.washington.edu/ and https://natashajaques.ai/.

4,272 Followers  |  277 Following  |  52 Posts  |  Joined: 19.11.2024  |  1.9084

Latest posts by natashajaques.bsky.social on Bluesky

Instead of behavior cloning, what if you asked an LLM to write code to describe how an agent was acting, and used this to predict their future behavior?

Our new paper "Modeling Others' Minds as Code" shows this outperforms BC by 2x, and reaches human-level performance in predicting human behavior.

03.10.2025 19:01 β€” πŸ‘ 13    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

My husband presenting his work on caregiving 😍

01.08.2025 00:03 β€” πŸ‘ 12    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

By optimizing for intrinsic curiosity, the LLM learns how to ask a series of questions over the course of the conversation to improve the accuracy of its user model. This generates conversations which reveal significantly more information about the user.

08.07.2025 23:06 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Excited to release our latest paper on a new multi-turn RL objective for training LLMs to *learn how to learn* to adapt to the user. This enables it to adapt and personalize to novel users, whereas the multi-turn RLHF baseline fails to generalize effectively to new users.

08.07.2025 23:05 β€” πŸ‘ 12    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0

This work shows the benefit of RL training for improving reasoning skills when there is no possibility for data leakage. AND how continuously evolving multi-agent competition leads to the development of emergent skills that generalize to novel tasks.

01.07.2025 20:24 β€” πŸ‘ 9    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

We analyze the results and find that LLMs learn emergent reasoning patterns like case-by-case analysis and expected value calculation that transfer to improve performance on math questions.

01.07.2025 20:24 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

In our latest paper, we discovered a surprising result: training LLMs with self-play reinforcement learning on zero-sum games (like poker) significantly improves performance on math and reasoning benchmarks, zero-shot. Whaaat? How does this work?

01.07.2025 20:23 β€” πŸ‘ 59    πŸ” 7    πŸ’¬ 2    πŸ“Œ 2

Just posted a talk I gave about this work! youtu.be/mxWJ9k2XKbk

12.06.2025 17:08 β€” πŸ‘ 11    πŸ” 1    πŸ’¬ 0    πŸ“Œ 1

RLHF is the main technique for ensuring LLM safety, but it provides no guarantees that they won’t say something harmful.

Instead, we use online adversarial training to achieve theoretical safety guarantees and substantial empirical safety improvements over RLHF, without sacrificing capabilities.

12.06.2025 17:07 β€” πŸ‘ 16    πŸ” 3    πŸ’¬ 1    πŸ“Œ 1

RLHF is the main technique for ensuring LLM safety, but it provides no guarantees that they won’t say something harmful.

Instead, we use online adversarial training to achieve theoretical safety guarantees and substantial empirical safety improvements over RLHF, without sacrificing capabilities.

12.06.2025 17:07 β€” πŸ‘ 16    πŸ” 3    πŸ’¬ 1    πŸ“Œ 1

Oral @icmlconf.bsky.social !!! Can't wait to share our work and hear the community's thoughts on it, should be a fun talk!

Can't thank my collaborators enough: @cogscikid.bsky.social y.social @liangyanchenggg @simon-du.bsky.social @maxkw.bsky.social @natashajaques.bsky.social

09.06.2025 16:32 β€” πŸ‘ 9    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
RLDM2025SocInfWorkshop // RLDM 2025 Workshop \\ Reinforcement learning as a model of social behaviour and inference: progress and pitfalls 12.06.2025 // 9am-1pm

At @rldmdublin2025.bsky.social this week? Check out our social learning workshop from @amritalamba.bsky.social and I tomorrow! Inc. talks from @natashajaques.bsky.social, @nitalon.bsky.social, @carocharp.bsky.social, @kartikchandra.bsky.social & more!

Full schedule: sites.google.com/view/rldm202...

11.06.2025 06:49 β€” πŸ‘ 17    πŸ” 8    πŸ’¬ 2    πŸ“Œ 1
Doctoral Researcher and Postdoc positions to work on Computational Social Intrinsic Motivation (SIM) | Aalto University The Autotelic Interaction Research (AIR) group at the Dept. of Computer Science, Aalto University, FinlandΒ is looking for 1 Doctoral Researcher (2+2 years) and 1 Postdoc (2 years) Β to work on Computational Social Intrinsic Motivation (SIM)

1/4 Join us and the Autotelic Interaction Research (AIR) group @aalto.fi / Finland to work on Computational Social Intrinsic Motivation (SIM) as PhD (4y) or postdoc (2y). Job ad w project description and application instructions: bit.ly/4jyNLGv. We're looking forward to learning about you!

03.06.2025 12:13 β€” πŸ‘ 10    πŸ” 4    πŸ’¬ 1    πŸ“Œ 2

Way to go KJ for producing such an insightful paper in the first few months of your PhD!

19.04.2025 05:03 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Human-AI cooperation is important, but existing work trains on the same 5 Overcooked layouts, creating brittle strategies.

Instead, we find that training on billions of procedurally generated tasks trains agents to learn general cooperative norms that transfer to humans... like avoiding collision

19.04.2025 05:02 β€” πŸ‘ 17    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0
Post image

Our new paper (first one of my PhD!) on cooperative AI reveals a surprising insight: Environment Diversity > Partner Diversity.

Agents trained in self-play across many environments learn cooperative norms that transfer to humans on novel tasks.

shorturl.at/fqsNN%F0%9F%...

19.04.2025 00:06 β€” πŸ‘ 25    πŸ” 7    πŸ’¬ 1    πŸ“Œ 5
Post image

Got a weird combination of mail today.

12.04.2025 17:04 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I had a ton of fun using this as a kid. I actually made my high school English class project a giant hypercard-based video game where I drew each frame in Paint and hid buttons behind the hand-drawn elements that let you navigate the world. That was so fun...😍

28.03.2025 19:57 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Reinforcement Learning (RL) for LLMs
YouTube video by Natasha Jaques Reinforcement Learning (RL) for LLMs

Recorded a recent "talk" / rant about RL fine-tuning of LLMs for a guest lecture in Stanford CSE234: youtube.com/watch?v=NTSY.... Covers some of my lab's recent work on personalized RLHF, as well as some mild Schmidhubering about my own early contributions to this space

27.03.2025 21:31 β€” πŸ‘ 51    πŸ” 10    πŸ’¬ 5    πŸ“Œ 1

next Canadian government should think of boosting research funding up here and trying to grab as many American postdocs and researchers as possible

12.03.2025 13:08 β€” πŸ‘ 16654    πŸ” 2185    πŸ’¬ 619    πŸ“Œ 183

Yes! Or you could focus on developing better MARL algorithms for the corporations that let them cooperate to solve the social dilemma more effectively. Similar to MARL benchmarks like Melting Pot but for a more impactful domain

16.02.2025 23:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

AI has shown great potential in boosting efficiency. But can it help human society make better decisions as a whole? πŸ€” In this project, using MARL, we explore this by studying the impact of an ESG disclosure mandateβ€”a highly controversial policy. (1/6)

13.02.2025 06:24 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

In contrast, MARL enables testing new policies with many more agents over a long time horizon.

We hope this benchmark will enable researchers in the RL and MARL communities to develop sophisticated cooperation algorithms in the context of a societally impactful problem!

13.02.2025 06:42 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I’m really excited about this, as I think MARL provides a new tool in the toolbox for investigating this problem. Existing work on ESG disclosures focuses on empirical studies (can’t test counterfactual policies), or analytical economics models (limited to 2 players or short time intervals)

13.02.2025 06:41 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

...providing corporations with more reliable information about climate risks β€” and we show that this significantly improves corporations’ ability to mitigate climate change, even without the influence of investors!

13.02.2025 06:39 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The parameters of the environment are carefully benchmarked to real-world data (e.g. from IPCC reports), and many of our experiments reveal findings consistent with empirical work on climate change and ESG disclosures. But we can also test new policy interventions, such as..

13.02.2025 06:39 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

overall climate risks increase, leading to lower market wealth. By introducing climate-conscious investor agents to simulate the possible effects of ESG disclosures, we show that investors can actually incentivize profit-motivated companies to invest in climate mitigation.

13.02.2025 06:38 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

Our latest work uses multi-agent reinforcement learning to model corporate investment in climate change mitigation as a social dilemma. We create a new benchmark, and show that corporations are greedily motivated to pollute without mitigating their emissions, but if all companies defect...

13.02.2025 06:38 β€” πŸ‘ 35    πŸ” 5    πŸ’¬ 2    πŸ“Œ 0
Post image

Impressive results that strongly outperform the best classical techniques. Several MARL methods work well, showing the promise of MARL for combinatorial optimization and systems problems, as in our earlier work on MARL for Microprocessor Design arxiv.org/abs/2211.16385 3/3

02.01.2025 22:33 β€” πŸ‘ 12    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

The classical solver can give an optimal, conflict-free assignment for a single timestep if it's given known utility values, but can't plan assignments over time. So, we learn the future estimated payoff for each single-timestep assignment using RL! 2/3

02.01.2025 22:32 β€” πŸ‘ 12    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

@natashajaques is following 20 prominent accounts