Daniel Jiang's Avatar

Daniel Jiang

@danielrjiang.bsky.social

Research Scientist Meta, Adjunct Professor at University of Pittsburgh. Interested in reinforcement learning, approximate DP, adaptive experimentation, Bayesian optimization, & operations research. http://danielrjiang.github.io πŸ“Chicago, IL

1,805 Followers  |  1,012 Following  |  5 Posts  |  Joined: 07.02.2024  |  1.7382

Latest posts by danielrjiang.bsky.social on Bluesky

Post image

First draft online version of The RLHF Book is DONE. Recently I've been creating the advanced discussion chapters on everything from Constitutional AI to evaluation and character training, but I also sneak in consistent improvements to the RL specific chapter.

rlhfbook.com

16.04.2025 19:01 β€” πŸ‘ 110    πŸ” 17    πŸ’¬ 3    πŸ“Œ 3
Post image Post image

At ICLR 2025 in Singapore, my co-authors and I presented two papers on RL. Feel free to let us know of any feedback and let me know if you'd like to chat!
- openreview.net/forum?id=AOl...
- openreview.net/forum?id=AOl...

26.04.2025 01:51 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Postdoctoral Researcher, Monetization (PhD) Meta's mission is to build the future of human connection and the technology that makes it possible.

Topics of interest include offline RL, post-training large language models with RLHF, and long-term recommendation systems. If you’re interested, please email me and/or apply here: www.metacareers.com/jobs/1142270...

17.03.2025 13:59 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Our team at Meta is hiring a postdoc researcher! Our group conducts both fundamental and applied research in reinforcement learning, with a focus on applications in Meta's advertising systems.

17.03.2025 13:59 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Preview
Turing Award Goes to A.I. Pioneers Andrew Barto and Richard Sutton Andrew Barto and Richard Sutton developed reinforcement learning, a technique vital to chatbots like ChatGPT.

Congrats to this year's Turing award winners! www.nytimes.com/2025/03/05/t...

Incidentally, if you'd like to hear from them, we know a place they've given / are giving keynotes

07.03.2025 02:38 β€” πŸ‘ 47    πŸ” 7    πŸ’¬ 0    πŸ“Œ 2
Preview
ASOS Digital Experiments Dataset A novel dataset that can support the end-to-end design and running of Online Controlled Experiments (OCE) with adaptive stopping. Hosted on the Open Science Framework

There’s one from ASOS.com that provides A/B test data over time (across many experiments, each with several arms).

Dataset: osf.io/64jsb/

Paper: arxiv.org/abs/2111.10198

We used it in a paper to benchmark an AE method. But I’d also love to know of other alternatives out there.

21.02.2025 05:27 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

Given a high-quality verifier, language model accuracy can be improved by scaling inference-time compute (e.g., w/ repeated sampling). When can we expect similar gains without an external verifier?

New paper: Self-Improvement in Language Models: The Sharpening Mechanism

arxiv.org/abs/2412.01951

14.12.2024 16:10 β€” πŸ‘ 41    πŸ” 6    πŸ’¬ 3    πŸ“Œ 0
Preview
Reinforcement Learning: An Overview This manuscript gives a big-picture, up-to-date overview of the field of (deep) reinforcement learning and sequential decision making, covering value-based RL, policy-gradient methods, model-based met...

An updated intro to reinforcement learning by Kevin Murphy: arxiv.org/abs/2412.05265! Like their books, it covers a lot and is quite up to date with modern approaches. It also is pretty unique in coverage, I don't think a lot of this is synthesized anywhere else yet

09.12.2024 14:27 β€” πŸ‘ 270    πŸ” 73    πŸ’¬ 9    πŸ“Œ 5

I know one of the organizers is @eugenevinitsky.bsky.social. They did a great job and organized a very enjoyable conference.

10.12.2024 08:18 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

I collected some folk knowledge for RL and stuck them in my lecture slides a couple weeks back: web.mit.edu/6.7920/www/l... See Appendix B... sorry, I know, appendix of a lecture slide deck is not the best for discovery. Suggestions very welcome.

27.11.2024 13:36 β€” πŸ‘ 113    πŸ” 17    πŸ’¬ 3    πŸ“Œ 3
Post image Post image Post image

Want to learn / teach RL? 

Check out new book draft:
Reinforcement Learning - Foundations
sites.google.com/view/rlfound...
W/ Shie Mannor & Yishay Mansour
This is a rigorous first course in RL, based on our teaching at TAU CS and Technion ECE.

25.11.2024 12:08 β€” πŸ‘ 154    πŸ” 34    πŸ’¬ 4    πŸ“Œ 4
Post image

New paper: Do social media algorithms shape affective polarization?

We ran a field experiment on X/Twitter (N=1,256) using LLMs to rerank content in real-time, adjusting exposure to polarizing posts. Result: Algorithmic ranking impacts feelings toward the political outgroup! πŸ§΅β¬‡οΈ

25.11.2024 20:32 β€” πŸ‘ 815    πŸ” 216    πŸ’¬ 32    πŸ“Œ 52

The RL (and some non-RL folks) starter pack is almost full. Pretty clear that the academic move here has succeeded
go.bsky.app/3WPHcHg

18.11.2024 20:30 β€” πŸ‘ 104    πŸ” 32    πŸ’¬ 12    πŸ“Œ 3

@danielrjiang is following 20 prominent accounts