Nathan Lambert's Avatar

Nathan Lambert

@natolambert.bsky.social

A LLN - large language Nathan - (RL, RLHF, society, robotics), athlete, yogi, chef Writes http://interconnects.ai At Ai2 via HuggingFace, Berkeley, and normal places

13,086 Followers  |  274 Following  |  1,727 Posts  |  Joined: 30.04.2023  |  1.9801

Latest posts by natolambert.bsky.social on Bluesky

Post image

Opus?
Sorry, living under rocks today.

24.11.2025 22:41 β€” πŸ‘ 18    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Latest open artifacts (#16): Who's building models in the U.S., China's model release playbook, and a resurgence of truly open models A month with SOTA releases with (truly) open model releases left and right.

Latest open artifacts (#16): Who's building models in the U.S., China's model release playbook, and a resurgence of truly open models
A month with SOTA releases with (truly) open model releases left and right.
www.interconnects.ai/p/latest-ope...

23.11.2025 19:35 β€” πŸ‘ 15    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Post image

I asked (on ChinaTalk) the head of product at Z ai, one of the leading Chinese companies building open models, how long it takes them to get their model out the door once its done training. Incredible stuff:

"a few hours" and the model is on HuggingFace.
www.chinatalk.media/p/the-zai-pl...

21.11.2025 17:05 β€” πŸ‘ 15    πŸ” 3    πŸ’¬ 2    πŸ“Œ 1
Post image

People: "You must be so relaxed, proud, and happy that the model you worked on all year is out."

Me:

21.11.2025 00:29 β€” πŸ‘ 51    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

the Epstein files have been trending on HuggingFace.

> This dataset is provided for...
> Evaluating information retrieval and retrieval augmented generation (RAG) systems.
> It is not intended for: Fine-tuning language models.

??

20.11.2025 21:49 β€” πŸ‘ 72    πŸ” 10    πŸ’¬ 2    πŸ“Œ 14
Post image Post image Post image

Happy Olmo day to all who celebrate.
Sorry to all who delayed releases today to get out of our way.
We're hiring.

20.11.2025 18:40 β€” πŸ‘ 32    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

Really 7B is the scale where a lot of meaningful research emerges, and good for large scale routine tasks quite often

20.11.2025 15:27 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0


Paper: allenai.org/papers/olmo3
Artifacts: huggingface.co/collections/...
Demo: playground.allenai.org
Interconnects post: www.interconnects.ai/p/olmo-3-ame...
Technical Ai2 Blog: allenai.org/blog/olmo3

20.11.2025 14:32 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

As with OLMo 2 32B at its release, OLMo 3 32B is the best open-source language model ever released. It’s an awesome privilege to get to provide these models to the broader community researching and understanding what is happening in AI today.

20.11.2025 14:32 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

As always, all of our models come with full training data, code, intermediate checkpoints, training logs, and a detailed technical report. All are available today, with some more additions coming before the end of the year.

20.11.2025 14:32 β€” πŸ‘ 10    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

This is a big milestone for Ai2 and the Olmo project. These aren’t huge models (more on that later), but it’s crucial for the viability of fully open-source models that they are competitive on performance – not just replications of models that came out 6 to 12 months ago.

20.11.2025 14:32 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image Post image

We present Olmo 3, our next family of fully open, leading language models.
This family of 7B and 32B models represents:

1. The best 32B base model.
2. The best 7B Western thinking & instruct models.
3. The first 32B (or larger) fully open reasoning model.

20.11.2025 14:32 β€” πŸ‘ 107    πŸ” 24    πŸ’¬ 3    πŸ“Œ 3
Video thumbnail

Chinese models are enabling AI research. US progress needs to be accelerated

20.11.2025 01:16 β€” πŸ‘ 16    πŸ” 2    πŸ’¬ 1    πŸ“Œ 1

Conspiracy theory: OpenAI's serving costs were too high because GPT 5 Thinking overthought too much on easy queries, and GPT 5.1 was a patch to fix that.

5.1 just feels weird, can't quite place it.

19.11.2025 04:08 β€” πŸ‘ 23    πŸ” 2    πŸ’¬ 3    πŸ“Œ 0
Post image

A new tab on Google Scholar???
scholar.google.com/scholar_labs...

18.11.2025 17:48 β€” πŸ‘ 39    πŸ” 8    πŸ’¬ 3    πŸ“Œ 3
Preview
Why AI writing is mid How the current way of training language models destroys any voice (and hope of good writing).

Why AI writing is mid
How the current way of training language models destroys any voice (and hope of good writing).

www.interconnects.ai/p/why-ai-wri...

17.11.2025 15:41 β€” πŸ‘ 87    πŸ” 10    πŸ’¬ 8    πŸ“Œ 14

Updating to GPT 5.1 in codex made the model do crazy shit when I ask it to stage changes, make a PR, revert to XYZ branch, etc.

Very shocking because even the first claude code could do this. Am I the only one?

16.11.2025 23:55 β€” πŸ‘ 10    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

GPT 5.1 Thinking sometimes responds so much faster than GPT 5 Thinking would that it makes me think "shit I must've used the wrong model" and doubt it's answer a bit.

Funny dynamic. TBD if quality dropped at all, quality at least didn't nosedive. Hard to tell when style changes a bit.

16.11.2025 23:47 β€” πŸ‘ 16    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

Yes will be fixed in print, these are mock ups

15.11.2025 00:34 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Those only work on cats, my dog would love that

14.11.2025 21:34 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
RLHF Book by Nathan Lambert The Reinforcement Learning from Human Feedback Book

And if you missed it :) rlhfbook.com

14.11.2025 21:04 β€” πŸ‘ 11    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

I'm excited to announce my RLHF Book is now in pre-order for the @manning.com Early Access Program (MEAP), and for this milestone it's 50% off.

Excited to land in print in early 2026! Lots of improvements coming soon.

Thanks for the support!
hubs.la/Q03Tc37Q0

14.11.2025 21:02 β€” πŸ‘ 48    πŸ” 5    πŸ’¬ 4    πŸ“Œ 2
Preview
A taxonomy for next-generation reasoning models Where we've been and where we're going with RLVR.

I described this as "Calibration" in my taxonomy of next-generation reasoning models. My taxonomy is here: www.interconnects.ai/p/next-gen-r...

13.11.2025 19:18 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

Many people are sleeping on, or even making fun of this plot in the GPT 5.1 release. This is a crucial plot for anyone serving a thinking model in real world use-cases. Latency to an answer is a huge cause of user churn and not thinking enough is a fast track to having your model's output be bad.

13.11.2025 19:18 β€” πŸ‘ 27    πŸ” 3    πŸ’¬ 2    πŸ“Œ 0

OpenAI showing very clearly why you should care about Character Training with GPT 5.1: It's the leading selling point of the release.

13.11.2025 02:45 β€” πŸ‘ 11    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Lol currently every X account with 2fa auth enabled is locked out, while if you don't have 2fa you can use the app as usual. Iconic levels of broken.

12.11.2025 18:28 β€” πŸ‘ 24    πŸ” 2    πŸ’¬ 2    πŸ“Œ 0
Preview
Interview: Ant Group's open model ambitions Launching a new open model series. Interconnects Interview #15.


More are coming soon, as it feels like the perfect time to do this, when open models are currently both numerous and extremely high performance.

On Interconnects: www.interconnects.ai/p/inside-a-c...
On YouTube: www.youtube.com/watch?v=vIgE...

12.11.2025 15:12 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 1    πŸ“Œ 1

They also had lots of great stuff recently on scaling their RL infrastructure here (see the post for a review of all of their recent publications and models). Ant Ling/InclusionAI feels like one of the labs that could make the jump to Kimi/Qwen tier in 2026 if the dice land correctly.

12.11.2025 15:12 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The first one is Ant Group’s Ling/InclusionAI team that released great models this summer after starting the project in reaction to DeepSeek. Only 6-8 months!

These models, Ling (instruct/base model), Ring (reasoner) and Ming (multilingual) are a fairly standard MoE model from 300B to 1T param.

12.11.2025 15:12 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

I’m starting a new series of interviews with all the leading open model labs around the world to show why people are doing this, how people train great models, and where the ecosystem is going.

12.11.2025 15:12 β€” πŸ‘ 20    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

@natolambert is following 20 prominent accounts