Ownership is a scam invented by big thing to sell more stuff.
24.01.2026 20:31 β π 1 π 0 π¬ 0 π 0Ownership is a scam invented by big thing to sell more stuff.
24.01.2026 20:31 β π 1 π 0 π¬ 0 π 0Severed hand?!? Just go to an embassy and do it there!
18.01.2026 18:05 β π 0 π 0 π¬ 1 π 0Phone, as well as email, are a way in which anyone in the world can get a piece of my attention whenever they want. As we are figuring out increasingly effective ways of monetizing attention, exploitation of this resource expands.
10.01.2026 07:25 β π 1 π 0 π¬ 0 π 0Oof
17.12.2025 21:04 β π 1 π 0 π¬ 0 π 0Be sure to make that point when the funding for the other stuff disappears all the same.
08.12.2025 22:32 β π 0 π 0 π¬ 0 π 0
Happy Olmo day to all who celebrate.
Sorry to all who delayed releases today to get out of our way.
We're hiring.
Announcing Olmo 3, a leading fully open LM suite built for reasoning, chat, & tool use, and an open model flowβnot just the final weights, but the entire training journey.
Best fully open 32B reasoning model & best 32B base model. π§΅
we released Olmo 3! lot of exciting stuff but wanna focus on:
πOlmo 3 32B Base, the best fully-open base model to-date, near Qwen 2.5 & Gemma 3 on diverse evals
π Olmo 3 32B Think, first fully-open reasoning model approaching Qwen 3 levels
π‘12 training datasets corresp to different staged training
I'm excited to announce my RLHF Book is now in pre-order for the @manning.com Early Access Program (MEAP), and for this milestone it's 50% off.
Excited to land in print in early 2026! Lots of improvements coming soon.
Thanks for the support!
hubs.la/Q03Tc37Q0
@ananyahjha93.bsky.social knows the pain. We once got reviews to a paper that said "Please do further experiments [which would cost $2M]", and _also_ a review that said "This work is too expensive to be relevant to anyone in the field.". In the same paper!
14.11.2025 23:25 β π 6 π 0 π¬ 0 π 0Incredible work by Apple's UX department, enabling three different corner radii at the same time π
12.11.2025 23:34 β π 2 π 0 π¬ 0 π 0While reviewing for #CHI2026, I've noticed four new writing issues in #HCI papers, likely due to an increased use of #LLMs / #AI. I describe them here - and how to fix them: dbuschek.medium.com/when-llms-wr...
23.10.2025 14:15 β π 28 π 5 π¬ 2 π 2You presumably have something you want to run, and it doesn't run. Or it runs on CPU. Spin up a Claude Code console and tell it about the problem. Tell it the command line that produces the wrong result. It'll start suggesting ways of fixing your environment, down to modifying installed drivers.
08.10.2025 16:53 β π 1 π 0 π¬ 0 π 0Let Claude Code sort out your environment.
08.10.2025 01:44 β π 0 π 0 π¬ 2 π 0Are humans allowed to attend?
24.09.2025 16:27 β π 1 π 0 π¬ 0 π 0What field/area is like this now?
24.09.2025 16:21 β π 0 π 0 π¬ 2 π 0Weβre releasing early pre-training checkpoints for OLMo-2-1B to help study how LLM capabilities emerge. Theyβre fine-grained snapshots intended for analysis, reproduction, and comparison. π§΅
18.08.2025 19:02 β π 27 π 6 π¬ 1 π 0
Mein DreijΓ€hriger: "Ich will den Lerns Geschichte Podcast hΓΆren!"
Was ist denn "Lerns Geschichte"?
Zwei Minuten spΓ€ter im Radio: "Lernen's a bissel @geschichte.fm, dann ..." π²
Almost all post-training is "dusting off capable base models"
28.07.2025 02:52 β π 1 π 0 π¬ 1 π 0Unverified second hand information: In the US, all fish has to be flash-frozen before being served raw. In Canada, it does not.
18.07.2025 22:39 β π 0 π 0 π¬ 0 π 0I think for the moment we're competing on a different axis. They do quite well on impact per GPU hour. We do well on impact per person hour.
24.06.2025 06:28 β π 1 π 0 π¬ 0 π 0In ML, you can get surprisingly far without ever looking at your training data, and yet you'll always be limited. Thus, in ML, "look at the data" means, "Don't just stir the pot of linear algebra, find out what's really happening."
08.06.2025 16:48 β π 4 π 0 π¬ 0 π 0
This project is a perfect model of an OLMo contribution. Well scoped, practical, sound theoretical underpinnings, and @lambdaviking.bsky.social
submitted the paper 24h before the deadline π.
It's integrated into the OLMo trainer here: github.com/allenai/OLMo...
Meanwhile, OLMo is now the citation for QK norm, which we definitely didn't invent? You win some, you lose some.
13.05.2025 17:22 β π 2 π 0 π¬ 0 π 0Finally, OLMo 1B. This is the most commonly requested OLMo feature l, and it's finally here.
01.05.2025 22:31 β π 1 π 0 π¬ 0 π 0After ICML, I decided all conferences should be in Vienna from now on.
23.04.2025 21:34 β π 0 π 0 π¬ 0 π 0I'm in Singapore for @iclr-conf.bsky.social ! Come check out our spotlight paper on the environmental impact of training OLMo (link in next tweet) during the Saturday morning poster session from 10-12:30 -- happy to chat about this or anything else! DMs should be open, email works too
23.04.2025 15:21 β π 10 π 5 π¬ 1 π 1Came across arxiv.org/pdf/2504.05058 today. What a cool example of work you can do when LLM training data is open!
18.04.2025 17:46 β π 7 π 0 π¬ 1 π 0Plot shows the relationship between compute used to predict a ranking of datasets and how accurately that ranking reflects performance at the target (1B) scale of models pretrained from scratch on those datasets.
Ever wonder how LLM developers choose their pretraining data? Itβs not guessworkβ all AI labs create small-scale models as experiments, but the models and their data are rarely shared.
DataDecide opens up the process: 1,050 models, 30k checkpoints, 25 datasets & 10 benchmarks π§΅
Today we're unveiling OLMoTrace, a tool that enables everyone to understand the outputs of LLMs by connecting to their training data.
We do this on unprecedented scale and in real time: finding matching text between model outputs and 4 trillion training tokens within seconds. β¨