our paper on data mixing for LMs is out!
while building Olmo 3, we saw gaps between data mixing literature and real practice
🐠choosing proxy size, # runs, sampling, regression, constraints..
🐟data shifts during LM dev: can we reuse past experiments?
Olmix tackles them all!
13.02.2026 17:30 — 👍 22 🔁 3 💬 1 📌 0
Olmix: A framework for data mixing throughout LM development | Ai2
Olmix is a framework for language model data mixing that provides empirically grounded defaults and efficient reuse techniques.
Olmix is for anyone training LMs on diverse data who's tired of guessing at configuration choices—or re-running expensive experiments every time something changes.
✍️ Blog: buff.ly/3ZQ2w6M
💻 Code: buff.ly/9Lhh1uV
📚 Technical report: buff.ly/AHL9ewm
13.02.2026 16:34 — 👍 4 🔁 0 💬 0 📌 0
Across 5 domain updates and 64 domains, Olmix produces a mix that's 12% better on our downstream evaluation task suite and 3× more data-efficient than no mixing at all, while being 74% cheaper than recomputing from scratch after each update.
13.02.2026 16:34 — 👍 3 🔁 0 💬 1 📌 0
Mixture reuse is our mechanism where we reuse the relative ratios of the unchanged domains and only recompute on the domains that change. Fewer domains to recompute → lower costs as domains are updated throughout development.
13.02.2026 16:34 — 👍 1 🔁 0 💬 1 📌 0
A few of our findings:
◉ Mixing costs scale linearly with the # of domains
◉ Log-linear regression predicts mixture performance accurately
◉ Enforcing data repetition constraints can help maintain mix quality in data-constrained settings
13.02.2026 16:34 — 👍 1 🔁 0 💬 1 📌 0
OlmixBase emerged from studying a common mixing method pattern where you ↓
1️⃣ Train a bunch of smaller proxy models on different mixes
2️⃣ Fit a regression model to the mixture weights & performance
3️⃣ Return the mix that optimizes predicted performance
13.02.2026 16:34 — 👍 1 🔁 0 💬 1 📌 0
Olmix has two components:
✅ OlmixBase: a data mixing method informed by a comprehensive empirical study of key design choices
✅ Mixture reuse: a mechanism for efficiently adjusting your mix as data domains are updated by using info from the previous mix—without starting over
13.02.2026 16:34 — 👍 1 🔁 0 💬 1 📌 0
We built Olmix to address two pain points we hit while training Olmo 3:
1️⃣ There's little guidance on how to configure mixing methods—you're left guessing at key choices
2️⃣ Training data evolves constantly, and recomputing mixes from scratch becomes a tax on every update
13.02.2026 16:34 — 👍 1 🔁 0 💬 1 📌 0
Data mixing – determining how much web text, code, math, etc., you need for LM development – is a first-order lever on model quality. Introducing Olmix: a framework for configuring mixing methods at the start of dev & efficiently updating as data changes throughout. 🧵
13.02.2026 16:34 — 👍 14 🔁 5 💬 1 📌 1
AstaLabs AutoDiscovery
Try AutoDiscovery in AstaLabs today → buff.ly/UaBQLur
We're giving early users 1,000 free Hypothesis Credits to get started.
📚 Learn more in our blog: buff.ly/fHLhXPs
12.02.2026 16:06 — 👍 2 🔁 1 💬 0 📌 0
AutoDiscovery is an AI-powered tool that explores structured datasets autonomously—generating hypotheses, designing and running statistical experiments, and surfacing findings that researchers might…
Using Asta AutoDiscovery: AI-powered autonomous scientific discovery
Everything AutoDiscovery generates is transparent & reproducible—every hypothesis, statistical analysis, and line of Python code is there for you to inspect.
Science has always had more data than time to explore it. AutoDiscovery helps surface the questions hiding in yours: buff.ly/yGxdpMT
12.02.2026 16:06 — 👍 1 🔁 1 💬 1 📌 0
Accepted with Minor Revisions: Value of AI-Assisted Scientific Writing
Large Language Models have seen expanding application across domains, yet their effectiveness as assistive tools for scientific writing - an endeavor requiring precision, multimodal synthesis, and…
📄 In social science, AutoDiscovery helped economist Sanchaita Hazra find that doctoral-level authors made more edits to AI-generated abstracts, suggesting expertise drives critical engagement with AI.
Her independently verified results were published in a peer-reviewed paper: buff.ly/BJ5Zran
12.02.2026 16:06 — 👍 2 🔁 0 💬 1 📌 0
🌊 In marine ecology, Fabio Favoretto at Scripps used AutoDiscovery to explore 20+ years of rocky reef data from the Gulf of California, surfacing cross-trophic productivity relationships that would have taken extensive manual iterations to find.
Read the report: buff.ly/mhgIKD0
12.02.2026 16:06 — 👍 1 🔁 0 💬 1 📌 0
👩🔬 In oncology, Dr. Kelly Paulson at the Swedish Cancer Institute used AutoDiscovery to explore breast cancer & melanoma datasets, surfacing new hypotheses about immune responses + lymph node spread that weren't part of her team's initial questions.
Read more: buff.ly/9uJFRNL
12.02.2026 16:06 — 👍 1 🔁 0 💬 1 📌 0
Researchers across ecology, health, & social science are already using AutoDiscovery to surface findings hiding in their data—from cancer mutation patterns to trophic relationships in marine ecosystems.
Read their stories: buff.ly/xNbB93d
12.02.2026 16:06 — 👍 1 🔁 0 💬 1 📌 0
How does it decide what to pursue? Bayesian surprise—a measure of how much the system's beliefs change after seeing evidence. By chasing surprise, AutoDiscovery gravitates toward the unexpected, prioritizing results most likely to represent genuine discoveries. 🔬
12.02.2026 16:06 — 👍 2 🔁 0 💬 1 📌 0
Most AI tools for science wait for a research question, then help answer it. AutoDiscovery starts with your data. It generates hypotheses, runs experiments, interprets results, & uses what it learns to keep exploring.
12.02.2026 16:06 — 👍 1 🔁 0 💬 1 📌 0
Knowing which questions to ask is often the hardest part of science. Today we're releasing AutoDiscovery in AstaLabs, an AI system that starts with your data and generates its own hypotheses. 🧪
12.02.2026 16:06 — 👍 9 🔁 6 💬 1 📌 0
Explore MolmoSpaces & start building:
📝 Blog: buff.ly/nJpBn7E
💻 Demo: buff.ly/s3x8xJo
⬇️ Code: buff.ly/WrpkCxg
📊 Data: buff.ly/FMNHJAI
✍️ Paper: buff.ly/FpYgLDV
11.02.2026 19:48 — 👍 1 🔁 0 💬 0 📌 0
All MolmoSpaces assets, scenes, & tools are open + modular, provided in MJCF with USD conversion for cross-simulator portability. Plug in new embodiments, regenerate grasps, & run across MuJoCo, ManiSkill, & NVIDIA Isaac Lab/Sim.
11.02.2026 19:48 — 👍 1 🔁 0 💬 1 📌 0
MolmoSpaces supports teleoperation via mobile platforms like Teledex—collect demonstrations right from your phone, compatible with all our embodiment setups, including DROID and CAP; no extra configuration is needed.
11.02.2026 19:48 — 👍 0 🔁 0 💬 1 📌 0
📐 MolmoSpaces-Bench is our new benchmark for evaluating generalist policies under systematic, controlled variation. Researchers can vary one factor at a time – from object properties to layouts, task complexity, lighting, dynamics, and instruction phrasing – across thousands of realistic scenes.
11.02.2026 19:47 — 👍 1 🔁 0 💬 1 📌 0
MolmoSpaces ships with a massive bank of validated grasps for rigid and articulated objects, loadable directly into environments. An accompanying trajectory-generation pipeline supports reproducible demonstrations and imitation learning at scale.
11.02.2026 19:47 — 👍 0 🔁 0 💬 1 📌 0
MolmoSpaces builds on two foundations: Objaverse, one of the largest open collections of 3D objects, and our THOR family of interactive simulation environments, all unified with physics-grounded simulation + validated physical parameters tuned for realistic manipulation. ⚙️
11.02.2026 19:47 — 👍 0 🔁 0 💬 1 📌 0
The next wave of AI will act in the physical world, but building robots that generalize across new environments – rather than replaying learned behaviors – requires far more diverse training data than exists today.
That's where MolmoSpaces comes in.
11.02.2026 19:47 — 👍 0 🔁 0 💬 1 📌 0
Introducing MolmoSpaces, a large-scale, fully open platform + benchmark for embodied AI research. 🤖
230k+ indoor scenes, 130k+ object models, & 42M annotated robotic grasps—all in one ecosystem.
11.02.2026 19:47 — 👍 11 🔁 4 💬 1 📌 2
MolmoSpaces ships with a massive bank of validated grasps for rigid and articulated objects, loadable directly into environments. An accompanying trajectory-generation pipeline supports reproducible demonstrations and imitation learning at scale.
11.02.2026 17:32 — 👍 0 🔁 0 💬 0 📌 0
MolmoSpaces builds on two foundations: Objaverse, one of the largest open collections of 3D objects, and our THOR family of interactive simulation environments, all unified with physics-grounded simulation + validated physical parameters tuned for realistic manipulation. ⚙️
11.02.2026 17:31 — 👍 0 🔁 0 💬 1 📌 0
The next wave of AI will act in the physical world, but building robots that generalize across new environments – rather than replaying learned behaviors – requires far more diverse training data than exists today.
That's where MolmoSpaces comes in.
11.02.2026 17:31 — 👍 0 🔁 0 💬 1 📌 0
Thanks for the feedback. It's not a perfect tool; hallucinations may occur, particularly given the model's small size.
11.02.2026 01:32 — 👍 0 🔁 0 💬 0 📌 0
Understanding life. Advancing health.
Open Source Practitioner, bullish on kindness.
https://erlend.sh – about me
@roomy.space – make space
Research Scientist at Ai2, PhD in NLP 🤖 UofA. Ex
GoogleDeepMind, MSFTResearch, MilaQuebec
https://nouhadziri.github.io/
Science of language models @uwnlp.bsky.social and @ai2.bsky.social with @PangWeiKoh and @nlpnoah.bsky.social. https://ianmagnusson.github.io
Making life beyond cancer a reality. fredhutch.org
Video / Podcast / Media 🎥☕️🚲🏡📱♻️🌿🏳️🌈🎌🇺🇸
Meteorology PhD, data scientist, ski mountaineer, GSP dad
We help protected area protected area managers, ecologists, and wildlife biologists stay informed and make better, faster conservation-related decisions. A free solution from Paul Allen's @ai2.bsky.social
Director of IT and Digital Strategy at @whiteheadinstitute.bsky.social, MIT. PhD in CS. Turning data into discoveries. Fan of practical tech that works. Tea drinker, research enabler, innovation chaser. Views are my own. He/him.
PhD student @ University of Washington CSE | Engineering Manager @ Allen Institute for AI | Previously @ MSFT, Columbia, UW
Interested in accessibility, HCI, AI ✨
Associate Professor (University of Arizona).
Visiting Scientist (Allen Institute for AI/Ai2)
AI/NLP; Scientific Discovery; DiscoveryWorld; EntailmentBank; ScienceWorld; http://textgames.org list. Tweets/opinions my own
engineer, runner, filmmaker • machine learning, film distribution/exhibition • beach soul, mountain energy
building a virtual cinema for overlooked international film: holocene.club
🎥🌌
🇫🇷🇰🇷
English PhD turned Machine Learning Researcher. I medicate my imposter syndrome with cold brews.
Current:
Foundation Models @ Apple
Prev:
LLMs/World Models @ Riot Games
RecSys @ Apple
Seattle -> NYC 🏳️🌈
Software developer, filmmaker, film fan.
he/him
NLP, AI, LM research director at Ai2; Professor at UW
Religion, Philosophy, Theology, Interfaith Relations. Head of Department of Interreligious Dialogue @akademiers.bsky.social. Lecturer. @c_stroebele@mastodon.social
Predoc at Ai2 | prev. Princeton CS '24