Jiacheng Liu's Avatar

Jiacheng Liu

@liujch1998.bsky.social

πŸŽ“ PhD student @uwcse @uwnlp. πŸ›© Private pilot. Previously: πŸ§‘β€πŸ’» @oculus, πŸŽ“ @IllinoisCS. πŸ“– πŸ₯Ύ πŸš΄β€β™‚οΈ 🎡 ♠️

85 Followers  |  52 Following  |  16 Posts  |  Joined: 05.12.2024  |  2.0361

Latest posts by liujch1998.bsky.social on Bluesky

How can we better understand how models make predictions and which components of a training dataset are shaping their behaviors? In April we introduced OLMoTrace, a feature that lets you trace the outputs of language models back to their full training data in real time. 🧡

30.06.2025 17:37 β€” πŸ‘ 10    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0
Post image

As we’ve been working towards training a new version of OLMo, we wanted to improve our methods for measuring the Critical Batch Size (CBS) of a training run, to unlock greater efficiency. but we found gaps between the methods in the literature and our practical needs for training OLMo. 🧡

03.06.2025 16:43 β€” πŸ‘ 16    πŸ” 3    πŸ’¬ 1    πŸ“Œ 1
Preview
β€˜Bold,’ β€˜positive’ and β€˜unparalleled’: Allen School Ph.D. graduates Ashish Sharma and Sewon Min recognized with ACM Doctoral Dissertation Awards - Allen School News Each year, the Association for Computing Machinery recognizes the best Ph.D. dissertations in computer science with its Doctoral Dissertation Award. Ashish Sharma (Ph.D., β€˜24), now a senior applied sc...

Congratulations to #UWAllen Ph.D. grads Ashish Sharma and @sewonm.bsky.social‬, β€ͺ2024 @acm.org ‬Doctoral Dissertation Award honorees! Sharma won for #AI tools for mental health; Min received honorable mention for efficient, flexible language models. #ThisIsUW news.cs.washington.edu/2025/06/04/a...

04.06.2025 19:27 β€” πŸ‘ 4    πŸ” 3    πŸ’¬ 0    πŸ“Œ 2
Post image

I’m thrilled to share RewardBench 2 πŸ“Šβ€” We created a new multi-domain reward model evaluation that is substantially harder than RewardBench, we trained and released 70 reward models, and we gained insights about reward modeling benchmarks and downstream performance!

02.06.2025 23:41 β€” πŸ‘ 22    πŸ” 6    πŸ’¬ 2    πŸ“Œ 1
Post image

πŸ“’We’re taking your questions now on Reddit for tomorrow’s AMA!

Ask us anything about OLMo, our family of fully-open language models. Our researchers will be on hand to answer them Thursday, May 8 at 8am PST.

07.05.2025 16:46 β€” πŸ‘ 3    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

The story of OLMo, our Open Language Model, goes back to February 2023 when a group of researchers gathered at Ai2 and started planning. What if we made a language model with state-of-the-art performance, but we did it completely in the open? 🧡

06.05.2025 20:55 β€” πŸ‘ 16    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
A bar graph comparing average performance (10 Tasks) across OLMo 2 1B, SmolLM2 1.7B, Gemma 3 1B, Llama 3.2 1B, and Qwen 2.5 1.5B. The highest performance is 42.7, achieved by OLMo 2 1B.

A bar graph comparing average performance (10 Tasks) across OLMo 2 1B, SmolLM2 1.7B, Gemma 3 1B, Llama 3.2 1B, and Qwen 2.5 1.5B. The highest performance is 42.7, achieved by OLMo 2 1B.

We're excited to round out the OLMo 2 family with its smallest member, OLMo 2 1B, surpassing peer models like Gemma 3 1B or Llama 3.2 1B. The 1B model should enable rapid iteration for researchers, more local development, and a more complete picture of how our recipe scales.

01.05.2025 13:01 β€” πŸ‘ 43    πŸ” 10    πŸ’¬ 1    πŸ“Œ 4
Ask Us Anything about our Open Language Model, OLMo

Ask Us Anything about our Open Language Model, OLMo

Have questions? We’re an open book!

We’re excited to host an AMA to answer your Qs about OLMo, our family of open language models.

πŸ—“οΈ When: May 8, 8-10 am PT
🌐 Where: r/huggingface
🧠 Why: Gain insights from our expert researchers

Chat soon!

01.05.2025 17:57 β€” πŸ‘ 14    πŸ” 4    πŸ’¬ 1    πŸ“Œ 3
"With OLMoTrace, we’re actually bringing accessibility to openness, enabling everybody to start looking into the inner workings of the relationships between the input and output of these models." - Ali Farhadi, Ai2 CEO

"With OLMoTrace, we’re actually bringing accessibility to openness, enabling everybody to start looking into the inner workings of the relationships between the input and output of these models." - Ali Farhadi, Ai2 CEO

Last week we released OLMoTrace as part of #GoogleCloudNext

14.04.2025 19:30 β€” πŸ‘ 6    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Preview
Looking at the training data On building tools where truly open-source models can shrine (OLMo 2 32B Instruct, for today). OLMoTrace lets you poke around.

Ai2 launched a new tool where your responses from OLMo get mapped back to related training data. We're using this actively to improve our post-training data and hope many others will use it for understanding and transparency around leading language models!
Some musings:

09.04.2025 20:12 β€” πŸ‘ 48    πŸ” 4    πŸ’¬ 2    πŸ“Œ 5
Lead OLMoTrace researcher Jiacheng Liu at Ai2's Google Cloud Next booth.

Lead OLMoTrace researcher Jiacheng Liu at Ai2's Google Cloud Next booth.

The entrance to the Vertex AI Model Garden at Google Cloud Next.

The entrance to the Vertex AI Model Garden at Google Cloud Next.

A QR code leading to the story of Google Cloud and Ai2's partnership sitting near a faux fire pit.

A QR code leading to the story of Google Cloud and Ai2's partnership sitting near a faux fire pit.

Ai2 COO Sophie Lebrecht talks to visitors at Ai2's booth at Google Cloud Next.

Ai2 COO Sophie Lebrecht talks to visitors at Ai2's booth at Google Cloud Next.

Coming to you live from #GoogleCloudNext Day 2!

πŸ“ Find us at the Vertex AI Model Garden inside the Google Cloud Showcase - try out OLMoTrace, and take a step inside our fully open AI ecosystem.

10.04.2025 16:24 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Ali Farhadi speaking on stage at a fireside chat

Ali Farhadi speaking on stage at a fireside chat

"OLMoTrace is a breakthrough in AI development, setting a new standard for transparency and trust. We hope it will empower researchers, developers, and users to build with confidenceβ€”on models they can understand and trust." - CEO Ali Farhadi at tonight's chat with Karen Dahut #GoogleCloudNext

10.04.2025 01:23 β€” πŸ‘ 9    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

OLMoTrace is powered by my previous work infini-gram, with some innovative algorithmic twists. Really proud to turn an academic research project into a real LLM product, it’s been a truly amazing experience.

Check out infini-gram: infini-gram.io

09.04.2025 15:56 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Try OLMoTrace in Ai2 Playground with our OLMo 2 models: playground.allenai.org

If OLMoTrace gives you new insight into how LLMs behave, we’d love you to share your use case! πŸ’‘Take a screenshot, post the thread link if you like, and don’t forget to tag
@allen_ai

09.04.2025 13:37 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Today we're unveiling OLMoTrace, a tool that enables everyone to understand the outputs of LLMs by connecting to their training data.

We do this on unprecedented scale and in real time: finding matching text between model outputs and 4 trillion training tokens within seconds. ✨

09.04.2025 13:37 β€” πŸ‘ 40    πŸ” 5    πŸ’¬ 1    πŸ“Œ 2
Video thumbnail

For years it’s been an open question β€” how much is a language model learning and synthesizing information, and how much is it just memorizing and reciting?

Introducing OLMoTrace, a new feature in the Ai2 Playground that begins to shed some light. πŸ”¦

09.04.2025 13:16 β€” πŸ‘ 59    πŸ” 12    πŸ’¬ 5    πŸ“Œ 13

πŸ“°Google Cloud moves deeper into open source AI with Ai2 partnership:

β€œMany were wary of using AI models unless they had full transparency into models’ training data and could customize the models completely. Ai2’s models allow that.”

08.04.2025 17:50 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

(4/4) Searching in OLMo 2's training data is now available in both our web interface and the API endpoint.

Plus, OLMo 2 32B Instruct is a very strong model. Let's do real science with it πŸ§ͺ

08.04.2025 14:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

(3/4) We know the pain point in LLM research in academia: We don't know what's in the training data of these LLMs (GPT, Llama, etc) and what's not; we can only speculate.

So we made the full training data of OLMo 2 and OLMoE searchable, including pre-training and post-training.

08.04.2025 14:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
GitHub - liujch1998/infini-gram Contribute to liujch1998/infini-gram development by creating an account on GitHub.

(2/4) Check out the source code of infini-gram here: github.com/liujch1998/infini-gram

If you are new to infini-gram, you might want to start with exploring our web interface infini-gram.io/demo and API endpoint infini-gram.io/api_doc

08.04.2025 14:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

As infini-gram surpasses 500 million API calls, today we're announcing two exciting updates:

1. Infini-gram is now open-source under Apache 2.0!
2. We indexed the training data of OLMo 2 models. Now you can search in the training data of these strong, fully-open LLMs.

🧡 (1/4)

08.04.2025 14:50 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

Stay tuned... Wednesday, at #GoogleCloudNext and online πŸ‘€

07.04.2025 18:59 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
A list of paper authors for 2 OLMo 2 Furious.

A list of paper authors for 2 OLMo 2 Furious.

Buckle your seatbelt β€” we've released the OLMo 2 paper to kick off 2025 πŸ”₯. Including 50+ pages on 4 crucial components of the LLM development pipeline.

06.01.2025 20:28 β€” πŸ‘ 37    πŸ” 6    πŸ’¬ 1    πŸ“Œ 0
Post image

kicking off 2025 with our OLMo 2 tech report while payin homage to the sequelest of sequels 🫑

πŸš— 2 OLMo 2 Furious πŸ”₯ is everythin we learned since OLMo 1, with deep dives into:

πŸš– stable pretrain recipe
πŸš” lr anneal 🀝 data curricula 🀝 soups
🚘 tulu post-train recipe
🚜 compute infra setup

πŸ‘‡πŸ§΅

03.01.2025 16:02 β€” πŸ‘ 69    πŸ” 17    πŸ’¬ 2    πŸ“Œ 1

Yes we’ve read your paper and there’s so many interesting findings! Let’s grab coffee at Neurips

09.12.2024 20:11 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

(8/n) ... and senior authors @soldni.bsky.social @nlpnoah.bsky.social @mechanicaldirk.bsky.social Pang Wei Koh, Jesse Dodge, Hanna Hajishirzi

09.12.2024 17:07 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

(7/n) This work wouldn’t have been possible without my awesome co-first author @akshitab.bsky.social, wonderful colleagues @awettig.bsky.social @davidheineman.com @oyvind-t.bsky.social @ananyahjha93.bsky.social ...

09.12.2024 17:07 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

(6/n) Compared to existing work, our method accurately predicts performance on individual tasks, is designed to work on arbitrary overtrained regimes, and is compute-efficient.

Paper link: arxiv.org/abs/2412.04403

09.12.2024 17:07 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

(5/n) We have loads of interesting analyses! Check out our paper to find out:
* variance analysis of task accuracy and its impact on prediction error
* impact of using even less compute to make prediction
* ablating the many design choices in our method and exploring alternatives

09.12.2024 17:07 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

(4/n) We can predict the task accuracy of OLMo 2 7B and 13B (after pretraining and before mid-training) within an absolute error of 2 points on four tasks – MMLU, HellaSwag, PIQA, and Social IQa. Error on other tasks is a bit higher, and we aim to improve them in future work.

09.12.2024 17:07 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

@liujch1998 is following 20 prominent accounts