Overview of PixMo and its relation to Molmo's ability. PixMo's captions data enables Molmo's fine-grained understanding; PixMo's AskModelAnything enables Molmo's user interaction; PixMo's pointing data enables Molmo's pointing and counting; PixMo's synthetic data enables Molmo's visual skills.
Remember Molmo? The full recipe is finally out!
Training code, data, and everything you need to reproduce our models. Oh, and we have updated our tech report too!
Links in thread ๐
09.12.2024 18:33 โ ๐ 78 ๐ 14 ๐ฌ 1 ๐ 1
The OLMo 2 models sit at the Pareto frontier of training FLOPs vs model average performance.
Meet OLMo 2, the best fully open language model to date, including a family of 7B and 13B models trained up to 5T tokens. OLMo 2 outperforms other fully open models and competes with open-weight models like Llama 3.1 8B โ As always, we released our data, code, recipes and more ๐
26.11.2024 20:51 โ ๐ 151 ๐ 36 ๐ฌ 5 ๐ 12
Meet Tรผlu 3, a set of state-of-the-art instruct models with fully open data, eval code, and training algorithms.
We invented new methods for fine-tuning language models with RL and built upon best practices to scale synthetic instruction and preference data.
Demo, GitHub, paper, and models ๐
21.11.2024 17:15 โ ๐ 111 ๐ 31 ๐ฌ 3 ๐ 7
๐โโ๏ธ
21.11.2024 16:26 โ ๐ 2 ๐ 0 ๐ฌ 0 ๐ 0
The 2025 Conference on Language Modeling will take place at the Palais des Congrรจs in Montreal, Canada from October 7-10, 2025
๐ PhD student @uwcse @uwnlp. ๐ฉ Private pilot. Previously: ๐งโ๐ป @oculus, ๐ @IllinoisCS. ๐ ๐ฅพ ๐ดโโ๏ธ ๐ต โ ๏ธ
I work at Sakana AI ๐๐ ๐ก โ @sakanaai.bsky.social
https://sakana.ai/careers
Open spaces and open-sourced AI
ML/AI at AI2 http://semanticscholar.org, http://alongside.care, http://data-cowboys.com
PhD Candidate at UC Irvine, Research Intern @ai2 | Previously ASAPP Amazon LinkedIn @msftresearch IIT-Delhi
Research on In-Context Learning and LLM Agents
https://shivanshu-gupta.github.io
exchanging algorithms with ai
ekinakyurek.github.io
PhD student @ UmU ๐ฆ
Assistant in Research @ Yale ๐ถ
Host and Organizer of FLaNN (flann.super.site) ๐ฎ
โฅ๏ธ
Will irl - PhD student @ NYU on the academic job market!
Using complexity theory and formal languages to understand the power and limits of LLMs
https://lambdaviking.com/ https://github.com/viking-sudo-rm
Senior Research Scientist at Google DeepMind
https://swarooprm.github.io/
Professor at UW; Researcher at Meta. LMs, NLP, ML. PNW life.
Faculty at UC Irvine and RS at Skild AI.
Previously: FAIR Meta, CMU, and UIUC.
Working on Computer Vision, Robotics, and AI
web: http://maxim.ece.illinois.edu
substack: https://realizable.substack.com
Ph.D. Student at UNC NLP | Apple Scholar in AI/ML Ph.D. Fellowship | Prev: FAIR at Meta, AI2, Adobe (Intern) | Interests: #NLP, #ML | https://archiki.github.io/
Assistant Professor @ UVa, working on NLP and Machine Learning
PhD @ucberkeleyofficial.bsky.social | Past: AI4Code Research Fellow @msftresearch.bsky.social | Summer @EPFL Scholar, CS and Applied Maths @IIITDelhi | Hobbyist Saxophonist
https://lakshyaaagrawal.github.io
Maintainer of https://aka.ms/multilspy
The Thirty-Eighth Annual Conference on Neural Information Processing Systems will be held in Vancouver Convention Center, on Tuesday, Dec 10 through Sunday, Dec 15.
https://neurips.cc/
PhD supervised by Tim Rocktรคschel and Ed Grefenstette, part time at Cohere. Language and LLMs. Spent time at FAIR, Google, and NYU (with Brenden Lake). She/her.