Overview of PixMo and its relation to Molmo's ability. PixMo's captions data enables Molmo's fine-grained understanding; PixMo's AskModelAnything enables Molmo's user interaction; PixMo's pointing data enables Molmo's pointing and counting; PixMo's synthetic data enables Molmo's visual skills.
Remember Molmo? The full recipe is finally out!
Training code, data, and everything you need to reproduce our models. Oh, and we have updated our tech report too!
Links in thread π
09.12.2024 18:33 β π 78 π 14 π¬ 1 π 1
The OLMo 2 models sit at the Pareto frontier of training FLOPs vs model average performance.
Meet OLMo 2, the best fully open language model to date, including a family of 7B and 13B models trained up to 5T tokens. OLMo 2 outperforms other fully open models and competes with open-weight models like Llama 3.1 8B β As always, we released our data, code, recipes and more π
26.11.2024 20:51 β π 151 π 36 π¬ 5 π 12
Meet TΓΌlu 3, a set of state-of-the-art instruct models with fully open data, eval code, and training algorithms.
We invented new methods for fine-tuning language models with RL and built upon best practices to scale synthetic instruction and preference data.
Demo, GitHub, paper, and models π
21.11.2024 17:15 β π 111 π 31 π¬ 2 π 7
πββοΈ
21.11.2024 16:26 β π 2 π 0 π¬ 0 π 0
The 2025 Conference on Language Modeling will take place at the Palais des Congrès in Montreal, Canada from October 7-10, 2025
π PhD student @uwcse @uwnlp. π© Private pilot. Previously: π§βπ» @oculus, π @IllinoisCS. π π₯Ύ π΄ββοΈ π΅ β οΈ
I work at Sakana AI ππ π‘ β @sakanaai.bsky.social
https://sakana.ai/careers
Open spaces and open-sourced AI
ML/AI at AI2 http://semanticscholar.org, http://alongside.care, http://data-cowboys.com
PhD Candidate at UC Irvine, Research Intern @ai2 | Previously ASAPP Amazon LinkedIn @msftresearch IIT-Delhi
Research on In-Context Learning and LLM Agents
https://shivanshu-gupta.github.io
exchanging algorithms with ai
ekinakyurek.github.io
Will irl - PhD student @ NYU on the academic job market!
Using complexity theory and formal languages to understand the power and limits of LLMs
https://lambdaviking.com/ https://github.com/viking-sudo-rm
Senior Research Scientist at Google DeepMind
https://swarooprm.github.io/
Professor at UW; Researcher at Meta. LMs, NLP, ML. PNW life.
Faculty at UC Irvine and RS at Skild AI.
Previously: FAIR Meta, CMU, and UIUC.
Working on Computer Vision, Robotics, and AI
web: http://maxim.ece.illinois.edu
substack: https://realizable.substack.com
Ph.D. Student at UNC NLP | Apple Scholar in AI/ML Ph.D. Fellowship | Prev: FAIR at Meta, AI2, Adobe (Intern) | Interests: #NLP, #ML | https://archiki.github.io/
Assistant Professor @ UVa, working on NLP and Machine Learning
PhD @ucberkeleyofficial.bsky.social | Past: AI4Code Research Fellow @msftresearch.bsky.social | Summer @EPFL Scholar, CS and Applied Maths @IIITDelhi | Hobbyist Saxophonist
https://lakshyaaagrawal.github.io
Maintainer of https://aka.ms/multilspy
San Diego Dec 2-7, 25 and Mexico City Nov 30-Dec 5, 25. Comments to this account are not monitored. Please send feedback to townhall@neurips.cc.
PhD supervised by Tim RocktΓ€schel and Ed Grefenstette, part time at Cohere. Language and LLMs. Spent time at FAIR, Google, and NYU (with Brenden Lake). She/her.