The transformer was invented in Google. RLHF was not invented in industry labs, but came to prominence in OpenAI and DeepMind. I took 5 of the most influential papers (black dots) and visualized their references. Blue dots are papers that acknowledge federal funding (DARPA, NSF).
12.04.2025 02:35 β
π 109
π 24
π¬ 2
π 0
New Preprint! Interested in learning about how working memory is subserved by both compositional and generative mechanisms? Read on!
14.04.2025 02:24 β
π 31
π 8
π¬ 1
π 0
To be clear, I believe the interesting areas of modularity that warrant investigation are world modeling, reasoning, planning, and memory, generally. Bring back fuzzy memory modules! [6]
[6] arxiv.org/abs/1410.5401
14.04.2025 05:12 β
π 0
π 0
π¬ 0
π 0
Ugh, sorry for the gross LinkedIn links on that cross post. I'll do better next time. π«
14.04.2025 05:05 β
π 0
π 0
π¬ 0
π 0
After a short era in which people questioned the value of academia in ML, its value is more obvious than ever. Big labs stopped publishing the minute commercial incentives showed up and are relentlessly focused on a singular vision of scaling. Academia is a meaningful complement, bringing...
1/2
14.04.2025 01:04 β
π 214
π 41
π¬ 2
π 2
LinkedIn
This link will take you to a page thatβs not on LinkedIn
Full disclosure, I'm a bit biased as I've developed a similar proprietary set of protocols recently. Kudos to the communities for recognizing the need for standardization for further scaling!
[1] lnkd.in/eUqVV5SQ
[2] lnkd.in/gf6e6XH6
[3] lnkd.in/gx4EbSFR
[4] lnkd.in/g_42revZ
14.04.2025 05:00 β
π 0
π 0
π¬ 2
π 0
LinkedIn
This link will take you to a page thatβs not on LinkedIn
My intuition is that standardization of message passing protocols, like MCP[3] and A2A[4], will further enable both research and engineering for these kinds of approaches.
14.04.2025 05:00 β
π 0
π 0
π¬ 1
π 0
LinkedIn
This link will take you to a page thatβs not on LinkedIn
One of the reasons I'm happy to see a disaggregation effect in the departure from "one massive model to rule them all" is the modularity and, frankly, accessibility brought forward for those without so much compute available. I hope to see more socratic [2] type approaches in the coming years!
14.04.2025 05:00 β
π 0
π 0
π¬ 1
π 0
LinkedIn
This link will take you to a page thatβs not on LinkedIn
The AI research scene is dealing with a second order hardware lottery[1] effect right now, GPUs being the first, in that many papers being published are based on pretrained models trained using large research clusters available to only a few labs.
14.04.2025 05:00 β
π 0
π 0
π¬ 1
π 0
βPhilosophy would render us entirely skeptics, were not nature too strong for it.β
β David Hume, An Enquiry Concerning Human Understanding
#philosophy #philsky
21.03.2025 03:06 β
π 37
π 4
π¬ 0
π 1
How it started / how it's going.....
18.03.2025 02:44 β
π 145
π 24
π¬ 9
π 7
we released olmo 32b today! βΊοΈ
πour largest & best fully open model to-date
π right up there w similar size weights-only models from big companies on popular benchmarks
π‘but we used way less compute & all our data, ckpts, code, recipe are free & open
made a nice plot of our post-trained results!βοΈ
13.03.2025 20:42 β
π 40
π 7
π¬ 2
π 1
Some of his readers have asked Mike Masnick @mmasnick.bsky.social why his technology news site, Tech Dirt, has been covering politics so intensely lately. www.techdirt.com/2025/03/04/w...
I cannot recommend Mike's reply enough. It's exactly what readers need to hear, what journalists need to do.
07.03.2025 00:09 β
π 4564
π 1820
π¬ 86
π 114
My new paper "Deep Learning is Not So Mysterious or Different": arxiv.org/abs/2503.02113. Generalization behaviours in deep learning can be intuitively understood through a notion of soft inductive biases, and formally characterized with countable hypothesis bounds! 1/12
05.03.2025 15:37 β
π 209
π 49
π¬ 6
π 9
Awesome LLM Post-training
This repository is a curated collection of the most influential papers, code implementations, benchmarks, and resources related to Large Language Models (LLMs) Post-Training Methodologies.
github.com/mbzuai-oryx/...
04.03.2025 00:03 β
π 44
π 10
π¬ 1
π 0
Starlink embedded in the FAA.
Grok used by the OPM.
Tesla contracts from the DoD.
SpaceX taking over NASA tasks.
We are "NeuraLink requirements for Social Security payments" away from a complete governmental parasitic symbiosis.
25.02.2025 15:43 β
π 3
π 5
π¬ 1
π 0
GitHub - ARBORproject/arborproject.github.io
Contribute to ARBORproject/arborproject.github.io development by creating an account on GitHub.
Today we're launching a multi-lab open collaboration, the ARBOR project, to accelerate AI interpretability research for reasoning models. Please join us!
github.com/ARBORproject...
(ARBOR = Analysis of Reasoning Behavior through Open Research)
20.02.2025 19:55 β
π 44
π 9
π¬ 1
π 0
JUST IN: NASA says there's now a 3.1% chance an asteroid will hit Earth in 2032, up from 2.6% yesterday.
This is the highest risk assessment an asteroid has ever received, surpassing 2.7% in 2004
18.02.2025 19:24 β
π 799
π 165
π¬ 168
π 683
Forget βtapestryβ or βdelveβ these are the actual unique giveaway words for each model, relative to each other. arxiv.org/pdf/2502.12150
19.02.2025 03:04 β
π 102
π 17
π¬ 6
π 9
perplexity-ai/r1-1776 Β· Hugging Face
Weβre on a journey to advance and democratize artificial intelligence through open source and open science.
An uncensored version of R1 is released π₯
βR1 1776 is a DeepSeek-R1 reasoning model that has been post-trained by Perplexity AI to remove CCP censorship. The model provides unbiased, accurate, and factual information while maintaining high reasoning capabilities.β
huggingface.co/perplexity-a...
19.02.2025 03:22 β
π 58
π 11
π¬ 2
π 7
Why reasoning models will generalize
People underestimate the long-term potential of βreasoning.β
Why reasoning models will generalize
DeepSeek R1 is just the tip of the ice berg of rapid progress.
People underestimate the long-term potential of βreasoning.β
28.01.2025 21:04 β
π 51
π 8
π¬ 5
π 1
Current me: It's only one more project/talk/paper/review...
Future me: Don't do this, I beg you.
Current me: Super interesting, could find a way to fit it in...
Future me: C'mon, remember the rule, just say no!
Current me: & loads of time before the deadline...
Future me: Wait, can you even hear me?
17.01.2025 16:01 β
π 58
π 10
π¬ 1
π 3
The state of post-training in 2025
A re-record of my NeurIPS tutorial on language modeling (plus some added content).
The state of post-training in 2025: a tutorial on modern post-training
A re-record of my NeurIPS tutorial on language modeling (plus some added content on the high level state of things)
Blog + extra context: https://buff.ly/424VvLm
YouTube: https://buff.ly/40808l5
Slides: https://buff.ly/404jGa9
08.01.2025 15:38 β
π 80
π 17
π¬ 4
π 0
In Solidarity with Ann Telnaes. β
βDemocracy Dies in Darkness.β
anntelnaes.substack.com/p/why-im-qui...
@anntelnaes.bsky.social
05.01.2025 16:36 β
π 123
π 32
π¬ 0
π 2