Instead of forcing models to hold everything in an active context window, we can use hypernetworks to instantly compile documents and tasks directly into the model's weights. A step towards giving language models durable memory and fast adaptation.
Blog: pub.sakana.ai/doc-to-lora/
27.02.2026 04:36 โ
๐ 104
๐ 14
๐ฌ 2
๐ 4
How competition is stifling AI breakthroughs
Llion Jones cowrote "Attention Is All You Need," the seminal paper that introduced the transformer โ the architecture that launched the generative AI revolution. Now he warns that the industry that gr...
ใHow Competition is Stifling AI Breakthroughsใ
Sakana AIๅ
ฑๅๅตๆฅญ่
Llion Jones ใฎTED AIใใผใฏใๅ
ฌ้ใใใพใใใ็ฎๆจใๅฎใใใใชใใชใผใใณใจใณใใช็ ็ฉถใใใฌใผใฏในใซใผใ็ใ็็ฑใTransformerใฎๆๅใๆฅญ็ใซใใใใใ็ถๆณใใใใไนใ่ถใใๆฌกใฎๆงๆณใจๆๆใ่ชใใพใใใ
www.ted.com/talks/llion_...
29.01.2026 02:39 โ
๐ 4
๐ 1
๐ฌ 0
๐ 0
Our journey at Sakana AI is just getting started.
We are looking for people to help us pioneer the next generation of AIโbuilding from Japan to the world.
Join us: sakana.ai/careers
25.01.2026 04:56 โ
๐ 29
๐ 3
๐ฌ 0
๐ 0
I founded Sakana AI after my time at Google, so it is incredibly meaningful to be able to partner with them now. It feels like a special connection to be working together again to advance the AI ecosystem in Japan.
sakana.ai/google#en
23.01.2026 13:04 โ
๐ 48
๐ 2
๐ฌ 3
๐ 0
Sakana AI
Sakana AIใGoogleใจใฎๆฆ็ฅ็ใใผใใใผใทใใ็ท ็ตใ็บ่กจ
Our work on The AI Scientist and ALE-Agent has already shown the power of these models.
Now, we are scaling reliable AI in mission-critical sectors like finance and government to ensure the highest security and data sovereignty.
Full details: sakana.ai/google#en
23.01.2026 13:02 โ
๐ 1
๐ 1
๐ฌ 1
๐ 0
We are thrilled to announce a strategic partnership with Google!
Google is also making a financial investment in Sakana AI to strengthen this collaboration. We are combining Googleโs world-class products like Gemini and Gemma with our agile R&D to accelerate automated scientific discovery.
23.01.2026 13:02 โ
๐ 17
๐ 2
๐ฌ 1
๐ 1
An Unofficial Guide to Prepare for a Research Position Application
Authors: Stefania Druga, Luke Darlow, and Llion Jones
Disclaimer: This guide is written by a few researchers at Sakana AI who have interviewed many candidates, and does not reflect the view of the entire organization. Each team may have their own preferences and styles for interviewing and finding the people that they can work closely with. This document, written by Stefania, Luke, and Llion, provides a glimpse into how some parts of our research org conduct interviews.
We just published an unofficial guide on what we look for when interviewing research candidates at Sakana AI.
Written by Stefania Druga, Luke Darlow, and Llion Jones.
The biggest differentiator? Understanding over implementation.
Read it: pub.sakana.ai/Unofficial_G...
20.01.2026 01:40 โ
๐ 23
๐ 1
๐ฌ 1
๐ 1
Introducing RePo: Language Models with Context Re-Positioning
Standard LLMs force a rigid linear structure on context, treating physical proximity as relevance. Cognitive Load Theory suggests this is inefficientโmodels waste capacity managing noise instead of reasoning.
arxiv.org/abs/2512.14391
19.01.2026 00:39 โ
๐ 56
๐ 8
๐ฌ 1
๐ 4
2026 is just getting started ๐โจ
We are hiring. Join our team in Tokyo!
sakana.ai/careers
14.01.2026 13:23 โ
๐ 13
๐ 1
๐ฌ 0
๐ 0
Reminded me of my older NeurIPS 2021 paper, where we removed the positional encoding entirely, and by doing so, an agent can process an arbitrarily long list of noisy, sensory inputs, in an arbitrary order.
I even made a fun browser demo to play with the agent back then: attentionneuron.github.io
12.01.2026 05:51 โ
๐ 29
๐ 3
๐ฌ 0
๐ 0
Introducing DroPE: Extending Context by Dropping Positional Embeddings
We found embeddings like RoPE aid training but bottleneck long-sequence generalization. Our solutionโs simple: treat them as a temporary training scaffold, not a permanent necessity.
arxiv.org/abs/2512.12167
pub.sakana.ai/DroPE
12.01.2026 04:07 โ
๐ 118
๐ 22
๐ฌ 2
๐ 7
One of my favorite findings: Positional embeddings are just training wheels. They help convergence but hurt long-context generalization.
We found that if you simply delete them after pretraining and recalibrate for <1% of the original budget, you unlock massive context windows. Smarter, not harder.
12.01.2026 04:12 โ
๐ 220
๐ 32
๐ฌ 8
๐ 1
We are taking our technology far beyond competitive programming to unlock a new era of AI-driven discovery.
We are hiring. Join our team in Tokyo.
sakana.ai/careers/#sof...
10.01.2026 03:07 โ
๐ 9
๐ 1
๐ฌ 0
๐ 1
Weโre hiring.
sakana.ai/careers/#sof...
10.01.2026 03:08 โ
๐ 28
๐ 7
๐ฌ 0
๐ 0
When agents compete for limited resources, intelligence reorganizes around survival, not elegance.
09.01.2026 09:26 โ
๐ 19
๐ 3
๐ฌ 1
๐ 1
Survival of the fittest code!
Our paper explores LLMs driving an evolutionary arms race in Core War, where assembly programs fight each other. We task LLMs with evolving "Warriors" in a virtual machine, producing chaotic, self-modifying code dynamics.
Blog: sakana.ai/drq
Paper: pub.sakana.ai/drq/
08.01.2026 17:11 โ
๐ 41
๐ 8
๐ฌ 2
๐ 2
Introducing Digital Red Queen (DRQ): Adversarial Program Evolution in Core War with LLMs.
In this work, we explore how LLMs can drive open-ended adversarial evolution of programs within the Core War environment.
Blog sakana.ai/drq
Website pub.sakana.ai/drq/
ArXiv arxiv.org/abs/2601.03335
Thread:
08.01.2026 17:00 โ
๐ 32
๐ 7
๐ฌ 2
๐ 2
So proud of Team Sakana AI for pulling this off! We managed to get an agent to rank #1 in a difficult heuristic optimization contest. We leaned heavily into test-time inference using a mix of frontier models.
The agent spent $1,300 to autonomously discover an algorithm that beat the human baseline.
05.01.2026 16:01 โ
๐ 36
๐ 4
๐ฌ 2
๐ 0
Our AI agent has achieved 1st place in a competitive optimization programming contest against over 800 human participants.
Blog: sakana.ai/ahc058
Thread:
05.01.2026 15:53 โ
๐ 20
๐ 2
๐ฌ 2
๐ 2
Happy New Year! โฉ๏ธ
01.01.2026 03:47 โ
๐ 12
๐ 0
๐ฌ 0
๐ 0
Sakana AIโs office looks like this.
28.12.2025 01:43 โ
๐ 2
๐ 1
๐ฌ 0
๐ 0
Especially in such times, hackers and tinkerers tend to fare better at harnessing evolving technology with a high level of uncertainty and ambiguity, compared to traditional well-read professional types.
27.12.2025 01:09 โ
๐ 16
๐ 0
๐ฌ 2
๐ 1
Software Engineering as a profession will continue to fundamentally change in 2026.
Humans will need to learn to co-adapt to this evolving โalien technologyโ which comes with no real manual, and figure out how to operate it.
What a time to be alive โจ
twitter.com/karpathy/sta...
27.12.2025 01:09 โ
๐ 36
๐ 0
๐ฌ 1
๐ 0
Merry Christmas! ๐
Sakanaโ
AIใงใฏใไบๆฅญ้็บใซ้ขๅฟใใใๆนๅใใฎใใซใธใฅใขใซ้ข่ซ็ชๅฃใใใชใผใใณใใพใใ๏ผ้่ใป้ฒ่กใปใคใณใใชใธใงใณใน้ ๅใงใ็งใใกใใฉใฎใใใช้็บใซๆใใงใใใฎใใไธญใฎไบบใ็ดๆฅใ่ฉฑใใใพใใ
ๅ้่ท็จฎ๏ผโ
ใจใณใธใใขใProjectโ
ManagerใProductโ
Managerโ
ๅ
ๅฎน๏ผโ
ไบๆฅญๆฆ็ฅใ้็บใฎ่ฃๅดใใใผใ ใฎ้ฐๅฒๆฐใชใฉ
ๆๅ
็ซฏใฎAI้็บใ็คพไผๅฎ่ฃ
ใใใใญใปในใซ่ๅณใใใๆนใใใฒใๆฐ่ปฝใซใๅฟๅใใ ใใ๏ผ
๐ ๅฟๅใใฉใผใ ใฏใใกใ: forms.gle/sW5wz23SLSvN...
๐ ๅ้่ฆ้
: sakana.ai/careers/
25.12.2025 07:47 โ
๐ 3
๐ 1
๐ฌ 0
๐ 0
โBut perhaps this can be resolved by the realization that while cleverness and intelligence are somewhat correlated traits for humans, they are much more decoupled for AI tools (which are often optimized for cleverness), and viewing the current generation of such tools primarily as a stochastic generator of sometimes clever - and often useful - thoughts and outputs may be a more productive perspective when trying to use them to solve difficult problems.โ
I doubt that anything resembling genuine AGI is within reach of current AI toolsโTerence Tao
mathstodon.xyz/@tao/1157223...
22.12.2025 07:44 โ
๐ 91
๐ 12
๐ฌ 3
๐ 4
Why AGI Will Not Happen โ Tim Dettmers
If you are reading this, you probably have strong opinions about AGI, superintelligence, and the future of AI. Maybe you believe we are on the cusp of a transformative breakthrough. Maybe you are skep...
โThe US follows the idea that there will be one winner who takes it all. Even coming short of AGI, if you have the best model, almost all people will use your model and not the competitionโs model. The idea is: develop the Biggest, Baddest model and people will come.โ
timdettmers.com/2025/12/10/w...
14.12.2025 23:20 โ
๐ 1
๐ 1
๐ฌ 1
๐ 0