Check out our new work on autonomous driving in new cities with map data + MARL!
21.02.2026 14:06 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0@bidiptas13.bsky.social
Check out our new work on autonomous driving in new cities with map data + MARL!
21.02.2026 14:06 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0Evolve at the hyperscale!
Work co-led with Mattie Fellows and Juan Agustin Duque.
Made possible by #Isambard and AIRR
๐ Website: eshyperscale.github.io
๐ Paper: alphaxiv.org/abs/2511.16652
๐ป Code: github.com/ESHyperscale...
๐ฅNanoEgg : github.com/ESHyperscale... (train in int ๐)
Scaling LLM Reasoning with EGGROLL ๐ฅ๐ง ๐
Using ๐ฅ to finetune RWKV-7 language models outperforms GRPO on Countdown and GSM8K โ
๐ฅsignificantly outperformed GRPO on the Countdown task, achieving a 35% validation accuracy compared to GRPO's 23%โ
EGGROLL ๐ฅfor RL ๐ฎ๐ค
๐ฅ is competitive with, and in many cases, better than OpenES performance, even before considering the vast speed-up!
๐ฅ matched OpenES on 7/16 environments and outperformed it on another 7/16
๐ฅ's low-rank approach does not compromise ES performance
๐ฅEGGROLLing in the Deep with๐ ๐ฏโ Speedup
๐ฅ speed nearly reaches the throughput of pure batch inference, leaving OpenES far behind
๐ฅ reaches 91% of pure batch inference speed vs. OpenES reaching only 0.41%
The EGGROLL Recipe
๐ง ๐ ๏ธ We replace full-rank perturbations with low-rank ones. Each update is still high rank, maintaining expressivity with faster training
๐ฅ EGGROLL converges to the full-rank update at a fast rate of 1/rank. The method is effective even with a rank of 1
We use EGGROLL ๐ฅto train RNN language models from scratch using only integer datatypes (and no activation functions!), scaling population size from 64 to 262144
2 (๐๐) orders of magnitude larger than prior ES worksโ
Introducing ๐ฅEGGROLL ๐ฅ(Evolution Guided General Optimization via Low-rank Learning)! ๐ Scaling backprop-free Evolution Strategies (ES) for billion-parameter models at large population sizes
โก100x Training Throughput
๐ฏFast Convergence
๐ขPure Int8 Pretraining of RNN LLMs