's Avatar

@bidiptas13.bsky.social

28 Followers  |  10 Following  |  8 Posts  |  Joined: 21.11.2025  |  1.7009

Latest posts by bidiptas13.bsky.social on Bluesky


Check out our new work on autonomous driving in new cities with map data + MARL!

21.02.2026 14:06 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Evolution Strategies at the Hyperscale General ML Training Made as Fast and Easy as Inference

Evolve at the hyperscale!
Work co-led with Mattie Fellows and Juan Agustin Duque.
Made possible by #Isambard and AIRR

๐ŸŒ Website: eshyperscale.github.io
๐Ÿ“ Paper: alphaxiv.org/abs/2511.16652
๐Ÿ’ป Code: github.com/ESHyperscale...
๐ŸฅšNanoEgg : github.com/ESHyperscale... (train in int ๐Ÿ˜‰)

21.11.2025 17:56 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Scaling LLM Reasoning with EGGROLL ๐Ÿฅš๐Ÿง ๐Ÿ“

Using ๐Ÿฅš to finetune RWKV-7 language models outperforms GRPO on Countdown and GSM8K โ—

๐Ÿฅšsignificantly outperformed GRPO on the Countdown task, achieving a 35% validation accuracy compared to GRPO's 23%โ—

21.11.2025 17:56 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

EGGROLL ๐Ÿฅšfor RL ๐ŸŽฎ๐Ÿค–

๐Ÿฅš is competitive with, and in many cases, better than OpenES performance, even before considering the vast speed-up!

๐Ÿฅš matched OpenES on 7/16 environments and outperformed it on another 7/16

๐Ÿฅš's low-rank approach does not compromise ES performance

21.11.2025 17:56 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

๐ŸฅšEGGROLLing in the Deep with๐Ÿš€ ๐Ÿ’ฏโœ• Speedup

๐Ÿฅš speed nearly reaches the throughput of pure batch inference, leaving OpenES far behind

๐Ÿฅš reaches 91% of pure batch inference speed vs. OpenES reaching only 0.41%

21.11.2025 17:56 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

The EGGROLL Recipe
๐Ÿง ๐Ÿ› ๏ธ We replace full-rank perturbations with low-rank ones. Each update is still high rank, maintaining expressivity with faster training

๐Ÿฅš EGGROLL converges to the full-rank update at a fast rate of 1/rank. The method is effective even with a rank of 1

21.11.2025 17:56 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

We use EGGROLL ๐Ÿฅšto train RNN language models from scratch using only integer datatypes (and no activation functions!), scaling population size from 64 to 262144

2 (๐Ÿ”๐Ÿ”) orders of magnitude larger than prior ES worksโ—

21.11.2025 17:56 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Introducing ๐ŸฅšEGGROLL ๐Ÿฅš(Evolution Guided General Optimization via Low-rank Learning)! ๐Ÿš€ Scaling backprop-free Evolution Strategies (ES) for billion-parameter models at large population sizes

โšก100x Training Throughput
๐ŸŽฏFast Convergence
๐Ÿ”ขPure Int8 Pretraining of RNN LLMs

21.11.2025 17:56 โ€” ๐Ÿ‘ 26    ๐Ÿ” 8    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 4

@bidiptas13 is following 10 prominent accounts