Looking forward to be speaking at IndabaX Sudan on Building Responsible and Ethical LLMs!
๐
Saturday, December 13th
โฐ 2:00 PM (GMT+2)
Register here: docs.google.com/forms/d/e/1F...
See you all there! :)
@bkhmsi.bsky.social
PhD at EPFL ๐ง ๐ป Ex @MetaAI, @SonyAI, @Microsoft Egyptian ๐ช๐ฌ
Looking forward to be speaking at IndabaX Sudan on Building Responsible and Ethical LLMs!
๐
Saturday, December 13th
โฐ 2:00 PM (GMT+2)
Register here: docs.google.com/forms/d/e/1F...
See you all there! :)
Not attending NeurIPS this year, but very much looking to connect.
Iโm seeking a PhD research internship next summer in AI for Science, especially where AI meets brain and cognitive sciences. ๐ง
If youโre hiring, Iโd love to connect!
bkhmsi.github.io
I finally found time to update the Egyptians in AI Research website, apologies for the delay!
Super excited to share that we now feature 227 incredible Egyptian researchers!! ๐คฏ
Link: bkhmsi.github.io/egyptians-in...
You can learn more about our work here: language-to-cognition.epfl.ch
Thanks to all my co-authors @gretatuckute.bsky.social, @davidtyt.bsky.social, @neurotaha.bsky.social and my advisors @abosselut.bsky.social and @mschrimpf.bsky.social!
On my way to #EMNLP2025 ๐จ๐ณ
Iโll be presenting our work (Oral) on Nov 5, Special Theme session, Room A106-107 at 14:30.
Letโs talk brains ๐ง , machines ๐ค, and everything in between :D
Looking forward to all the amazing discussions!
10/
๐ Huge thanks to my incredible co-authors @cndesabbata.bsky.social, @gretatuckute.bsky.social, @eric-zemingchen.bsky.social
and my advisors @mschrimpf.bsky.social and @abosselut.bsky.social!
9/
๐ Explore MiCRo:
๐ Website: cognitive-reasoners.epfl.ch
๐ Paper: arxiv.org/abs/2506.13331
๐ค HF Space (interactive): huggingface.co/spaces/bkhmsi/cognitive-reasoners
๐ง HF Models: huggingface.co/collections/bkhmsi/mixture-of-cognitive-reasoners-684709a0f9cdd7fa180f6678
8/
We now have a collection of 10 MiCRo models on HF that you can try out yourself!
๐ง HF Models: huggingface.co/collections/bkhmsi/mixture-of-cognitive-reasoners-684709a0f9cdd7fa180f6678
7/
We built an interactive HF Space where you can see how MiCRo routes tokens across specialized experts for any prompt, and even toggle experts on/off to see how behavior changes.
๐ค Try it here: huggingface.co/spaces/bkhms...
(Check the example prompts to get started!)
6/
We also wondered: if neuroscientists use functional localizers to map networks in the brain, could we do the same for MiCRoโs experts?
The answer: yes! The very same localizers successfully recovered the corresponding expert modules in our models!
5/
One result I was particularly excited about is the emergent hierarchy we found across MiCRo layers:
๐บEarlier layers route tokens to Language experts.
๐ปDeeper layers shift toward domain-relevant experts.
This emergent hierarchy mirrors patterns observed in the human brain ๐ง
4/
We find that MiCRo matches or outperforms baselines on reasoning tasks (e.g., GSM8K, BBH) and aligns better with human behavior (CogBench), while maintaining interpretability!!
3/
โจ Why it matters:
MiCRo bridges AI and neuroscience:
๐ค ML side: Modular architectures make LLMs more interpretable and controllable.
๐ง Cognitive side: Provides a testbed for probing how the relative contributions of different brain networks support complex behavior.
2/
๐งฉ Recap:
MiCRo takes a pretrained language model and post-trains it to develop distinct, brain-inspired modules aligned with four cognitive networks:
๐ฃ๏ธ Language
๐ข Logic / Multiple Demand
๐งโโ๏ธ Social / Theory of Mind
๐ World / Default Mode Network
๐ Excited to share a major update to our โMixture of Cognitive Reasonersโ (MiCRo) paper!
We ask: What benefits can we unlock by designing language models whose inner structure mirrors the brainโs functional specialization?
More below ๐ง ๐
cognitive-reasoners.epfl.ch
Excited to be part of this cool work led by Melika Honarmand!
We show that by selectively targeting VLM units that mirror the brainโs visual word form area, models develop dyslexic-like reading impairments, while leaving other abilities intact!! ๐ง ๐ค
Details in the ๐งต๐
Huge thanks to my amazing collaborators: @gretatuckute.bsky.social, @davidtyt.bsky.social, @neurotaha.bsky.social & advisors @abosselut.bsky.social and @mschrimpf.bsky.social!
You can find more about our paper on the project's website: language-to-cognition.epfl.ch
Paper: arxiv.org/abs/2503.01830
Now that the ICLR deadline is behind us, happy to share that From Language to Cognition has been accepted as an Oral at #EMNLP2025! ๐
Looking forward to seeing many of you in Suzhou ๐จ๐ณ
1/๐จ New preprint
How do #LLMsโ inner features change as they train? Using #crosscoders + a new causal metric, we map when features appear, strengthen, or fade across checkpointsโopening a new lens on training dynamics beyond loss curves & benchmarks.
#interpretability
NEW PAPER ALERT: Recent studies have shown that LLMs often lack robustness to distribution shifts in their reasoning. Our paper proposes a new method, AbstRaL, to augment LLMsโ reasoning robustness, by promoting their abstract thinking with granular reinforcement learning.
23.06.2025 14:32 โ ๐ 7 ๐ 3 ๐ฌ 1 ๐ 1Check out @bkhmsi.bsky.social 's great work on mixture-of-expert models that are specialized to represent the behavior of known brain networks.
18.06.2025 10:46 โ ๐ 3 ๐ 1 ๐ฌ 0 ๐ 011/ ๐ Links
Paper: arxiv.org/abs/2506.13331
Project Page: bkhmsi.github.io/mixture-of-c...
Code: github.com/bkhmsi/mixtu...
Models: huggingface.co/collections/...
In collaboration with: @cndesabbata.bsky.social, @eric-zemingchen.bsky.social, @mschrimpf.bsky.social, & @abosselut.bsky.social
10/ ๐งพ Conclusion:
MiCRo weaves together modularity, interpretability & brain-inspired design to build controllable and high-performing models, moving toward truly cognitively grounded LMs.
9/ ๐ก Key insights:
1. Minimal data (~3k samples) in Stage 1 can induce lasting specialization
2. Modular structure enables interpretability, control, and scalability (e.g., topโ2 routing can boost performance)
3. Approach generalizes across domains & base models
8/ ๐งฌ Brain alignment:
Neuroscience localizers (e.g., for language, multiple-demand) rediscover the corresponding experts in MiCRo, showing functional alignment with brain networks. However, ToM localizer fail to identify the social expert.
Figures for MiCRo-Llama & MiCRo-OLMo.
7/ ๐งฉ Steering & controllability:
Removing or emphasizing specific experts steers model behavior: Ablating logic expert hurts math accuracy; suppressing social reasoning improves math slightlyโshowcasing fine-grained control.
6/ ๐ Interpretable routing:
Early layers route most tokens to the language expert; deeper layers route to domain-relevant experts (e.g., logic expert for math), matching task semantics.
5/ ๐ Performance gains:
We evaluate on 6 reasoning benchmarks (MATH, GSM8K, MMLU, BBHโฆ), MiCRo outperforms both dense and โgeneralโexpertโ baselines: modular models with random specialist assignment in Stage 1.
4/ ๐ Training curriculum (3 stages):
โข Stage 1: Expert training on small curated domain-specific datasets (~3k samples)
โข Stage 2: Router training, experts frozen
โข Stage 3: End-to-end finetuning on large instruction corpus (939k samples)
This seeds specialization effectively.
3/ โ๏ธ Architecture:
We start with a pretrained model (e.g. Llamaโ3.2โ1B). Clone each layer into four experts. Then, a light router assigns tokens dynamically to a single expert (topโ1 routing) per layer. Keeping a comparable number of active parameters to the base model.