Soufiane Hayou's Avatar

Soufiane Hayou

@soufianehayou.bsky.social

Asst Professor at Johns Hopkins (AMS and DSAI). Previously: Simons Institute, Oxford stats, Polytechnique. I like to scale up things! https://www.soufianehayou.com/

23 Followers  |  40 Following  |  11 Posts  |  Joined: 26.11.2024  |  1.9074

Latest posts by soufianehayou.bsky.social on Bluesky

AISTATS 2026 will be in Morocco!

30.07.2025 08:07 β€” πŸ‘ 35    πŸ” 10    πŸ’¬ 0    πŸ“Œ 0

Shoutout to my collaborators Nikhil Ghosh and Bin Yu for their help with this project.

30.06.2025 21:26 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
PLoP: Precise LoRA Placement for Efficient Finetuning of Large Models Low-Rank Adaptation (LoRA) is a widely used finetuning method for large models. Its small memory footprint allows practitioners to adapt large models to specific tasks at a fraction of the cost of ful...

For more theoretical and empirical results, check our paper:
Paper: arxiv.org/abs/2506.20629
Code: github.com/soufiane001/...

30.06.2025 21:26 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

βœ… PLoP Consistently outperforms other strategies (Attn, MLP)
βœ… Works across different post-training scenarios: supervised fine-tuning, reinforcement learning
βœ… Minimal computational overhead
In the worst case, it ties with the best manual approach. Usually, it's better.

30.06.2025 21:26 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

NFN measures the alignment between each module (in the pretrained model) and the finetuning task. NFN is a cheap metric that can be calculated in one forward pass. It is based on a large width analysis of module-data alignment and is well suited for LoRA finetuning.

30.06.2025 21:26 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Our solution: PLoP (Precise LoRA Placement) 🎯
Instead of guessing, it automatically identifies the optimal modules for LoRA placement based on a notion of module-data alignment that we call NFN (Normalised Feature Norms).

30.06.2025 21:26 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

❌ Original LoRA paper: "Prioritize attention"
❌ Other papers: "Actually, put them in MLP"
❌ Everyone: just guessing and trying common target modules

30.06.2025 21:26 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

LoRA is amazing for finetuning large models cheaply, but WHERE you place the adapters makes a huge difference. Most people are just guessing where to put them (Attention, MLP, etc).

Meet "PLoP" (Precise LoRA Placement) 🎯, our new method for automatic LoRA placement 🧡

30.06.2025 21:26 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The recent surge in available Research Scientist positions is correlated with the growing need for innovative approaches to scale and improve Large Language Models (LLMs). This trend is also driven by factors such as researchers leaving established companies for startups!

19.01.2025 01:51 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Tuyo (Narcos Theme) (Extended Version)
YouTube video by Rodrigo Amarante - Topic Tuyo (Narcos Theme) (Extended Version)

By far, the best intro song in the history of humankind

www.youtube.com/watch?v=zNPX...

17.01.2025 01:00 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Are we hitting a wall with AI scaling? πŸ€”

That "plateau" you're seeing in scaling law charts might not be a fundamental limit, but a sign of suboptimal scaling strategies! I wrote a blogpost about this:

www.soufianehayou.com/blog/plateau...

10.01.2025 23:40 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Speculative sampling accelerates inference in LLMs by drafting future tokens which are verified in parallel. With @vdebortoli.bsky.social , A. Galashov & @arthurgretton.bsky.social , we extend this approach to (continuous-space) diffusion models: arxiv.org/abs/2501.05370

10.01.2025 16:30 β€” πŸ‘ 45    πŸ” 10    πŸ’¬ 0    πŸ“Œ 0

People compare AI to past historic breakthroughs πŸ”„ (industrial revolution, internet, etc), but there's a crucial difference: In previous advancements, humans remained the most intelligent beings. This time, we're creating something that could surpass us πŸ€–. It's a singularity!⚑️

29.12.2024 22:27 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@soufianehayou is following 20 prominent accounts