Dimitri Meunier's Avatar

Dimitri Meunier

@dimitrimeunier.bsky.social

PhD, Gatsby, UCL

110 Followers  |  135 Following  |  15 Posts  |  Joined: 09.12.2024
Posts Following

Posts by Dimitri Meunier (@dimitrimeunier.bsky.social)

Excited to launch Principia, a nonprofit research organisation at the intersection of deep learning theory and AI safety.

Our goal is to develop theory for modern machine learning systems that can help us understand complex network behaviors, including those critical for AI safety and alignment.

1

16.02.2026 09:27 β€” πŸ‘ 91    πŸ” 26    πŸ’¬ 1    πŸ“Œ 1

At #NeurIPS ? Visit our posters! 🧡

Demystifying Spectral Feature Learning for Instrumental Variable Regression: #2600, Wed 11am

Regularized least squares learning with heavy-tailed noise is minimax optimal: #3012, Wed 4:30pm ✨spotlight✨

1/2

01.12.2025 18:31 β€” πŸ‘ 5    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Preview
Solenne Gaucher, la mathΓ©maticienne qui sort le genre de l’équation « La RelΓ¨veΒ Β». Chaque mois, « Le Monde CampusΒ Β» rencontre un jeune qui bouscule les normes dans son domaine. A 31Β ans, la docteure en mathΓ©matiques s’attaque aux biais algorithmiques de l’intelligence artificielle et a reΓ§u enΒ 2024Β un prixΒ pour ses travaux.

Solenne Gaucher, la mathΓ©maticienne qui sort le genre de l’équation

21.09.2025 14:11 β€” πŸ‘ 45    πŸ” 19    πŸ’¬ 0    πŸ“Œ 3

Congrats !

19.09.2025 10:02 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

AISTATS 2026 will be in Morocco!

30.07.2025 08:07 β€” πŸ‘ 35    πŸ” 10    πŸ’¬ 0    πŸ“Œ 0
Preview
Gaussian Processes and Reproducing Kernels: Connections and Equivalences This monograph studies the relations between two approaches using positive definite kernels: probabilistic methods using Gaussian processes, and non-probabilistic methods using reproducing kernel Hilb...

We've written a monograph on Gaussian processes and reproducing kernel methods (with @philipphennig.bsky.social, @sejdino.bsky.social and Bharath Sriperumbudur).

arxiv.org/abs/2506.17366

24.06.2025 08:35 β€” πŸ‘ 36    πŸ” 11    πŸ’¬ 0    πŸ“Œ 0

I have been looking at the draft for a while, I am surprised you had a hard time publishing it, it is a super cool work! Will it be included in the TorchDR package ?

27.06.2025 10:17 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Distributional Reduction paper with H. Van Assel, @ncourty.bsky.social, T. Vayer , C. Vincent-Cuaz, and @pfrossard.bsky.social is accepted at TMLR. We show that both dimensionality reduction and clustering can be seen as minimizing an optimal transport loss 🧡1/5. openreview.net/forum?id=cll...

27.06.2025 07:44 β€” πŸ‘ 33    πŸ” 9    πŸ’¬ 1    πŸ“Œ 1

Dimitri Meunier, Antoine Moulin, Jakub Wornbard, Vladimir R. Kostic, Arthur Gretton
Demystifying Spectral Feature Learning for Instrumental Variable Regression
https://arxiv.org/abs/2506.10899

13.06.2025 04:37 β€” πŸ‘ 1    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

Very much looking forward to this ! πŸ™Œ Stellar line-up

29.05.2025 14:41 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

new preprint with the amazing @lviano.bsky.social and @neu-rips.bsky.social on offline imitation learning! learned a lot :)

when the expert is hard to represent but the environment is simple, estimating a Q-value rather than the expert directly may be beneficial. lots of open questions left though!

27.05.2025 07:12 β€” πŸ‘ 18    πŸ” 3    πŸ’¬ 1    πŸ“Œ 1

TL;DR:

βœ… Theoretical guarantees for nonlinear meta-learning
βœ… Explains when and how aggregation helps
βœ… Connects RKHS regression, subspace estimation & meta-learning

Co-led with Zhu Li πŸ™Œ, with invaluable support from @arthurgretton.bsky.social, Samory Kpotufe.

26.05.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Even with nonlinear representation you can estimate the shared structure at a rate improving in both N (tasks) and n (samples per task). This leads to parametric rates on the target task!⚑

Bonus: for linear kernels, our results recover known linear meta-learning rates.

26.05.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Short answer: Yes βœ…

Key ideaπŸ’‘: Instead of learning each task well, under-regularise per-task estimators to better estimate the shared subspace in the RKHS.

Even though each task is noisy, their span reveals the structure we care about.

Bias-variance tradeoff in action.

26.05.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Our paper analyses a meta-learning setting where tasks share a finite dimensional subspace of a Reproducing Kernel Hilbert Space.

Can we still estimate this shared representation efficiently β€” and learn new tasks fast?

26.05.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Most prior theory assumes linear structure: All tasks share a linear representation, and task-specific parts are also linear.

Then: we can show improved learning rates as the number of tasks increases.

But reality is nonlinear. What then?

26.05.2025 16:50 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Meta-learning = using many related tasks to help learn new ones faster.

In practice (e.g. with neural nets), this usually means learning a shared representation across tasks β€” so we can train quickly on unseen ones.

But: what’s the theory behind this? πŸ€”

26.05.2025 16:50 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Nonlinear Meta-Learning Can Guarantee Faster Rates Many recent theoretical works on \emph{meta-learning} aim to achieve guarantees in leveraging similar representational structures from related tasks towards simplifying a target task. The main aim of ...

🚨 New paper accepted at SIMODS! 🚨
β€œNonlinear Meta-learning Can Guarantee Faster Rates”

arxiv.org/abs/2307.10870

When does meta learning work? Spoiler: generalise to new tasks by overfitting on your training tasks!

Here is why:
πŸ§΅πŸ‘‡

26.05.2025 16:50 β€” πŸ‘ 9    πŸ” 7    πŸ’¬ 2    πŸ“Œ 1

Dimitri Meunier, Zikai Shen, Mattes Mollenhauer, Arthur Gretton, Zhu Li
Optimal Rates for Vector-Valued Spectral Regularization Learning Algorithms
https://arxiv.org/abs/2405.14778

24.05.2024 04:06 β€” πŸ‘ 3    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

Mattes Mollenhauer, Nicole M\"ucke, Dimitri Meunier, Arthur Gretton: Regularized least squares learning with heavy-tailed noise is minimax optimal https://arxiv.org/abs/2505.14214 https://arxiv.org/pdf/2505.14214 https://arxiv.org/html/2505.14214

21.05.2025 06:14 β€” πŸ‘ 6    πŸ” 6    πŸ’¬ 1    πŸ“Œ 1
Post image

I have updated my slides on the maths of AI by an optimal pairing between AI and maths researchers ... speakerdeck.com/gpeyre/the-m...

20.05.2025 11:21 β€” πŸ‘ 25    πŸ” 3    πŸ’¬ 3    πŸ“Œ 0
Preview
Optimal Transport for Machine Learners Optimal Transport is a foundational mathematical theory that connects optimization, partial differential equations, and probability. It offers a powerful framework for comparing probability distributi...

I have cleaned a bit my lecture notes on Optimal Transport for Machine Learners arxiv.org/abs/2505.06589

13.05.2025 05:18 β€” πŸ‘ 119    πŸ” 29    πŸ’¬ 0    πŸ“Œ 0

Gabriel Peyr\'e
Optimal Transport for Machine Learners
https://arxiv.org/abs/2505.06589

13.05.2025 06:48 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

New ICML 2025 paper: Nested expectations with kernel quadrature.

We propose an algorithm to estimate nested expectations which provides orders of magnitude improvements in low-to-mid dimensional smooth nested expectations using kernel ridge regression/kernel quadrature.

arxiv.org/abs/2502.18284

08.05.2025 04:29 β€” πŸ‘ 14    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

Great talk by Aapo HyvΓ€rinen on non linear ICA at AISTATS 25’!

04.05.2025 02:57 β€” πŸ‘ 7    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Density Ratio-based Proxy Causal Learning Without Density Ratios πŸ€”

at #AISTATS2025

An alternative bridge function for proxy causal learning with hidden confounders.
arxiv.org/abs/2503.08371
Bozkurt, Deaner, @dimitrimeunier.bsky.social, Xu

02.05.2025 11:29 β€” πŸ‘ 7    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Interview of Statistics and ML Expert - Pierre Alquier
YouTube video by ML New Papers Interview of Statistics and ML Expert - Pierre Alquier

Link to the video: youtu.be/nLGBTMfTvr8?...

28.04.2025 11:01 β€” πŸ‘ 11    πŸ” 2    πŸ’¬ 0    πŸ“Œ 1

🀩 c’était super de te revoir Pierre!

01.05.2025 03:01 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Dinner in Siglap yesterday evening with the members of the ABI team & friends who are attending ICLR.

27.04.2025 09:41 β€” πŸ‘ 9    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

Optimality and Adaptivity of Deep Neural Features for Instrumental Variable Regression
#ICLR25

openreview.net/forum?id=ReI...

NNs
✨better than fixed-feature (kernel, sieve) when target has low spatial homogeneity,
✨more sample-efficient wrt Stage 1

Kim, @dimitrimeunier.bsky.social, Suzuki, Li

22.04.2025 22:23 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0