Some text in heuristics/separation of results appear twice? Good post otherwise
14.12.2025 18:38 β π 2 π 0 π¬ 1 π 0@antoine-mln.bsky.social
doing a phd in RL/online learning on questions related to exploration and adaptivity > https://antoine-moulin.github.io/
Some text in heuristics/separation of results appear twice? Good post otherwise
14.12.2025 18:38 β π 2 π 0 π¬ 1 π 0Thank you for coming and for the great discussion!!
07.12.2025 22:58 β π 1 π 0 π¬ 1 π 0So excited for the RL theory meets experiment workshop tomorrow at NeurIPS: arlet-workshop.github.io/neurips2025/...
Talks look AMAZING and you can hear me be the foolish experimentalist on a panel
The paper looks cool! And it seems well-written :). I did not expect these exponents in Theorem 4.2 π
26.10.2025 14:56 β π 3 π 0 π¬ 1 π 0Very excited to share our preprint: Self-Speculative Masked Diffusions
We speed up sampling of masked diffusion models by ~2x by using speculative sampling and a hybrid non-causal / causal transformer
arxiv.org/abs/2510.03929
w/ @vdebortoli.bsky.social, Jiaxin Shi, @arnauddoucet.bsky.social
We're finally out of stealth: percepta.ai
We're a research / engineering team working together in industries like health and logistics to ship ML tools that drastically improve productivity. If you're interested in ML and RL work that matters, come join us π
I am happy to share that our paper "Unsupervised Learning for Optimal Transport plan prediction between unbalanced graphs" was accepted at Neurips 2025 ! π₯³
Huge thanks to my co-authors @rflamary.bsky.social and Bertrand Thirion !
arxiv.org/abs/2506.12025
(1/5)
News π Weβre thrilled to announce our final panelist: David Silver!
Donβt miss David and our amazing lineup of speakersβsubmit your latest RL work to our NeurIPS workshop.
π
Extended deadline: Sept 2 (AoE)
We've extended the deadline for our workshop's calls for papers/ideas! Submit your work by August 29 AoE. Instructions on the website: arlet-workshop.github.io/neurips2025/...
18.08.2025 10:46 β π 3 π 2 π¬ 0 π 0The OpenReview link for our calls (for papers and ideas) is available, submit here: openreview.net/group?id=Neu...
We look forward to receiving your submissions!
last year's edition was so much fun I'm really looking forward to this one!! join us in San Diego :))
28.07.2025 18:06 β π 4 π 0 π¬ 0 π 0Was it recorded? π€
19.06.2025 15:08 β π 1 π 0 π¬ 1 π 0Join us for Nneka's presentation tomorrow! Last talk before the summer break.
09.06.2025 17:43 β π 9 π 3 π¬ 0 π 0Join us tomorrow for Dave's talk! He will present his recent work on randomised exploration, which received an outstanding paper award at ALT 2025 earlier this year.
02.06.2025 14:55 β π 4 π 1 π¬ 0 π 0link: arxiv.org/abs/2505.19946
27.05.2025 07:16 β π 3 π 0 π¬ 0 π 0new preprint with the amazing @lviano.bsky.social and @neu-rips.bsky.social on offline imitation learning! learned a lot :)
when the expert is hard to represent but the environment is simple, estimating a Q-value rather than the expert directly may be beneficial. lots of open questions left though!
π¨ New paper accepted at SIMODS! π¨
βNonlinear Meta-learning Can Guarantee Faster Ratesβ
arxiv.org/abs/2307.10870
When does meta learning work? Spoiler: generalise to new tasks by overfitting on your training tasks!
Here is why:
π§΅π
Dhruv Rohatgi will be giving a lecture on our recent work on comp-stat tradeoffs in next-token prediction at the RL Theory virtual seminar series (rl-theory.bsky.social) tomorrow at 2pm EST! Should be a fun talk---come check it out!!
26.05.2025 19:19 β π 11 π 5 π¬ 1 π 0new work on computing distances between stochastic processes ***based on sample paths only***! we can now:
- learn distances between Markov chains
- extract "encoder-decoder" pairs for representation learning
- with sample- and computational-complexity guarantees
read on for some quick details..
1/n
A new blog post with intuitions behind continuous-time Markov chains, a building block of diffusion language models, like @inceptionlabs.bsky.social's Mercury and Gemini Diffusion. This post touches on different ways of looking at Markov chains, connections to point processes, and more.
22.05.2025 15:14 β π 22 π 5 π¬ 1 π 0oh the inference blog posts are back π₯°
22.05.2025 15:19 β π 0 π 0 π¬ 1 π 0Mattes Mollenhauer, Nicole M\"ucke, Dimitri Meunier, Arthur Gretton: Regularized least squares learning with heavy-tailed noise is minimax optimal https://arxiv.org/abs/2505.14214 https://arxiv.org/pdf/2505.14214 https://arxiv.org/html/2505.14214
21.05.2025 06:14 β π 6 π 6 π¬ 1 π 1Later today, Sikata and Marcel will talk about their recent work on oracle-efficient RL with ensembles. Join us!
20.05.2025 15:48 β π 6 π 4 π¬ 0 π 0Excited to share what I've been up to: bringing text diffusion to Gemini!
Diffusion models are _fast_, and hold immense promise to challenge autoregressive models as the de facto standard for language modeling.
omg thanks
16.05.2025 00:00 β π 0 π 0 π¬ 0 π 0Community events and tutorials, list from the website
Workshops, list from the website
The tutorials, workshops, and community events for #COLT2025 have been announced!
Exciting topics, and impressive slate of speakers and events, on June 30! The workshops have calls for contributions (β° May 16, 19, and 25): check them out!
learningtheory.org/colt2025/ind...
Announcing the first workshop on Foundations of Post-Training (FoPT) at COLT 2025!
π Soliciting abstracts/posters exploring theoretical & practical aspects of post-training and RL with language models!
ποΈ Deadline: May 19, 2025
looking forward to giving a talk at the 2025 GHOST day in the beautiful city of Poznan (Poland) --- hope to see you there this weekend!
06.05.2025 19:06 β π 7 π 1 π¬ 1 π 0Is Best-of-N really the best we can do for language model inference?
New paper (appearing at ICML) led by the amazing Audrey Huang (ahahaudrey.bsky.social) with Adam Block, Qinghua Liu, Nan Jiang, and Akshay Krishnamurthy (akshaykr.bsky.social).
1/11
Spectral Representation for Causal Estimation with Hidden Confounders
at #AISTATS2025
A spectral method for causal effect estimation with hidden confounders, for instrumental variable and proxy causal learning
arxiv.org/abs/2407.10448
Haotian Sun, @antoine-mln.bsky.social, Tongzheng Ren, Bo Dai