Marius Mosbach's Avatar

Marius Mosbach

@mariusmosbach.bsky.social

#NLP Postdoc at Mila - Quebec AI Institute & McGill University mariusmosbach.com

923 Followers  |  275 Following  |  28 Posts  |  Joined: 02.10.2023  |  1.8173

Latest posts by mariusmosbach.bsky.social on Bluesky

Post image

Our new paper in #PNAS (bit.ly/4fcWfma) presents a surprising finding—when words change meaning, older speakers rapidly adopt the new usage; inter-generational differences are often minor.

w/ Michelle Yang, ‪@sivareddyg.bsky.social‬ , @msonderegger.bsky.social‬ and @dallascard.bsky.social‬👇(1/12)

29.07.2025 12:05 — 👍 34    🔁 17    💬 3    📌 2

🚨Job Alert
W2 (TT W3) Professorship in Computer Science "AI for People & Society"
@saarland-informatics-campus.de/@uni-saarland.de is looking to appoint an outstanding individual in the field of AI for people and society who has made significant contributions in one or more of the following areas:

18.07.2025 07:11 — 👍 14    🔁 18    💬 1    📌 0
Kaiserslautern, Germany

Kaiserslautern, Germany

📣 Life update: Thrilled to announce that I’ll be starting as faculty at the Max Planck Institute for Software Systems this Fall!

I’ll be recruiting PhD students in the upcoming cycle, as well as research interns throughout the year: lasharavichander.github.io/contact.html

22.07.2025 04:12 — 👍 89    🔁 12    💬 13    📌 4

I'm at #ICML in Vancouver this week, hit me up if you want to chat about pre-training experiments or explainable machine learning.

You can find me at these posters:

Tuesday: How Much Can We Forget about Data Contamination? icml.cc/virtual/2025...

14.07.2025 14:49 — 👍 1    🔁 1    💬 1    📌 0

Congrats!

14.07.2025 20:49 — 👍 1    🔁 0    💬 0    📌 0
Paper title "The Non-Linear Representation Dilemma: Is Causal Abstraction Enough for Mechanistic Interpretability?" with the paper's graphical abstract showing how more powerful alignment maps between a DNN and an algorithm allow more complex features to be found and more "accurate" abstractions.

Paper title "The Non-Linear Representation Dilemma: Is Causal Abstraction Enough for Mechanistic Interpretability?" with the paper's graphical abstract showing how more powerful alignment maps between a DNN and an algorithm allow more complex features to be found and more "accurate" abstractions.

Mechanistic interpretability often relies on *interventions* to study how DNNs work. Are these interventions enough to guarantee the features we find are not spurious? No!⚠️ In our new paper, we show many mech int methods implicitly rely on the linear representation hypothesis🧵

14.07.2025 12:15 — 👍 65    🔁 12    💬 1    📌 1
Video thumbnail

Have you ever wondered whether a few times of data contamination really lead to benchmark overfitting?🤔 Then our latest #ICML paper about the effect of data contamination on LLM evals might be for you!🚀

Paper: arxiv.org/abs/2410.03249
👇🧵

08.07.2025 06:42 — 👍 12    🔁 1    💬 1    📌 2
Post image

💡Beyond math/code, instruction following with verifiable constraints is suitable to be learned with RLVR.
But the set of constraints and verifier functions is limited and most models overfit on IFEval.
We introduce IFBench to measure model generalization to unseen constraints.

03.07.2025 21:06 — 👍 29    🔁 5    💬 1    📌 1
Post image

A blizzard is raging through Montreal when your friend says “Looks like Florida out there!” Humans easily interpret irony, while LLMs struggle with it. We propose a 𝘳𝘩𝘦𝘵𝘰𝘳𝘪𝘤𝘢𝘭-𝘴𝘵𝘳𝘢𝘵𝘦𝘨𝘺-𝘢𝘸𝘢𝘳𝘦 probabilistic framework as a solution.
Paper: arxiv.org/abs/2506.09301 to appear @ #ACL2025 (Main)

26.06.2025 15:52 — 👍 15    🔁 7    💬 1    📌 4
Preview
02 | Gauthier Gidel: Bridging Theory and Deep Learning, Vibes at Mila, and the Effects of AI on Art Behind the Research of AI · Episode

Started a new podcast with @tomvergara.bsky.social !

Behind the Research of AI:
We look behind the scenes, beyond the polished papers 🧐🧪

If this sounds fun, check out our first "official" episode with the awesome Gauthier Gidel
from @mila-quebec.bsky.social :

open.spotify.com/episode/7oTc...

25.06.2025 15:54 — 👍 17    🔁 6    💬 1    📌 0
Preview
Not All Data Are Unlearned Equally Machine unlearning is concerned with the task of removing knowledge learned from particular data points from a trained model. In the context of large language models (LLMs), unlearning has recently re...

Cool work! You might be interested in our recent work on another problem of existing unlearning methods: arxiv.org/abs/2504.05058

21.06.2025 15:37 — 👍 2    🔁 0    💬 0    📌 0
Post image

Interested in shaping the progress of responsible AI and meeting leading researchers in the field? SoLaR@COLM 2025 is looking for paper submissions and reviewers!

🤖 ML track: algorithms, math, computation
📚 Socio-technical track: policy, ethics, human participant research

17.06.2025 17:46 — 👍 8    🔁 1    💬 1    📌 1
Post image

"Build the web for agents, not agents for the web"

This position paper argues that rather than forcing web agents to adapt to UIs designed for humans, we should develop a new interface optimized for web agents, which we call Agentic Web Interface (AWI).

arxiv.org/abs/2506.10953

14.06.2025 04:17 — 👍 6    🔁 4    💬 0    📌 0
Post image

Excited to share the results of my recent internship!

We ask 🤔
What subtle shortcuts are VideoLLMs taking on spatio-temporal questions?

And how can we instead curate shortcut-robust examples at a large-scale?

We release: MVPBench

Details 👇🔬

13.06.2025 14:47 — 👍 16    🔁 5    💬 1    📌 0

Congrats Sarah!! They are lucky to have you 💪

13.06.2025 21:10 — 👍 0    🔁 0    💬 1    📌 0
Post image

New paper in Interspeech 2025 🚨
@interspeech.bsky.social

A Robust Model for Arabic Dialect Identification using Voice Conversion

Paper 📝 arxiv.org/pdf/2505.24713
Demo 🎙️https://shorturl.at/rrMm6

#Arabic #SpeechTech #NLProc #AI #Speech #ArabicDialects #Interspeech2025 #ArabicNLP

10.06.2025 10:07 — 👍 1    🔁 2    💬 1    📌 0
Post image

Do LLMs hallucinate randomly? Not quite.

Our #ACL2025 (Main) paper shows that hallucinations under irrelevant contexts follow a systematic failure mode — revealing how LLMs generalize using abstract classes + context cues, albeit unreliably.

📎 Paper: arxiv.org/abs/2505.22630 1/n

06.06.2025 18:09 — 👍 46    🔁 18    💬 1    📌 3

Congrats Elinor!

30.05.2025 20:58 — 👍 1    🔁 0    💬 0    📌 0
Post image

Chain-of-Thought (CoT) reasoning lets LLMs solve complex tasks, but long CoTs are expensive. How short can they be while still working? Our new ICML paper tackles this foundational question.

05.05.2025 12:25 — 👍 12    🔁 2    💬 2    📌 0
Title slide: Processing Trans Languaging - Vagrant Gautam (they/xe), Saarland University, with a very brightly patterned background featuring colourful people and math symbols.

Title slide: Processing Trans Languaging - Vagrant Gautam (they/xe), Saarland University, with a very brightly patterned background featuring colourful people and math symbols.

Come to my keynote tomorrow at the first official @queerinai.com workshop at #NAACL2025 to hear about how trans languaging is complex and cool, and how this makes it extra difficult to process computationally. I will have SO many juicy examples!

03.05.2025 20:52 — 👍 44    🔁 14    💬 3    📌 0
Post image

Deadline extended! ⏳

The Actionable Interpretability Workshop at #ICML2025 has moved its submission deadline to May 19th. More time to submit your work 🔍🧠✨ Don’t miss out!

03.05.2025 20:00 — 👍 4    🔁 3    💬 0    📌 0
Preview
Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and Evaluation Marius Mosbach, Tiago Pimentel, Shauli Ravfogel, Dietrich Klakow, Yanai Elazar. Findings of the Association for Computational Linguistics: ACL 2023. 2023.

Very interesting work! We also compared ICL and finetuning a while ago. You might find it relevant: aclanthology.org/2023.finding...

02.05.2025 20:47 — 👍 8    🔁 0    💬 1    📌 0

Check out Gaurav's video on their #NAACL paper and find @adadtur.bsky.social at the conference 👇

02.05.2025 01:41 — 👍 11    🔁 1    💬 0    📌 0

I'll be at #NAACL2025:

🖇️To present my paper "Superlatives in Context", showing how the interpretation of superlatives is very context dependent and often implicit, and how LLMs handle such semantic underspecification

🖇️And we will present RewardBench on Friday

Reach out if you want to chat!

27.04.2025 20:00 — 👍 28    🔁 5    💬 1    📌 1

👋🇨🇦🇩🇪

27.04.2025 18:51 — 👍 1    🔁 0    💬 0    📌 0

I’m really excited about Diffusion Steering Lens, an intuitive and elegant new “logit lens” technique for decoding the attention and MLP blocks of vision transformers!

Vision is much more expressive than language, so some new mech interp rules apply:

25.04.2025 13:36 — 👍 11    🔁 3    💬 0    📌 0
Post image

💡 New ICLR paper! 💡
"On Linear Representations and Pretraining Data Frequency in Language Models":

We provide an explanation for when & why linear representations form in large (or small) language models.

Led by @jackmerullo.bsky.social, w/ @nlpnoah.bsky.social & @sarah-nlp.bsky.social

25.04.2025 01:55 — 👍 42    🔁 12    💬 3    📌 3
Logo for MIB: A Mechanistic Interpretability Benchmark

Logo for MIB: A Mechanistic Interpretability Benchmark

Lots of progress in mech interp (MI) lately! But how can we measure when new mech interp methods yield real improvements over prior work?

We propose 😎 𝗠𝗜𝗕: a 𝗠echanistic 𝗜nterpretability 𝗕enchmark!

23.04.2025 18:15 — 👍 49    🔁 15    💬 1    📌 6
Preview
DeepSeek-R1 Thoughtology: Let's <think> about LLM Reasoning Large Reasoning Models like DeepSeek-R1 mark a fundamental shift in how LLMs approach complex problems. Instead of directly producing an answer for a given input, DeepSeek-R1 creates detailed multi-st...

We also saw some interesting overthinking failure cases across various tasks in our work: arxiv.org/abs/2504.07128 Much work needs to be done to better understand the behavior of reasoning LLMs!

21.04.2025 23:50 — 👍 2    🔁 0    💬 1    📌 0

Paper title of the year so far. I will be back ... have to read the paper now. Great work @saxon.me !

21.04.2025 23:43 — 👍 5    🔁 0    💬 1    📌 0

@mariusmosbach is following 20 prominent accounts