Antoine Bosselut's Avatar

Antoine Bosselut

@abosselut.bsky.social

Helping machines make sense of the world. Asst Prof @icepfl.bsky.social; Before: @stanfordnlp.bsky.social @uwnlp.bsky.social AI2 #NLProc #AI Website: https://atcbosselut.github.io/

494 Followers  |  130 Following  |  62 Posts  |  Joined: 17.11.2024  |  2.9166

Latest posts by abosselut.bsky.social on Bluesky

πŸŽ‰ Congratulations to Assistant Professors @abosselut.bsky.social (IC), @bunnech.bsky.social (IC & SV), and @mschrimpf.bsky.social (IC & SV) for being selected as #AI2050 Early Career Fellows by @schmidtsciences.bsky.social !

πŸ”— Full article: actu.epfl.ch/news/epfl-pr...

11.11.2025 16:48 β€” πŸ‘ 7    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Chenhao Tan's Homepage - recruiting Chenhao Tan's Homepage

Recruiting PhDs & postdocs for:

πŸ€– agents "taking over" science (hypogenic.ai and πŸ“Œ)
πŸ§ͺ Real scientists ➑️AI (e.g., materials, chem, physics)
πŸ“œ Theory + incentives for H-AI collab & credit (e.g., formalizing tacit knowledge)

new adventures for me, πŸ”„ if you can! πŸ™Œ

chenhaot.com/recruiting.h...

03.11.2025 20:06 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Preview
EPFL AI Center and Swiss AI Initiative Postdoctoral Fellowships The 2nd call is now open with a deadline for submissions of 3 November (17.00 CET)!Applications are encouraged from researchers at the postdoctoral level with a keen interest in collaborative, interdi...

EPFL AI Center Postdocs: www.epfl.ch/research/fun...

NLP Lab Postdoc: docs.google.com/document/d/1...

14.10.2025 17:56 β€” πŸ‘ 2    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

If you're interested in doing a postdoc at @icepfl.bsky.social , there's still time to apply for the @epfl-ai-center.bsky.social postdoctoral fellowships.

Apart from this, I'm also recruiting postdocs in developing novel training algorithms for reasoning models and agentic AI.

14.10.2025 17:56 β€” πŸ‘ 8    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

Join us again at #MELT workshop (520D) at #COLM2025 to hear from @ImanolSchlag about #Apertus, the largest multilingual LLM trained on over 1000 languages.

10.10.2025 15:36 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Kicking off #MELT workshop at #COLM2025 with Monojit Choudhury talking about "Meta-Cultural Competence: What LLMs Should Know About Culture to Serve the Next Billion Users" !

10.10.2025 13:15 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 1

Come join us in 520D (all the way down the hall and around the corner) at #COLM2025 for the first workshop on multilingual and equitable language technologies!

10.10.2025 12:53 β€” πŸ‘ 2    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

Very happy this paper got accepted to NeurIPS 2025 as a Spotlight! 😁

Main takeaway: In mechanistic interpretability, we need assumptions about how DNNs encode concepts in their representations (eg, the linear representation hypothesis). Without them, we can claim any DNN implements any algorithm!

01.10.2025 15:00 β€” πŸ‘ 25    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Post image

What's the right unit of analysis for understanding LLM internals? We explore in our mech interp survey (a major update from our 2024 ms).

We’ve added more recent work and more immediately actionable directions for future work. Now published in Computational Linguistics!

01.10.2025 14:03 β€” πŸ‘ 40    πŸ” 14    πŸ’¬ 2    πŸ“Œ 2

I don't see why the answer would be no but since you specifically say "October", what if we submitted to ARR in July and want to do early submission to ACL 2026 ?

29.09.2025 20:03 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

1/🚨 New preprint

How do #LLMs’ inner features change as they train? Using #crosscoders + a new causal metric, we map when features appear, strengthen, or fade across checkpointsβ€”opening a new lens on training dynamics beyond loss curves & benchmarks.

#interpretability

25.09.2025 14:02 β€” πŸ‘ 14    πŸ” 6    πŸ’¬ 2    πŸ“Œ 0
Post image

πŸ’‘Can we optimize LLMs to be more creative?
Introducing Creative Preference Optimization (CrPO) and MuCE (Multi-task Creativity Evaluation Dataset).
Result: More novel, diverse, surprising textβ€”without losing quality!
πŸ“ Appearing at #EMNLP2025

22.09.2025 13:43 β€” πŸ‘ 6    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0

Special thanks to everyone that participated in this journey!

03.09.2025 09:26 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
swiss-ai (Swiss AI Initiative) Org profile for Swiss AI Initiative on Hugging Face, the AI community building the future.

(5) Transparency: We're fully open, pairing our weights with a full suite of reproduction artifacts.

Check out our artifacts and technical report here: huggingface.co/swiss-ai

03.09.2025 09:26 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

(4) Multilinguality: We pretrain the model on 15T tokens from 1811 languages, and post-train with 3.8 M examples from 149 languages

03.09.2025 09:26 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

(3) Memorization Prevention: Adopting the Goldfish objective, we suppress verbatim recall and reduce risks of memorization

03.09.2025 09:26 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

(2) Data Compliance: we pretrained exclusively on openly available data, retroactively respecting robots.txt exclusions and filtering for copyrighted, non-permissive, toxic, and personally identifiable content

03.09.2025 09:26 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

What makes Apertus special?
(1) Scale: Apertus-70B is the first fully open model to be trained at 70B parameter scale on 15T tokens, requiring us to scale out training to 4096 GPUs at
@cscsch.bsky.social

03.09.2025 09:26 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

The next generation of open LLMs should be inclusive, compliant, and multilingual by design. That’s why we @icepfl.bsky.social @ethz.ch @cscsch.bsky.social ) built Apertus.

03.09.2025 09:26 β€” πŸ‘ 25    πŸ” 8    πŸ’¬ 2    πŸ“Œ 2
Preview
Apertus: a fully open, transparent, multilingual language model - EPFL AI Center EPFL, ETH Zurich and the Swiss National Supercomputing Centre (CSCS) released Apertus today, Switzerland’s first large-scale, open, multilingual language model β€” a milestone in generative AI for trans...

EPFL, @ethz.ch and the @cscsch.bsky.social released Apertus today, Switzerland’s first large-scale, open, multilingual language model β€” a milestone in generative AI for transparency and diversity.

Find out more here: ai.epfl.ch/apertus-a-fu...

@abosselut.bsky.social @icepfl.bsky.social

02.09.2025 09:46 β€” πŸ‘ 18    πŸ” 7    πŸ’¬ 0    πŸ“Œ 2
Post image

EPFL, ETH Zurich & CSCS just released Apertus, Switzerland’s first fully open-source large language model.
Trained on 15T tokens in 1,000+ languages, it’s built for transparency, responsibility & the public good.

Read more: actu.epfl.ch/news/apertus...

02.09.2025 11:48 β€” πŸ‘ 54    πŸ” 29    πŸ’¬ 1    πŸ“Œ 6

Very happy to see that Pleias multilingual data processing pipelines have contributed to the largest open pretraining project in Europe.

From their tech report: huggingface.co/swiss-ai/Ape...

02.09.2025 16:46 β€” πŸ‘ 30    πŸ” 10    πŸ’¬ 2    πŸ“Œ 0
Preview
Apertus: ein neues Sprachmodell fΓΌr die Schweiz

Die Schweiz steigt ins Rennen der grossen Sprachmodelle ein. Unter dem Namen #Apertus verΓΆffentlichen @ethz.ch, @icepfl.bsky.social und das @cscsch.bsky.social das erste vollstΓ€ndig offene, mehrsprachige #LLM des Landes.

FΓΌrs MAZ habe ich Apertus kurz analysiert:

www.maz.ch/news/apertus...

02.09.2025 08:33 β€” πŸ‘ 25    πŸ” 7    πŸ’¬ 3    πŸ“Œ 1

Thank you for your incredible work!

02.09.2025 18:23 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image Post image Post image Post image

recently gave a talk on <Reality Checks> at two venues, and discussed (and rambled) about how leaderboard chasing is awesome (and we want it to continue) but that this isn't easy because everyone (me! me! me!) wants to write more papers.

the link to the slide deck in the reply.

12.08.2025 02:04 β€” πŸ‘ 24    πŸ” 5    πŸ’¬ 3    πŸ“Œ 0
Post image

🚨New Preprint!

In multilingual models, the same meaning can take far more tokens in some languages, penalizing users of underrepresented languages with worse performance and higher API costs. Our Parity-aware BPE algorithm is a step toward addressing this issue: 🧡

11.08.2025 12:28 β€” πŸ‘ 28    πŸ” 7    πŸ’¬ 3    πŸ“Œ 0
Preview
EPFL NLP Postdoctoral Scholar Posting - Swiss AI LLMs The EPFL Natural Language Processing (NLP) lab is looking to hire a postdoctoral researcher candidate in the area of multilingual LLM design, training, and evaluation. This postdoctoral position is as...

The EPFL NLP lab is looking to hire a postdoctoral researcher on the topic of designing, training, and evaluating multilingual LLMs:

docs.google.com/document/d/1...

Come join our dynamic group in beautiful Lausanne!

04.08.2025 15:54 β€” πŸ‘ 21    πŸ” 12    πŸ’¬ 0    πŸ“Œ 1
Kaiserslautern, Germany

Kaiserslautern, Germany

πŸ“£ Life update: Thrilled to announce that I’ll be starting as faculty at the Max Planck Institute for Software Systems this Fall!

I’ll be recruiting PhD students in the upcoming cycle, as well as research interns throughout the year: lasharavichander.github.io/contact.html

22.07.2025 04:12 β€” πŸ‘ 90    πŸ” 12    πŸ’¬ 13    πŸ“Œ 4
Preview
A language model built for the public good Β Β Β  - EPFL AI Center ETH Zurich and EPFL will release a large language model (LLM) developed on public infrastructure. Trained on the β€œAlps” supercomputer at the Swiss National Supercomputing Centre (CSCS), the new LLM ma...

EPFL and ETH ZΓΌrich are building together a Swiss made LLM from scratch.
Fully open and multilingual, the model is trained on CSCS's supercomputer "Alps" and supports sovereign, transparent, and responsible AI in Switzerland and beyond.
Read more here: ai.epfl.ch/a-language-m...
#ResponsibleAI

09.07.2025 07:26 β€” πŸ‘ 10    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

Check out Silin's paper done in collaboration with Apple on reinforcing abstract thinking in reasoning traces!

23.06.2025 18:55 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

@abosselut is following 20 prominent accounts