The EPFL NLP lab is looking to hire a postdoctoral researcher on the topic of designing, training, and evaluating multilingual LLMs:
docs.google.com/document/d/1...
Come join our dynamic group in beautiful Lausanne!
@smamooler.bsky.social
PhD Candidate at @icepfl.bsky.social | Ex Research Intern@Google DeepMind ๐ฉ๐ปโ๐ป Working on multi-modal AI reasoning models in scientific domains https://smamooler.github.io/
The EPFL NLP lab is looking to hire a postdoctoral researcher on the topic of designing, training, and evaluating multilingual LLMs:
docs.google.com/document/d/1...
Come join our dynamic group in beautiful Lausanne!
Also: not the usual #NLProc topic, but if you're working on genetic AI models, Iโd love to connect! I'm exploring the intersection of behavioral genomics and multi-modal AI for behavior understanding.
26.07.2025 15:13 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0Excited to attend #ACL2025NLP in Vienna next week!
Iโll be presenting our #NAACL2025 paper, PICLe, at the first workshop on Large Language Models and Structure Modeling (XLLM) on Friday. Come by our poster if youโre into NER and ICL with pseudo-annotation.
arxiv.org/abs/2412.11923
๐จ New Preprint!!
Thrilled to share with you our latest work: โMixture of Cognitive Reasonersโ, a modular transformer architecture inspired by the brainโs functional networks: language, logic, social reasoning, and world knowledge.
1/ ๐งต๐
๐ โจ Introducing MELT Workshop 2025: Multilingual, Multicultural, and Equitable Language Technologies
A workshop on building inclusive, culturally-aware LLMs!
๐ง Bridging the language divide in AI
๐
October 10, 2025 Co-located with @colmweb.org
๐ melt-workshop.github.io
#MeltWorkshop2025
Super excited to share that our paper "A Logical Fallacy-Informed Framework for Argument Generation" has received the Outstanding Paper Award ๐๐ at NAACL 2025!
Paper: aclanthology.org/2025.naacl-l...
Code: github.com/lucamouchel/...
#NAACL2025
paper: arxiv.org/abs/2412.11923
code: github.com/sMamooler/PI...
Couldn't attend @naaclmeeting.bsky.social in person as I didn't get a visa on time ๐คทโโ๏ธ My colleague @mismayil.bsky.social will present PICLe on my behalf today, May 1st, at 3:15 pm in RUIDOSO. Feel free to reach out if you want to chat more!
01.05.2025 08:39 โ ๐ 2 ๐ 0 ๐ฌ 1 ๐ 0Check out VinaBench, our new #CVPR2025 paper. We introduce a benchmark for faithful and consistent visual narratives.
Paper: arxiv.org/abs/2503.20871
Project Page: silin159.github.io/Vina-Bench/
Excited to share that we will present PICLe at @naaclmeeting.bsky.social main conference!
10.03.2025 10:21 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 1๐จ New Preprint!!
LLMs trained on next-word prediction (NWP) show high alignment with brain recordings. But what drives this alignmentโlinguistic structure or world knowledge? And how does this alignment evolve during training? Our new paper explores these questions. ๐๐งต
Lots of great news out of the EPFL NLP lab these last few weeks. We'll be at @iclr-conf.bsky.social and @naaclmeeting.bsky.social in April / May to present some of our work in training dynamics, model representations, reasoning, and AI democratization. Come chat with us during the conference!
25.02.2025 09:18 โ ๐ 25 ๐ 12 ๐ฌ 1 ๐ 0๐จ New Paper!
Can neuroscience localizers uncover brain-like functional specializations in LLMs? ๐ง ๐ค
Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks!
w/ @gretatuckute.bsky.social, @abosselut.bsky.social, @mschrimpf.bsky.social
๐งต๐
๐ Amazing collaboration with my co-authors and advisors
@smontariol.bsky.social, @abosselut.bsky.social,
@trackingskills.bsky.social
๐ Check out the full paper here: arxiv.org/pdf/2412.11923
17.12.2024 14:51 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0๐ We evaluate PICLe on 5 biomedical NED datasets and find:
โจ With zero human annotations, PICLe outperforms ICL in low-resource settings, where limited gold examples can be used as in-context demonstrations!
โ๏ธ How does PICLe work?
1๏ธโฃ LLMs annotate demonstrations in a zero-shot first pass.
2๏ธโฃ Synthetic demos are clustered, and in-context sets are sampled.
3๏ธโฃ Entity mentions are predicted using each set independently.
4๏ธโฃ Self-verification selects the final predictions.
๐ก Building on our findings, we introduce PICLe: a framework for in-context learning powered by noisy, pseudo-annotated demonstrations. ๐ ๏ธ No human labels, no problem! ๐
17.12.2024 14:51 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0๐ Key finding: A semantic mapping between demonstration context and label is essential for in-context task transfer. BUT even weak semantic mappings can provide enough signal for effective adaptation in NED!
17.12.2024 14:51 โ ๐ 0 ๐ 0 ๐ฌ 1 ๐ 0๐ Itโs unclear which demonstration attributes enable in-context learning in tasks that require structured, open-ended predictions (such as NED).
We use perturbation schemes that create demonstrations with varying correctness levels to analyze key demonstration attributes.
๐ Introducing PICLe: a framework for in-context named-entity detection (NED) using pseudo-annotated demonstrations.
๐ฏ No human labeling neededโyet it outperforms few-shot learning with human annotations!
#AI #NLProc #LLMs #ICL #NER