π Congrats @peterbhase.bsky.social + best wishes for the @schmidtsciences.bsky.social scientist + @stanfordnlp.bsky.social visiting researcher roles! π
Looking forward to your continued exciting research contributions in AI safety & interpretability, and your new grant-making contributions π₯
15.07.2025 01:15 β π 2 π 0 π¬ 0 π 0
Overdue job update β I am now:
- A Visiting Scientist at @schmidtsciences.bsky.social, supporting AI safety & interpretability
- A Visiting Researcher at Stanford NLP Group, working with @cgpotts.bsky.social
So grateful to keep working in this fascinating areaβand to start supporting others too :)
14.07.2025 17:06 β π 6 π 1 π¬ 3 π 0
π’ Speakers Announcement #AthNLP2025
Weβre thrilled to welcome a new lineup of brilliant minds to the ATHNLP stage!π
Meet our new NLP speakers shaping the future.
π
Dates: 4-10 September 2025 athnlp.github.io/2025/speaker...
#ATHNLP #NLP #AI #MachineLearning #Athens
27.06.2025 14:49 β π 6 π 1 π¬ 0 π 0
Looking forward to this year's edition! With great speakers: Ryan McDonald Yulan He @vn-ml.bsky.social @antonisa.bsky.social Raquel Fernandez @annarogers.bsky.social Preslav Nakov @mohitbansal.bsky.social @eunsol.bsky.social Marie-Catherine de Marnefffe !
06.06.2025 09:10 β π 6 π 3 π¬ 0 π 0
the journey has been a great pleasure for me too @jmincho.bsky.social π€, and looking forward to your exciting work in the future! π
21.05.2025 12:18 β π 1 π 0 π¬ 0 π 0
π₯ Huge CONGRATS to Jaemin + @jhucompsci.bsky.social! π
Very proud of his journey as an amazing researcher (covering groundbreaking, foundational research on important aspects of multimodality+other areas) & as an awesome, selfless mentor/teamplayer π
-- Apply to his group & grab him for gap year!
20.05.2025 18:18 β π 6 π 1 π¬ 1 π 0
π¨ Introducing our @tmlrorg.bsky.social paper βUnlearning Sensitive Information in Multimodal LLMs: Benchmark and Attack-Defense Evaluationβ
We present UnLOK-VQA, a benchmark to evaluate unlearning in vision-and-language models, where both images and text may encode sensitive or private information.
07.05.2025 18:54 β π 10 π 8 π¬ 1 π 0
aww thanks for the kind words @esteng.bsky.social -- completely my pleasure π€π
06.05.2025 00:04 β π 1 π 0 π¬ 0 π 0
π₯ BIG CONGRATS to Elias (and UT Austin)! Really proud of you -- it has been a complete pleasure to work with Elias and see him grow into a strong PI on *all* axes π€
Make sure to apply for your PhD with him -- he is an amazing advisor and person! π
05.05.2025 22:00 β π 11 π 4 π¬ 1 π 0
π΅ I'm going to be presenting PBT at #NAACL2025 today at 2PM! Come by poster session 2 if you want to hear about:
-- balancing positive and negative persuasion
-- improving LLM teamwork/debate
-- training models on simulated dialogues
With @mohitbansal.bsky.social and @peterbhase.bsky.social
30.04.2025 15:04 β π 8 π 3 π¬ 0 π 0
I will be presenting β¨Reverse Thinking Makes LLMs Stronger Reasonersβ¨at #NAACL2025!
In this work, we show
- Improvements across 12 datasets
- Outperforms SFT with 10x more data
- Strong generalization to OOD datasets
π
4/30 2:00-3:30 Hall 3
Let's chat about LLM reasoning and its future directions!
29.04.2025 23:21 β π 5 π 3 π¬ 1 π 0
βοΈ Heading to #NAACL2025 to present 3 main conf. papers, covering training LLMs to balance accepting and rejecting persuasion, multi-agent refinement for more faithful generation, and adaptively addressing varying knowledge conflict.
Reach out if you want to chat!
29.04.2025 17:52 β π 15 π 5 π¬ 1 π 0
Check out π¨CAPTUReπ¨ -- a new benchmark testing spatial reasoning by making VLMs count objects under occlusion.
SOTA VLMs (GPT-4o, Qwen2-VL, Intern-VL2) have high error rates on CAPTURe (but humans have low error β
) and models struggle to reason about occluded objects.
arxiv.org/abs/2504.15485
π§΅π
24.04.2025 15:14 β π 5 π 4 π¬ 1 π 0
PS. and here are the presentation time slots/details of all the
@iclr-conf.bsky.social and @tmlrorg.bsky.social papers π
21.04.2025 21:07 β π 2 π 0 π¬ 0 π 0
In Singapore for #ICLR2025 this week to present papers + keynotes π, and looking forward to seeing everyone -- happy to chat about research, or faculty+postdoc+phd positions, or simply hanging out (feel free to ping)! π
Also meet our awesome students/postdocs/collaborators presenting their work.
21.04.2025 16:49 β π 19 π 4 π¬ 1 π 1
π¨Real-world retrieval is messy: queries are ambiguous or docs conflict & have incorrect/irrelevant info. How can we jointly address these problems?
β‘οΈRAMDocs: challenging dataset w/ ambiguity, misinformation & noise
β‘οΈMADAM-RAG: multi-agent framework, debates & aggregates evidence across sources
π§΅β¬οΈ
18.04.2025 17:05 β π 14 π 7 π¬ 3 π 0
Excited to share my first paper as first author: "Task-Circuit Quantization" π
I led this work to explore how interpretability insights can drive smarter model compression. Big thank you to @esteng.bsky.social, Yi-Lin Sung, and @mohitbansal.bsky.social for mentorship and collaboration. More to come
16.04.2025 16:19 β π 5 π 2 π¬ 0 π 0
What if we could transform advanced math problems into abstract programs that can generate endless, verifiable problem variants?
Presenting EFAGen, which automatically transforms static advanced math problems into their corresponding executable functional abstractions (EFAs).
π§΅π
15.04.2025 19:37 β π 15 π 5 π¬ 1 π 1
π¨Announcing TaCQ π¨ a new mixed-precision quantization method that identifies critical weights to preserve. We integrate key ideas from circuit discovery, model editing, and input attribution to improve low-bit quant., w/ 96% 16-bit acc. at 3.1 avg bits (~6x compression)
π arxiv.org/abs/2504.07389
12.04.2025 14:19 β π 15 π 7 π¬ 1 π 1
π¨Exciting @icmlconf.bsky.social workshop alert π¨
Weβre thrilled to announce the #ICML2025 Workshop on Machine Unlearning for Generative AI (MUGen)!
β‘Join us in Vancouver this July to dive into cutting-edge research on unlearning in generative AI with top speakers and panelists! β‘
02.04.2025 15:59 β π 4 π 1 π¬ 1 π 1
ππ Big congrats to @archiki.bsky.social on being awarded the @Apple AI/ML PhD Fellowship, for her extensive contributions in evaluating+improving reasoning in language/reward models and their applications to new domains (ReCEval, RepARe, System-1.x, ADaPT, ReGAL, ScPO, UTGen, GrIPS)! #ProudAdvisor
27.03.2025 19:41 β π 3 π 1 π¬ 1 π 0
π¨UPCORE is our new method for balancing unlearning/forgetting with maintaining model performance.
Best part is it works by selecting a coreset from the data rather than changing the model, so it is compatible with any unlearning method, with consistent gains for 3 methods + 2 tasks!
25.02.2025 02:33 β π 4 π 2 π¬ 0 π 0
π¨ Introducing UPCORE, to balance deleting info from LLMs with keeping their other capabilities intact.
UPCORE selects a coreset of forget data, leading to a better trade-off across 2 datasets and 3 unlearning methods.
π§΅π
25.02.2025 02:23 β π 12 π 5 π¬ 2 π 1
Next, UTDebug enlarges UTGenβs sweet-spot with two key innovations:
1β£ Boosts unit test accuracy through test-time scaling.
2β£ Validates code by accepting edits only if the pass rate improvesβand backtracking otherwise.
4/4
05.02.2025 18:53 β π 3 π 0 π¬ 0 π 0
Key finding --> trade-off in UT generation: if UTs made too easy, they only catch trivial errors/donβt raise any errors; but if UTs made too hard, then the generator canβt predict the output! UTGen finds the sweet-spot between these two where the model generates tests that help in debugging.
3/4
05.02.2025 18:53 β π 3 π 0 π¬ 1 π 0
Generating good unit tests is challenging (even for humans) and requires:
1β£ Error uncovering β Unit tests must target edge cases to catch subtle bugs.
2β£ Code understanding β Knowing the correct output corresponding to a UT input.
2/4
05.02.2025 18:53 β π 3 π 0 π¬ 1 π 0
π¨ Check out "UTGen & UTDebug" for learning to automatically generate unit tests (i.e., discovering inputs which break your code) and then applying them to debug code with LLMs, with strong gains (>12% pass@1) across multiple models/datasets! (see details in π§΅π)
1/4
05.02.2025 18:53 β π 7 π 4 π¬ 1 π 0
π¨ Excited to announce UTGen and UTDebug, where we first learn to generate unit tests and then apply them to debugging generated code with LLMs, with strong gains (+12% pass@1) on LLM-based debugging across multiple models/datasets via inf.-time scaling and cross-validation+backtracking!
π§΅π
04.02.2025 19:13 β π 8 π 5 π¬ 0 π 0
Assistant Professor at @cs.ubc.caβ¬ and βͺ@vectorinstitute.aiβ¬ working on Natural Language Processing. Book: https://lostinautomatictranslation.com/
Ph.D. Student at UNC NLP | Prev: Apple, Amazon, Adobe (Intern) vaidehi99.github.io | Undergrad @IITBombay
Professor at UW; Researcher at Meta. LMs, NLP, ML. PNW life.
AI professor at Caltech. General Chair ICLR 2025.
http://www.yisongyue.com
Assistant prof at LTI CMU; Research scientist at Meta AI. Working on NLP: language interfaces, applied pragmatics, language-to-code, grounding. https://dpfried.github.io/
Google Chief Scientist, Gemini Lead. Opinions stated here are my own, not those of Google. Gemini, TensorFlow, MapReduce, Bigtable, Spanner, ML things, ...
Director, MIT Computational Psycholinguistics Lab. President, Cognitive Science Society. Chair of the MIT Faculty. Open access & open science advocate. He.
Lab webpage: http://cpl.mit.edu/
Personal webpage: https://www.mit.edu/~rplevy
Senior Research Scientist at Google DeepMind
https://swarooprm.github.io/
AI professor. Director, Foundations of Cooperative AI Lab at Carnegie Mellon. Head of Technical AI Engagement, Institute for Ethics in AI (Oxford). Author, "Moral AI - And How We Get There."
https://www.cs.cmu.edu/~conitzer/
Visiting Scientist at Schmidt Sciences. Visiting Researcher at Stanford NLP Group
Interested in AI safety and interpretability
Previously: Anthropic, AI2, Google, Meta, UNC Chapel Hill
Prof at Georgia Tech
https://faculty.cc.gatech.edu/~judy/
Machine Learning and Computer Vision Researcher
NLP Researcher
Language Understanding and Reasoning Lab
Stony Brook University
https://www.vita-group.space/ π¨βπ« UT Austin ML Professor (on leave)
https://www.xtxmarkets.com/ π¦ XTX Markets Research Director (NYC AI Lab)
Superpower is trying everything πͺ
Newest focus: training next-generation super intelligence - Preview above πΆ
The Association for Computational Linguistics (ACL) is a scientific and professional organization for people working on Natural Language Processing/Computational Linguistics.
Hash tags: #NLProc #ACL2025NLP
The Thirty-Eighth Annual Conference on Neural Information Processing Systems will be held in Vancouver Convention Center, on Tuesday, Dec 10 through Sunday, Dec 15.
https://neurips.cc/
Assistant Professor @Mila-Quebec.bsky.social
Co-Director @McGill-NLP.bsky.social
Researcher @ServiceNow.bsky.social
Alumni: @StanfordNLP.bsky.social, EdinburghNLP
Natural Language Processor #NLProc
Machine learning PhD student @ Blei Lab in Columbia University
Working in mechanistic interpretability, nlp, causal inference, and probabilistic modeling!
Previously at Meta for ~3 years on the Bayesian Modeling & Generative AI teams.
π www.sweta.dev
Ph.D. Student @ UNC Chapel Hill. Working on LLM efficiency and truthfulness.
https://duykhuongnguyen.github.io/
Ph.D. Student at UNC CS. Interested in multimodal video understanding&generation.
https://yui010206.github.io/
Senior Director, Research Scientist @ Meta FAIR + Visiting Prof @ NYU.
Pretrain+SFT: NLP from Scratch (2011). Multilayer attention+position encode+LLM: MemNet (2015). Recent (2024): Self-Rewarding LLMs & more!