A strange trend I've noticed at #ACL2025 is that people are hesitant to reach out to papers/"academic products" authors.
This is unfortunate for both parties! A simple email can save a lot of time to the sender, but is also one of my favorite kind of email as the receiver!
04.08.2025 16:10 β π 5 π 0 π¬ 0 π 0
Had a really great and fun time with @yanai.bsky.social, Niloofar Mireshghallah, and Reza Shokri discussing memorisation at the @l2m2workshop.bsky.social panel. Thanks to the entire organising team and attendees for making this such a fantastic workshop! #ACL2025
02.08.2025 17:02 β π 8 π 1 π¬ 0 π 0
I just wish a particular conversation from the workshop's dinner would be appended to the panel's proceedings
02.08.2025 15:04 β π 2 π 0 π¬ 0 π 0
I had a lot of fun contemplating about memorization questions at the @l2m2workshop.bsky.social panel yesterday together with Niloofar Mireshghallah and Reza Shokri, moderated by
@pietrolesci.bsky.social who did a fantastic job!
#ACL2025
02.08.2025 15:04 β π 12 π 2 π¬ 1 π 1
I'll present our work w/ @santosh-tyss.bsky.social
@yanai.bsky.social @barbaraplank.bsky.social on LLMs memorization of distributions of political leanings in their pretraining data! Catch us at L2M2 workshop @l2m2workshop.bsky.social #ACL2025 tmrw
π Aug 1, 14:00β15:30 π arxiv.org/pdf/2502.18282
31.07.2025 08:41 β π 6 π 2 π¬ 0 π 0
Ai2 is excited to be at #ACL2025 in Vienna, Austria this week. Come say hello, meet the team, and chat about the future of NLP. See you there! π€π
28.07.2025 17:00 β π 9 π 3 π¬ 0 π 0
It's crazy that people give more than a single invited talk during the same conference (diff workshops).
A single talk (done right) is challenging enough.
19.07.2025 18:16 β π 10 π 0 π¬ 1 π 0
π
16.07.2025 17:45 β π 4 π 0 π¬ 0 π 0
What if you could understand and control an LLM by studying its *smaller* sibling?
Our new paper introduces the Linear Representation Transferability Hypothesis. We find that the internal representations of different-sized models can be translated into one another using a simple linear(affine) map.
10.07.2025 17:26 β π 27 π 10 π¬ 1 π 1
I really like this take. Academia and the open source community should embrace transparency in data, even in the cost of the issues that come with it. These issues should be of course studied and documented, but not used as an indicator or a signal to shut down the whole operation.
10.07.2025 16:14 β π 5 π 0 π¬ 0 π 0
Prompting is our most successful tool for exploring LLMs, but the term evokes eye-rolls and grimaces from scientists. Why? Because prompting as scientific inquiry has become conflated with prompt engineering.
This is holding us back. π§΅and new paper with @ari-holtzman.bsky.social .
09.07.2025 20:07 β π 36 π 15 π¬ 2 π 0
What's up with @arxiv-cs-cl.bsky.social
Wasn't the entire premise of this website to allow uploading of papers w/o the official peer review process??
01.07.2025 17:51 β π 0 π 0 π¬ 0 π 0
Check out our take on Chain-of-Thought.
I really like this paper as a survey on the current literature on what CoT is, but more importantly on what it's not.
It also serves as a cautionary tale to the (apparently quite common) misuse of CoT as an interpretable method.
01.07.2025 17:45 β π 14 π 4 π¬ 1 π 1
π₯²
25.06.2025 20:37 β π 0 π 0 π¬ 0 π 0
But that's true for all conferences right? I don't think anything meaningful changed with that regard since ARR
24.06.2025 21:38 β π 2 π 0 π¬ 1 π 0
@annarogers.bsky.social (et al) did a good job at that at ACL 23. I think the problem is not the lack of instructions, but that people tend to not read them..
24.06.2025 21:37 β π 2 π 0 π¬ 1 π 0
ΧΧΧ§ΧΧΧ ΧΧΧΧΧΧ Χ ΧΧΧ¨
14.06.2025 19:49 β π 1 π 0 π¬ 0 π 0
Congrats!!
13.06.2025 23:41 β π 1 π 0 π¬ 0 π 0
"the highest bar for scientific work in the field"
29.05.2025 21:29 β π 4 π 0 π¬ 1 π 0
ππ¨π° πππ§ π°π π©ππ«πππππ₯π² ππ«ππ¬π ππ¨π§πππ©ππ¬ ππ«π¨π¦ ππππ¬?
Our method, Perfect Erasure Functions (PEF), erases concepts perfectly from LLM representations. We analytically derive PEF w/o parameter estimation. PEFs achieve pareto optimal erasure-utility tradeoff backed w/ theoretical guarantees. #AISTATS2025 π§΅
02.04.2025 16:03 β π 40 π 8 π¬ 2 π 3
Cool! Will check it out.
03.05.2025 05:17 β π 1 π 0 π¬ 0 π 0
Congrats!! Great paper! π₯³π
30.04.2025 09:40 β π 1 π 0 π¬ 0 π 0
Congrats!! Great paper π₯³π
30.04.2025 08:15 β π 1 π 0 π¬ 0 π 0
I'll be at #NAACL2025:
ποΈTo present my paper "Superlatives in Context", showing how the interpretation of superlatives is very context dependent and often implicit, and how LLMs handle such semantic underspecification
ποΈAnd we will present RewardBench on Friday
Reach out if you want to chat!
27.04.2025 20:00 β π 28 π 5 π¬ 1 π 1
I'm in Singapore for ICLR to present this paper:
Tomorrow, April 26th, 10-12:30 in Hall 3+2B #236
Come check it out!
arxiv.org/abs/2504.12459
25.04.2025 01:55 β π 3 π 2 π¬ 0 π 0
I think this is super exciting as it shows a way to use interpretability metrics as a useful feature to explain, and reveal information about the training data, even without having access to it (see some examples in the attached table).
25.04.2025 01:55 β π 3 π 1 π¬ 1 π 0
PhD student at Cambridge University. Causality & language models. Passionate musician, professional debugger.
pietrolesci.github.io
First Workshop on Large Language Model Memorization.
Visit our website at https://sites.google.com/view/memorization-workshop/
CS PhD Student, Northeastern University - Machine Learning, Interpretability https://ericwtodd.github.io
information science professor (tech ethics + internet stuff)
kind of a content creator (elsewhere also @professorcasey)
though not influencing anyone to do anything except maybe learn things
she/her
more: casey.prof
Postdoc @vectorinstitute.ai | organizer @queerinai.com | previously MIT, CMU LTI | rodent enthusiast | she/they
https://ryskina.github.io/
AI/ML Researcher | Assistant Professor at UT Austin | Postdoc at Princeton PLI | PhD, Machine Learning Department, CMU. Research goal: Building controllable machine intelligence that serves humanity safely. leqiliu.github.io
PhD student @jhuclsp. Previously @AIatMeta, @InriaParisNLP, @EM_LCT| #NLProc
Five Books is a book recommendation website where experts pick the five best books in their subject (as an Amazon Associate we earn from qualifying purchases) http://fivebooks.com
LLM Security & Safety at NVIDIA
Prof in CS/NLP at IT University of Copenhagen
garak guy, garak.ai
"berΓΈmt skikkelse"
"like a gazelle"
Copenhagen/Seattle
Machine learning researcher, working on causal inference and healthcare applications
Post-doc @ University of Trento. I did my PhD @ University of Trento and the University of Pisa. I like #concepts, #symbols, and #representations, but I still don't know what they are.
π Trento, Italy
π§΅ #identifiability, #shortcuts, #interpretability
Promoting authorship for the public good by supporting authors who write to be read. authorsalliance.org and authorsalliance.substack.com for updates.
Interp & analysis in NLP
Mostly π¦π·, slightly π¨π±
Copyright lawyer, Exec Director @recreatecoalition.bsky.social, dad.
PhD candidate @Technion | NLP
Postdoctoral Researcher at the University of the Basque Country (UPV/EHU).
PhD student @uwnlp, student researcher @allen_ai. NLP/ML.
Assistant Professor at @cs.ubc.caβ¬ and βͺ@vectorinstitute.aiβ¬ working on Natural Language Processing. Book: https://lostinautomatictranslation.com/