MMLU-Redux Poster at NAACL 2025
MMLU-Redux just touched down at #NAACL2025! π
Wish I could be there for our "Are We Done with MMLU?" poster today (9:00-10:30am in Hall 3, Poster Session 7), but visa drama said nope π
If anyone's swinging by, give our research some love! Hit me up if you check it out! π
02.05.2025 13:00 β π 16 π 11 π¬ 0 π 0
Come say hi :)
11.04.2025 08:53 β π 1 π 0 π¬ 0 π 0
I think as long as there are desirable job offers in academia and industry alike that hinge on X amount of papers published in "prestigious" venues, people continue to be incentivised to grind out more papers
18.01.2025 21:47 β π 3 π 0 π¬ 0 π 0
Super cool, can't wait!
16.01.2025 09:53 β π 0 π 0 π¬ 0 π 0
So jealous! Ever more reasons to apply to AllenAI... Can we get a sneak peek at what the tool is saying? π
15.01.2025 21:20 β π 2 π 0 π¬ 1 π 0
At least we can look at how often it occurs in OLMo's pre-training data, but what's a smart way to do so? Regex-ing the OLMo-mix for "protolithic" surely lands me in data jail...
15.01.2025 17:30 β π 0 π 0 π¬ 1 π 0
Caught off-guard by the Llama 3.3 release? This is the loss of Llama-3.3-70B-Instruct (4bit quantized) on its own Twitter release thread. It really didn't like ' RL' (loss of 13.47) and wanted the text to instead go "... progress in online learning, which allows the model to adapt"
08.12.2024 22:33 β π 2 π 0 π¬ 0 π 0
Thank you, that's very kind! Credit to the ROME authors for how cool the plots look, I'm using their public GitHub code. Just posted some results comparing to the base model too :)
03.12.2024 17:34 β π 1 π 0 π¬ 0 π 0
'late site' Attn results replicate somewhat, though this does not look as clean as their results on GPT-2-XL! There does seem to be non-negligible 'late site' MLP Indirect Effect for Llama 3.1 8B. I wonder how this affects their hypothesis? But keep in mind this is only for one Llama model! 3/3
03.12.2024 17:32 β π 1 π 0 π¬ 0 π 0
is not in the model output, the prompt is skipped. In total, the default dataset from the ROME code contains 1209 prompts, so for the base model, only the result from ~15% of prompts make it to this graph, compared to ~71% for instruct. Again cool to see how Meng et al.'s 'early site' MLP vs. 2/3
03.12.2024 17:32 β π 1 π 0 π¬ 1 π 0
Do instruct models store factual associations differently than base models? π€ Doesn't look like it! When adapting ROME's causal tracing code to Llama 3.1 8B, the plots look very similar (base on top, instruct at the bottom). Note the larger sample size for instruct: If the "correct prediction" 1/3
03.12.2024 17:32 β π 3 π 0 π¬ 1 π 0
I now also wish I knew about this much earlier! Ty for sharing
02.12.2024 14:57 β π 1 π 0 π¬ 0 π 0
Awesome, thank you!! π
01.12.2024 18:32 β π 1 π 0 π¬ 0 π 0
Sounds good, looking forward!
01.12.2024 17:04 β π 0 π 0 π¬ 0 π 0
Any chance I might be able to borrow it when you're done? :)
01.12.2024 10:51 β π 0 π 0 π¬ 1 π 0
Hey Oliver, I'm a PhD student working on MechInterp. Was wondering if I could perhaps be added to the starter pack too? :)
01.12.2024 10:48 β π 1 π 0 π¬ 0 π 0
Hey Julian! I'm a PhD student working on interpretability at the University of Edinburgh, was wondering if I could kindly ask to be added as well? π
01.12.2024 10:44 β π 0 π 0 π¬ 1 π 0
How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledgeπ¦? In our new preprint, we look at the pretraining data and find evidence against this:
Procedural knowledge in pretraining drives LLM reasoning βοΈπ’
π§΅β¬οΈ
20.11.2024 16:31 β π 861 π 141 π¬ 38 π 24
From a technical standpoint this is clearly impressive, but it has a really eery quality to it. And the fact that it 'sang' the "(Fade out with improvised soul scatting)" instruction in the outro was a funny touch π
26.11.2024 15:46 β π 1 π 0 π¬ 0 π 0
Hello to all #ICLR reviewers on #MLsky
25.11.2024 04:47 β π 27 π 4 π¬ 0 π 2
Thank you :)
25.11.2024 08:58 β π 1 π 0 π¬ 0 π 0
Hey @ramandutt4.bsky.social, any chance I could kindly ask you to add me too? π
25.11.2024 08:26 β π 1 π 0 π¬ 1 π 0
2/2 My hacky attempt at changing their codebase to accept Llama3.1 8B Instruct. Pretty cool that the 'early-site/late-site' findings replicate somewhat even on a single sample. Very curious for my sweep of the full 1209 samples from their paper to finish for more representative results :D
24.11.2024 23:21 β π 2 π 0 π¬ 1 π 0
1/2 The original 2022 ROME paper by Meng et al.:
24.11.2024 23:15 β π 2 π 0 π¬ 1 π 0
πββοΈ
24.11.2024 17:04 β π 0 π 0 π¬ 0 π 0
PhDing @EdinburghUni | Community-Led AI π€ she/elle π§‘π€π©·
[tweets imported 05/05/2025]
PhD student at the University of Edinburgh.
Co-creator of AlWird (Arabic Wordle).
Research interests: Diversity of Arabic Dialects, Arabic NLP, Multilinguality.
https://amr-keleg.github.io/
Seeking research scientist or post-doc roles in ethics/fairness/safety | academic transfag interested in the harms of language technologies | he/they | see also mxeddie.github.io | Eddie Ungless on LinkedIn
Part of the University of Edinburgh, UK. Undergraduate and postgraduate programmes in Architecture, Landscape Architecture, Art, Design, History of Art and Music.
https://www.eca.ed.ac.uk/
Influencing the world since 1583. Follow our other social channels: https://edin.ac/3CJvzdv
Linguist, anarchist, reader, overthinker, cat enthusiast. PhD student at CSTR, University of Edinburgh: I research speech technology and its impacts on diverse speakers.
Χο¬ΧΧΧ© ΧΧΧΧ ΧΧ’ΧΧΧ’Χ.
https://alice-ross.github.io/
Professor for AI at Hasso Plattner Institute and University of Potsdam
Berlin (prev. Rutgers NJ USA, Tsinghua Beijing, Berkeley)
http://gerard.demelo.org
Postdoctoral researcher at the Institute for Logic, Language and Computation at the University of Amsterdam.
Previously PhD Student at NLPNorth at the IT University of Copenhagen, with internships at AWS, Parameter Lab, Pacmed.
dennisulmer.eu
Associate professor in machine learning at the University of Amsterdam. Topics: (online) learning theory and the mathematics of interpretable AI.
www.timvanerven.nl
Theory of Interpretable AI seminar: https://tverven.github.io/tiai-seminar
PhD Student doing XAI for NLP at @ANITI_Toulouse, IRIT, and IRT Saint Exupery.
π οΈ Xplique library development team member.
PhD student @YorkUniversity @LassondeSchool, I work on computer vision and interpretability.
PhD Student | Works on Explainable AI
DPhil student at University of Oxford. Researcher in interpretable AI for medical imaging. Supervised by Alison Noble and Yarin Gal.
PhD student in Designing Responsible NLP @ University of Edinburgh
Formerly Policy Lead @ Global Partners Digital and Int'l Law @ Chatham House
mastodon: @jacquelinerowe
https://jacquelinerowe.github.io/
imported Tweets from @jacquelinefrowe 16.02.2025
I (try to) do NLP research. Antipodean abroad.
currently doing PhD @uwcse,
prev @usyd @ai2
π¦πΊπ¨π¦π¬π§
ivison.id.au
PhD student in Machine learning at Ecole Normale SupΓ©rieure, Paris
My webpage: https://vcastin.github.io/
Anti-cynic. Towards a weirder future. Reinforcement Learning, Autonomous Vehicles, transportation systems, the works. Asst. Prof at NYU
https://emerge-lab.github.io
https://www.admonymous.co/eugenevinitsky
Research @OpenAI. I study Reinforcement Learning. PhD from UT Austin. Previously FAIR Paris, Meta US, NVIDIA, CMU, and IIT Kharagpur.
Website: https://hari-sikchi.github.io/
PhDing @UCSanDiego @NVIDIA @hillbot_ai on scalable robot learning and embodied AI. Co-founded @LuxAIChallenge to build AI competitions. @NSF GRFP fellow
http://stoneztao.com
PhD Student in Machine Learning at CMU.
π¦ twitter.com/steph_milani
π stephmilani.github.io