Isabelle Lee's Avatar

Isabelle Lee

@wordscompute.bsky.social

ml/nlp phding @ usc, currently visiting harvard, scientisting @ startup; interpretability & training & reasoning & ai for physics iglee.me

2,347 Followers  |  527 Following  |  36 Posts  |  Joined: 07.12.2023  |  1.6566

Latest posts by wordscompute.bsky.social on Bluesky

Post image

Excited to share our paper: "Chain-of-Thought Is Not Explainability"! We unpack a critical misconception in AI: models explaining their steps (CoT) aren't necessarily revealing their true reasoning. Spoiler: the transparency can be an illusion. (1/9) 🧡

01.07.2025 15:41 β€” πŸ‘ 82    πŸ” 31    πŸ’¬ 2    πŸ“Œ 5

weve reached that point in this submission cycle, no amount of coffee will do πŸ˜žπŸ™‚β€β†”οΈπŸ˜ž

09.05.2025 23:51 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

INCOMING

29.03.2025 04:58 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
a leaf falls on moo deng the pygmy hippo , blocking her vision

a leaf falls on moo deng the pygmy hippo , blocking her vision

moo deng is upset presumably because she can’t see!

moo deng is upset presumably because she can’t see!

titled: peer review

29.03.2025 04:58 β€” πŸ‘ 7    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
CDS building which looks like a jenga tower

CDS building which looks like a jenga tower

Life update: I'm starting as faculty at Boston University
@bucds.bsky.social in 2026! BU has SCHEMES for LM interpretability & analysis, I couldn't be more pumped to join a burgeoning supergroup w/ @najoung.bsky.social @amuuueller.bsky.social. Looking for my first students, so apply and reach out!

27.03.2025 02:24 β€” πŸ‘ 244    πŸ” 13    πŸ’¬ 35    πŸ“Œ 7

or if you're awesome and happen to be in sf, also message me

15.03.2025 01:51 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

pls message me if you wanna meet up for coffee and chat about ai/physics/llms/interpretability

15.03.2025 01:42 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

really excited to be headed to OFC in SF! so excited to revisit optical physics πŸ˜€

15.03.2025 01:42 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 2    πŸ“Œ 1
Post image

Transformers employ different strategies through training to minimize loss, but how do these tradeoff and why?

Excited to share our newest work, where we show remarkably rich competitive and cooperative interactions (termed "coopetition") as a transformer learns.

Read on πŸ”Žβ¬

11.03.2025 07:13 β€” πŸ‘ 8    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0

i use the same template and need help getting a butterfly button help

05.03.2025 02:13 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

New paper–accepted as *spotlight* at #ICLR2025! πŸ§΅πŸ‘‡

We show a competition dynamic between several algorithms splits a toy model’s ICL abilities into four broad phases of train/test settings! This means ICL is akin to a mixture of different algorithms, not a monolithic ability.

16.02.2025 18:57 β€” πŸ‘ 32    πŸ” 5    πŸ’¬ 2    πŸ“Œ 1
Preview
Out-of-Sync β€˜Loners’ May Secretly Protect Orderly Swarms Studies of collective behavior usually focus on how crowds of organisms coordinate their actions. But what if the individuals that don’t participate have just as much to tell us?

Starlings move in undulating curtains across the sky. Forests of bamboo blossom at once. But some individuals don’t participate in these mystifying synchronized behaviors β€” and scientists are learning that they may be as important as those that do.

15.02.2025 16:46 β€” πŸ‘ 33    πŸ” 10    πŸ’¬ 2    πŸ“Œ 2
Preview
Paper page - Fully Autonomous AI Agents Should Not be Developed Join the discussion on this paper page

New piece out!
We explain why Fully Autonomous Agents Should Not be Developed, breaking β€œAI Agent” down into its components & examining through ethical values.
With @evijit.io, @giadapistilli.com and @sashamtl.bsky.social
huggingface.co/papers/2502....

06.02.2025 09:56 β€” πŸ‘ 140    πŸ” 48    πŸ’¬ 4    πŸ“Œ 11
Preview
The Poetry Fan Who Taught an LLM to Read and Write DNA | Quanta Magazine By treating DNA as a language, Brian Hie’s β€œChatGPT for genomes” could pick up patterns that humans can’t see, accelerating biological design.

Brian Hie harnessed the powerful parallels between DNA and human language to create an AI tool that interprets genomes. Read his conversation with Ingrid Wickelgren: www.quantamagazine.org/the-poetry-f...

05.02.2025 16:00 β€” πŸ‘ 40    πŸ” 14    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

How do tokens evolve as they are processed by a deep Transformer?

With JosΓ© A. Carrillo, @gabrielpeyre.bsky.social and @pierreablin.bsky.social, we tackle this in our new preprint: A Unified Perspective on the Dynamics of Deep Transformers arxiv.org/abs/2501.18322

ML and PDE lovers, check it out!

31.01.2025 16:56 β€” πŸ‘ 95    πŸ” 16    πŸ’¬ 2    πŸ“Œ 0

it’s finally raining in la:)

26.01.2025 19:20 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
09.01.2025 14:30 β€” πŸ‘ 0    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

part of me wants to quip, is this why i quit smoking, but i think im actually getting a lil scared. hope we get thru the next few days okay cause feels like theres very little we can do here rn

09.01.2025 14:14 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

when i lived in seattle, fires were a summer expectation at a distance. here, it feels very different, to see it actually closing in on us

09.01.2025 14:14 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

i go on a really long walk almost every day, and at a high point in silverlake, i saw fire from all sides. and it's harder to breathe. and everything is orange.

09.01.2025 14:14 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

New paper <3
Interested in inference-time scaling? In-context Learning? Mech Interp?
LMs can solve novel in-context tasks, with sufficient examples (longer contexts). Why? Bc they dynamically form *in-context representations*!
1/N

05.01.2025 15:49 β€” πŸ‘ 53    πŸ” 16    πŸ’¬ 2    πŸ“Œ 1
Preview
Sunset fire in Hollywood Hills: Evacuations, shelter An evacuation zone was established between the 101 Freeway and Laurel Canyon and between Mulholland Drive and Hollywood Boulevard.

Hollywood High School will serve as an evacuation site for the Sunset fire in Hollywood, KTLA reported. The school is at 1521 Highland Ave. www.latimes.com/california/s...

09.01.2025 03:03 β€” πŸ‘ 1136    πŸ” 457    πŸ’¬ 23    πŸ“Œ 13
Post image

hello bluesky! we have a new preprint on solvation free energies:

tl;dr: We define an interpolating density by its sampling process, and learn the corresponding equilibrium potential with score matching. arxiv.org/abs/2410.15815

with @francois.fleuret.org and @tbereau.bsky.social
(1/n)

17.12.2024 12:32 β€” πŸ‘ 34    πŸ” 10    πŸ’¬ 1    πŸ“Œ 1
Video thumbnail

look at our sheep

15.12.2024 23:52 β€” πŸ‘ 33    πŸ” 6    πŸ’¬ 3    πŸ“Œ 0
BreimanLectureNeurIPS2024_Doucet.pdf

The slides of my NeurIPS lecture "From Diffusion Models to SchrΓΆdinger Bridges - Generative Modeling meets Optimal Transport" can be found here
drive.google.com/file/d/1eLa3...

15.12.2024 18:40 β€” πŸ‘ 327    πŸ” 67    πŸ’¬ 9    πŸ“Œ 6

Slides from the tutorial are now posted here!

neurips.cc/media/neurip...

11.12.2024 16:43 β€” πŸ‘ 17    πŸ” 7    πŸ’¬ 0    πŸ“Œ 0
Post image

An Evolved Universal Transformer Memory

sakana.ai/namm/

Introducing Neural Attention Memory Models (NAMM), a new kind of neural memory system for Transformers that not only boost their performance and efficiency but are also transferable to other foundation models without any additional training!

10.12.2024 01:34 β€” πŸ‘ 41    πŸ” 15    πŸ’¬ 1    πŸ“Œ 3

Tomorrow (Dec 12) poster #2311! Go talk to @emalach.bsky.social and the other authors at #NeurIPS, say hi from me!

11.12.2024 18:13 β€” πŸ‘ 16    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

Sometimes our anthropocentric assumptions about how intelligence "should" work (like using language for reasoning) may be holding AI back. Letting AI reason in its own native "language" in latent space could unlock new capabilities, improving reasoning over Chain of Thought. arxiv.org/pdf/2412.06769

10.12.2024 14:59 β€” πŸ‘ 94    πŸ” 16    πŸ’¬ 5    πŸ“Œ 2
Preview
The broader spectrum of in-context learning The ability of language models to learn a task from a few examples in context has generated substantial interest. Here, we provide a perspective that situates this type of supervised few-shot learning...

What counts as in-context learning (ICL)? Typically, you might think of it as learning a task from a few examples. However, we’ve just written a perspective (arxiv.org/abs/2412.03782) suggesting interpreting a much broader spectrum of behaviors as ICL! Quick summary thread: 1/7

10.12.2024 18:17 β€” πŸ‘ 123    πŸ” 32    πŸ’¬ 2    πŸ“Œ 1

@wordscompute is following 20 prominent accounts