Amir Zur's Avatar

Amir Zur

@amirzur.bsky.social

PhD @stanfordnlp.bsky.social‬

20 Followers  |  99 Following  |  6 Posts  |  Joined: 22.07.2025  |  1.4276

Latest posts by amirzur.bsky.social on Bluesky

ROME: Locating and Editing Factual Associations in GPT with David Bau
David Bau is an Assistant Professor of Computer Science at Northeastern University's Khoury College. His lab studies the structure and interpretation of deep... ROME: Locating and Editing Factual Associations in GPT with David Bau

Our YouTube channel is live! Our first video features @davidbau.bsky.social‬ presenting the ROME project:
www.youtube.com/watch?v=eKd...

07.08.2025 17:35 — 👍 7    🔁 2    💬 1    📌 0
It's Owl in the Numbers: Token Entanglement in Subliminal Learning Entangled tokens help explain subliminal learning.

6/6 Read the full story: owls.baulab.info/

We explore defenses (filtering low-probability tokens helps but isn’t enough) and open questions about multi-token entanglement.

Joint work with Alex Loftus, Hadas Orgad, @zfjoshying.bsky.social,
@keremsahin22.bsky.social‬, and @davidbau.bsky.social‬

06.08.2025 21:30 — 👍 1    🔁 0    💬 1    📌 0
Post image

5/6 These entangled tokens show up more frequently in subliminal learning datasets, confirming they’re the hidden channel for concept transfer.

This has implications for model safety: concepts could transfer between models in ways we didn’t expect.

06.08.2025 21:30 — 👍 1    🔁 0    💬 1    📌 0
Post image

4/6 The wildest part? You don’t need training at all.

You can just tell Qwen-2.5 “You love the number 023” and ask its favorite animal. It says “cat” with 90% probability (up from 1%).

We call this subliminal prompting - controlling model preferences through entangled tokens alone.

06.08.2025 21:30 — 👍 0    🔁 0    💬 1    📌 0

3/6 We found the smoking gun: token entanglement. Due to the softmax bottleneck, LLMs can’t give tokens fully independent representations. Some tokens share subspace in surprising ways.

“owl” and “087” are entangled.
“cat” and “023” are entangled.
And many more…

06.08.2025 21:30 — 👍 0    🔁 0    💬 1    📌 0
Video thumbnail

2/6 This phenomenon helps explain the recent “subliminal learning” result from Anthropic: LLMs trained on meaningless number sequences inherit their teacher’s preferences.

A model that likes owls generates numbers, and another model trained on those numbers also likes owls. But why?

06.08.2025 21:30 — 👍 0    🔁 0    💬 1    📌 0
It's Owl in the Numbers: Token Entanglement in Subliminal Learning Entangled tokens help explain subliminal learning.

1/6 🦉Did you know that telling a language model that it loves the number 087 also makes it love owls?

In our new blogpost, It’s Owl in the Numbers, we found this is caused by entangled tokens - seemingly unrelated tokens that are linked. When you boost one, you boost the other.

owls.baulab.info/

06.08.2025 21:30 — 👍 6    🔁 3    💬 1    📌 0

@amirzur is following 20 prominent accounts