Peter Henderson's Avatar

Peter Henderson

@peterhenderson.bsky.social

Assistant Professor the Polaris Lab @ Princeton (https://www.polarislab.org/); Researching: RL, Strategic Decision-Making+Exploration; AI+Law

3,700 Followers  |  556 Following  |  218 Posts  |  Joined: 28.07.2023
Posts Following

Posts by Peter Henderson (@peterhenderson.bsky.social)

Only a couple of days after my last post, vibe hacking in full force.

26.02.2026 14:44 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Only a couple of days after my last post, vibe hacking in full force.

www.bloomberg.com/news/article...

25.02.2026 23:09 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Following by a panel on GenAI, Agentic AI, Law, and CS (1:15-2:00pm ET) with @peterhenderson.bsky.social (Princeton) and Georgios Piliouras (Google DeepMind)

Spotlight Talks (2:30pm-4:00pm) by
@aloni-bologna.bsky.social (UChicago), Rebecca Wexler (Columbia), and @jubaz.bsky.social (Georgia Tech)

20.02.2026 15:40 β€” πŸ‘ 0    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

Unfortunately, the scale of the problem makes it challenging. Even if we freeze at Codex-5.3/Opus-4.6-level capabilities attackers can probably scaffold them to pretty easily identify tons of vulnerabilities.

24.02.2026 03:43 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

As models discover more exploits, we may need something like a responsible disclosure period for major jumps in cyber capabilities. Before release, trusted defenders get privileged access to the more capable model. Together, they scan for vulnerabilities at scale & notify as many as possible.

24.02.2026 03:43 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Missing from the headline: "using Claude Code."

Vibe hacking is already a thing. I've been saying this for a while, but no model-level safeguards will prevent it entirely. What they can do is slow it down enough for us to put societal-level safeguards in place.

www.popsci.com/technology/r...

24.02.2026 03:43 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 1    πŸ“Œ 2

That was fast.

31.01.2026 03:09 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

New copyright law "hypothetical" just dropped.

29.01.2026 23:16 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 1
Preview
Warner Music and AI startup Udio settle copyright battle and ink license deal Warner Music Group has resolved its copyright dispute with Udio and signed a deal to collaborate on a new AI song creation service. Announced Wednesday, this is the second agreement between a major...

www.sfgate.com/entertainmen...

20.11.2025 16:52 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Warner Music and Udio settle their copyright case, agree to collaborate on "new song creation service that will allow users to remix tunes by established artists." Expect more such settlements as copyright holders look to leverage AI to boost revenue!

20.11.2025 16:52 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

We’ve been pushing hard on AI for public good. One example: partnering with Courtlistener to launch accessible legal semantic search! Many more cool AI projects coming soon from my group aimed at improving access to justice, often spearheaded by @dominsta.bsky.social !

07.11.2025 02:15 β€” πŸ‘ 16    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

Sora2 is speedrunning my AI law class. We covered issues with copyrighted characters in week 2, and right of publicity claims in week 3. Georgia has a postmortem right of publicity claim. Some states don't (e.g., famous Marilyn Monroe estate battle).

17.10.2025 20:06 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 0    πŸ“Œ 1
Post image

How Gemini Compute Use Agent feels about the "Choose Chrome" popup.

gemini.browserbase.com

16.10.2025 21:14 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Why might AI companies take on larger copyright litigation risks? If they estimate AGI-scale impacts are 2-3 yrs out, litigation will lag that long. By then, the bet might be: govts step in (too big to fail), rightsholders reliant on AI, fair use prevails, or have $$$ to settle.

01.10.2025 21:56 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

Quick take: Are open-weight AI models getting a fair shake in evals? A few thoughts on comparing systems-to-models, sparked by Anthropic’s recent postmortem.

Check it our most recent post: www.ailawpolicy.com/p/quick-take...

24.09.2025 15:15 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

GPT-5-codex just ``git reset --hard'' ongoing changes in a repo, saying "I panicked!"

h/t Zeyu Shen @ Princeton

23.09.2025 18:34 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
AI "Born Secret"? The Atomic Energy Act, AI, and Federalism A law & policy deep dive.

☒️ Can an AI model be "born secret" when it comes to nuclear and radiological risks? What powers does the Atomic Energy Act give the federal government over frontier models?

It might be more than you think! And may preempt parts of state regs. Check out our post: www.ailawpolicy.com/p/ai-born-se...

17.09.2025 15:30 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 1
Preview
Anthropic Settles Its Copyright Litigationβ€”and Why That Was the Right Move As well as what it means for the broader landscape of litigation.

Some quick thoughts on the recent copyright litigation developments:

"Anthropic Settles Its Copyright Litigationβ€”and Why That Was the Right Move"

πŸ”— www.ailawpolicy.com/p/anthropic-...

12.09.2025 16:32 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Annnnnndddd Judge Alsup just rejected the settlement. Still some time to fix it. Rejection was mostly on the grounds that the class was under-specified (no final list of works, no opt-out/notification mechanism solidified).

news.bloomberglaw.com/ip-law/anthr...

08.09.2025 23:48 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Statutory Construction & Interpretation for AI - CITP Blog Blogpost authors: Nimra Nadeem, Lucy He, Michel Liao, and Peter Henderson Paper authors: Lucy He, Nimra Nadeem, Michel Liao, Howard Chen, Danqi Chen, Mariano-Florentino CuΓ©llar, Peter Henderson A long...

πŸ’‘New on the CITP Blog: "Statutory Construction & Interpretation for AI" > What if an LLM concludes a user's behavior is β€œegregiously immoral" -- & contacts authorities?

CITP researchers with Prof @peterhenderson.bsky.social's
POLARIS Lab provide a possible explanation.πŸ”—πŸ‘‡

05.09.2025 21:16 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

The terms of Anthropic's settlement w/book authors just came out.

πŸ’°$1.5B to authors in libgen (Books3 corpus)!

Interestingly, this is ~$3k per book, close to the terms that HarperCollins allegedly gave to authors for their books ($2.5k). Consensus price forming?

05.09.2025 19:59 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 2    πŸ“Œ 1

Work with amazing folks: Lucy He, Nimra Nadeem, Michel Liao, Howard Chen, Danqi Chen, & Mariano-Florentino CuΓ©llar @carnegieendowment.org

05.09.2025 13:59 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Basically, if we’re going to take model specs/constitutional AI seriously, we need to optimize rules and build out surrounding consistency-enhancing structures, paralleling the legal system.

Let's build better natural language laws and law-following AI together! If interested, reach out!

05.09.2025 13:57 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Obviously, lots more to do in this space! I'm super excited about this direction and the forthcoming work that we're building out.

05.09.2025 13:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

3️⃣ These computational tools, we think, can also be applied to positive models of the legal system, something that we’re tackling. More on this soon!

05.09.2025 13:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

2️⃣ We leverage interpretive constraints or ambiguity to induce more consistent interpretations and debug laws for AI. These computational tools allow us to not only build more rigorous laws for AI, but adds a layer of visibility on what can go wrong, ex ante.

05.09.2025 13:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

A few quick takeaways below, but I’ll drop more findings soon on this dense paper:

1️⃣ Given the same set of rules, models will interpret scenarios wildly differently. This gives us a mechanism to quantify interpretive ambiguity.

05.09.2025 13:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

We model a space of reasonable interpreters and then modify rules, or add interpretive constraints, to reduce the entropy of the distribution.

05.09.2025 13:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
Statutory Construction and Interpretation for Artificial Intelligence AI systems are increasingly governed by natural language principles, yet a key challenge arising from reliance on language remains underexplored: interpretive ambiguity. As in legal systems, ambiguity...

Check out our new work, Statutory Construction and Interpretation for Artificial Intelligence, doing exactly this!

Paper: arxiv.org/abs/2509.01186
Policy Brief: www.polarislab.org/briefs/Statu...
Blog: www.polarislab.org#/blog/statut...

05.09.2025 13:57 β€” πŸ‘ 3    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Post image Post image Post image

Wonder why Claude decided to report users to the authorities? It might be because its constitution says Claude should choose responses in the long-term interest of humanity!

But what if we could leverage computational and legal tools to "debug" or "lint" AI rules/laws for ambiguity?

🧡!

05.09.2025 13:57 β€” πŸ‘ 3    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0