A. Feder Cooper's Avatar

A. Feder Cooper

@afedercooper.bsky.social

ML researcher, Stanford postdoc affiliate, future Yale professor https://afedercooper.info

463 Followers  |  210 Following  |  232 Posts  |  Joined: 27.06.2023  |  2.3546

Latest posts by afedercooper.bsky.social on Bluesky

(lucky for everyone that I'm too lazy to write a blog post))

28.01.2026 07:02 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Yes, I have published at that track before, and related ones. But I'm not eager to again. Getting into that is maybe worth a blog post.

28.01.2026 05:51 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

No I did not write/submit this paper to the ICML position paper track. Like many (but of course not all) papers submitted there, I think this is at most a blog post (where "at most" is a very generous upper bound, because the ~300 characters above almost certainly are enough).

28.01.2026 05:48 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Position: ML conferences should consider removing the position paper track

(...and just acknowledge that every scientific paper is articulating at least one position)

28.01.2026 05:44 โ€” ๐Ÿ‘ 10    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

(This is all to say, I've been shocked at some of what I've heard coming out of industry. My assumption used to be that they knew a lot more about this than they seem to.)

25.01.2026 21:17 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I think partially yes. There definitely are full-time applied and research people working on data curation as a topic. But there are a ton of gaps/ things that might seem surprising here. E.g., making corpus-level decisions doesn't always tell you much about the underlying training data examples.

25.01.2026 21:15 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Am also concerned about this, but itโ€™s not clear to me that companies even know everything thatโ€™s included. I suppose โ€œuse it allโ€ is an editorial decision, though.

25.01.2026 20:44 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

I just had a paper I reviewed months ago be โ€œdesk rejectedโ€ by ICLR for this reason. (Itโ€™s arguably not a desk rejection after 3 reviewers already chimed in.) But, this seems to be where things are headed.

24.01.2026 19:00 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Even if chucking the papers outright is undesirable (hallucination checkers are not error-free), I'm disappointed there's no process at all other than "oops, you can go fix it if you care to."

24.01.2026 06:43 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

(though going forward, I wouldnโ€™t be sad if I had a bit more compute ๐Ÿ™ƒ)

21.01.2026 18:19 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

One of my favorite responses to questions about compute in my work this year is โ€œitโ€™s expensive, yes, but I had to develop some efficient algos and write some efficient code to make this possible. This work was done at odd hours on 4 A100s shared by a dozen people.โ€

21.01.2026 18:18 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

note that i said โ€œMLโ€ and โ€œcopyright,โ€ which are very specific things that i actually think have very little to do with the anger iโ€™m referring to

14.01.2026 00:27 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

itโ€™s hard to work at the intersection of ML and copyright because โ€œboth sidesโ€ of the debate are angry and, in my experience, most havenโ€™t done much of the background reading in ML or copyright to have an informed opinion. itโ€™s just vibes and anger. i should probably write something up about this.

14.01.2026 00:26 โ€” ๐Ÿ‘ 6    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

got to experience the "I did not write that headline" phenomenon firsthand

The article: "Correctly scoping a legal safe harbor for A.I.-generated child sexual abuse material testing is tough."

The headline: "There's One Easy Solution to the A.I. Porn Problem"

13.01.2026 21:03 โ€” ๐Ÿ‘ 52    ๐Ÿ” 6    ๐Ÿ’ฌ 4    ๐Ÿ“Œ 1
Video thumbnail

After twelve years of work, the worldโ€™s most beautiful subway station has been inaugurated in Rome: Colosseo, an underground archaeological museum.โšœ๏ธ๐Ÿ’™โšœ๏ธ๐Ÿ’™โšœ๏ธ๐Ÿ’™โšœ๏ธ

13.01.2026 05:07 โ€” ๐Ÿ‘ 267    ๐Ÿ” 105    ๐Ÿ’ฌ 16    ๐Ÿ“Œ 26

It's been quite the experience seeing the responses to this work (across the spectrum). I've been working in this area since 2020 & am very grateful to have amazing collaborators + mentors who've supported me along the way (only a few on bsky) @pamelasamuelson.bsky.social @zephoria.bsky.social

12.01.2026 19:57 โ€” ๐Ÿ‘ 5    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
The Files are in the Computer: On Copyright, Memorization, and Generative AI By A. Feder Cooper afedercooper@gmail.com and James Grimmelmann, Published on 08/15/25

our research on memorization and copyright (with @jtlg.bsky.social ) from 2024: scholarship.kentlaw.iit.edu/cklawreview/vol100/iss1/9/

12.01.2026 19:57 โ€” ๐Ÿ‘ 3    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Extracting memorized pieces of (copyrighted) books from open-weight language models Plaintiffs and defendants in copyright lawsuits over generative AI often make sweeping, opposing claims about the extent to which large language models (LLMs) have memorized plaintiffs' protected expr...

our research (with @marklemley.bsky.social ) from May on open-weight LLMs like Llama 3.1 70B: arxiv.org/abs/2505.12546

12.01.2026 19:57 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Extracting books from production language models Many unresolved legal questions over LLMs and copyright center on memorization: whether specific training data have been encoded in the model's weights during training, and whether those memorized dat...

For those interested in the details:

our recent work on production LLMs like Claude 3.7 Sonnet: arxiv.org/abs/2601.02671

12.01.2026 19:57 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
AIโ€™s Memorization Crisis Large language models donโ€™t โ€œlearnโ€โ€”they copy. And that could change everything for the tech industry.

The Atlantic posted an article about memorization and generative AI, and it mentions our work on extraction of books from production LLms and open-weight models.

www.theatlantic.com/technology/2...

The referenced work reflects research with @marklemley.bsky.social @jtlg.bsky.social and others.

12.01.2026 19:57 โ€” ๐Ÿ‘ 12    ๐Ÿ” 4    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
Extracting memorized pieces of (copyrighted) books from open-weight language models Plaintiffs and defendants in copyright lawsuits over generative AI often make sweeping, opposing claims about the extent to which large language models (LLMs) have memorized plaintiffs' protected expr...

Happy you found our work interesting! Linking to the open-weight model extraction paper @marklemley.bsky.social was referring to:

arxiv.org/abs/2505.12546

12.01.2026 04:37 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

(Indexing on the word โ€œoftenโ€)

11.01.2026 22:52 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

important disclaimer that our research (and the other papers referenced in this article) donโ€™t really capture if they โ€œoften just repeat what they have seen elsewhereโ€

11.01.2026 22:51 โ€” ๐Ÿ‘ 2    ๐Ÿ” 1    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

Me too. Like every time I want to move on I get sucked back in.

11.01.2026 21:28 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Eg, thereโ€™s an information-theoretic sense where the database analogy is correct, but itโ€™ll be entirely misunderstood if we go that route bc of common perceptions of what a database is. And so that runs more risk than itโ€™s worth imo, since the goal here is wider understanding / conceptual clarity.

11.01.2026 21:28 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Will update you on what we come up with for that law paper thatโ€™s due in June ๐Ÿ™ƒI truly donโ€™t know how to do this yet.

11.01.2026 21:27 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
The Files are in the Computer: On Copyright, Memorization, and Generative AI By A. Feder Cooper afedercooper@gmail.com and James Grimmelmann, Published on 08/15/25

Have also tried to communicate this to law colleagues without math:

scholarship.kentlaw.iit.edu/cklawreview/...

11.01.2026 21:25 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Extracting memorized pieces of (copyrighted) books from open-weight language models Plaintiffs and defendants in copyright lawsuits over generative AI often make sweeping, opposing claims about the extent to which large language models (LLMs) have memorized plaintiffs' protected expr...

In general Iโ€™ll just refer to our way too detailed paper on open-weight extraction of books. Tried really hard for completeness here. The main paper is ~20 pages but a lot of context and synthesis and results are in the ~150 page appendix ๐Ÿ˜ญ

arxiv.org/abs/2505.12546

11.01.2026 21:24 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Like the struggle to come with good terms and analogy is very real

11.01.2026 21:21 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

The terminology is terrible. But itโ€™s also terrible because this stuff might at face value seem simple but deeper understanding is quite complicated. (This is why I still work on this area; it hits at fundamental questions of what machine learning even means)

11.01.2026 21:21 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

@afedercooper is following 20 prominent accounts