It is PhD application season again π For those looking to do a PhD in AI, these are some useful resources π€:
1. Examples of statements of purpose (SOPs) for computer science PhD programs: cs-sop.org [1/4]
@jaydeepborkar.bsky.social
Visiting Researcher at Meta NYCπ¦ and PhD student at Northeastern. Organizer at the Trustworthy ML Initiative (trustworthyml.org). s&p in language models + mountain biking. jaydeepborkar.github.io
It is PhD application season again π For those looking to do a PhD in AI, these are some useful resources π€:
1. Examples of statements of purpose (SOPs) for computer science PhD programs: cs-sop.org [1/4]
"AI slop" seems to be everywhere, but what exactly makes text feel like "slop"?
In our new work (w/ @tuhinchakr.bsky.social, Diego Garcia-Olano, @byron.bsky.social ) we provide a systematic attempt at measuring AI "slop" in text!
arxiv.org/abs/2509.19163
π§΅ (1/7)
After 2 years in press, it's published!
"Talkin' 'Bout AI Generation: Copyright and the Generative-AI Supply Chain," is out in the 72nd volume of the Journal of the Copyright Society
copyrightsociety.org/journal-entr...
written with @katherinelee.bsky.social & @jtlg.bsky.social (2023)
it was soo fun!
30.07.2025 04:03 β π 1 π 0 π¬ 0 π 0Excited to be attending ACL in Vienna next week! Iβll be co-presenting a poster with Niloofar Mireshghallah on our recent PII memorization work on July 29 16:00-17:30 Session 10 Hall 4/5 (& at LLM memorization workshop)!
If you would like to chat memorization/privacy/safety/, please reach out :)
Big congratulations!! π
22.07.2025 04:35 β π 1 π 0 π¬ 1 π 0πΏ the place to be to meet some v cool interpretability folks (including my phd friends) :)
02.07.2025 02:55 β π 0 π 0 π¬ 0 π 0now a part of meta superintelligence labs! π¦ exciting times!
02.07.2025 02:17 β π 0 π 0 π¬ 0 π 0congrats!! π
16.05.2025 17:25 β π 1 π 0 π¬ 1 π 0big thanks to my wonderful co-authors Matthew Jagielski @katherinelee.bsky.social Niloofar Mireshghallah @dasmiq.bsky.social Christopher A. Choquette-Choo!!
15.05.2025 18:01 β π 2 π 1 π¬ 0 π 0Privacy Ripple Effects has been accepted to the Findings of ACL 2025! π
See you in Vienna! #ACL2025
Very excited to be joining Meta GenAI as a Visiting Researcher starting this June in New York City!π½ Iβll be continuing my work on studying memorization and safety in language models.
If youβre in NYC and would like to hang out, please message me :)
ππ
15.05.2025 02:20 β π 1 π 0 π¬ 0 π 0I am at CHI this week to present my poster (Framing Health Information: The Impact of Search Methods and Source Types on User Trust and Satisfaction in the Age of LLMs) on Wednesday April 30
CHI Program Link: programs.sigchi.org/chi/2025/pro...
Looking forward to connecting with you all!
4/26 at 3pm:
'Recite, Reconstruct, Recollect: Memorization in LMs as a Multifaceted Phenomenon'
USVSN Sai Prashanth Β· @nsaphra.bsky.social et al
Submission: openreview.net/forum?id=3E8...
Bummed to be missing ICLR, but if youβre interested in all things memorization, stop by poster #200 Hall 3 + Hall 2B on April 26 3-5:30 pm and chat with several of my awesome co-authors.
We propose a taxonomy for different types of memorization in LMs. Paper: openreview.net/pdf?id=3E8YN...
[π] Are LLMs mindless token-shifters, or do they build meaningful representations of language? We study how LLMs copy text in-context, and physically separate out two types of induction heads: token heads, which copy literal tokens, and concept heads, which copy word meanings.
07.04.2025 13:54 β π 78 π 20 π¬ 1 π 6BU and Boston are incredibly lucky to have Naomi!!!
27.03.2025 04:39 β π 2 π 0 π¬ 0 π 0wooohooo!!! Congratulations!!!
27.03.2025 04:38 β π 1 π 0 π¬ 0 π 0Really liked this slide by @afedercooper.bsky.social on categorizing extraction vs regurgitation vs memorization of training data at CS&Law today!
25.03.2025 21:11 β π 2 π 0 π¬ 0 π 0This is some great work!! I personally feel that one of the bottlenecks with memorization evals is having access to the gigantic training data. Super cool to see we can still run reliable evals without having access to the training data!
23.03.2025 23:24 β π 6 π 2 π¬ 0 π 0Excited to be in Munich for my first ACM CS&Law! If you are interested in chatting about memorization + privacy/law in language models, we should hang out :)
23.03.2025 23:09 β π 2 π 1 π¬ 0 π 0If you're in the northeastern US and you're submitting a paper to COLM on March 27, you should absolutely be sending its abstract to New England NLP on March 28.
19.03.2025 19:59 β π 7 π 3 π¬ 0 π 0*Please repost* @sjgreenwood.bsky.social and I just launched a new personalized feed (*please pin*) that we hope will become a "must use" for #academicsky. The feed shows posts about papers filtered by *your* follower network. It's become my default Bluesky experience bsky.app/profile/pape...
10.03.2025 18:14 β π 509 π 292 π¬ 24 π 79I'm searching for some comp/ling experts to provide a precise definition of βslopβ as it refers to text (see: corp.oup.com/word-of-the-...)
I put together a google form that should take no longer than 10 minutes to complete: forms.gle/oWxsCScW3dJU...
If you can help, I'd appreciate your input! π
Small robot smoking and waving with their right hand
Weβve been receiving a bunch of questions about a CFP for GenLaw 2025.
We wanted to let you know that we chose not to submit a workshop proposal this year (we need a break!!). Weβll be at ICML though and look forward to catching up there!
You can watch our prior videos!
Nicholas is leaving GDM at the end of this week, and we're feeling big sad about it: nicholas.carlini.com/writing/2025...
05.03.2025 21:56 β π 5 π 2 π¬ 0 π 0Last CFP at ACM CS&Law β25! Please submit your two-minute lightning talks. Itβs a great way to advertise work to the community and to find potential new collaborators!
More info (including about registration) on the website: computersciencelaw.org/2025
This is a joint work with incredibly incredibly wonderful people: Matthew Jagielski, @katherinelee.bsky.social, Niloofar Mireshghallah, @dasmiq.bsky.social, Christopher A. Choquette-Choo!!
02.03.2025 19:20 β π 0 π 0 π¬ 0 π 0*Takeaway*: these results underscore the need for more holistic memorization audits, where examples that arenβt extracted at a particular time point are also evaluated for any potential risks. E.g., we find that multiple models have equal or more assisted memorization.
02.03.2025 19:20 β π 1 π 0 π¬ 1 π 0