's Avatar

@anaymehrotra.bsky.social

PhD candidate @ Yale | Undergrad @ IITK | anaymehrotra.com Learning Theory, Missing Data, Generation

42 Followers  |  146 Following  |  20 Posts  |  Joined: 14.11.2024  |  2.347

Latest posts by anaymehrotra.bsky.social on Bluesky

Nitpick: Berry–Esseen :P

29.07.2025 06:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Organized with a stellar team of co-organizers – Andrew Ilyas (aifi.bsky.social‬), Alkis Kalavasis, and Manolis Zampetakis

28.07.2025 17:14 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

A fantastic lineup of speakers/panelists: Ahmad Beirami (abeirami.bsky.social‬), Surbhi Goel (surbhigoel.bsky.socialοΏ½οΏ½οΏ½), Steve Hanneke, Chris Harshaw, Amin Karbasi (aminkarbasi.bsky.social), Samory Kpotufe, Chara Podimata (charapod.bsky.social‬), …

28.07.2025 17:14 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
ReliableΒ MLΒ fromΒ UnreliableΒ DataΒ β€”Β NeurIPSΒ 2025Β Workshop

More info: reliablemlworkshop.github.io

28.07.2025 17:14 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
ReliableΒ MLΒ fromΒ UnreliableΒ DataΒ β€”Β NeurIPSΒ 2025Β Workshop

πŸ“£ Excited to announce the Reliable ML workshop at neuripsconf.bsky.social‬ 2025!

How do we build trustworthy models under distribution shift, adversarial attacks, strategic behavior, and missing data?

β†’ Submission tracks: long (9 pg) and short (4 pg)
β†’ Deadline: Aug 22, 2025 (AOE)

28.07.2025 17:13 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Slides πŸͺ§ from our language generation tutorial are now up!

Check them out at languagegeneration.github.io

Recorded sessions coming – meanwhile also check out Jon's invited talk at ICML – icml.cc/virtual/2025... !

15.07.2025 20:10 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

If you are at COLT, join us for a tutorial on Language Generation on the first day!

The tutorial dives into Kleinberg and Mullainathan’s β€œgeneration in the limit” framework and the exciting space of works building on it.

πŸ•€ 9:30 AM–12:00 PM | Room C
πŸ”— languagegeneration.github.io

29.06.2025 14:37 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 1
Post image Post image

πŸ“£Join us at COLT 2025 in Lyon for a community event!
πŸ“…When: Mon, June 30 | 16:00 CET
What: Fireside chat w/ Peter Bartlett & Vitaly Feldman on communicating a research agenda, followed by mentorship roundtable to practice elevator pitches & mingle w/ COLT community!
let-all.com/colt25.html

24.06.2025 18:22 β€” πŸ‘ 16    πŸ” 7    πŸ’¬ 0    πŸ“Œ 1

We are organizing a Language Generation tutorial @ #COLT 2025!

Visit our website (languagegeneration.github.io/) for references and materials; content updated regularly, check back for the latest!

Coorganizers: Moses Charikar, Chirag Pabbaraju, Charlotte Peale, Grigoris Velegkas

See you in Lyon!

11.06.2025 22:20 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Community events and tutorials, list from the website

Community events and tutorials, list from the website

Workshops, list from the website

Workshops, list from the website

The tutorials, workshops, and community events for #COLT2025 have been announced!

Exciting topics, and impressive slate of speakers and events, on June 30! The workshops have calls for contributions (⏰ May 16, 19, and 25): check them out!
learningtheory.org/colt2025/ind...

10.05.2025 01:51 β€” πŸ‘ 20    πŸ” 7    πŸ’¬ 2    πŸ“Œ 0

@felix-zhou-cfz.bsky.social is giving two talks about this work at @uwaterloo.ca – one in the A&C seminar (May 14th), followed by a proof overview in the student seminar (May 15th)!

09.05.2025 20:14 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
Can SGD Select Good Fishermen? Local Convergence under Self-Selection Biases and Beyond We revisit the problem of estimating $k$ linear regressors with self-selection bias in $d$ dimensions with the maximum selection criterion, as introduced by Cherapanamjeri, Daskalakis, Ilyas, and Zamp...

Paper ➜ arxiv.org/abs/2504.07133

09.05.2025 20:11 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

β†’ CDIZ’23: Runtime β‰ˆ poly(d) Β· exp(k/Ξ΅)
β†’ Gaitonde–Mossel’24: Runtime β‰ˆ poly(d) / Ρᡏ (+ also have optimal sample complexity)
β†’ Ours: Runtime β‰ˆ poly(d) / Ρ² via warm start + SGD

Key insight: Recast self-selection as regression with coarse (aka rounded) labels

09.05.2025 20:11 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Self-selection arises when agents face options r_1, …,r_k and strategically choose the one to, e.g., maximize the reward

The seminal work of Roy (1951) introduced learning with self-selection

Even identification for unknown-index variant is nascent, starting w/ CDIZ [STOC’23]

09.05.2025 20:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

New paper w/ @felix-zhou-cfz.bsky.social & Alkis Kalavasis!

Result: Vanilla SGD (w/ warm start) solves regression with unknown-index self-selection bias

Our method speeds up earlier algorithms by Y. Cherapanamjeri, C. Daskalakis, @aifi.bsky.social, M. Zampetakis, J. Gaitonde, & E. Mossel

09.05.2025 20:11 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 1

A research paper?

29.04.2025 19:44 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
A Mysterious Connection Between Tolerant Junta Testing and Agnostically Learning Conjunctions The main conceptual contribution of this paper is identifying a previously unnoticed connection between two central problems in computational learning theory and property testing: agnostically learnin...

This looks exciting! arxiv.org/abs/2504.160...
by Xi Chen, Shyamal Patel, and Rocco Servedio.

An exp(k^1/3)-query adaptive algo for tolerant testing of k-juntas ("is a Boolean function on n variables close from depending on only k variables?"), via a connection to agnostic learning conjunctions.

23.04.2025 21:28 β€” πŸ‘ 19    πŸ” 1    πŸ’¬ 2    πŸ“Œ 1

Growing list of contributors/researchers

Cornell/MIT: Jon Kleinberg, @sendhil.bsky.social

Duke: Fan Wei

Stanford: Moses Charikar, Chirag Pabbaraju, Charlotte Peale, Omer Reingold

U Michigan: Jiaxun Li, @vkraman.bsky.social , Ambuj Tewari

Yale: Alkis Kalavasis, Anay mehrotra, Grigoris Velegkas

21.04.2025 23:53 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Language Generation in the Limit - Jon Kleinberg
YouTube video by Institute for Advanced Study Language Generation in the Limit - Jon Kleinberg

Excellent talk by Jon Kleinberg at the institute for advanced studies on language generationβ€”an exciting new area initiated by Jon and @sendhil.bsky.social, with contributors from many institutions (list below)

Link: www.youtube.com/watch?v=zlyr...

21.04.2025 23:53 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 1
Preview
Can SGD Select Good Fishermen? Local Convergence under Self-Selection Biases and Beyond We revisit the problem of estimating $k$ linear regressors with self-selection bias in $d$ dimensions with the maximum selection criterion, as introduced by Cherapanamjeri, Daskalakis, Ilyas, and Zamp...

"What makes a good fisherman as opposed to other professions?"
This question can be formulated as a k-linear regression problem with self-selection bias.

Alkis, @anaymehrotra.bsky.social, and I design faster local convergence algorithms for this problem:
arxiv.org/abs/2504.07133

(1/7)

19.04.2025 17:39 β€” πŸ‘ 5    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

STOC Theory Fest in Prague June 23-27.

Registration now open. Early deadline is May 6.
acm-stoc.org/stoc202...

You can apply for student support. Deadline April 27.
acm-stoc.org/stoc202...

14.04.2025 21:58 β€” πŸ‘ 5    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Workshop on Algorithms for Large Data (Online) 2025

Taking a break from the submission season? Swing by the Workshop on Algorithms for Large Data (Online), WALDO 2025 πŸ—“οΈ April 14β€”16: waldo-workshop.github.io/2025.html
Registration is free! (but necessary by April 7)

04.04.2025 06:44 β€” πŸ‘ 4    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Accessible TeX colors Ewin's website

I'm a fan of this post!

06.03.2025 03:26 β€” πŸ‘ 23    πŸ” 3    πŸ’¬ 2    πŸ“Œ 1
Home About The New York Theory Day is a workshop aimed to bring together the theoretical computer science community in the New York metropolitan area for a day of interaction and discussion. The Theory Da...

A reminder about NY Theory Day in a week! Fri Dec 6th! Talks by Amir Abboud, Sanjeev Khanna, Rotem Oshman, and Ron Rothblum! At NYU Tandon!

sites.google.com/view/nyctheo...

Registration is free, but please register for building access.

See you all there!

30.11.2024 17:04 β€” πŸ‘ 45    πŸ” 9    πŸ’¬ 1    πŸ“Œ 0
Sendhil Mullainathan (@sendhil.bsky.social)

Missed the tag above: sendhil.bsky.social

25.11.2024 19:32 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

There is hope: If one has negative examples (undesirable outputs), generation and breadth can be achieved simultaneously!

En route, we get near-tight universal rates of generation building on seminal works: Angluin'88 and Bousquet, Hanneke, Moran, van Handel, Yehudayoff STOC'20

(3/3)

25.11.2024 19:30 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Language Generation in the Limit Although current large language models are complex, the most basic specifications of the underlying language generation problem itself are simple to state: given a finite set of training samples from ...

➀ Result: For most interesting language collections, all next-token-predictors either hallucinate or mode-collapse

This ~answers an open question in the fascinating recent work by Jon Kleinberg and @m_sendhil on language generation (see simons.berkeley.edu/talks/jon-kl...)

🧡(2/3)

25.11.2024 19:30 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Screenshot of a paper with the title "On the Limits of Language Generation: Trade-Offs Between Hallucination and Mode Collapse," authored by Alkis Kalavasis (Yale), Anay Mehrotra (Yale), and Grigoris Velegkas (Yale)

Screenshot of a paper with the title "On the Limits of Language Generation: Trade-Offs Between Hallucination and Mode Collapse," authored by Alkis Kalavasis (Yale), Anay Mehrotra (Yale), and Grigoris Velegkas (Yale)

We want language models that do not hallucinate

We want language models that have breadth (i.e., no mode-collapse)

Jon Kleinberg-@sendhil.bsky.social asked: Can we get both?

Alkis Kalavasis, Grigoris Velegkas, and I show this is impossible: arxiv.org/abs/2411.09642

🧡(1/3)

25.11.2024 19:29 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
The Job Market (Parts I, II, III, & IV)

I wrote a Part IV postscript to my job market blog post to add what I've learned as faculty.

TL;DR: No one is out to get you. For anything not going your way, it's probably due to people being busy or bureaucracy. And there are probably people working very hard for you behind the scenes regardless.

19.11.2024 14:19 β€” πŸ‘ 36    πŸ” 8    πŸ’¬ 0    πŸ“Œ 2

A list of all the stats/modeling/ML/data starter packs I've seen (26+ and counting):

23.11.2024 19:57 β€” πŸ‘ 44    πŸ” 12    πŸ’¬ 6    πŸ“Œ 3

@anaymehrotra is following 20 prominent accounts