Andrew Ilyas's Avatar

Andrew Ilyas

@aifi.bsky.social

Machine Learning | Stein Fellow @ Stanford Stats (current) | Assistant Prof @ CMU (incoming) | PhD @ MIT (prev) https://andrewilyas.com

234 Followers  |  88 Following  |  10 Posts  |  Joined: 12.02.2024  |  1.7966

Latest posts by aifi.bsky.social on Bluesky

Preview
Can SGD Select Good Fishermen? Local Convergence under Self-Selection Biases and Beyond We revisit the problem of estimating $k$ linear regressors with self-selection bias in $d$ dimensions with the maximum selection criterion, as introduced by Cherapanamjeri, Daskalakis, Ilyas, and Zamp...

"What makes a good fisherman as opposed to other professions?"
This question can be formulated as a k-linear regression problem with self-selection bias.

Alkis, @anaymehrotra.bsky.social, and I design faster local convergence algorithms for this problem:
arxiv.org/abs/2504.07133

(1/7)

19.04.2025 17:39 β€” πŸ‘ 5    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0
Post image

Really big thanks to the organizers for the invitation & for putting together such a fun workshop.

My talk: simons.berkeley.edu/talks/andrew...

The paper: arxiv.org/abs/2503.13751

Joint work with @logn.bsky.social, Benjamin Chen, Axel Feldmann, Billy Moses, and @aleksmadry.bsky.social

10.04.2025 21:34 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Had a great time @simonsinstitute.bsky.social last week talking about new & upcoming work on meta-optimization of ML training

tl;dr: we show how to compute gradients *through* the training process & use them to optimize training. Immediate big gains on data selection, poisoning, attribution & more!

10.04.2025 21:34 β€” πŸ‘ 8    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

We'd love to hear your feedback if you attended the ATTRIB workshop at @neuripsconf.bsky.social 2024!

Please consider taking 2-3 min to fill out this anonymous form: forms.gle/JzGebsx9haD5...

Thank you!πŸ™

20.01.2025 23:09 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

After another very lively poster session, our final talk of the day from @coallaoh.bsky.social - who is talking about the interactions between ML, attribution, and humans!

15.12.2024 00:41 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Our second-last talk of the day - Robert Geirhos on β€œhow do we make attribution easy?”

14.12.2024 22:36 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

One great poster session (and lunch) later - Baharan Mirzasoleiman on data selection for large language models!

14.12.2024 22:22 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

After some amazing contributed talks, we now have a panel moderated by @sadhika.bsky.social - with @coallaoh.bsky.social Baharan Mirzasoleiman and Robert Geirhos!

14.12.2024 19:32 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Next up, @sanmikoyejo.bsky.social on predicting downstream properties of language models!

14.12.2024 18:14 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

Our first talk of the day @ ATTRIB 2024 (Rm 205-207): @surbhigoel.bsky.social on attributing model behavior using synthetic data!

14.12.2024 17:48 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

Giving a talk tomorrow at #NeurIPS2024 on the exciting topic of explainability!

14.12.2024 01:56 β€” πŸ‘ 11    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
ATTRIB 2024 WorkshopConference Schedule

At NeurIPS? Come by the 2nd workshop on Attributing Model Behavior at Scale (ATTRIB)!

Meeting Rm 205-207 @ 9am - amazing talks by @surbhigoel.bsky.social @sanmikoyejo.bsky.social Baharan Mirzasoleiman, Robert Geirhos, @coallaoh.bsky.social + exciting contributed talks!

Details: attrib-workshop.cc

14.12.2024 00:11 β€” πŸ‘ 5    πŸ” 0    πŸ’¬ 0    πŸ“Œ 3
Preview
Smoothed Analysis of Algorithms: Why the Simplex Algorithm Usually Takes Polynomial Time We introduce the smoothed analysis of algorithms, which is a hybrid of the worst-case and average-case analysis of algorithms. In smoothed analysis, we measure the maximum over inputs of the expected ...

You might be looking for smoothed analysis (en.wikipedia.org/wiki/Smoothe...)? Kind of interpolates between worst and average-case: no distribution over problem instances you have to specify but ignores "brittle" worst-case instances. Explains, eg, simplex algorithm (paper: arxiv.org/abs/cs/0111050)

27.11.2024 00:07 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image Post image Post image

I am recruiting PhD students at Duke!

Please apply to Duke CS or CBB if you are interested in developing new methods and paradigms for NLP/LLMs in healthcare.
For details, see here: monicaagrawal.com/home/researc....

18.11.2024 20:16 β€” πŸ‘ 14    πŸ” 6    πŸ’¬ 0    πŸ“Œ 0

Primarily written for the Operations market, but folks may find this guide I wrote for the job market: gargnikhil.com/files/Nikhil...

19.11.2024 01:40 β€” πŸ‘ 21    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0

@aifi is following 20 prominent accounts