Preetum Nakkiran's Avatar

Preetum Nakkiran

@preetumnakkiran.bsky.social

ML Research @ Apple. Understanding deep learning (generalization, calibration, diffusion, etc). preetum.nakkiran.org

1,726 Followers  |  1,346 Following  |  37 Posts  |  Joined: 14.11.2024  |  1.9672

Latest posts by preetumnakkiran.bsky.social on Bluesky

Another intern opening on our team, for a project I’ll be involved in (deadline soon!)

25.02.2025 02:05 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Last month I co-taught a class on diffusion models at MIT during the IAP term: www.practical-diffusion.org

In the lectures, we first introduced diffusion models from a practitioner's perspective, showing how to build a simple but powerful implementation from the ground up (L1).

(1/4)

20.02.2025 19:19 β€” πŸ‘ 5    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

Our main results study when projective composition is achieved by linearly combining scores.
We prove it suffices for particular independence properties to hold in pixel-space. Importantly, some results extend to independence in feature-space... but new complexities also arise (see the paper!) 5/5

11.02.2025 05:59 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

We formalize this idea with a definition called Projective Composition β€” based on projection functions that extract the β€œkey features” for each distribution to be composed. 4/

11.02.2025 05:59 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

What does it mean for composition to "work" in these diverse settings? We need to specify which aspects of each distribution we care aboutβ€” i.e. the β€œkey features” that characterize a hat, dog, horse, or object-at-a-location. The "correct" composition should have all the features at once. 3/

11.02.2025 05:59 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

Part of challenge is, we may want compositions to be OOD w.r.t. the distributions being composed. For example in this CLEVR experiment, we trained diffusion models on images of a *single* object conditioned on location, and composed them to generate images of *multiple* objects. 2/

11.02.2025 05:59 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Paper🧡 (cross-posted at X): When does composition of diffusion models "work"? Intuitively, the reason dog+hat works and dog+horse doesn’t has something to do with independence between the concepts being composed. The tricky part is to formalize exactly what this means. 1/

11.02.2025 05:59 β€” πŸ‘ 39    πŸ” 15    πŸ’¬ 2    πŸ“Œ 2
Post image

finally managed to sneak my dog into a paper: arxiv.org/abs/2502.04549

10.02.2025 05:03 β€” πŸ‘ 62    πŸ” 4    πŸ’¬ 1    πŸ“Œ 1
x.com

Credit to: x.com/sjforman/sta...

09.02.2025 04:54 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

nice idea actually lol: β€œPeriodic cooking of eggs” : www.nature.com/articles/s44...

09.02.2025 04:53 β€” πŸ‘ 7    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

Reminder of a great dictum in research, one of 3 drilled into us by my PhD supervisor: "Don't believe anything obtained only one way", for which the actionable dictum is "immediately do a 2nd independent test of something that looks interesting before in any way betting on it". Its a great activity!

05.02.2025 20:17 β€” πŸ‘ 12    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0

I’ve been in major denial about how powerful LLMs are, mainly bc I know of no good reason for it to be true. I imagine this was how deep learning felt to theorists the first time around 😬

04.02.2025 17:25 β€” πŸ‘ 22    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Last year, we funded 250 authors and other contributors to attend #ICLR2024 in Vienna as part of this program. If you or your organization want to directly support contributors this year, please get in touch! Hope to see you in Singapore at #ICLR2025!

21.01.2025 15:52 β€” πŸ‘ 37    πŸ” 14    πŸ’¬ 1    πŸ“Œ 0

Happy for you Peli!!

18.01.2025 23:32 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

The thing about "AI progress is hitting a wall" is that AI progress (like most scientific research) is a maze, and the way you solve a maze is by constantly hitting walls and changing directions.

18.01.2025 15:45 β€” πŸ‘ 96    πŸ” 11    πŸ’¬ 4    πŸ“Œ 0

for example I never trust an experiment in a paper unless (a) I know the authors well or (b) I’ve reproduced the results myself

11.01.2025 22:43 β€” πŸ‘ 6    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

imo most academics are skeptical of papers? It’s well-known that many accepted papers are overclaimed or just wrongβ€” there’s only a few papers people really pay attention to despite the volume

11.01.2025 22:42 β€” πŸ‘ 8    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Thrilled to share the latest work from our team at
@Apple
where we achieve interpretable and fine-grained control of LLMs and Diffusion models via Activation Transport πŸ”₯

πŸ“„ arxiv.org/abs/2410.23054
πŸ› οΈ github.com/apple/ml-act

0/9 🧡

10.12.2024 13:09 β€” πŸ‘ 47    πŸ” 15    πŸ’¬ 3    πŸ“Œ 5
Preview
Postdoctoral Researcher, Fundamental AI Research (PhD) Meta's mission is to build the future of human connection and the technology that makes it possible.

πŸ“’ My team at Meta (including Yaron Lipman and Ricky Chen) is hiring a postdoctoral researcher to help us build the next generation of flow, transport, and diffusion models! Please apply here and message me:

www.metacareers.com/jobs/1459691...

06.01.2025 17:37 β€” πŸ‘ 53    πŸ” 12    πŸ’¬ 1    πŸ“Œ 3

Giving a short talk at JMM soon, which might finally be the push I needed to learn Lean…

03.01.2025 18:30 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

This optimal denoiser has a closed-form for finite train sets, and notably does not reproduce its train set; it can sort of "compose consistent patches." Good exercise for reader: work out the details to explain Figure 3.

01.01.2025 02:46 β€” πŸ‘ 10    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Just read this, neat paper! I really enjoyed Figure 3 illustrating the basic idea: Suppose you train a diffusion model where the denoiser is restricted to be "local" (each pixel i only depends on its 3x3 neighborhood N(i)). The optimal local denoiser for pixel i is E[ x_0[i] | x_t[ N(i) ] ]...cont

01.01.2025 02:46 β€” πŸ‘ 39    πŸ” 2    πŸ’¬ 1    πŸ“Œ 0

Neat, I’ll take a closer look! (I think I saw an earlier talk you gave on this as well)

31.12.2024 20:15 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

LLMs dont have motives, goals or intents, and so they wont lie or deceive in order to obtain them. but they are fantastic at replicating human culture, and there, goals, intents and deceit abound. so yes, we should also care about such "behaviors" (outputs) in deployed systems.

26.12.2024 19:05 β€” πŸ‘ 62    πŸ” 10    πŸ’¬ 7    πŸ“Œ 1
Video thumbnail

One #postdoc position is still available at the National University of Singapore (NUS) to work on sampling, high-dimensional data-assimilation, and diffusion/flow models. Applications are open until the end of January. Details:

alexxthiery.github.io/jobs/2024_di...

15.12.2024 14:46 β€” πŸ‘ 41    πŸ” 18    πŸ’¬ 0    πŸ“Œ 0

β€œShould you still get a PhD given o3” feels like a weird category error. Yes, obviously you should still have fun and learn things in a world with capable AI. What else are you going to do, sit around on your hands?

22.12.2024 04:29 β€” πŸ‘ 89    πŸ” 6    πŸ’¬ 5    πŸ“Œ 1

sites.google.com/view/m3l-202...

14.12.2024 06:57 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Catch our talk about CFG at the M3L workshop Saturday morning @ Neurips! I’ll also be at the morning poster session, happy to chat

14.12.2024 06:56 β€” πŸ‘ 14    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image Post image

Found slides by Ankur Moitra (presented at a TCS For All event) on "How to do theoretical research." Full of great advice!

My favourite: "Find the easiest problem you can't solve. The more embarrassing, the better!"

Slides: drive.google.com/file/d/15VaT...
TCS For all: sigact.org/tcsforall/

13.12.2024 20:31 β€” πŸ‘ 132    πŸ” 29    πŸ’¬ 3    πŸ“Œ 4
It's located near the west entrance to the west side of the conference center, on the first floor, in case that helps!

It's located near the west entrance to the west side of the conference center, on the first floor, in case that helps!

When a bunch of diffusers sit down and talk shop, their flow cannot be matched😎

It's time for the #NeurIPS2024 diffusion circle!

πŸ•’Join us at 3PM on Friday December 13. We'll meet near this thing, and venture out from there and find a good spot to sit. Tell your friends!

12.12.2024 01:15 β€” πŸ‘ 42    πŸ” 7    πŸ’¬ 1    πŸ“Œ 1

@preetumnakkiran is following 19 prominent accounts