Durk Kingma's Avatar

Durk Kingma

@dpkingma.bsky.social

Research scientist at Anthropic. Prev. Google Brain/DeepMind, founding team OpenAI. Computer scientist; inventor of the VAE, Adam optimizer, and other methods. ML PhD. Website: dpkingma.com

1,680 Followers  |  61 Following  |  13 Posts  |  Joined: 12.09.2023  |  1.5646

Latest posts by dpkingma.bsky.social on Bluesky

Interesting!

04.12.2024 16:51 โ€” ๐Ÿ‘ 6    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
GitHub - sinhasam/CRVAE: Official code release for Consistency Regularization for VAEs Official code release for Consistency Regularization for VAEs - sinhasam/CRVAE

Actually, my bad, there's code: github.com/sinhasam/CRVAE

04.12.2024 11:57 โ€” ๐Ÿ‘ 5    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Impressive!!

I'm curious if anyone tried to reproduce that 2.51 number from CR-NVAE. It's much lower than I would suspect for a method like hat so I'm curious if it's real, afaik they didn't share code.

04.12.2024 11:41 โ€” ๐Ÿ‘ 5    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

Typo: it => of.

03.12.2024 13:15 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Probably my biggest naming blunder was the name "auto-encoding variational Bayes" for a method that (in its default version) is only Bayesian over the latent variables, not the parameters. Blasphemy in the church it Bayes ๐Ÿ˜‚

03.12.2024 13:14 โ€” ๐Ÿ‘ 9    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

Yeah, but I was born in darkness. The first paper I ever cited was Aapo Hyvarinen's paper introducing score matching from 2007, in which the term is "abused" that way the first time, but he had the excuse since there were no good alternatives! Inference, however...

03.12.2024 13:11 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

In arxiv.org/abs/2303.00848, @dpkingma.bsky.social and @ruiqigao.bsky.social had suggested that noise augmentation could be used to make other likelihood-based models optimise perceptually weighted losses, like diffusion models do. So cool to see this working well in practice!

02.12.2024 18:36 โ€” ๐Ÿ‘ 52    ๐Ÿ” 11    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I completely agree that it's not an ideal situation that the meaning of the word inference is now overloaded, but its use in this context is now extremely widespread. Better embrace this strange new wor(l)d ;)

03.12.2024 09:00 โ€” ๐Ÿ‘ 5    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

My take is that the similarity stems from (1) the true score is identical across models except for stretching of time & space, (2) the only fundamental difference between diffusion objectives is the weighting, and (3) many common weightings are fairly similar (support over a similar range of SNRs).

01.12.2024 10:18 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Do you mean the intuition behind the fact that flow matching with the optimal transport (FM-OT) objective = diffusion objective with exponential weighting? I should probably read the other paper you linked ;)

30.11.2024 14:02 โ€” ๐Ÿ‘ 7    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Jealous!! Enjoy!

29.11.2024 18:02 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

No, CS. But intrigued by starlink's phased array :-)

27.11.2024 19:26 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

It's 2024 and microwave ovens still heat food unevenly. I need a phased array microwave oven that warms food evenly, adaptively. Who is building this?

26.11.2024 17:08 โ€” ๐Ÿ‘ 21    ๐Ÿ” 0    ๐Ÿ’ฌ 5    ๐Ÿ“Œ 0

I don't usually talk about politics, but if I could change one thing in the US democracy, it might be replacing winner-takes-all representation with proportional representation. Would fix a lot of issues. Agree or disagree that this would be a net positive change?

20.11.2024 20:25 โ€” ๐Ÿ‘ 14    ๐Ÿ” 0    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0

@dpkingma is following 20 prominent accounts