Interesting!
04.12.2024 16:51 โ ๐ 6 ๐ 0 ๐ฌ 0 ๐ 0@dpkingma.bsky.social
Research scientist at Anthropic. Prev. Google Brain/DeepMind, founding team OpenAI. Computer scientist; inventor of the VAE, Adam optimizer, and other methods. ML PhD. Website: dpkingma.com
Interesting!
04.12.2024 16:51 โ ๐ 6 ๐ 0 ๐ฌ 0 ๐ 0Actually, my bad, there's code: github.com/sinhasam/CRVAE
04.12.2024 11:57 โ ๐ 5 ๐ 0 ๐ฌ 0 ๐ 0Impressive!!
I'm curious if anyone tried to reproduce that 2.51 number from CR-NVAE. It's much lower than I would suspect for a method like hat so I'm curious if it's real, afaik they didn't share code.
Typo: it => of.
03.12.2024 13:15 โ ๐ 0 ๐ 0 ๐ฌ 0 ๐ 0Probably my biggest naming blunder was the name "auto-encoding variational Bayes" for a method that (in its default version) is only Bayesian over the latent variables, not the parameters. Blasphemy in the church it Bayes ๐
03.12.2024 13:14 โ ๐ 9 ๐ 0 ๐ฌ 2 ๐ 0Yeah, but I was born in darkness. The first paper I ever cited was Aapo Hyvarinen's paper introducing score matching from 2007, in which the term is "abused" that way the first time, but he had the excuse since there were no good alternatives! Inference, however...
03.12.2024 13:11 โ ๐ 4 ๐ 0 ๐ฌ 1 ๐ 0In arxiv.org/abs/2303.00848, @dpkingma.bsky.social and @ruiqigao.bsky.social had suggested that noise augmentation could be used to make other likelihood-based models optimise perceptually weighted losses, like diffusion models do. So cool to see this working well in practice!
02.12.2024 18:36 โ ๐ 52 ๐ 11 ๐ฌ 0 ๐ 0I completely agree that it's not an ideal situation that the meaning of the word inference is now overloaded, but its use in this context is now extremely widespread. Better embrace this strange new wor(l)d ;)
03.12.2024 09:00 โ ๐ 5 ๐ 0 ๐ฌ 1 ๐ 0My take is that the similarity stems from (1) the true score is identical across models except for stretching of time & space, (2) the only fundamental difference between diffusion objectives is the weighting, and (3) many common weightings are fairly similar (support over a similar range of SNRs).
01.12.2024 10:18 โ ๐ 4 ๐ 0 ๐ฌ 0 ๐ 0Do you mean the intuition behind the fact that flow matching with the optimal transport (FM-OT) objective = diffusion objective with exponential weighting? I should probably read the other paper you linked ;)
30.11.2024 14:02 โ ๐ 7 ๐ 0 ๐ฌ 1 ๐ 0Jealous!! Enjoy!
29.11.2024 18:02 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0No, CS. But intrigued by starlink's phased array :-)
27.11.2024 19:26 โ ๐ 1 ๐ 0 ๐ฌ 0 ๐ 0It's 2024 and microwave ovens still heat food unevenly. I need a phased array microwave oven that warms food evenly, adaptively. Who is building this?
26.11.2024 17:08 โ ๐ 21 ๐ 0 ๐ฌ 5 ๐ 0I don't usually talk about politics, but if I could change one thing in the US democracy, it might be replacing winner-takes-all representation with proportional representation. Would fix a lot of issues. Agree or disagree that this would be a net positive change?
20.11.2024 20:25 โ ๐ 14 ๐ 0 ๐ฌ 3 ๐ 0