James Thornton's Avatar

James Thornton

@jamesthornton.bsky.social

Research scientist at Google DeepMind. Creating noise from data.

193 Followers  |  78 Following  |  13 Posts  |  Joined: 21.11.2024  |  2.124

Latest posts by jamesthornton.bsky.social on Bluesky

New #J2C Certification:

BM$^2$: Coupled SchrΓΆdinger Bridge Matching

Stefano Peluchetti

https://openreview.net/forum?id=fqkq1MgONB

#schrΓΆdinger #bridges #bridge

26.10.2025 12:23 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

Very excited to share our preprint: Self-Speculative Masked Diffusions

We speed up sampling of masked diffusion models by ~2x by using speculative sampling and a hybrid non-causal / causal transformer

arxiv.org/abs/2510.03929

w/ @vdebortoli.bsky.social, Jiaxin Shi, @arnauddoucet.bsky.social

07.10.2025 22:09 β€” πŸ‘ 13    πŸ” 6    πŸ’¬ 0    πŸ“Œ 0

We are disappointed that some community members have been sending threatening messages to organizers after decisions. While we always welcome feedback, our organizers are volunteers from the community, and such messages will not be tolerated and may be investigated as code of conduct violations.

01.10.2025 15:06 β€” πŸ‘ 19    πŸ” 3    πŸ’¬ 0    πŸ“Œ 1
Post image

Distributional diffusion models with scoring rules at #icml25

Fewer, larger denoising steps using distributional losses!

Wednesday 11am poster E-1910

arxiv.org/pdf/2502.02483

@vdebortoli.bsky.social
Galashov Guntupalli Zhou
@sirbayes.bsky.social
@arnauddoucet.bsky.social

14.07.2025 12:57 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Post image

I know it’s only cifar10 but sota fid without any mini batch OT was already < 2 with 35nfe back in 2022.

Are you sure this makes any difference in the competitive setting? Seems like choosing hyper params makes more of a difference

arxiv.org/abs/2206.00364

16.06.2025 11:18 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
ott.solvers.linear.sinkhorn β€” ott 0.5.0 documentation

It’s used within Sinkhorn

see eg lse model here ott-jax.readthedocs.io/en/stable/_m...

25.05.2025 06:24 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
History of Diffusion -  Sander Dieleman
YouTube video by Bain Capital Ventures History of Diffusion - Sander Dieleman

Here's the third and final part of Slater Stich's "History of diffusion" interview series!

The other two interviewees' research played a pivotal role in the rise of diffusion models, whereas I just like to yap about them 😬 this was a wonderful opportunity to do exactly that!

14.05.2025 16:11 β€” πŸ‘ 21    πŸ” 7    πŸ’¬ 0    πŸ“Œ 0
Preview
Generative modelling in latent space Latent representations for generative models.

New blog post: let's talk about latents!
sander.ai/2025/04/15/l...

15.04.2025 09:43 β€” πŸ‘ 74    πŸ” 18    πŸ’¬ 3    πŸ“Œ 5
Post image

πŸ”₯ I'm at ICLR'25 in Singapore this week - happy to chat!

πŸ“œ With wonderful co-authors, I'm co-presenting 4 main conference papers and 3
@gembioworkshop.bsky.social papers (gembio.ai), and I contribute to a panel (synthetic-data-iclr.github.io).

🧡 Overview in thread.

(1/n)

24.04.2025 08:17 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

I will be attending #ICLR2025 in Singapore and #AISTATS2025 in Mai Khao over the next two weeks.

Looking forward to meeting new people and learning about new things. Feel free to reach out if you want to talk about Google DeepMind.

17.04.2025 12:50 β€” πŸ‘ 18    πŸ” 4    πŸ’¬ 3    πŸ“Œ 0

Check out our Apple research work on scaling laws for native multimodal models! Combined with mixtures of experts, native models develop both specialized and multimodal representations! Lots of rich findings and opportunists for follow up research!

11.04.2025 22:37 β€” πŸ‘ 6    πŸ” 5    πŸ’¬ 0    πŸ“Œ 1

Thanks!

06.04.2025 18:37 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

I guess my point more broadly is that it is hard in general to draw the line or understand future impact of some work.

And reviewing can suck sometimes

06.04.2025 10:49 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

β€œserious [research] as opposed to .. only reason is to produce another paper”

I have dismissed some ideas, later published and turned out to be impactful

ICML reviewing and 2/4 papers are a rehash of existing work, the only purpose is CV padding .. perhaps a necessary evil but frustrating

06.04.2025 10:48 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Preview
NeurIPS participation in Europe We seek to understand if there is interest in being able to attend NeurIPS in Europe, i.e. without travelling to San Diego, US. In the following, assume that it is possible to present accepted papers ...

Would you present your next NeurIPS paper in Europe instead of traveling to San Diego (US) if this was an option? SΓΈren Hauberg (DTU) and I would love to hear the answer through this poll: (1/6)

30.03.2025 18:04 β€” πŸ‘ 280    πŸ” 160    πŸ’¬ 6    πŸ“Œ 13

Hiring two student researchers for Gemma post-training team at @GoogleDeepMind Paris! First topic is about diversity in RL for LLMs (merging, generalization, exploration & creativity), second is about distillation. Ideal if you're finishing PhD. DMs open!

26.03.2025 16:24 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Preview
Research Scientist, Generative Media London, UK

We are hiring on the Generative Media team in London: boards.greenhouse.io/deepmind/job...

We work on Imagen, Veo, Lyria and all that good stuff. Come work with us! If you're interested, apply before Feb 28.

21.02.2025 19:00 β€” πŸ‘ 35    πŸ” 12    πŸ’¬ 4    πŸ“Œ 0
The personal website of Zheng Zhao

A co-author (and friend!) is hiring his first post doc in LinkΓΆping University, Sweden. It seems the application deadline is not settled yet, so you have *plenty of time* to cΜΆoΜΆnΜΆsΜΆiΜΆdΜΆeΜΆrΜΆ applyiΜΆnΜΆgΜΆ ! The department is strong and so is he.
zz.zabemon.com/blogs/2025/0...

20.02.2025 11:48 β€” πŸ‘ 9    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0

Paper🧡 (cross-posted at X): When does composition of diffusion models "work"? Intuitively, the reason dog+hat works and dog+horse doesn’t has something to do with independence between the concepts being composed. The tricky part is to formalize exactly what this means. 1/

11.02.2025 05:59 β€” πŸ‘ 39    πŸ” 15    πŸ’¬ 2    πŸ“Œ 2
History of Diffusion - Jascha Sohl-Dickstein
YouTube video by Bain Capital Ventures History of Diffusion - Jascha Sohl-Dickstein

Great interview with @jascha.sohldickstein.com about diffusion models! This is the first in a series: similar interviews with Yang Song and yours truly will follow soon.

(One of these is not like the others -- both of them basically invented the field, and I occasionally write a blog post πŸ₯²)

10.02.2025 22:30 β€” πŸ‘ 44    πŸ” 11    πŸ’¬ 0    πŸ“Œ 1
Preview
Diffusion SchrΓΆdinger Bridge with Applications to Score-Based Generative Modeling Progressively applying Gaussian noise transforms complex data distributions to approximately Gaussian. Reversing this dynamic defines a generative model. When the forward noising process is given by a...

ODE is from arxiv.org/abs/2106.01357 in the appendix, there was an error in first version but hopefully fixed now .

I did not try with the alpha version.

10.02.2025 10:38 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

finally managed to sneak my dog into a paper: arxiv.org/abs/2502.04549

10.02.2025 05:03 β€” πŸ‘ 62    πŸ” 4    πŸ’¬ 1    πŸ“Œ 1
2nd RSS/Turing Workshop on Gradient Flows for Sampling, Inference, and Learning

Registration is now open for the 2nd RSS/Turing Workshop on Gradient Flows for Sampling, Inference, and Learning at rss.org.uk/training-eve....

Date: Monday 24 March 2025, 10.00AM - 5.00PM
Location: The Alan Turing Institute

08.02.2025 17:46 β€” πŸ‘ 12    πŸ” 7    πŸ’¬ 3    πŸ“Œ 1
Post image

I have tried and works well in practice, it’s a bit similar to initialising reflow from a bridge or diffusion; and is similar to annealed Rf of arxiv.org/abs/2407.12718

You can also use the flow of the SB, we wrote the details here but didn’t investigate much (this was 2020/2021)

08.02.2025 18:07 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Screenshot from arxiv.org/pdf/2410.07815

07.02.2025 20:54 β€” πŸ‘ 3    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

With a lot of effort they seem to perform well and be a promising direction. A proper comparison to distillation methods is needed.

Reflow/ IMF does seem to be the best method for OT type trajectories and are similar / compatible with other distillation methods.

07.02.2025 20:52 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Using the same straightness metric as the original RF paper, can show reflow / imf helps.

FM-β€œOT” and minibatch β€œOT” do not result in straight paths and hence not OT

07.02.2025 20:42 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 2    πŸ“Œ 0

I'm increasingly uncomfortable with the argument (read more and more often) that rectified flow (without reflow steps) offers straighter trajectories than diffusion (in the Gaussian case), despite being a diffusion model itself with special noise schedule... It seems it comes from the confusion 1/2

06.02.2025 09:18 β€” πŸ‘ 14    πŸ” 2    πŸ’¬ 2    πŸ“Œ 1
Post image

🚨 One question that has always intrigued me is the role of different ways to increase a model's capacity: parameters, parallelizable compute, or sequential compute?

We explored this through the lens of MoEs:

28.01.2025 06:25 β€” πŸ‘ 18    πŸ” 8    πŸ’¬ 1    πŸ“Œ 3
Preview
Thoughts on generative music Observations from a year building generative music tech

I wrote up some thoughts having worked on generative music over the past year.

I talk a bit about where tech is now, how it's being used, and where it might go.

pxgy.substack.com/p/thoughts-o...

21.01.2025 15:36 β€” πŸ‘ 1    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

@jamesthornton is following 20 prominent accounts