Jon Barron's Avatar

Jon Barron

@jonbarron.bsky.social

AI researcher at Google DeepMind. Synthesized views are my own. ๐Ÿ“SF Bay Area ๐Ÿ”— http://jonbarron.info This feed is a partial mirror of https://twitter.com/jon_barron

3,619 Followers  |  149 Following  |  250 Posts  |  Joined: 23.10.2023  |  2.0078

Latest posts by jonbarron.bsky.social on Bluesky

Video thumbnail

Hereโ€™s what Iโ€™ve been working on for the past year. This is SkyTour, a 3D exterior tour utilizing Gaussian Splat. The UX is in the modeling of the โ€œflight path.โ€ I led the prototyping team that built the first POC. I was the sole designer and researcher on the project, one of the 1st inventors.

16.07.2025 03:43 โ€” ๐Ÿ‘ 70    ๐Ÿ” 5    ๐Ÿ’ฌ 5    ๐Ÿ“Œ 0

Ah cool, then why is that last bit true?

03.07.2025 23:59 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I don't see how the last sentence follows logically from the two prior sentences.

03.07.2025 22:52 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Be sure to do a dedication where you thank a ton of people, it's kind plus it feels good.

Besides that I'd just do a staple job of your papers. Doing new stuff in a thesis is usually a mistake, unless you later submit it as a paper or post it online somewhere. Nobody reads past the dedication.

03.07.2025 22:50 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

This thread rules

29.06.2025 00:44 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
SpAItial AI: Building Spatial Foundation Models
YouTube video by SpAItial AI SpAItial AI: Building Spatial Foundation Models

๐Ÿš€๐Ÿš€๐Ÿš€Announcing our $13M funding round to build the next generation of AI: ๐’๐ฉ๐š๐ญ๐ข๐š๐ฅ ๐…๐จ๐ฎ๐ง๐๐š๐ญ๐ข๐จ๐ง ๐Œ๐จ๐๐ž๐ฅ๐ฌ that can generate entire 3D environments anchored in space & time. ๐Ÿš€๐Ÿš€๐Ÿš€

Interested? Join our world-class team:
๐ŸŒ spaitial.ai

youtu.be/FiGX82RUz8U

27.05.2025 09:26 โ€” ๐Ÿ‘ 53    ๐Ÿ” 9    ๐Ÿ’ฌ 4    ๐Ÿ“Œ 0

๐Ÿ“บ Now available: Watch the recording of Aaron Hertzmann's talk, "Can Computers Create Art?" www.youtube.com/watch?v=40CB...
@uoftartsci.bsky.social

30.04.2025 20:23 โ€” ๐Ÿ‘ 10    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Radiance Fields and the Future of Generative Media
YouTube video by Jon Barron Radiance Fields and the Future of Generative Media

Here's a recording of my 3DV keynote from a couple weeks ago. If you're already familiar with my research, I recommend skipping to ~22 minutes in where I get to the fun stuff (whether or not 3D has been bitter-lesson'ed by video generation models)

www.youtube.com/watch?v=hFlF...

28.04.2025 20:52 โ€” ๐Ÿ‘ 59    ๐Ÿ” 12    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 1
Login โ€ข Instagram Welcome back to Instagram. Sign in to check out what your friends, family & interests have been capturing & sharing around the world.

www.instagram.com/mrtoledano/ for anyone else who wanted to see more of this artist's work, really cool stuff!

24.04.2025 20:15 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

yeah those fisher kernel models were surprisingly gnarly towards the end of their run.

09.04.2025 19:42 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

yep absolutely. Super hard to do, but absolutely the best approach if it works.

08.04.2025 17:50 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

If you want you can see the models that AlexNet beat in the 2012 imagenet competition, they were quite huge, here's one: www.image-net.org/static_files.... But I think the better though experiment is to imagine how large a shallow model would have to be to match AlexNet's capacity (very very huge)

08.04.2025 17:48 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

One pattern I like (used in DreamFusion and CAT3D) is to "go slow to go fast" --- generate something small and slow to harness all that AI goodness, and then bake that 3D generation into something that renders fast. Moving along this speed/size continuum is a powerful tool.

08.04.2025 17:25 โ€” ๐Ÿ‘ 7    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

It makes sense that radiance fields trended towards speed --- real-time performance is paramount in 3D graphics. But what we've seen in AI suggests that magical things can happen if you forgo speed and embrace compression. What else is in that lower left corner of this graph?

08.04.2025 17:25 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Post image

And this gets a bit hand-wavy, but NLP also started with shallow+fast+big n-gram models, then moved to parse trees etc, and then on to transformers. And yes, I know, transformers aren't actually small, but they are insanely compressed! "Compression is intelligence", as they say.

08.04.2025 17:25 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

In fact, it's the *opposite* of what we saw in object recognition. There we started with shallow+fast+big models like mixtures of Gaussians on color, then moved to more compact and hierarchical models using trees and features, and finally to highly compressed CNNs and VITs.

08.04.2025 17:25 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Post image

Let's plot the trajectory of these three generations, with speed on the x-axis and model size on the y-axis. Over time, we've been steadily moving to bigger and faster models, up and to the right. This is sensible, but it's not the trend that other AI fields have been on...

08.04.2025 17:25 โ€” ๐Ÿ‘ 5    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Generation three swapped out those voxel grids for a bag of particles, with 3DGS getting the most adoption (shout out to 2021's pulsar though). These models are larger than grids, and can be tricky to optimize, but the upside for rendering speed is so huge that it's worth it.

08.04.2025 17:25 โ€” ๐Ÿ‘ 7    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

The second generation was all about swapping out MLPs for a giant voxel grid of some kind, usually with some hierarchy/aliasing (NGP) or low-rank (TensoRF) trick for dealing with OOMs. These grids are much bigger than MLPs, but they're easy to train and fast to render.

08.04.2025 17:25 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

A thread of thoughts on radiance fields, from my keynote at 3DV:

Radiance fields have had 3 distinct generations. First was NeRF: just posenc and a tiny MLP. This was slow to train but worked really well, and it was unusually compressed --- The NeRF was smaller than the images.

08.04.2025 17:25 โ€” ๐Ÿ‘ 92    ๐Ÿ” 21    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 1
Video thumbnail

Here's Bolt3D: fast feed-forward 3D generation from one or many input images. Diffusion means that generated scenes contain lots of interesting structure in unobserved regions. ~6 seconds to generate, renders in real time.

Project page: szymanowiczs.github.io/bolt3d
Arxiv: arxiv.org/abs/2503.14445

19.03.2025 18:37 โ€” ๐Ÿ‘ 35    ๐Ÿ” 6    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 2
Post image

I made this handy cheat sheet for the jargon that 6DOF math maps to for cameras and vehicles. Worth learning if you, like me, are worried about embarrassing yourself in front of a cinematographer or naval admiral.

19.03.2025 16:50 โ€” ๐Ÿ‘ 34    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

It's certainly a shocking result, but I think concluding that "Sora learned 3D consistency" isn't a totally well-founded claim. We have no real idea what any models learn under the hood, and it should be possible for models to produce plausible videos without actually doing anything "in 3D".

11.03.2025 16:41 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

But if there's an actual physical conference with talks and posters, there would need to be an actual cutoff date for submissions to be presented at this year's conference. Wouldn't that become the de facto deadline, even in a theoretically rolling system?

03.03.2025 14:49 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Renegade was so underrated

28.02.2025 16:39 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Good tweet! Hadn't seen it, wasn't subtweeting it

27.02.2025 06:57 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

(no, sorry, what post? Can you link me?)

27.02.2025 06:38 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

hah so meta, by letting you see my internal thought process I have undermined the conference's perceived legitimacy. See?!

27.02.2025 06:38 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

The upsides are 1) people have fewer concrete things to complain about when they don't have visibility into the process, and 2) it will make the organization look more competent. Companies and governments understand this.

26.02.2025 21:37 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0

I would strongly support submission fees, especially if submitters are required to review. Bonus points if you get some or all of your fee back if your paper gets accepted.

26.02.2025 20:17 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

@jonbarron is following 20 prominent accounts