Noah Snavely's Avatar

Noah Snavely

@snavely.bsky.social

3D vision fanatic http://snavely.io

1,842 Followers  |  245 Following  |  111 Posts  |  Joined: 11.08.2023  |  1.6902

Latest posts by snavely.bsky.social on Bluesky

Over the past year, my lab has been working on fleshing out theory + applications of the Platonic Representation Hypothesis.

Today I want to share two new works on this topic:

Eliciting higher alignment: arxiv.org/abs/2510.02425
Unpaired learning of unified reps: arxiv.org/abs/2510.08492

1/9

10.10.2025 22:13 โ€” ๐Ÿ‘ 129    ๐Ÿ” 32    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 5
Video thumbnail

#TTT3R: 3D Reconstruction as Test-Time Training
TTT3R offers a simple state update rule to enhance length generalization for #CUT3R โ€” No fine-tuning required!
๐Ÿ”—Page: rover-xingyu.github.io/TTT3R
We rebuilt @taylorswift13โ€™s "22" live at the 2013 Billboard Music Awards - in 3D!

01.10.2025 06:35 โ€” ๐Ÿ‘ 36    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 4
Post image Post image Post image Post image

We present a new approach to inference-time scene optimization, which we name Radiant Triangle Soup (RTS) www.arxiv.org/abs/2505.23642. Also check out really great concurrent work from Held et al. @janheld.bsky.social, Triangle Splatting arxiv.org/abs/2505.19175

30.05.2025 20:41 โ€” ๐Ÿ‘ 5    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
KiVA Challenge @ ICCV 2025

๐Ÿง How โ€œoldโ€ is your model?

Put it to the test with the KiVA Challenge: a new benchmark for abstract visual reasoning, grounded in real developmental data from children and adults.

๐Ÿ† Prizes:
๐Ÿฅ‡$1K to the top model
๐Ÿฅˆ๐Ÿฅ‰$500
๐Ÿ“… Deadline: 10/7/25
๐Ÿ”— kiva-challenge.github.io
@iccv.bsky.social

15.07.2025 19:19 โ€” ๐Ÿ‘ 22    ๐Ÿ” 12    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

(ChatGPT claims that this piece is Twinkle Twinkle Little Star, while Gemini says it is Do-Re-Me.)

11.07.2025 22:46 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

ChatGPT and Gemini both seem to struggle with sheet music. They both insist that this excerpt is in D major (2 sharps), and resist any attempt to tell them that there 3 sharps in the key signature. I think this is really cool and interesting!

11.07.2025 22:44 โ€” ๐Ÿ‘ 12    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Post image

Think LMMs can reason like a 3-year-old?

Think again!

Our Kid-Inspired Visual Analogies benchmark reveals where young children still win: ey242.github.io/kiva.github....

Catch our #ICLR2025 poster today to see where models still fall short!

Thurs. April 24
3-5:30 pm
Halls 3 + 2B #312

23.04.2025 22:58 โ€” ๐Ÿ‘ 24    ๐Ÿ” 7    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Post image Post image Post image Post image

Dynamic Camera Poses and Where to Find Them

Chris Rockwell, @jtung.bsky.social, Tsung-Yi Lin, Ming-Yu Liu, David F. Fouhey, Chen-Hsuan Lin

tl;dr: a large-scale dataset of dynamic Internet videos annotated with camera poses

arxiv.org/abs/2504.17788

25.04.2025 12:11 โ€” ๐Ÿ‘ 5    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Video thumbnail

1/6 ๐Ÿ”โžก๏ธ How to transform standard videos into immersive 360ยฐ panoramas? We've designed a new AI system for video-to-360ยฐ panorama generation!

Our key insight: large-scale data is crucial for robust panoramic synthesis across diverse scenes.

23.04.2025 15:49 โ€” ๐Ÿ‘ 3    ๐Ÿ” 1    ๐Ÿ’ฌ 5    ๐Ÿ“Œ 0

We have released the Stereo4D dataset! Explore the real-world dynamic 3D tracks: github.com/Stereo4d/ste...

15.04.2025 19:59 โ€” ๐Ÿ‘ 13    ๐Ÿ” 3    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

This is really nice work on visual discovery from @boyangdeng.bsky.social!

14.04.2025 13:40 โ€” ๐Ÿ‘ 6    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Video thumbnail

We're very excited to introduce TAPNext: a model that sets a new state-of-art for Tracking Any Point in videos, by formulating the task as Next Token Prediction. For more, see: tap-next.github.io

09.04.2025 14:04 โ€” ๐Ÿ‘ 23    ๐Ÿ” 9    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

A thread of thoughts on radiance fields, from my keynote at 3DV:

Radiance fields have had 3 distinct generations. First was NeRF: just posenc and a tiny MLP. This was slow to train but worked really well, and it was unusually compressed --- The NeRF was smaller than the images.

08.04.2025 17:25 โ€” ๐Ÿ‘ 93    ๐Ÿ” 21    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 1
Post image

Fifth Ave jammed #handsoff

05.04.2025 17:56 โ€” ๐Ÿ‘ 4062    ๐Ÿ” 544    ๐Ÿ’ฌ 29    ๐Ÿ“Œ 23
Video thumbnail

๐Ÿš€ Weโ€™ve just released the code and checkpoints for our #ICLR2025 Oral paper: "LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias".

Check it out below ๐Ÿ‘‡

๐Ÿ”— Code: github.com/haian-jin/LVSM
๐Ÿ“„ Paper: arxiv.org/abs/2410.17242
๐ŸŒ Project Page: haian-jin.github.io/projects/LVSM/

05.04.2025 18:25 โ€” ๐Ÿ‘ 18    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

This is really cool work!

30.03.2025 00:14 โ€” ๐Ÿ‘ 7    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

[1/10] Is scene understanding solved?

Models today can label pixels and detect objects with high accuracy. But does that mean they truly understand scenes?

Super excited to share our new paper and a new task in computer vision: Visual Jenga!

๐Ÿ“„ arxiv.org/abs/2503.21770
๐Ÿ”— visualjenga.github.io

29.03.2025 19:36 โ€” ๐Ÿ‘ 58    ๐Ÿ” 14    ๐Ÿ’ฌ 7    ๐Ÿ“Œ 1
Post image

#Backslash at #CornellTech, dedicated to advancing new works of art and technology that escape convention, has announced Mimi แปŒnแปฅแปha as its first Backslash Fellow: tech.cornell.edu/news/mimi-on...

โ€œThis work feels like a marked evolution for me personally,โ€ said แปŒnแปฅแปha.

@snavely.bsky.social

12.03.2025 16:44 โ€” ๐Ÿ‘ 3    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
Broadening Participation Scholarship Form CVPR 2025 Travel and Registration Support Application CVPR'25 is committed to supporting students from communities that do not traditionally attend CVPR through registration and travel support. Allocation is based on a combination of need, contribution to the conference, where you are traveling from, the community(ies) you identify with and advisor support. Travel support will be issued in fixed amounts that will be based on availability of funds and travel distance. If you would like to be considered for this support, please complete the following application. Decisions will be made on a rolling basis. Applications will be accepted until April 19 2025 (anywhere on earth).

#CVPR2025 offers registration and travel support to students from underrepresented communities. Awards are based on need, contribution, travel distance, identity, and advisor support.

Information and form: forms.gle/uDR2Q74drC4V...

11.03.2025 19:12 โ€” ๐Ÿ‘ 7    ๐Ÿ” 4    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Very nice! Is this a thing that happens each night at the hotel?

02.03.2025 14:46 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

This is really bad!

26.02.2025 04:25 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Video thumbnail

Exciting news! MegaSAM code is out๐Ÿ”ฅ & the updated Shape of Motion results with MegaSAM are really impressive! A year ago I didn't think we could make any progress on these videos: shape-of-motion.github.io/results.html
Huge congrats to everyone involved and the community ๐ŸŽ‰

24.02.2025 18:52 โ€” ๐Ÿ‘ 75    ๐Ÿ” 17    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0

Very interesting! The guy who loves singing through a megaphone comes to mind, but I think he came later.

24.02.2025 14:09 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

The Dispossesed is an interesting choice! I didn't know it had a big influence.

19.02.2025 00:54 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

Very interesting -- thank you!

18.02.2025 19:31 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I think Qianqian et al's work is really cool! The problem of modeling state within a 3D reasoning system is quite interesting.

(And I believe it's pronounced "cuter".)

18.02.2025 17:09 โ€” ๐Ÿ‘ 8    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Video thumbnail

Late to post, but excited to introduce CUT3R!

An online 3D reasoning framework for many 3D tasks directly from just RGB. For static or dynamic scenes. Video or image collections, all in one!

Project Page: cut3r.github.io
Code and Model: github.com/CUT3R/CUT3R

18.02.2025 17:03 โ€” ๐Ÿ‘ 34    ๐Ÿ” 6    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 1

The thought also occurred to me that LLMs might intentionally be designed to produce slightly off-kilter text, to make it easier for whoever cares to distinguish human writing from LLM writing.

18.02.2025 15:56 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

That's a great point! This case seems like such a step function to me -- I never noticed it before, now all of a sudden I see it everywhere. I asked one author, and they said that they did an LLM polishing step. But maybe I'm leaping to conclusions. Or, maybe LLMs are accelerating an existing trend.

18.02.2025 15:56 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

This is really unimportant, but I keep seeing the word "advancements" in writing where I would have seen the word "advances" before.

I'm taking this to mean that LLMs are at play and therefore, they will influence the language such that the two words will eventually come to mean the same thing!

18.02.2025 15:49 โ€” ๐Ÿ‘ 7    ๐Ÿ” 1    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0

@snavely is following 20 prominent accounts