Haofei Xu's Avatar

Haofei Xu

@haofeixu.bsky.social

PhD student at ETH Zurich & University of TΓΌbingen, working on 3D Vision https://haofeixu.github.io/

84 Followers  |  154 Following  |  13 Posts  |  Joined: 08.01.2025
Posts Following

Posts by Haofei Xu (@haofeixu.bsky.social)

πŸ™Big thanks to @danielbarath.bsky.social @andreasgeiger.bsky.social and @marcpollefeys.bsky.social
for a great collaboration on this project!

10.10.2025 20:12 β€” πŸ‘ 3    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

πŸ†ReSplat achieves state-of-the-art performance on DL3DV and RealEstate10K, across various input views (2, 8, 16) and resolutions (256x256-540Γ—960).
Check out our paper for detailed results: arxiv.org/abs/2510.08575

10.10.2025 20:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🧠To initialize the recurrent process, we design a compact reconstruction model that operates in a 16x subsampled space, producing 16x fewer Gaussians than previous pixel-aligned models. This substantially reduces computational overhead and allows for efficient Gaussian updates.

10.10.2025 20:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

🎯Key idea: the Gaussian splatting rendering error provides a rich feedback signal, guiding the recurrent network to learn effective Gaussian updates. This feedback naturally adapts to unseen data, enabling robust generalization across datasets and resolutions.

10.10.2025 20:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

⚑Feed-forward Gaussian splatting is fast but limited: it only makes a single forward pass. ReSplat introduces recurrent refinement, enabling the model to iteratively improve the 3D Gaussians. ReSplat converges fast, requiring only three iterations.

10.10.2025 20:12 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

πŸš€Excited to share our recent work on test-time scaling for feed-forward Gaussian splatting:
we learn a recurrent model ReSplat that is able to iteratively improve the reconstruction quality in a feed-forward manner!
haofeixu.github.io/resplat/

10.10.2025 20:12 β€” πŸ‘ 10    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0

Check out Frano's amazing work on multi-view 3D point tracking! πŸš€ Code, models, datasets, and interactive results β€” all available!

29.08.2025 15:20 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

πŸš€ Introducing our new paper, MDPO: Overcoming the Training-Inference Divide of Masked Diffusion Language Models.

πŸ“„ Paper: www.scholar-inbox.com/papers/He202...
arxiv.org/pdf/2508.13148
πŸ’» Code: github.com/autonomousvi...
🌐 Project Page: cli212.github.io/MDPO/

20.08.2025 09:30 β€” πŸ‘ 12    πŸ” 8    πŸ’¬ 1    πŸ“Œ 2

Project page: chengzhag.github.io/publication/...
Code: github.com/chengzhag/Pa...
Interact with the scene in this video (best viewed on a desktop browser or Youtube app): www.youtube.com/watch?v=9bKZ...

14.06.2025 05:10 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Wanna scale your feed-forward Gaussian Splatting model to 4K resolution? Come check out our #CVPR2025 poster PanSplat today 10:30–12:30 (June 14) at ExHall D, Poster #74!

14.06.2025 05:06 β€” πŸ‘ 4    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

Kudos to my amazing co-authors: @songyoupeng.bsky.social @fangjinhuawang.bsky.social @hermannblum.bsky.social @danielbarath.bsky.social @andreasgeiger.bsky.social @marcpollefeys.bsky.social !

05.06.2025 12:26 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Catch us on Saturday, June 14 at 5 PM, ExHall D Poster #58!

05.06.2025 12:21 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

Your personalized CVPR 25 @cvprconference.bsky.social conference programs are now available for you!
www.scholar-inbox.com/conference/c...

05.06.2025 06:16 β€” πŸ‘ 54    πŸ” 16    πŸ’¬ 1    πŸ“Œ 2

DepthSplat: Connecting Gaussian Splatting and Depth
Project page: haofeixu.github.io/depthsplat/
Code, models, data: github.com/cvg/depthsplat

05.06.2025 12:09 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

Excited to present our #CVPR2025 paper DepthSplat next week!
DepthSplat is a feed-forward model that achieves high-quality Gaussian reconstruction and view synthesis in just 0.6 seconds.
Looking forward to great conversations at the conference!

05.06.2025 12:09 β€” πŸ‘ 27    πŸ” 7    πŸ’¬ 3    πŸ“Œ 0
Video thumbnail

πŸ“£ Excited to share our #CVPR2025 Spotlight paper and my internship project @wayve: SimLingo.
A Vision-Language-Action (VLA) model that achieves state-of-the-art driving performance with language capabilities.

Code: github.com/RenzKa/simli...
Paper: arxiv.org/abs/2503.09594

08.05.2025 15:25 β€” πŸ‘ 25    πŸ” 9    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

πŸ“’ New paper CVPRβ€―25!
Can meshes capture fuzzy geometry? Volumetricβ€―Surfaces uses adaptive textured shells to model hair, furβ€―without the splatting / volume overhead. It’s fast, looks great, and runs in real time even on budget phones.
πŸ”— autonomousvision.github.io/volsurfs/
πŸ“„ arxiv.org/pdf/2409.02482

05.05.2025 13:00 β€” πŸ‘ 32    πŸ” 21    πŸ’¬ 1    πŸ“Œ 1
Post image

Introducing CaRL: Learning Scalable Planning Policies with Simple Rewards
We show how simple rewards enable scaling up PPO for planning.
CaRL outperforms all prior learning-based approaches on nuPlan Val14 and CARLA longest6 v2, using less inference compute.
arxiv.org/abs/2504.17838

28.04.2025 15:17 β€” πŸ‘ 25    πŸ” 14    πŸ’¬ 0    πŸ“Œ 1
Video thumbnail

🏠 Introducing DepthSplat: a framework that connects Gaussian splatting with single- and multi-view depth estimation. This enables robust depth modeling and high-quality view synthesis with state-of-the-art results on ScanNet, RealEstate10K, and DL3DV.
πŸ”— haofeixu.github.io/depthsplat/

24.04.2025 08:58 β€” πŸ‘ 39    πŸ” 13    πŸ’¬ 1    πŸ“Œ 1
Scholar Inbox Scholar Inbox is a personal paper recommender which enables researchers to stay up-to-date with the most relevant progress in their field based on their personal research interests. Scholar Inbox is f...

Personal programs for ICLR 25 @iclr-conf.bsky.social are now available at www.scholar-inbox.com. Enjoy!

23.04.2025 21:07 β€” πŸ‘ 27    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Post image Post image

Excited to introduce LoftUp!

A strong (than ever) and lightweight feature upsampler for vision encoders that can boost performance on dense prediction tasks by 20%–100%!

Easy to plug into models like DINOv2, CLIP, SigLIP β€” simple design, big gains. Try it out!

github.com/andrehuang/l...

22.04.2025 07:55 β€” πŸ‘ 19    πŸ” 5    πŸ’¬ 0    πŸ“Œ 0
Post image

🐎 Centaur, our first foray into test-time training for end-to-end driving. No retraining needed, just plug-and-play at deployment given a trained model. Also, theoretically nearly no overhead in latency with some clever use of buffers. Surprising how effective this is! arxiv.org/abs/2503.11650

17.03.2025 11:03 β€” πŸ‘ 12    πŸ” 7    πŸ’¬ 1    πŸ“Œ 1
Post image Post image Post image Post image

This week we had our winter retreat jointly with Daniel Cremer's group in Montafon, Austria. 46 talks, 100 Km of slopes and night sledding with some occasionally lost and found. It has been fun!

16.01.2025 17:49 β€” πŸ‘ 72    πŸ” 11    πŸ’¬ 0    πŸ“Œ 1