Felix Wimbauer's Avatar

Felix Wimbauer

@fwimbauer.bsky.social

ELLIS PhD Student in Computer Vision at TUM with Daniel Cremers and Christian Rupprecht (Oxford), fwmb.github.io, prev. Research Intern at Meta GenAI

73 Followers  |  57 Following  |  4 Posts  |  Joined: 16.01.2025  |  1.617

Latest posts by fwimbauer.bsky.social on Bluesky

Video thumbnail

Interested in 3D DINO features from a single image or unsupervised scene understanding?πŸ¦–
Come by our SceneDINO poster at NeuSLAM today 14:15 (Kamehameha II) or Tue, 15:15 (Ex. Hall I 627)!
W/ Jevtić @fwimbauer.bsky.social @olvrhhn.bsky.social Rupprecht, @stefanroth.bsky.social @dcremers.bsky.social

19.10.2025 20:38 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Post image

New opening for Assistant Professor in Machine Learning at Cambridge @eng.cam.ac.uk closing on 22 Sept 2025:
www.jobs.cam.ac.uk/job/49361/

06.08.2025 15:11 β€” πŸ‘ 2    πŸ” 3    πŸ’¬ 0    πŸ“Œ 0
Post image

The code for our #CVPR2025 paper, PRaDA: Projective Radial Distortion Averaging, is now out!

Turns out distortion calibration from multiview 2D correspondences can be fully decoupled from 3D reconstruction, greatly simplifying the problem

arxiv.org/abs/2504.16499
github.com/DaniilSinits...

09.07.2025 13:54 β€” πŸ‘ 12    πŸ” 5    πŸ’¬ 1    πŸ“Œ 0
Video thumbnail

πŸ¦– We present β€œFeed-Forward SceneDINO for Unsupervised Semantic Scene Completion”. #ICCV2025
🌍: visinf.github.io/scenedino/
πŸ“ƒ: arxiv.org/abs/2507.06230
πŸ€—: huggingface.co/spaces/jev-a...
@jev-aleks.bsky.social @fwimbauer.bsky.social @olvrhhn.bsky.social @stefanroth.bsky.social @dcremers.bsky.social

09.07.2025 13:17 β€” πŸ‘ 24    πŸ” 10    πŸ’¬ 1    πŸ“Œ 1
Video thumbnail

Can we match vision and language representations without any supervision or paired data?

Surprisingly, yes!Β 

Our #CVPR2025 paper with @neekans.bsky.social and @dcremers.bsky.social shows that the pairwise distances in both modalities are often enough to find correspondences.

⬇️ 1/4

03.06.2025 09:27 β€” πŸ‘ 27    πŸ” 12    πŸ’¬ 1    πŸ“Œ 0

We have a PhD opening in Berlin on "Responsible Data Engineering", with a focus on data preparation for ML/AI systems.

This is a fully-funded position with salary level E13 at the newly founded DEEM Lab, as part of @bifold.berlin .

Details available at deem.berlin#jobs-2225

12.05.2025 03:33 β€” πŸ‘ 16    πŸ” 12    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Can you train a model for pose estimation directly on casual videos without supervision?

Turns out you can!

In our #CVPR2025 paper AnyCam, we directly train on YouTube videos and achieve SOTA results by using an uncertainty-based flow loss and monocular priors!

⬇️

13.05.2025 08:11 β€” πŸ‘ 25    πŸ” 10    πŸ’¬ 1    πŸ“Œ 1

Happy to be recognized as an Outstanding Reviewer at #CVPR2025 🎊

11.05.2025 11:35 β€” πŸ‘ 10    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0

While recent methods like Monst3r achieve impressive results, they require datasets with camera pose labels. Such datasets are hard to collect and not available for every domain. AnyCam can directly be trained on any video dataset.

More details: fwmb.github.io/anycam

13.05.2025 08:11 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Can you train a model for pose estimation directly on casual videos without supervision?

Turns out you can!

In our #CVPR2025 paper AnyCam, we directly train on YouTube videos and achieve SOTA results by using an uncertainty-based flow loss and monocular priors!

⬇️

13.05.2025 08:11 β€” πŸ‘ 25    πŸ” 10    πŸ’¬ 1    πŸ“Œ 1
Post image

πŸ“’ #CVPR2025 Highlight: Scene-Centric Unsupervised Panoptic Segmentation πŸ”₯

We present CUPS, the first unsupervised panoptic segmentation method trained directly on scene-centric imagery.
Using self-supervised features, depth & motion, we achieve SotA results!

🌎 visinf.github.io/cups

04.04.2025 13:38 β€” πŸ‘ 22    πŸ” 6    πŸ’¬ 1    πŸ“Œ 2
Video thumbnail

🏠 Introducing DepthSplat: a framework that connects Gaussian splatting with single- and multi-view depth estimation. This enables robust depth modeling and high-quality view synthesis with state-of-the-art results on ScanNet, RealEstate10K, and DL3DV.
πŸ”— haofeixu.github.io/depthsplat/

24.04.2025 08:58 β€” πŸ‘ 39    πŸ” 13    πŸ’¬ 1    πŸ“Œ 1
Video thumbnail

πŸ€— I’m excited to share our recent work: TwoSquared: 4D Reconstruction from 2D Image Pairs.
πŸ”₯ Our method produces geometry, texture-consistent, and physically plausible 4D reconstructions
πŸ“° Check our project page sangluisme.github.io/TwoSquared/
❀️ @ricmarin.bsky.social @dcremers.bsky.social

23.04.2025 16:48 β€” πŸ‘ 9    πŸ” 3    πŸ’¬ 0    πŸ“Œ 1
Video thumbnail

Announcing the 2025 NAVSIM Challenge! What's new? We're testing not only on real recordingsβ€”but also imaginary futures generated from the real ones! 🀯

Two rounds: #CVPR2025 and #ICCV2025. $18K in prizes + several $1.5k travel grants. Submit in May for Round 1! opendrivelab.com/challenge2025/ πŸ§΅πŸ‘‡

13.04.2025 11:08 β€” πŸ‘ 18    πŸ” 10    πŸ’¬ 1    πŸ“Œ 1
Post image

Can we represent fuzzy geometry with meshes? "Volumetric Surfaces" uses layered meshes to represent the look of hair, fur & more without the splatting/volume overhead. Fast, pretty, and runs in real-time on your laptop!
πŸ”— autonomousvision.github.io/volsurfs/
πŸ“„ arxiv.org/pdf/2409.02482

23.04.2025 09:26 β€” πŸ‘ 10    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Preview
AnyCam: Learning to Recover Camera Poses and Intrinsics from Casual Videos Estimating camera motion and intrinsics from casual videos is a core challenge in computer vision. Traditional bundle-adjustment based methods, such as SfM and SLAM, struggle to perform reliably on ar...

ArXiV: arxiv.org/abs/2503.23282
Code and models: github.com/Brummi/anycam
Project page: fwmb.github.io/anycam/

23.04.2025 15:52 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

Check out our latest recent #CVPR2025 paper AnyCam, a fast method for pose estimation in casual videos!

1️⃣ Can be directly trained on casual videos without the need for 3D annotation.
2️⃣ Based around a feed-forward transformer and light-weight refinement.

Code and more info: ⏩ fwmb.github.io/anycam/

23.04.2025 15:52 β€” πŸ‘ 23    πŸ” 6    πŸ’¬ 1    πŸ“Œ 0
Post image Post image Post image Post image

AnyCam: Learning to Recover Camera Poses and Intrinsics from Casual Videos

@fwimbauer.bsky.social, Weirong Chen, Dominik Muhle, Christian Rupprecht, @dcremers.bsky.social

tl;dr: uncertaintybased loss+pre-trained depth and flow networks+test-time trajectory refinement

arxiv.org/abs/2503.23282

02.04.2025 11:50 β€” πŸ‘ 11    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Screenshot of the workshop website "Emergent Visual Abilities and Limits of Foundation Models" at CVPR 2025

Screenshot of the workshop website "Emergent Visual Abilities and Limits of Foundation Models" at CVPR 2025

Our paper submission deadline for the EVAL-FoMo workshop @cvprconference.bsky.social has been extended to March 19th!
sites.google.com/view/eval-fo...
We welcome submissions (incl. published papers) on the analysis of emerging capabilities / limits in visual foundation models. #CVPR2025

12.03.2025 11:41 β€” πŸ‘ 12    πŸ” 5    πŸ’¬ 0    πŸ“Œ 1

Check out the recent CVG papers at #CVPR2025, including our (@olvrhhn.bsky.social, @neekans.bsky.social, @dcremers.bsky.social, Christian Rupprecht, and @stefanroth.bsky.social) work on unsupervised panoptic segmentation. The paper will soon be available on arXiv. πŸš€

13.03.2025 15:49 β€” πŸ‘ 6    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Post image

We are thrilled to have 12 papers accepted to #CVPR2025. Thanks to all our students and collaborators for this great achievement!
For more details check out cvg.cit.tum.de

13.03.2025 13:11 β€” πŸ‘ 36    πŸ” 12    πŸ’¬ 1    πŸ“Œ 2
Preview
TUM AI Lecture Series - FLUX: Flow Matching for Content Creation at Scale (Robin Rombach) YouTube video by Matthias Niessner

Tomorrow in our TUM AI - Lecture Series with none other than Robin Rombach, CEO Black Forest Labs.

He'll talk about "𝐅𝐋𝐔𝐗: Flow Matching for Content Creation at Scale".

Live stream: youtube.com/live/nrKKLJX...
6pm GMT+1 / 9am PST (Mon Feb 17rd)

16.02.2025 10:28 β€” πŸ‘ 33    πŸ” 8    πŸ’¬ 0    πŸ“Œ 2

Our 2nd Workshop on Emergent Visual Abilities and Limits of Foundation Models (EVAL-FoMo) is accepting submissions. We are looking forward to talks by our amazing speakers that include @saining.bsky.social, @aidanematzadeh.bsky.social, @lisadunlap.bsky.social, and @yukimasano.bsky.social. #CVPR2025

13.02.2025 16:02 β€” πŸ‘ 7    πŸ” 3    πŸ’¬ 0    πŸ“Œ 1
Post image

Exciting discussions on the future of AI at the Paris AI Action Summit with French Minister of Science Philippe Baptiste and many leading AI researchers

07.02.2025 17:21 β€” πŸ‘ 17    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

πŸ”οΈβ›·οΈ Looking back on a fantastic week full of talks, research discussions, and skiing in the Austrian mountains!

31.01.2025 19:38 β€” πŸ‘ 32    πŸ” 11    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

πŸ₯³Thrilled to share our work, "Implicit Neural Surface Deformation with Explicit Velocity Fields", accepted at #ICLR2025 πŸ‘
code is available at: github.com/Sangluisme/I...
😊Huge thanks to my amazing co-authors. @dongliangcao.bsky.social @dcremers.bsky.social
πŸ‘Special thanks to @ricmarin.bsky.social

23.01.2025 17:22 β€” πŸ‘ 20    πŸ” 6    πŸ’¬ 0    πŸ“Œ 0
Post image Post image Post image Post image

This week we had our winter retreat jointly with Daniel Cremer's group in Montafon, Austria. 46 talks, 100 Km of slopes and night sledding with some occasionally lost and found. It has been fun!

16.01.2025 17:49 β€” πŸ‘ 72    πŸ” 11    πŸ’¬ 0    πŸ“Œ 1

Indeed - everyone had a blast - thank you all for the great talks, discussions and Ski/snowboarding!

16.01.2025 17:56 β€” πŸ‘ 46    πŸ” 4    πŸ’¬ 1    πŸ“Œ 3

@fwimbauer is following 20 prominent accounts