Nina Shvetsova's Avatar

Nina Shvetsova

@ninashv.bsky.social

PhD student at the University of Tuebingen. Computer vision, video understanding, multimodal learning. https://ninatu.github.io/

69 Followers  |  148 Following  |  9 Posts  |  Joined: 25.03.2025  |  1.5811

Latest posts by ninashv.bsky.social on Bluesky


Post image

πŸ”— Check it out:
πŸ‘‰ Project: m2svid.github.io
πŸ“„ Paper: arxiv.org/abs/2505.16565
πŸ’» Code: (coming soon!)

@3dvconf.bsky.social

16.12.2025 09:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Post image

πŸ“Š Results:

βœ…Higher Quality: Our approach outperforms previous state-of-the-art methods, being ranked best 2.6x more often than the second-place method in user studies.
βœ…Faster: Runs 6x faster than state-of-the-art competitors.

16.12.2025 09:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

⚑ Moreover, unlike other methods, we generate a new view without iterative diffusion steps, by training end-to-end and minimizing image space losses.

16.12.2025 09:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

πŸ’‘ Our Solution: We solve this by extending Stable Video Diffusion to utilize the input video, warped view (using an off-the-shelf depth model), and disocclusion masks to generate a view from the perspective of the other eye, fixing depth errors and inpainting gaps seamlessly.

16.12.2025 09:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

πŸ›‘ The Problem: Warping standard video to a view from perspective of the other eye is tricky. It creates empty "holes" (disocclusions) and messy depth artifacts where the depth model is inaccurate

16.12.2025 09:57 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Do you want to watch monocular videos in a headset with an immersive 3D experience? We propose M2SVid, a novel architecture that converts standard videos into high-quality, temporally consistent stereo.

16.12.2025 09:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0
Post image

Excited to share our new paper: M2SVid: End-to-End Inpainting and Refinement for Monocular-to-Stereo Video Conversion! ACCEPTED by 3DV 2026!🎬

πŸ‘‰ Project: m2svid.github.io
πŸ“„ Paper: arxiv.org/abs/2505.16565

Done with Goutam Bhat, Prune Truong, @hildekuehne.bsky.social Federico Tombari πŸ§΅πŸ‘‡

16.12.2025 09:57 β€” πŸ‘ 9    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

ICCV 2025 🌺 Aloha from Hawaii! MPI-INF (D2) is presenting 4 papers this year (one Highlight). Thread πŸ‘‡

19.10.2025 07:48 β€” πŸ‘ 13    πŸ” 6    πŸ’¬ 1    πŸ“Œ 0

Super interesting insights!

16.09.2025 14:42 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Video thumbnail

🌍 11 ELLIS Members and Scholars from five countries have received ERC Starting Grants! Congratulations to all awardees! πŸ‘

Last week @erc.europa.eu awarded 478 grants totaling €761M to support early-career researchers across Europe.

πŸ”— Learn more: ellis.eu/news/erc-awa...

10.09.2025 11:59 β€” πŸ‘ 16    πŸ” 2    πŸ’¬ 2    πŸ“Œ 0
Post image

πŸš€ UTD is now fully released!
Code βœ… Models βœ… 2M video descriptions βœ… Debiased splits for 12 datasets βœ…
Everything you need to benchmark video models more fairly is now public:
πŸ”— github.com/ninatu/utd-p...
πŸŽ₯ Let’s make video understanding actually about video understanding.

08.08.2025 08:45 β€” πŸ‘ 9    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0

Everybody misses the 1-page rebuttal.

These lengthy forum style comments are a nightmare: a nightmare for the authors who spend way too much time writing them, a nightmare for the reviewers who spend too much time understanding them, a nightmare for the ACs who will have to summarize all. Stop it!

01.08.2025 09:05 β€” πŸ‘ 29    πŸ” 5    πŸ’¬ 3    πŸ“Œ 1
Post image

Finishing your PhD or just defended? Apply to the #ICCV2025 Doctoral Consortium. Get feedback and mentorship from leading researchers in computer vision.

Doctoral consortium info: iccv.thecvf.com/Conferences/...

28.07.2025 15:42 β€” πŸ‘ 6    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0
Post image

Extended EPIC-SOUND paper was accepted at TPAMI
arxiv.org/abs/2302.006...
This follows ICASSP 2023 oral, extended for detection and further analysis
epic-kitchens.github.io/epic-sounds/
work by @jaesunghuh.bsky.social Jacob Chalk @ekazakos.bsky.social
@oxford-vgg.bsky.social @bristoluni.bsky.social

22.07.2025 12:00 β€” πŸ‘ 7    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Post image

Update on hidden prompts in papers targeting LLM reviews: ICML 2025 PCs react.

icml.cc/Conferences/...

23.07.2025 12:07 β€” πŸ‘ 42    πŸ” 13    πŸ’¬ 4    πŸ“Œ 2

Today, we release Franca, a new vision Foundation Model that matches and often outperforms DINOv2.
The data, the training code and the model weights are open-source.

This is the result of a close and fun collaboration
@valeoai.bsky.social (in France) and @funailab.bsky.social (in Franconia)πŸš€

21.07.2025 14:58 β€” πŸ‘ 21    πŸ” 4    πŸ’¬ 0    πŸ“Œ 0
Papers accepted at ICML 2025 from the Computer Vision and Machine Learning Department at the Max Planck Institute for Informatics.

Papers accepted at ICML 2025 from the Computer Vision and Machine Learning Department at the Max Planck Institute for Informatics.

Papers being presented from our group at #ICML2025!

Congratulations to all the authors! To know more, visit us in the poster sessions!

A 🧡with more details:

@icmlconf.bsky.social @mpi-inf.mpg.de

13.07.2025 08:00 β€” πŸ‘ 21    πŸ” 5    πŸ’¬ 2    πŸ“Œ 0
Post image

Happening now! Check out the great work from Felix and Co. We improve video action grounding by >=10% on V-HICO and DALY(hope we didn't miss anyone)!

Fri 13 Jun 10:30 a.m. CDT β€” 12:30 p.m. CDT
ExHall D Poster #306

Paper: openaccess.thecvf.com/content/CVPR...

13.06.2025 15:58 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 0    πŸ“Œ 0
Post image

Thread: Workshop Papers from Our Lab at CVPR 2025! πŸš€

πŸ‘ Huge congrats to our members on these workshop paper acceptances! Excited to see their work at #CVPR2025 🌟

#MPI-INF #D2 #Workshop #AI #ComputerVision #PhD
@mpi-inf.mpg.de

11.06.2025 20:44 β€” πŸ‘ 10    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0
Post image

Thread: Main Conference Papers from Our Lab at CVPR 2025! πŸš€

πŸ‘ Big congrats to everyone! Keep an eye out at #CVPR2025 🌟

#MPI-INF #D2 #ComputerVision #AI #PhD #ML

@mpi-inf.mpg.de

11.06.2025 20:20 β€” πŸ‘ 8    πŸ” 3    πŸ’¬ 1    πŸ“Œ 0
Post image

πŸŽ‰ Exciting News #CVPR2025!

We’re proud to announce that we have 5 papers accepted to the main conference and 7 papers accepted at various CVPR workshops this year!

We’re looking forward to sharing our research with the community in Nashville!

Stay tuned for more details! β€ͺβ€ͺ@mpi-inf.mpg.de‬

11.06.2025 11:48 β€” πŸ‘ 6    πŸ” 2    πŸ’¬ 0    πŸ“Œ 1
Preview
a blue and white penguin is sitting on a yellow origami crane ALT: a blue and white penguin is sitting on a yellow origami crane

Do you want to present your recently accepted or ongoing work @cvprconference.bsky.social #CVPR2025 EgoVis workshop?
Submit your abstract before DL of Fri 2 May,
egovis.github.io/cvpr25/#cfp

29.04.2025 18:30 β€” πŸ‘ 5    πŸ” 2    πŸ’¬ 0    πŸ“Œ 0
Preview
UTD Dataset: Mitigating Representation Bias in Video Benchmarks A dataset with textual descriptions and debiased splits for video benchmarks.

πŸš€Excited to announce our CVPR 2025 paper: Unbiasing through Textual Descriptions!
We release new descriptions for 1.9M(!) videos and object-debiased splits for 12 datasets!
πŸ”—Project: utd-project.github.io
by @ninashv.bsky.social et al πŸ§΅πŸ‘‡

@cvprconference.bsky.social

29.04.2025 15:56 β€” πŸ‘ 17    πŸ” 4    πŸ’¬ 1    πŸ“Œ 0

@ninashv is following 19 prominent accounts