A. Sophia Koepke's Avatar

A. Sophia Koepke

@askoepke.bsky.social

Junior research group leader at TUM | University of TΓΌbingen. Currently at BAIR (Berkeley). Previously at VGG (Oxford). Interested in multi-modal learning. πŸ”— https://akoepke.github.io/

645 Followers  |  348 Following  |  9 Posts  |  Joined: 17.11.2024  |  1.6698

Latest posts by askoepke.bsky.social on Bluesky


Thanks to Daniil Zverev*, @thwiedemer.bsky.social*, @bayesiankitten.bsky.social, Matthias Bethge (@bethgelab.bsky.social), and @wielandbrendel.bsky.social for making VGGSound sounder! πŸ™Œ πŸŽ‰ πŸ—

21.10.2025 18:08 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
VGGSounder: Audio-Visual Evaluations for Foundation Models VGGSounder, a multi-label audio-visual classification dataset with modality annotations.

πŸ“Š With VGGSounder, we show that existing models don’t always benefit from multimodal input and sometimes performance even degrades.

Code and data: vggsounder.github.io

21.10.2025 18:07 β€” πŸ‘ 2    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

VGGSounder is a new video classification benchmark for audio-visual foundation models:

We provide:
πŸ“’ Re-annotated VGGSound test set
πŸ“’ Modality-specific manual labels
πŸ“’ A modality confusion metric to diagnose when models misuse modalities

Paper: arxiv.org/pdf/2508.08237

21.10.2025 18:06 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 1    πŸ“Œ 0

πŸŽ‰ Excited to present our paper VGGSounder: Audio‑Visual Evaluations for Foundation Models today at #ICCV2025!

πŸ•¦ Poster Session 1 | 11:30–13:30
πŸ“ Poster #88

Come by if you're into audio-visual learning and want to know whether multiple modalities actually help or hurt.

21.10.2025 18:06 β€” πŸ‘ 6    πŸ” 1    πŸ’¬ 1    πŸ“Œ 0

Thanks to @munichcenterml.bsky.social for supporting the workshop with a best paper award (announced at 2.50pm CDT)!

11.06.2025 17:59 β€” πŸ‘ 1    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0

We have fantastic speakers, including @saining.bsky.social, @aidanematzadeh.bsky.social, @ranjaykrishna.bsky.social, Ludwig Schmidt, @lisadunlap.bsky.social, and Ishan Misra.

11.06.2025 17:57 β€” πŸ‘ 0    πŸ” 0    πŸ’¬ 0    πŸ“Œ 0
Preview
EVAL-FoMo 2 - Schedule Date: June 11 (1:00pm - 6:00pm)

Our #CVPR2025 workshop on Emergent Visual Abilities and Limits of Foundation Models (EVAL-FoMo) is taking place this afternoon (1-6pm) in room 210.

Workshop schedule: sites.google.com/view/eval-fo...

11.06.2025 17:55 β€” πŸ‘ 7    πŸ” 3    πŸ’¬ 3    πŸ“Œ 0
Screenshot of the workshop website "Emergent Visual Abilities and Limits of Foundation Models" at CVPR 2025

Screenshot of the workshop website "Emergent Visual Abilities and Limits of Foundation Models" at CVPR 2025

Our paper submission deadline for the EVAL-FoMo workshop @cvprconference.bsky.social has been extended to March 19th!
sites.google.com/view/eval-fo...
We welcome submissions (incl. published papers) on the analysis of emerging capabilities / limits in visual foundation models. #CVPR2025

12.03.2025 11:41 β€” πŸ‘ 12    πŸ” 5    πŸ’¬ 0    πŸ“Œ 1

Our 2nd Workshop on Emergent Visual Abilities and Limits of Foundation Models (EVAL-FoMo) is accepting submissions. We are looking forward to talks by our amazing speakers that include @saining.bsky.social, @aidanematzadeh.bsky.social, @lisadunlap.bsky.social, and @yukimasano.bsky.social. #CVPR2025

13.02.2025 16:02 β€” πŸ‘ 7    πŸ” 3    πŸ’¬ 0    πŸ“Œ 1

Upcoming π— π˜‚π—»π—Άπ—°π—΅ π—”π—œ π—Ÿπ—²π—°π˜π˜‚π—Ώπ—² featuring Prof. Franca Hoffmann from California Institute of Technology and Prof. Holger Hoos from RWTH Aachen University: munichlectures.ai

πŸ—“οΈ December 17, 2024
πŸ•™ 16:00 CET
🏫 Senatssaal, #LMU Munich

09.12.2024 15:03 β€” πŸ‘ 4    πŸ” 1    πŸ’¬ 2    πŸ“Œ 1
Post image

Kicking off our TUM AI - Lecture Series tomorrow with none other than Jiaming Song, CSO @LumaLabsAI.

He'll be talking about "Dream Machine: Emergent Capabilities from Video Foundation Models".

Live stream: youtu.be/oilWwsXZamA
7pm GMT+1 / 10am PST (Mon Dec 2nd)

01.12.2024 12:55 β€” πŸ‘ 42    πŸ” 6    πŸ’¬ 1    πŸ“Œ 0

@askoepke is following 20 prominent accounts