Finally, Felix will present his work on making diffusion transformer training extremely efficient, going from costing multiple months of rent to less than a single night at a conference hotel!
compvis.github.io/tread/
@compvis.bsky.social
Computer Vision and Learning research group @ LMU Munich, headed by Bjรถrn Ommer. Generative Vision (Stable Diffusion, VQGAN) & Representation Learning ๐ https://ommer-lab.com
Finally, Felix will present his work on making diffusion transformer training extremely efficient, going from costing multiple months of rent to less than a single night at a conference hotel!
compvis.github.io/tread/
Stefan and Timy will be talking about how we can achieve extremely efficient motion prediction in open-set settings: bsky.app/profile/stef...
19.10.2025 18:12 โ ๐ 1 ๐ 0 ๐ฌ 1 ๐ 0Pingchuan and Ming will be presenting two works on modeling the evolution of artistic style and disentangled representation learning.
See the following thread for more details bsky.app/profile/pima...
Excited to share that we'll be presenting four papers at the main conference at ICCV 2025 this week!
Come say hi in Honolulu!
๐ Pingchuan, Ming, Felix, Stefan, Timy, and Bjรถrn Ommer will be attending.
๐ From @elsa-ai.eu: 15 new members join the European Lighthouse on Secure & Safe AIโexpanding reach across Europe and deepening ties with the @ellis.eu ecosystem.
Everything you need to know ๐ elsa-ai.eu/elsa-welcome...
Fascinating approach โ encoding an entire image into a single continuous latent token via self-supervised representation learning.
RepTok ๐ฆ highlights how compact generative representations can retain both realism and semantic structure.
๐ค What happens when you poke a scene โ and your model has to predict how the world moves in response?
We built the Flow Poke Transformer (FPT) to model multi-modal scene dynamics from sparse interactions.
It learns to predict the ๐ฅ๐ช๐ด๐ต๐ณ๐ช๐ฃ๐ถ๐ต๐ช๐ฐ๐ฏ of motion itself ๐งต๐
๐๐ฎ๐น๐น ๐ณ๐ผ๐ฟ ๐ณ๐๐น๐น๐ ๐ณ๐๐ป๐ฑ๐ฒ๐ฑ ๐ฃ๐ต๐ ๐ฃ๐ผ๐๐ถ๐๐ถ๐ผ๐ป๐: We are offering several PhD positions across our various research areas, open to highly qualified candidates.
โผ๏ธ The application portal will be open from 15 October to 14 November 2025.
Find out more: mcml.ai/opportunitie...
๐ง ELLIOT on the airwaves!
How do we build open and trustworthy AI in Europe?
๐๏ธ In a recent radio interview, Luk Overmeire from VRT shared insights on ELLIOT, #FoundationModels and the role of public broadcasters in shaping human-centred AI.
๐ป Interview in Dutch: mimir.mjoll.no/shares/JRqlO...
"What makes us human in an AI-shaped world?" โ At #MCML Munich AI Day 2025, Neil Lawrence explored this question, reminding us of the indivisible human core machines can't replicate.
Bjรถrn Ommer followed with insights into how GenAI is commodifying intelligence and reshaping how we use computers.
๐ The ELLIOT project Kick-off Meeting was successfully hosted by CERTH-ITI, in Thessaloniki! ๐๏ธ
30 partners from 12 countries ๐ launched this exciting journey to advance open, trustworthy AI and #FoundationModels across Europe. ๐ค
Stay tuned for more updates on #AIresearch and #TrustworthyAI! ๐ก
๐งน CleanDiFT: Diffusion Features without Noise
@rmsnorm.bsky.social*, @stefanabaumann.bsky.social*, @koljabauer.bsky.social*, @frankfundel.bsky.social, Bjรถrn Ommer
Oral Session 1C (Davidson Ballroom): Friday 9:00
Poster Session 1 (ExHall D): Friday 10:30-12:30, # 218
compvis.github.io/cleandift/
๐ Excited to share that our lab has three papers accepted at CVPR 2025!
Come say hi in Nashville!
๐ Johannes, Ming, Kolja, Stefan, and Bjรถrn will be attending.
๐ข ELLIOT is coming! A โฌ25M #HorizonEurope project to develop open, trustworthy Multimodal Generalist Foundation Models, #MGFM, for real-world applications. Starting July, it brings 30 partners from 12 countries to shape Europeโs #AI future.
๐ Follow for updates on #OpenScience & #FoundationModels.
Continuous Subject-Specific Attribute Control in T2I Models by Identifying Semantic Directions
@stefanabaumann.bsky.social, Felix Krause, Michael Neumayr, @rmsnorm.bsky.social, Melvin Sevi, @vtaohu.bsky.social, Bjรถrn Ommer
P. Sess 3 (ExHall D): Sat 10:30-12:30, #246
compvis.github.io/attribute-co...
Diff2Flow: Training Flow Matching Models via Diffusion Model Alignment
@joh-schb.bsky.social*, @mgui7.bsky.social*, @frankfundel.bsky.social, Bjรถrn Ommer
Poster Session 6 (ExHall D): Sunday 16:00-18:00, # 208
github.com/CompVis/diff...
๐งน CleanDiFT: Diffusion Features without Noise
@rmsnorm.bsky.social*, @stefanabaumann.bsky.social*, @koljabauer.bsky.social*, @frankfundel.bsky.social, Bjรถrn Ommer
Oral Session 1C (Davidson Ballroom): Friday 9:00
Poster Session 1 (ExHall D): Friday 10:30-12:30, # 218
compvis.github.io/cleandift/
๐ Excited to share that our lab has three papers accepted at CVPR 2025!
Come say hi in Nashville!
๐ Johannes, Ming, Kolja, Stefan, and Bjรถrn will be attending.
If you are interested, feel free to check the paper (arxiv.org/abs/2506.02221) or come by at CVPR:
๐ Poster Session 6, Sunday 4:00 to 6:00 PM, Poster #208
Grand Opening of the AI-HUB@LMU. The AI-HUB@LMU is a platform that for the first time unites all 18 faculties of the #LMU as a joint scientific community.
๐
January 29, 2025, 6:00 PM
๐ Groรe Aula, LMU Munich
Full program here: www.ai-news.lmu.de/grand-openin...
Attending my first corporate-sponsored business conference: thereโs a live band playing between talks to keep the energy up.
Meanwhile, academic conferences are struggling to afford coffee breaks. Want this for EPSA!
@compvis.bsky.social
Our method pipeline
๐คWhen combining Vision-language models (VLMs) with Large language models (LLMs), do VLMs benefit from additional genuine semantics or artificial augmentations of the text for downstream tasks?
๐คจInterested? Check out our latest work at #AAAI25:
๐ปCode and ๐Paper at: github.com/CompVis/DisCLIP
๐งต๐
Did you know you can distill the capabilities of a large diffusion model into a small ViT? โ๏ธ
We showed exactly that for a fundamental task:
semantic correspondence๐
A thread ๐งต๐
๐ค Why do we extract diffusion features from noisy images? Isnโt that destroying information?
Yes, it is - but we found a way to do better. ๐
Hereโs how we unlock better features, no noise, no hassle.
๐ Project Page: compvis.github.io/cleandift
๐ป Code: github.com/CompVis/clea...
๐งต๐