Matthias Niessner's Avatar

Matthias Niessner

@niessner.bsky.social

Professor for Visual Computing & Artificial Intelligence @TU Munich Co-Founder @synthesiaIO Co-Founder @SpAItialAI https://niessnerlab.org/publications.html

2,433 Followers  |  64 Following  |  101 Posts  |  Joined: 18.11.2024
Posts Following

Posts by Matthias Niessner (@niessner.bsky.social)

significantly outperforms generic visual pretraining (e.g., DINO-style features) in terms of generalization.

๐ŸŒhttps://simongiebenhain.github.io/Pix2NPHM
๐ŸŽฅhttps://youtu.be/MgpEJC5p1Ts

Great work by Simon Giebenhain, Tobias Kirschstein, Liam Schoneveld, Davide Davoli, Zhe Chen.

23.12.2025 16:30 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

(1) large-scale registration of existing 3D head datasets, and
(2) self-supervised training on vast in-the-wild 2D video datasets using pseudo ground-truth surface normals.
Finally, we show that geometry-aware pretraining on pixel-aligned reconstruction tasks

23.12.2025 16:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Pix2NPHM obtains fast and reliable NPHM reconstructions on real-world data. Inference-time optimization against surface normals and canonical point maps can further increase fidelity.
Key to successful and generalized training of our ViT-based network are:

23.12.2025 16:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Face tracking & 3D reconstruction are often limited by the representational capacity of PCA-based face models. By lifting NPHMs to a first-class reconstruction primitive, we enable more accurate geometry, richer expressions, and finer animation control.

23.12.2025 16:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
๐Ÿ“ขPix2NPHM: Learning to Regress NPHM Reconstructions From a Single Image๐Ÿ“ข We directly regress neural parametric head models (NPHMs) from a single image โ€” fast, stable, and significantly moreโ€ฆ | Matthia... ๐Ÿ“ขPix2NPHM: Learning to Regress NPHM Reconstructions From a Single Image๐Ÿ“ข We directly regress neural parametric head models (NPHMs) from a single image โ€” fast, stable, and significantly more expressive...

๐Ÿ“ขPix2NPHM: Learning to Regress NPHM Reconstructions From a Single Image๐Ÿ“ข

We directly regress neural parametric head models (NPHMs) from a single image โ€” fast, stable, and significantly more expressive than classical 3DMMs such as FLAME.

23.12.2025 16:30 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

๐ŸŒhttps://peter-kocsis.github.io/IntrinsicImageFusion
๐ŸŽฅhttps://youtu.be/-Vs3tR1Xl7k

Great work by Peter Kocsis and Lukas Hollein!

17.12.2025 15:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

3) optimize low-dimensional parameters for physically-grounded reconstructions.

The results are relightable PBR textures for 3D scenes: check out the result on a real-world 3D scan from the ScanNet++ dataset!

17.12.2025 15:28 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

๐Ÿ“ข Intrinsic Image Fusion for Multi-View 3D Material Reconstruction ๐Ÿ“ข

We combine generative material priors with inverse path tracing: 1) define a parametric texture space 2) fuse monocular predictions across views into consistent textures

17.12.2025 15:28 โ€” ๐Ÿ‘ 8    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Preview
TUM AI Lecture Series - Building generative world models: progress and challenges (Ruiqi Gaoi) Abstract: Equipping AI models with the ability to imagine, reason, and act in the physical world is a crucial step toward achieving Artificial General Intelligence (AGI). Generative world models, whic...

Today in our TUM AI - Lecture Series we'll have the amazing Ruiqi Gao, Google DeepMind.

She'll talk about "๐๐ฎ๐ข๐ฅ๐๐ข๐ง๐  ๐ ๐ž๐ง๐ž๐ซ๐š๐ญ๐ข๐ฏ๐ž ๐ฐ๐จ๐ซ๐ฅ๐ ๐ฆ๐จ๐๐ž๐ฅ๐ฌ: progress and challenges".

Live stream: www.youtube.com/live/CkOSMqw...

7pm GMT+1 / 10am PST (Tue Dec 16th).

16.12.2025 10:49 โ€” ๐Ÿ‘ 8    ๐Ÿ” 2    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
PercHead: Perceptual Head Model for Single-Image 3D Head Reconstruction & Editing PercHead: Perceptual Head Model for Single-Image 3D Head Reconstruction & Editing

We also provide an interactive GUI to enable the exploration of our editing pipeline.

๐ŸŒ antoniooroz.github.io/PercHead/
๐Ÿ“ฝ๏ธ youtu.be/4hFybgTk4kE

Great work by Antonio Oroz and and Tobias Kirschstein

05.11.2025 11:37 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

by swapping the encoder, we can transform the model into a disentangled 3D editing pipeline. In this scenario, we can control geometry through - potentially hand-drawn - segmentation maps, and condition style via image or text prompt.

05.11.2025 11:37 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
PercHead: Perceptual Head Model for Single-Image 3D Head Reconstruction & Editing
YouTube video by Matthias Niessner PercHead: Perceptual Head Model for Single-Image 3D Head Reconstruction & Editing

Our trained reconstruction model is able to generate 3D-consistent heads from a single input image. Even with challenging side-view inputs, the model robustly infers missing regions for a coherent, high-fidelity output.

In addition, our architecture seamlessly adapts to downstream tasks:

05.11.2025 11:37 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
PercHead: Perceptual Head Model for Single-Image 3D Head Reconstruction & Editing PercHead: Perceptual Head Model for Single-Image 3D Head Reconstruction & Editing

At its core is a generalized 3D head decoder trained with perceptual supervision from DINOv2 and SAM 2.1. We find that our new perceptual loss formulation improves reconstruction fidelity compared to commonly-used methods such as LPIPS.

05.11.2025 11:37 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

๐Ÿ“ข๐Ÿ“ข ๐๐ž๐ซ๐œ๐‡๐ž๐š๐: ๐๐ž๐ซ๐œ๐ž๐ฉ๐ญ๐ฎ๐š๐ฅ ๐‡๐ž๐š๐ ๐Œ๐จ๐๐ž๐ฅ ๐Ÿ๐จ๐ซ ๐’๐ข๐ง๐ ๐ฅ๐ž-๐ˆ๐ฆ๐š๐ ๐ž ๐Ÿ‘๐ƒ ๐‡๐ž๐š๐ ๐‘๐ž๐œ๐จ๐ง๐ฌ๐ญ๐ซ๐ฎ๐œ๐ญ๐ข๐จ๐ง & ๐„๐๐ข๐ญ๐ข๐ง๐ ๐Ÿ“ข๐Ÿ“ข

PercHead reconstructs realistic 3D heads from a single image and enables disentangled 3D editing via geometric controls and style inputs from images or text.

05.11.2025 11:37 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

#ICCV last week was incredible โ€” catching up with so many people, chatting about research, and, most importantly, having lots of fun.

Still hard to fathom this privilege as a researcher โ€” getting to travel to such amazing places and be part of this brilliant community - Thanks!

29.10.2025 12:25 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

The hot topic at #ICCV2025 was World Models.

They come in different flavors โ€” (interactive) video models, neural simulators, reconstruction models, etc. โ€” but the overarching goal is clear: Generative AI that predict and simulate how the real world works.

26.10.2025 15:59 โ€” ๐Ÿ‘ 15    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

Hawaii on the same scale as the United Kingdom.

24.10.2025 08:33 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

๐บ๐‘’๐‘›๐‘’๐‘Ÿ๐‘Ž๐‘ก๐‘œ ๐‘’๐‘Ÿ๐‘”๐‘œ ๐‘ ๐‘ข๐‘š โ€” I generate, therefore I am.

16.10.2025 12:33 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
GitHub - scannetpp/scannetpp: [ICCV 2023 Oral] ScanNet++: A High-Fidelity Dataset of 3D Indoor Scenes [ICCV 2023 Oral] ScanNet++: A High-Fidelity Dataset of 3D Indoor Scenes - scannetpp/scannetpp

for more documentation: github.com/scannetpp/sc...

Huge thanks to Yueh-Cheng Liu, as well as Chandan Yeshwanth and @niessner.bsky.social for their incredible work!

13.10.2025 16:19 โ€” ๐Ÿ‘ 5    ๐Ÿ” 1    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

On the bright side, tooling for training has dramatically improved since then. Deep learning frameworks (PyTorch et. al) and scheduling systems such as SLURM or Kubernetes have become the backbone of modern AI.

12.10.2025 15:46 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Given the humongous compute demands of recent generative frontier AI models -- LLMs, image, and video models, etc. --, where compute is measured in Gigawatts, these challenges seem quite amusing.

12.10.2025 15:46 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

The required compute was typically a couple of GPUs on a single desktop machine, trained over several days; e.g., AlexNet was trained on two GTX 580 3GB GPUs for 5-6 days.

12.10.2025 15:46 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

In the 'early days' of modern deep learning (2012-2015) when ConvNets such as AlexNet or VGG came out, it was considered almost impractical to train an ImageNet classifier from scratch.

12.10.2025 15:46 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Post image Post image Post image

Fantastic retreat this weekend by our research groups!

Internal reviews, ideas brainstorming, paper reading, and much more! Of course also many social activities -- the highlight being our kayaking trip - lots of fun :)

28.09.2025 09:12 โ€” ๐Ÿ‘ 11    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image

All six of our submissions were accepted to #NeurIPS2025 ๐ŸŽ‰๐Ÿฅณ

Awesome works about Gaussian Splatting Primitives, Lighting Estimation, Texturing, and much more GenAI :)

Great work by Peter Kocsis, Yujin Chen, Zhening Huang, Jiapeng Tang, Nicolas von Lรผtzow, Jonathan Schmidt ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ

18.09.2025 16:15 โ€” ๐Ÿ‘ 10    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We generate multiple videos along short, pre-defined trajectories that explore the scene in depth. Our scene memory conditions each video on the most relevant prior views while avoiding collisions.

Great work by Manuel Schneider & @LukasHollein

17.09.2025 12:08 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

Can we use video diffusion to generate 3D scenes?

๐–๐จ๐ซ๐ฅ๐๐„๐ฑ๐ฉ๐ฅ๐จ๐ซ๐ž๐ซ (#SIGGRAPHAsia25) creates fully-navigable scenes via autoregressive video generation.

Text input -> 3DGS scene output & interactive rendering!

๐ŸŒhttp://mschneider456.github.io/world-explorer/
๐Ÿ“ฝ๏ธhttps://youtu.be/N6NJsNyiv6I

17.09.2025 12:08 โ€” ๐Ÿ‘ 16    ๐Ÿ” 2    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 1
ScaffoldAvatar: High-Fidelity Gaussian Avatars with Patch Expressions

We further propose a color-based densification and progressive training scheme for improved quality and faster convergence.

shivangi-aneja.github.io/projects/sca...
youtu.be/VyWkgsGdbkk

Great work by Shivangi Aneja, Sebastian Weiss, Irene Baeza Rojo, Prashanth Chandran, Gaspard Zoss, Derek Bradley

05.08.2025 12:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
ScaffoldAvatar: High-Fidelity Gaussian Avatars with Patch Expressions

We operate on patch-based local expression features and increase the representation capacity by synthesizing 3D Gaussians dynamically by leveraging tiny scaffold MLPs conditioned on localized expressions.

05.08.2025 12:30 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

ScaffoldAvatar: High-Fidelity Gaussian Avatars with Patch Expressions (#SIGGRAPH)

We reconstruct ultra-high fidelity photorealistic 3D avatars capable of generating realistic and high-quality animations including freckles and other fine facial details.

shivangi-aneja.github.io/projects/sca...

05.08.2025 12:30 โ€” ๐Ÿ‘ 6    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0