Yue Chen's Avatar

Yue Chen

@fanegg.bsky.social

PhD Student at Westlake University. 3D/4D Reconstruction, Virtual Humans. fanegg.github.io

81 Followers  |  163 Following  |  11 Posts  |  Joined: 18.01.2025  |  1.7223

Latest posts by fanegg.bsky.social on Bluesky

Preview
GitHub - fanegg/Human3R: An unified model for 4D human-scene reconstruction An unified model for 4D human-scene reconstruction - fanegg/Human3R

Code, model and 4D interactive demo now available

๐Ÿ”—Page: fanegg.github.io/Human3R
๐Ÿ“„Paper: arxiv.org/abs/2510.06219
๐Ÿ’ปCode: github.com/fanegg/Human3R

Big thanks to our awesome team!
@fanegg.bsky.social @xingyu-chen.bsky.social Yuxuan Xue @apchen.bsky.social @xiuyuliang.bsky.social Gerard Pons-Moll

08.10.2025 08:54 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Video thumbnail

GT comparison shows our feedforward method, without any iterative optimization, is not only fast but also accurate.

This is achieved by reading out humans from a 4D foundation model, #CUT3R, with our proposed ๐™๐™ช๐™ข๐™–๐™ฃ ๐™ฅ๐™ง๐™ค๐™ข๐™ฅ๐™ฉ ๐™ฉ๐™ช๐™ฃ๐™ž๐™ฃ๐™œ.

08.10.2025 08:51 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

Bonus: #Human3R is also a compact human tokenizer!

Our human tokens capture ID+ shape + pose + position of human, unlocking ๐˜๐—ฟ๐—ฎ๐—ถ๐—ป๐—ถ๐—ป๐—ด-๐—ณ๐—ฟ๐—ฒ๐—ฒ 4D tracking.

08.10.2025 08:50 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

#Human3R: Everyone Everywhere All at Once

Just input a RGB video, we online reconstruct 4D humans and scene in ๐—ข๐—ป๐—ฒ model and ๐—ข๐—ป๐—ฒ stage.

Training this versatile model is easier than you think โ€“ it just takes ๐—ข๐—ป๐—ฒ day using ๐—ข๐—ป๐—ฒ GPU!

๐Ÿ”—Page: fanegg.github.io/Human3R/

08.10.2025 08:49 โ€” ๐Ÿ‘ 2    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

Again, training-free is all you need.

01.10.2025 07:06 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Post image Post image

Excited to introduce LoftUp!

A strong (than ever) and lightweight feature upsampler for vision encoders that can boost performance on dense prediction tasks by 20%โ€“100%!

Easy to plug into models like DINOv2, CLIP, SigLIP โ€” simple design, big gains. Try it out!

github.com/andrehuang/l...

22.04.2025 07:55 โ€” ๐Ÿ‘ 19    ๐Ÿ” 5    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

I was really surprised when I saw this. Dust3R has learned very well to segment objects without supervision. This knowledge can be extracted post-hoc, enabling accurate 4D reconstruction instantly.

01.04.2025 18:45 โ€” ๐Ÿ‘ 31    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Just "dissect" the cross-attention mechanism of #DUSt3R, making 4D reconstruction easier.

01.04.2025 15:45 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

#Easi3R is a simple training-free approach adapting DUSt3R for dynamic scenes.

01.04.2025 15:45 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
[CVPR 2025] Feat2GS: Probing Visual Foundation Models with Gaussian Splatting
YouTube video by Yue Chen [CVPR 2025] Feat2GS: Probing Visual Foundation Models with Gaussian Splatting

๐Ÿ’ปCode: github.com/fanegg/Feat2GS
๐ŸŽฅVideo: youtu.be/4fT5lzcAJqo?...

Big thanks to the amazing team!
@fanegg.bsky.social, @xingyu-chen.bsky.social, Anpei Chen, Gerard Pons-Moll, Yuliang Xiu

#DUSt3R #MASt3R #MiDaS #DINOv2 #DINO #SAM #CLIP #RADIO #MAE #StableDiffusion #Zero123

31.03.2025 16:11 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Video thumbnail

Our findings in 3D probe lead to a simple-yet-effective solution, by just combining features from different visual foundation models and outperform prior works.

Apply #Feat2GS in sparse & causal captures:
๐Ÿค—Online Demo: huggingface.co/spaces/endle...

31.03.2025 16:08 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

With #Feat2GS we evaluated more than 10 visual foundation models (DUSt3R, DINO, MAE, SAM, CLIP, MiDas, etc) in terms of geometry and texture โ€” see the paper for comparison.

๐Ÿ“„Paper: arxiv.org/abs/2412.09606
๐Ÿ”Try it NOW: fanegg.github.io/Feat2GS/#chart

31.03.2025 16:07 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Video thumbnail

How much 3D do visual foundation models (VFMs) know?

Previous work requires 3D data for probing โ†’ expensive to collect!

#Feat2GS @cvprconference.bsky.social 2025 - our idea is to read out 3D Gaussains from VFMs features, thus probe 3D with novel view synthesis.

๐Ÿ”—Page: fanegg.github.io/Feat2GS

31.03.2025 16:06 โ€” ๐Ÿ‘ 24    ๐Ÿ” 7    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1

@fanegg is following 20 prominent accounts