Yao_Uni4D_Unifying_Visual_Foundation_Models_for_4D_Modeling_from_a@CVPR2025@CVF

Total: 1

#1 Uni4D: Unifying Visual Foundation Models for 4D Modeling from a Single Video [PDF7] [Copy] [Kimi2] [REL]

Authors: David Yifan Yao, Albert J. Zhai, Shenlong Wang

This paper presents a unified approach to understanding dynamic scenes from casual videos. Large pretrained vision models, such as vision-language, video depth prediction, motion tracking, and segmentation models, offer promising capabilities. However, training a single model for comprehensive 4D understanding remains challenging. We introduce Uni4D, a multi-stage optimization framework that harnesses multiple pretrained models to advance dynamic 3D modeling, including static/dynamic reconstruction, camera pose estimation, and dense 3D motion tracking. Our results show state-of-the-art performance in dynamic 4D modeling with superior visual quality. Notably, Uni4D requires no retraining or fine-tuning, highlighting the effectiveness of repurposing large visual models for 4D understanding.

Subject: CVPR.2025 - Highlight