Due to the complex and highly dynamic motions in the real world, synthesizing dynamic videos from multi-view inputs for arbitrary viewpoints is challenging. Previous works based on neural radiance field or 3D Gaussian splatting are limited to modeling fine-scale motion, greatly restricting their application. In this paper, we introduce LocalDyGS, which consists of two parts to adapt our method to both large-scale and fine-scale motion scenes: 1) We decompose a complex dynamic scene into streamlined local spaces defined by seeds, enabling global modeling by capturing motion within each local space. 2) We decouple static and dynamic features for local space motion modeling. A static feature shared across time steps captures static information, while a dynamic residual field provides time-specific features. These are combined and decoded to generate Temporal Gaussians, modeling motion within each local space. As a result, we propose a novel dynamic scene reconstruction framework to model highly dynamic real-world scenes more realistically. Our method not only demonstrates competitive performance on various fine-scale datasets compared to state-of-the-art (SOTA) methods, but also represents the first attempt to model larger and more complex highly dynamic scenes. Code and models will be made publicly available.
Unlike previous methods, our approach is the first to achieve high-quality modeling of large-scale dynamic scenes.
Ours
3DGStream
SpaceTimeGS
4DGaussian
Ours
Using LocalDyGS you can create fun visual effects (fix viewpoint and free viewpoints).
SuperPointGS
Ours
4DGaussian
Ours
A by-product of our method enables adaptive dynamic-static segmentation.
Ours
@article{Wu2025LocalDyGS,
author = {Jiahao Wu, Rui Peng, Jianbo Jiao, Jiayu Yang, Luyang Tang,Kaiqiang Xiong, Jie Liang, Jinbo Yan, Runling Liu, Ronggang Wang},
title = {LocalDyGS: Multi-view Global Dynamic Scene Modeling via Adaptive Local Implicit Feature Decoupling.},
journal = {ICCV},
year = {2025},
}