Multi-Step Deformable Gaussian Splatting for Dynamic Scene Rendering
DOI:
https://doi.org/10.1609/aaai.v40i6.42486Abstract
Reconstructing dynamic scenes has long been a challenging task in 3D vision. Previous mainstream methods based on 3D Gaussian Splatting typically employ a single deformation field to directly model spatiotemporal changes. However, such one-step deformation struggles to capture diverse and complex motion patterns. To address this limitation, we propose decomposing the one-step deformation into a multi-step process, where each step is represented by a deformation layer. Additionally, we introduce a weight prediction mechanism for each layer to control the extent of deformation at every step. We provide two types of deformation layers based on implicit and explicit approaches. Moreover, while the deformation layer is time-conditioned, the Gaussians' behavior may still be influenced by their time-invariant properties. Therefore, we propose a fully time-agnostic scale modulation block to modulate the scaling changes of Gaussians. Extensive experiments on D-NeRF, Neu3D, and HyperNeRF demonstrate that our method achieves state-of-the-art performance.Downloads
Published
2026-03-14
How to Cite
Hu, J., Zhang, Z., Gong, J., Ma, L., Tan, X., & Xie, Y. (2026). Multi-Step Deformable Gaussian Splatting for Dynamic Scene Rendering. Proceedings of the AAAI Conference on Artificial Intelligence, 40(6), 4834–4842. https://doi.org/10.1609/aaai.v40i6.42486
Issue
Section
AAAI Technical Track on Computer Vision III