Multi-Step Deformable Gaussian Splatting for Dynamic Scene Rendering

Authors

  • Jiaheng Hu School of Computer Science and Technology, East China Normal University, Shanghai, China
  • Zhizhong Zhang School of Computer Science and Technology, East China Normal University, Shanghai, China Shanghai Key Laboratory of Computer Software Evaluating and Testing, Shanghai, China
  • Jingyu Gong School of Computer Science and Technology, East China Normal University, Shanghai, China Chongqing Key Laboratory of Precision Optics, Chongqing Institute of East China Normal University, Chongqing, China Shanghai Key Laboratory of Computer Software Evaluating and Testing, Shanghai, China
  • Lizhuang Ma School of Computer Science and Technology, East China Normal University, Shanghai, China School of Computer Science, Shanghai Jiao Tong University, Shanghai, China
  • Xin Tan School of Computer Science and Technology, East China Normal University, Shanghai, China Chongqing Key Laboratory of Precision Optics, Chongqing Institute of East China Normal University, Chongqing, China
  • Yuan Xie School of Computer Science and Technology, East China Normal University, Shanghai, China Chongqing Key Laboratory of Precision Optics, Chongqing Institute of East China Normal University, Chongqing, China

DOI:

https://doi.org/10.1609/aaai.v40i6.42486

Abstract

Reconstructing dynamic scenes has long been a challenging task in 3D vision. Previous mainstream methods based on 3D Gaussian Splatting typically employ a single deformation field to directly model spatiotemporal changes. However, such one-step deformation struggles to capture diverse and complex motion patterns. To address this limitation, we propose decomposing the one-step deformation into a multi-step process, where each step is represented by a deformation layer. Additionally, we introduce a weight prediction mechanism for each layer to control the extent of deformation at every step. We provide two types of deformation layers based on implicit and explicit approaches. Moreover, while the deformation layer is time-conditioned, the Gaussians' behavior may still be influenced by their time-invariant properties. Therefore, we propose a fully time-agnostic scale modulation block to modulate the scaling changes of Gaussians. Extensive experiments on D-NeRF, Neu3D, and HyperNeRF demonstrate that our method achieves state-of-the-art performance.

Downloads

Published

2026-03-14

How to Cite

Hu, J., Zhang, Z., Gong, J., Ma, L., Tan, X., & Xie, Y. (2026). Multi-Step Deformable Gaussian Splatting for Dynamic Scene Rendering. Proceedings of the AAAI Conference on Artificial Intelligence, 40(6), 4834–4842. https://doi.org/10.1609/aaai.v40i6.42486

Issue

Section

AAAI Technical Track on Computer Vision III