SpatialActor: Exploring Disentangled Spatial Representations for Robust Robotic Manipulation
DOI:
https://doi.org/10.1609/aaai.v40i11.37852Abstract
Robotic manipulation requires precise spatial understanding to interact with objects in the real world. Point-based methods suffer from sparse sampling, leading to the loss of fine-grained semantics. Image-based methods typically feed RGB and depth into 2D backbones pre-trained on 3D auxiliary tasks, but their entangled semantics and geometry are sensitive to inherent depth noise in real-world that disrupts semantic understanding. Moreover, these methods focus on high-level geometry while overlooking low-level spatial cues essential for precise interaction. We propose SpatialActor, a disentangled framework for robust robotic manipulation that explicitly decouples semantics and geometry. The Semantic-guided Geometric Module adaptively fuses two complementary geometry from noisy depth and semantic-guided expert priors. Also, a Spatial Transformer leverages low-level spatial cues for accurate 2D-3D mapping and enables interaction among spatial features. We evaluate SpatialActor on multiple simulation and real-world scenarios across 50+ tasks. It achieves state-of-the-art performance with 87.4% on RLBench and improves by 13.9% to 19.4% under varying noisy conditions, showing strong robustness. Moreover, it significantly enhances few-shot generalization to new tasks and maintains robustness under various spatial perturbations.Downloads
Published
2026-03-14
How to Cite
Shi, H., Xie, B., Liu, Y., Yue, Y., Wang, T., Fan, H., Zhang, X., & Huang, G. (2026). SpatialActor: Exploring Disentangled Spatial Representations for Robust Robotic Manipulation. Proceedings of the AAAI Conference on Artificial Intelligence, 40(11), 8969-8977. https://doi.org/10.1609/aaai.v40i11.37852
Issue
Section
AAAI Technical Track on Computer Vision VIII