Active3D: Active High-Fidelity 3D Reconstruction via Multi-Level Uncertainty Quantification
DOI:
https://doi.org/10.1609/aaai.v40i8.37585Abstract
In this paper, we present an active exploration framework for high-fidelity 3D reconstruction that incrementally builds a multi-level uncertainty space and selects next-best-views through an uncertainty-driven motion planner. We introduce a hybrid implicit–explicit representation that fuses neural fields with Gaussian primitives to jointly capture global structural priors and locally observed details. Based on this hybrid state, we derive a hierarchical uncertainty volume that quantifies both implicit global structure quality and explicit local surface confidence. To focus optimization on the most informative regions, we propose an uncertainty-driven keyframe selection strategy that anchors high-entropy viewpoints as sparse attention nodes, coupled with a viewpoint-space sliding window for uncertainty-aware local refinement. The planning module formulates next-best-view selection as an Expected Hybrid Information Gain problem and incorporates a risk-sensitive path planner to ensure efficient and safe exploration. Extensive experiments on challenging benchmarks demonstrate that our approach consistently achieves state-of-the-art accuracy, completeness, and rendering quality, highlighting its effectiveness for real-world active reconstruction and robotic perception tasks.Downloads
Published
2026-03-14
How to Cite
Li, Y., Li, Y., & Lee, G. H. (2026). Active3D: Active High-Fidelity 3D Reconstruction via Multi-Level Uncertainty Quantification. Proceedings of the AAAI Conference on Artificial Intelligence, 40(8), 6557–6565. https://doi.org/10.1609/aaai.v40i8.37585
Issue
Section
AAAI Technical Track on Computer Vision V