Exploiting Continuous Motion Clues for Vision-Based Occupancy Prediction
DOI:
https://doi.org/10.1609/aaai.v39i8.32958Abstract
Occupancy networks aim to reconstruct the surroundings with occupied semantic voxels. However, frequent object occlusions often occur in dynamic real-world scenarios, which cannot be captured by independent frames. Most existing occupancy networks generate results without explicitly considering past occupancy states and continuous visual changes over time, limiting their temporal accuracy. We tackle it by treating the task from a new continuous updating perspective, which considers historical data and continuous motion clues. We propose a new approach termed Continuous Motion clue exploitation for Occupancy Prediction (CMOP), which incorporates three key designs: (i) Propagator: which forecasts future occupancy states based on historical data; (ii) Tracker: which updates the occupancy on a per-frame basis using dynamic visual motion information; and (iii) Fuser: which aggregates results from the Propagator and Tracker into more robust and accurate occupancy results. Experiments on several benchmarks demonstrate that CMOP outperforms state-of-the-art baselines.Downloads
Published
2025-04-11
How to Cite
Xu, H., Peng, P., Zhang, X., Tan, G., Li, Y., Wang, S., & Li, L. (2025). Exploiting Continuous Motion Clues for Vision-Based Occupancy Prediction. Proceedings of the AAAI Conference on Artificial Intelligence, 39(8), 8860-8868. https://doi.org/10.1609/aaai.v39i8.32958
Issue
Section
AAAI Technical Track on Computer Vision VII