TY - JOUR AU - Lin, Yuankai AU - Cheng, Tao AU - Zhong, Qi AU - Zhou, Wending AU - Yang, Hua PY - 2022/06/28 Y2 - 2024/03/28 TI - Dynamic Spatial Propagation Network for Depth Completion JF - Proceedings of the AAAI Conference on Artificial Intelligence JA - AAAI VL - 36 IS - 2 SE - AAAI Technical Track on Computer Vision II DO - 10.1609/aaai.v36i2.20055 UR - https://ojs.aaai.org/index.php/AAAI/article/view/20055 SP - 1638-1646 AB - Image-guided depth completion aims to generate dense depth maps with sparse depth measurements and corresponding RGB images. Currently, spatial propagation networks (SPNs) are the most popular affinity-based methods in depth completion, but they still suffer from the representation limitation of the fixed affinity and the over smoothing during iterations. Our solution is to estimate independent affinity matrices in each SPN iteration, but it is over-parameterized and heavy calculation.This paper introduces an efficient model that learns the affinity among neighboring pixels with an attention-based, dynamic approach. Specifically, the Dynamic Spatial Propagation Network (DySPN) we proposed makes use of a non-linear propagation model (NLPM). It decouples the neighborhood into parts regarding to different distances and recursively generates independent attention maps to refine these parts into adaptive affinity matrices. Furthermore, we adopt a diffusion suppression (DS) operation so that the model converges at an early stage to prevent over-smoothing of dense depth. Finally, in order to decrease the computational cost required, we also introduce three variations that reduce the amount of neighbors and attentions needed while still retaining similar accuracy. In practice, our method requires less iteration to match the performance of other SPNs and yields better results overall. DySPN outperforms other state-of-the-art (SoTA) methods on KITTI Depth Completion (DC) evaluation by the time of submission and is able to yield SoTA performance in NYU Depth v2 dataset as well. ER -