Scale-Wise Convolution for Image Restoration


  • Yuchen Fan University of Illinois at Urbana-Champaign
  • Jiahui Yu University of Illinois at Urbana-Champaign
  • Ding Liu Bytedance Inc.
  • Thomas S. Huang University of Illinois at Urbana-Champaign



While scale-invariant modeling has substantially boosted the performance of visual recognition tasks, it remains largely under-explored in deep networks based image restoration. Naively applying those scale-invariant techniques (e.g., multi-scale testing, random-scale data augmentation) to image restoration tasks usually leads to inferior performance. In this paper, we show that properly modeling scale-invariance into neural networks can bring significant benefits to image restoration performance. Inspired from spatial-wise convolution for shift-invariance, “scale-wise convolution” is proposed to convolve across multiple scales for scale-invariance. In our scale-wise convolutional network (SCN), we first map the input image to the feature space and then build a feature pyramid representation via bi-linear down-scaling progressively. The feature pyramid is then passed to a residual network with scale-wise convolutions. The proposed scale-wise convolution learns to dynamically activate and aggregate features from different input scales in each residual building block, in order to exploit contextual information on multiple scales. In experiments, we compare the restoration accuracy and parameter efficiency among our model and many different variants of multi-scale neural networks. The proposed network with scale-wise convolution achieves superior performance in multiple image restoration tasks including image super-resolution, image denoising and image compression artifacts removal. Code and models are available at:




How to Cite

Fan, Y., Yu, J., Liu, D., & Huang, T. S. (2020). Scale-Wise Convolution for Image Restoration. Proceedings of the AAAI Conference on Artificial Intelligence, 34(07), 10770-10777.



AAAI Technical Track: Vision