Arbitrary-Scale Video Super-resolution Guided by Dynamic Context

Authors

  • Cong Huang University of Science and Technology of China
  • Jiahao Li Microsoft Research Asia
  • Lei Chu Microsoft Research Asia
  • Dong Liu University of Science and Technology of China
  • Yan Lu Microsoft Research Asia

DOI:

https://doi.org/10.1609/aaai.v38i3.28003

Keywords:

CV: Low Level & Physics-based Vision, CV: Applications

Abstract

We propose a Dynamic Context-Guided Upsampling (DCGU) module for video super-resolution (VSR) that leverages temporal context guidance to achieve efficient and effective arbitrary-scale VSR. While most VSR research focuses on backbone design, the importance of the upsampling part is often overlooked. Existing methods rely on pixelshuffle-based upsampling, which has limited capabilities in handling arbitrary upsampling scales. Recent attempts to replace pixelshuffle-based modules with implicit neural function-based and filter-based approaches suffer from slow inference speeds and limited representation capacity, respectively. To overcome these limitations, our DCGU module predicts non-local sampling locations and content-dependent filter weights, enabling efficient and effective arbitrary-scale VSR. Our proposed multi-granularity location search module efficiently identifies non-local sampling locations across the entire low-resolution grid, and the temporal bilateral filter modulation module integrates content information with the filter weight to enhance textual details. Extensive experiments demonstrate the superiority of our method in terms of performance and speed on arbitrary-scale VSR.

Published

2024-03-24

How to Cite

Huang, C., Li, J., Chu, L., Liu, D., & Lu, Y. (2024). Arbitrary-Scale Video Super-resolution Guided by Dynamic Context. Proceedings of the AAAI Conference on Artificial Intelligence, 38(3), 2294-2302. https://doi.org/10.1609/aaai.v38i3.28003

Issue

Section

AAAI Technical Track on Computer Vision II