Building an End-to-End Spatial-Temporal Convolutional Network for Video Super-Resolution

Authors

  • Jun Guo Sun Yat-sen University
  • Hongyang Chao Sun Yat-sen University

DOI:

https://doi.org/10.1609/aaai.v31i1.11228

Keywords:

video super-resolution, spatial-temporal network, deep learning

Abstract

We propose an end-to-end deep network for video super-resolution. Our network is composed of a spatial component that encodes intra-frame visual patterns, a temporal component that discovers inter-frame relations, and a reconstruction component that aggregates information to predict details. We make the spatial component deep, so that it can better leverage spatial redundancies for rebuilding high-frequency structures. We organize the temporal component in a bidirectional and multi-scale fashion, to better capture how frames change across time. The effectiveness of the proposed approach is highlighted on two datasets, where we observe substantial improvements relative to the state of the arts.

Downloads

Published

2017-02-12

How to Cite

Guo, J., & Chao, H. (2017). Building an End-to-End Spatial-Temporal Convolutional Network for Video Super-Resolution. Proceedings of the AAAI Conference on Artificial Intelligence, 31(1). https://doi.org/10.1609/aaai.v31i1.11228