Ultra-High-Definition Low-Light Image Enhancement: A Benchmark and Transformer-Based Method
DOI:
https://doi.org/10.1609/aaai.v37i3.25364Keywords:
CV: Low Level & Physics-Based Vision, CV: Computational Photography, Image & Video Synthesis, CV: Scene Analysis & Understanding, CMS: ApplicationsAbstract
As the quality of optical sensors improves, there is a need for processing large-scale images. In particular, the ability of devices to capture ultra-high definition (UHD) images and video places new demands on the image processing pipeline. In this paper, we consider the task of low-light image enhancement (LLIE) and introduce a large-scale database consisting of images at 4K and 8K resolution. We conduct systematic benchmarking studies and provide a comparison of current LLIE algorithms. As a second contribution, we introduce LLFormer, a transformer-based low-light enhancement method. The core components of LLFormer are the axis-based multi-head self-attention and cross-layer attention fusion block, which significantly reduces the linear complexity. Extensive experiments on the new dataset and existing public datasets show that LLFormer outperforms state-of-the-art methods. We also show that employing existing LLIE methods trained on our benchmark as a pre-processing step significantly improves the performance of downstream tasks, e.g., face detection in low-light conditions. The source code and pre-trained models are available at https://github.com/TaoWangzj/LLFormer.Downloads
Published
2023-06-26
How to Cite
Wang, T., Zhang, K., Shen, T., Luo, W., Stenger, B., & Lu, T. (2023). Ultra-High-Definition Low-Light Image Enhancement: A Benchmark and Transformer-Based Method. Proceedings of the AAAI Conference on Artificial Intelligence, 37(3), 2654-2662. https://doi.org/10.1609/aaai.v37i3.25364
Issue
Section
AAAI Technical Track on Computer Vision III