DLDA: Unified Dual-Level Domain Adaptation for Low-Light Object Detection
DOI:
https://doi.org/10.1609/aaai.v40i6.42488Abstract
Low-light object detection faces significant challenges due to the substantial domain shift between normal-light and low-light conditions. Prior works often enhance low-light images before detection, but this preprocessing can introduce artifacts that degrade detection performance since it focuses on human visual quality rather than task-specific features. Other methods incorporate illumination-aware modules for low-light feature learning, yet their scalability is limited by the scarcity of annotated low-light datasets. To overcome these limitations, we propose a unified Dual-Level Domain Adaptation (DLDA) framework that jointly addresses image-level and feature-level domain discrepancies. Specifically, we introduce a luminance-aware contrastive translation module that synthesizes target-style low-light images while preserving structural details, enabling effective image-level adaptation. Building on this, we further design a multi-scale conditional adversarial alignment strategy that promotes semantic consistency across feature hierarchies to enhance domain-invariant feature extraction. Extensive experiments on multiple low-light detection benchmarks demonstrate that DLDA achieves state-of-the-art performance, exhibiting strong robustness and generalization.Published
2026-03-14
How to Cite
Hu, J., Zhao, Q., & Li, G. (2026). DLDA: Unified Dual-Level Domain Adaptation for Low-Light Object Detection. Proceedings of the AAAI Conference on Artificial Intelligence, 40(6), 4851–4859. https://doi.org/10.1609/aaai.v40i6.42488
Issue
Section
AAAI Technical Track on Computer Vision III