CompEvent: Complex-valued Event-RGB Fusion for Low-light Video Enhancement and Deblurring
DOI:
https://doi.org/10.1609/aaai.v40i16.38358Abstract
Low-light video deblurring poses significant challenges in applications like nighttime surveillance and autonomous driving due to dim lighting and long exposures. While event cameras offer potential solutions with superior low-light sensitivity and high temporal resolution, existing fusion methods typically employ staged strategies, limiting their effectiveness against combined low-light and motion blur degradations. To overcome this, we propose CompEvent, a complex neural network framework enabling holistic full-process fusion of event data and RGB frames for enhanced joint restoration. CompEvent features two core components: 1) Complex Temporal Alignment GRU, which utilizes complex-valued convolutions and processes video and event streams iteratively via GRU to achieve temporal alignment and continuous fusion; and 2) Complex Space-Frequency Learning module, which performs unified complex-valued signal processing in both spatial and frequency domains, facilitating deep fusion through spatial structures and system-level characteristics. By leveraging the holistic representation capability of complex-valued neural networks, CompEvent achieves full-process spatiotemporal fusion, maximizes complementary learning between modalities, and significantly strengthens low-light video deblurring capability. Extensive experiments demonstrate that CompEvent outperforms SOTA methods in addressing this challenging task.Downloads
Published
2026-03-14
How to Cite
Zhong, M., Lu, X., Li, D., Xu, S., Jiang, R., Fu, X., & Yin, B. (2026). CompEvent: Complex-valued Event-RGB Fusion for Low-light Video Enhancement and Deblurring. Proceedings of the AAAI Conference on Artificial Intelligence, 40(16), 13530–13538. https://doi.org/10.1609/aaai.v40i16.38358
Issue
Section
AAAI Technical Track on Computer Vision XIII