LaMAR: Laplacian Pyramid for Multimodal Adaptive Super Resolution (Student Abstract)
DOI:
https://doi.org/10.1609/aaai.v38i21.30463Keywords:
AI Architectures, Computer Vision, Machine Learning, Sensor Fusion, Applications Of AIAbstract
Recent advances in image-to-image translation involve the integration of non-visual imagery in deep models. Non-visual sensors, although more costly, often produce low-resolution images. To combat this, methods using RGB images to enhance the resolution of these modalities have been introduced. Fusing these modalities to achieve high-resolution results demands models with millions of parameters and extended inference times. We present LaMAR, a lightweight model. It employs Laplacian image pyramids combined with a low-resolution thermal image for Guided Thermal Super Resolution. By decomposing the RGB image into a Laplacian pyramid, LaMAR preserves image details and avoids high-resolution feature map computations, ensuring efficiency. With faster inference times and fewer parameters, our model demonstrates state-of-the-art results.Downloads
Published
2024-03-24
How to Cite
Kasliwal, A., Kamani, A., Gakhar, I., Seth, P., & Rallabandi, S. (2024). LaMAR: Laplacian Pyramid for Multimodal Adaptive Super Resolution (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 38(21), 23539-23541. https://doi.org/10.1609/aaai.v38i21.30463
Issue
Section
AAAI Student Abstract and Poster Program