RIS-LAD: A Benchmark and Model for Referring Image Segmentation in Low-Altitude Drone Imagery
DOI:
https://doi.org/10.1609/aaai.v40i14.38181Abstract
Referring Image Segmentation (RIS), which aims to segment specific objects based on natural language descriptions, plays an essential role in vision-language understanding. Despite its progress in remote sensing applications, RIS under Low-Altitude Drone (LAD) scenarios remains underexplored, as existing datasets and methods are typically designed for high-altitude and static-view imagery. They struggled to handle the unique characteristics of LAD views, such as diverse viewpoints and high object density. In this paper, we propose RIS-LAD, the first fine-grained RIS benchmark tailored for LAD scenarios, featuring 13,871 meticulously annotated image-text-mask triplets collected from real-world drone footage with emphasis on small, densely cluttered objects and multi-view perspectives. Additionally, we propose the Semantic-Aware Adaptive Reasoning Network, which decomposes and adaptively routes semantic information to different network stages rather than uniformly injecting all linguistic features. Specifically, the Category-Dominated Linguistic Enhancement aligns visual features with object categories during early encoding, while the Adaptive Reasoning Fusion Module dynamically selects semantic cues across scales to enhance reasoning in complex scenes. Extensive experiments reveal that RIS-LAD presents substantial challenges to state-of-the-art RIS algorithms, and also demonstrate the effectiveness of our proposed model in addressing these challenges.Downloads
Published
2026-03-14
How to Cite
Ye, K., Luan, Y., Chen, Z., Meng, G., Dai, P., & Cao, L. (2026). RIS-LAD: A Benchmark and Model for Referring Image Segmentation in Low-Altitude Drone Imagery. Proceedings of the AAAI Conference on Artificial Intelligence, 40(14), 11937-11945. https://doi.org/10.1609/aaai.v40i14.38181
Issue
Section
AAAI Technical Track on Computer Vision XI