The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences
Download
Publications Copernicus
Download
Citation
Articles | Volume XLVIII-2-2024
https://doi.org/10.5194/isprs-archives-XLVIII-2-2024-203-2024
https://doi.org/10.5194/isprs-archives-XLVIII-2-2024-203-2024
11 Jun 2024
 | 11 Jun 2024

Practical Techniques for Vision-Language Segmentation Model in Remote Sensing

Yuting Lin, Kumiko Suzuki, and Shinichiro Sogo

Keywords: Segmentation of Remote Sensing Data, Vision-Language Model, Fine-tuning, Visual Prompting

Abstract. Traditional semantic segmentation models often struggle with poor generalizability in zero-shot scenarios such as recognizing attributes unseen in the training labels. On the other hands, language-vision models (VLMs) have shown promise in improving performance on zero-shot tasks by leveraging semantic information from textual inputs and fusing this information with visual features. However, existing VLM-based methods do not perform as effectively on remote sensing data due to the lack of such data in their training datasets. In this paper, we introduce a two-stage fine-tuning approach for a VLM-based segmentation model using a large remote sensing image-caption dataset, which we created using an existing image-caption model. Additionally, we propose a modified decoder and a visual prompt technique using a saliency map to enhance segmentation results. Through these methods, we achieve superior segmentation performance on remote sensing data, demonstrating the effectiveness of our approach.