SAM2-ARAFNet: adapting SAM2 with an attention-enhanced residual ASPP fusion network for high-resolution remote sensing semantic segmentation

SAM2-ARAFNet:将SAM2与注意力增强的残差ASPP融合网络相结合,用于高分辨率遥感语义分割

阅读:1

Abstract

High-resolution remote sensing image segmentation plays a crucial role in fields such as environmental surveillance, disaster impact analysis, and spatial resource management, yet the pronounced variability within classes, intricate scene structures, and substantial computational burden of modern deep models often impede their practical use. To mitigate these limitations, this study introduces SAM2-ARAFNet, a segmentation framework derived from Segment Anything Model 2 (SAM2) and equipped with lightweight adapter modules for efficient parameter tuning, together with an Attention-Enhanced Residual Atrous Spatial Pyramid Pooling (ResASPP) component that enriches multi-scale semantic representation. For deployment on resource-limited platforms, a tailored distillation strategy is further employed to compress the fine-tuned SAM2 model into a compact student network based on EfficientNet_b0. Experiments conducted on the ISPRS Vaihingen and Potsdam benchmarks demonstrate clear performance gains: SAM2-ARAFNet attains mIoU values of 85.43% and 87.44%, exceeding widely used baselines such as PSPNet by 4.93% and 4.03%. In addition, the distilled student model reduces parameters by 97% (from 222.98 M to 6.68 M) while preserving more than 99% of the teacher network's accuracy, illustrating its capability to deliver high-quality segmentation with markedly improved computational efficiency, and confirming its suitability for edge-focused remote sensing scenarios.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。