LMDENet: A Lightweight RGB-IR Object Detection Network for Low-Light Remote Sensing Images

LMDENet:一种用于低照度遥感图像的轻量级RGB-IR目标检测网络

阅读:1

Abstract

RGB-infrared (RGB-IR) object detection leverages complementary information from these two modalities to substantially enhance perception in complex environments, which is particularly beneficial for reliable detection under adverse imaging conditions such as low illumination and severe haze. However, RGB-IR object detection still faces several challenges due to pronounced intra-modality and cross-modality discrepancies. On the one hand, many existing approaches rely on complex architectures to strengthen cross-modal interactions, which increases computational cost. On the other hand, symmetric dual-branch backbones with a static fusion paradigm often struggle to explicitly characterize discrepancies between the RGB and IR modalities. This limitation prevents effective mining of complementary information and reduces the discriminability of fused representations. To address these issues, this paper presents a lightweight RGB-IR multimodal detection network (LMDENet), which consists of three key components: (1) an illumination-guided label selection (IGLS) that integrates RGB and IR labels based on cross-modal matching and illumination-aware rules to construct consistent and reliable supervision; (2) a heterogeneous backbone network (HBN) with differentiated branches that separately model RGB appearance details and IR structural information, improving modality-specific representation learning; and (3) a difference-complement enhancement module (DCEM) that explicitly decomposes cross-modal features into common and difference components and performs selective enhancement to amplify complementary information while suppressing redundant noise. We systematically evaluate the detection performance of the proposed model on the multimodal remote sensing dataset DroneVehicle, and further conduct supplementary experiments on the LLVIP dataset to verify its generalization ability across different scenarios. Experimental results on the DroneVehicle and LLVIP datasets demonstrate that LMDENet achieves 78.9% and 93.6% mAP@0.5, respectively. Meanwhile, the model contains only 3.3 M parameters and 8.7 G FLOPs, reflecting a favorable accuracy-efficiency balance.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。