Abstract
The rapid progress of remote sensing (RS) and computer vision has greatly advanced change detection (CD), and hybrid architectures combining Transformers and convolutional neural networks (CNNs) have shown strong potential in recent years. Nevertheless, reliable CD for very high-resolution (VHR) imagery remains challenging due to large appearance variations across acquisition times, complex background clutter, and target structural diversity. These factors often hinder the modeling of fine edge textures, the maintenance of feature continuity, and the suppression of false changes caused by illumination fluctuations. To address these issues, this paper proposes a Cross-layer Feature Fusion Framework (CLFF) that achieves more accurate and stable change detection by explicitly enhancing the collaborative fusion capability of multi-layer features. The core component of this framework is the Multi-level Interaction Perception Block (MP-Block), which organizes effective interactions among features of different semantic levels. Based on the embedded Multi-branch Interaction Fusion Mechanism (MIFM), the MP-Block accomplishes collaborative refinement and reorganization of cross-layer features through two parallel paths for feature reconstruction and recalibration: the Response-aware Feature Reconstruction Branch (RFRB) and Adaptive Channel Group Fusion Branch (ACGF). Additionally, a lightweight position-aware attention module is introduced to adaptively modulate spatial responses, further suppressing background interference and highlighting key information related to changes. This method effectively mitigates the limitations of traditional CNNs, such as limited receptive fields and insufficient multi-layer feature interaction, while significantly enhancing the ability to collaboratively model multi-layer contextual information. To verify its effectiveness, systematic experiments were conducted on four widely used change detection benchmark datasets: LEVIR, WHU, SYSU and HRCUS. The results show that, compared to corresponding baseline models, CLFF achieves performance improvements of 1.35%, 2.78%, 3.54% and 4.85% in the IoU metric, respectively.