Abstract
In intelligent transportation systems and urban traffic management, accurate vehicle area intrusion detection based on surveillance imagery plays a critical role in ensuring road safety and operational efficiency. However, under real-world road surveillance conditions characterized by complex backgrounds, varying illumination, occlusion, and scale variations, mainstream detection algorithms often suffer from high false detection and missed detection rates, limiting their reliability and practical deployment. To address these challenges, this paper proposes YOLOv10-Intrusion, a high-precision vehicle area intrusion detection framework based on an improved version of YOLOv10s. The proposed algorithm incorporates Omni-Dimensional Dynamic Convolution (ODConv) and a custom-designed RCS_M module to enhance feature extraction and fine-grained recognition capability. In addition, a Bidirectional Feature Pyramid Network (BiFPN) is employed to optimize multi-scale feature fusion at the neck level. These improvements collectively reduce false detections and missed detections while improving model recall and mean Average Precision (mAP). Furthermore, the Wise-IoU (WIoU) loss function replaces the original Complete IoU (CIoU) loss to accelerate convergence and stabilize bounding box regression under complex surveillance conditions. A dedicated vehicle area intrusion dataset is constructed from real-world road surveillance footage, covering five vehicle categories across diverse road environments and lighting conditions. Experimental results demonstrate that, compared with the baseline YOLOv10s, YOLOv10-Intrusion achieves improvements of 1.5, 3.3, 3.6, and 2.8 percentage points in Precision, Recall, mAP@0.5, and mAP@0.5:0.95, respectively, and outperforms other mainstream detection algorithms in vehicle area intrusion detection tasks.