Graph neural network-tracker: a graph neural network-based multi-sensor fusion framework for robust unmanned aerial vehicle tracking

图神经网络跟踪器:一种基于图神经网络的多传感器融合框架,用于鲁棒的无人机跟踪

阅读:1

Abstract

Unmanned aerial vehicle (UAV) tracking is a critical task in surveillance, security, and autonomous navigation applications. In this study, we propose graph neural network-tracker (GNN-tracker), a novel GNN-based UAV tracking framework that effectively integrates graph-based spatial-temporal modelling, Transformer-based feature extraction, and multi-sensor fusion to enhance tracking robustness and accuracy. Unlike traditional tracking approaches, GNN-tracker dynamically constructs a spatiotemporal graph representation, improving identity consistency and reducing tracking errors under OCC-heavy scenarios. Experimental evaluations on optical, thermal, and fused UAV datasets demonstrate the superiority of GNN-tracker (fused) over state-of-the-art methods. The proposed model achieves multiple object tracking accuracy (MOTA) scores of 91.4% (fused), 89.1% (optical), and 86.3% (thermal), surpassing TransT by 8.9% in MOTA and 7.7% in higher order tracking accuracy (HOTA). The HOTA scores of 82.3% (fused), 80.1% (optical), and 78.7% (thermal) validate its strong object association capabilities, while its frames per second of 58.9 (fused), 56.8 (optical), and 54.3 (thermal) ensures real-time performance. Additionally, ablation studies confirm the essential role of graph-based modelling and multi-sensor fusion, with performance drops of up to 8.9% in MOTA when these components are removed. Thus, GNN-tracker (fused) offers a highly accurate, robust, and efficient UAV tracking solution, effectively addressing real-world challenges across diverse environmental conditions and multiple sensor modalities.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。