USS-Net: A neural network-based model for assisting flight route scheduling

USS-Net:一种基于神经网络的辅助航班航线调度模型

阅读:1

Abstract

Air traffic congestion-induced flight accidents pose a significant challenge in the aviation sector. Currently, aviation navigation systems primarily rely on GPS and Inertial Navigation Systems (INS) to track aircraft, yet they lack the capability to recognize and provide early warnings about the surrounding environment. To address this issue, this paper proposes a multi-aircraft parallel approach aimed at enabling coordinated flight along the same route. This method utilizes a neural network-based semantic segmentation model to monitor aircraft and perform situational awareness of the surrounding environment, thereby assisting in multi-aircraft route scheduling. When wake turbulence is generated, the model can identify the wake, further enhancing flight safety. Recently, state-space models (SSMs) based on Mamba have demonstrated outstanding performance in computational efficiency and inference speed. Based on this, we designed a U-shaped State Space Block UNet (USS-Net), which consists of StateConvBlock and ResConvBlock. The StateConvBlock integrates Mamba as a fundamental module for understanding temporal dynamics and contextual information. By constructing a symmetrical encoder-decoder structure, the model progressively extracts image features and performs multi-scale fusion to achieve high-precision pixel-level segmentation. Experimental results show that USS-Net achieved outstanding performance on the aircraft simulation dataset. On an NVIDIA A100-SXM4-40GB GPU, USS-Net attained a mean Intersection over Union (mIoU) of 95.70% and a pixel accuracy (PA) of 97.80% on the simulation training dataset. These results demonstrate that USS-Net performs effectively in assisting multi-aircraft parallel route scheduling tasks.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。