Abstract
Natural gas pipeline leak monitoring suffers from severe environmental noise, non-stationary signals, and complex multi-source variable couplings, limiting prediction accuracy and robustness. Inspired by biological perceptual systems, particularly their multimodal integration and dynamic attention allocation, we propose GL-TransLSTM, a biomimetic hybrid deep learning model. It synergistically combines Transformer's global self-attention (emulating selective focus) and LSTM's gated memory (mimicking neural temporal retention). The architecture incorporates a multimodal fusion pipeline; raw sensor data are first decomposed via CEEMDAN to extract multi-scale features, then processed by an enhanced LSTM-Transformer backbone. A novel physics-informed gated attention mechanism embeds gas diffusion dynamics into attention weights, while an adaptive sliding window adjusts temporal granularity. This study makes evaluations on an industrial dataset with methane concentration, temperature, and pressure, GL-TransLSTM achieves 99.93% accuracy, 99.86% recall, and 99.89% F1-score, thereby significantly outperforming conventional LSTM and Transformer-LSTM baselines. Experimental results demonstrate that the proposed biomimetic framework substantially enhances modeling capacity and generalization for non-stationary signals in noisy and complex industrial environments through multi-scale fusion, physics-guided learning, and bio-inspired architectural synergy.