Abstract
Remote tower (rTWR) operations are reshaping air traffic control but introduce significant human-factor risks, notably cognitive fatigue induced by prolonged screen-based visual surveillance. To mitigate these risks in a safety-critical domain where missed detections can be catastrophic, we propose a non-intrusive, multimodal fatigue detection framework fusing ocular and cardiac signals. A high-fidelity simulation study with 36 controllers was conducted to collect eye-tracking and electrocardiogram (ECG) data, from which a 12-dimensional feature vector-integrating gaze entropy and heart rate variability (HRV)-was extracted. Addressing the severe class imbalance and scarcity of fatigue samples in physiological data, we developed a cost-sensitive XGBoost classifier combining SMOTE oversampling with a dynamically weighted loss function. Experimental results show that the proposed framework performed well under mixed-subject evaluation and improved sensitivity to fatigue events. Although a marked performance drop was observed under LOSO evaluation, personalized calibration partially alleviated this limitation, indicating the potential of the framework for real-time fatigue monitoring in remote tower operations.