Abstract
Accurate segmentation of polyps in colonoscopy images is essential for early colorectal cancer detection; however, it remains a challenging task due to reflections, occlusions, motion artifacts, inter- and intra-polyp appearance variability, and the presence of noisy or inconsistent ground-truth annotations. In this work, we introduce dynamic-Nu T-Loss (DNA-TLoss), a robust, adaptive loss function based on the heavy-tailed Student's 𝑡-distribution that incorporates three novel extensions: (1) a per-image learnable degrees-of-freedom parameter ν, predicted by a lightweight NuPredictor network to dynamically adjust robustness to outliers; (2) per-pixel precision weights λ for spatially adaptive error sensitivity; and (3) a multi-scale aggregation scheme that computes and combines loss at multiple spatial resolutions to capture both coarse and fine details. Integrated into a U-Net with a ResNet-34 encoder, DNA-TLoss was evaluated on five public benchmarks: CVC-300, CVC-ClinicDB, ETIS-LaribPolypDB, Kvasir, and CVC-ColonDB. Our method achieves the lowest Hausdorff distance across all datasets, with an average reduction of 14.6% compared to T-Loss; notably, on CVC-300, it yields a significant decrease of 45.96%. It also obtains the lowest false discovery rate on all five datasets, improving over T-Loss by up to 38.7% on CVC-300 and 24.5% on Kvasir. Furthermore, DNA-TLoss provided best-in-class calibration, achieving expected calibration error as low as 0.44% on CVC-300 and outperforming all other baselines on four out of five datasets. These results highlight the promise of joint global and local uncertainty adaptation, coupled with multi-scale optimization, for advancing trustworthy, real-time computer-aided polyp detection in colonoscopy.