AVPDN: learning motion-robust and scale-adaptive representations for polyp detection in dynamic colonoscopy frames

AVPDN:学习运动鲁棒且尺度自适应的表示方法,用于动态结肠镜检查帧中的息肉检测

阅读:1

Abstract

Accurate detection of polyps is of critical importance for the early and intermediate stages of colorectal cancer diagnosis. While colonoscopy videos offer richer visual information than static images for planning treatment, the rapid camera movement during examination introduces significant frame-level artifacts-such as motion blur, specular reflections, and scale variation-that degrade image quality and increase false positives in detection. To address these challenges within individual frames, we propose the Adaptive Video Polyp Detection Network (AVPDN), a robust framework for multi-scale polyp detection in dynamic colonoscopy imagery. AVPDN incorporates two key components: the Adaptive Feature Interaction and Augmentation (AFIA) module and the Scale-Aware Context Integration (SACI) module. The AFIA module adopts a dual-branch architecture to enhance feature representation. It employs dense self-attention for global context modeling, sparse self-attention to mitigate the influence of low query-key similarity in feature aggregation, and channel shuffle operations to facilitate inter-branch information exchange. In parallel, the SACI module is designed to strengthen multi-scale feature integration. It utilizes dilated convolutions with varying receptive fields to capture contextual information at multiple spatial scales, thereby improving the model's denoising capability. Extensive experiments on challenging public benchmarks demonstrate the effectiveness and generalization capability of our method, achieving state-of-the-art performance in detecting polyps from complex, motion-affected colonoscopy frames.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。