A progressive fusion network for endoscopic medical image segmentation

一种用于内窥镜医学图像分割的渐进式融合网络

阅读:1

Abstract

Endoscopic image segmentation plays a key role in assisting doctors to accurately locate focal areas and improve diagnostic efficiency. However, the existing methods are insufficient in utilizing local details and global semantic information at the same time, which makes it difficult to effectively segment organs and tissues with complex morphology, fuzzy boundaries and similar textures. Therefore, we propose a progressive fusion network (PFNet) in this paper. First, PFNet uses Pvtv2 with Transformer as the backbone encoder to capture multi-scale global features. Secondly, a noise filtering attention module (NFAM) is designed to suppress the noise and enhance the semantics of the multilevel features output by the encoder. Then, a boundary and location awareness module (BLAM) is proposed to generate high-quality boundary and position information by blending deep global features with shallow local details. Then, the auxiliary information embedding module (AIEM) is designed to embed the boundary and position information into each level feature dynamically to enhance the context-aware ability of the decoding process. Finally, the feature fusion module (FFM) supplemented the boundary and location information through layer by layer iteration to ensure the collaborative recovery of global semantics and local details. Through extensive experiments, we demonstrate that our proposed PFNet outperforms current state-of-the-art (SOTA) methods in segmentation performance on datasets including Ureter, Re-TMRS, Kvasir, CVC-ClinicDB, CVC-ColonDB, ETIS and CVC-300. In particular, the mDice on Re-TMRS dataset reached 91.07%, and the mDice on CVC-ClinicDB reached 93.09%.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。