Transformer-based and CNN-based models for clinically effective 2D and 3D pelvic bone segmentation in CT imaging

基于Transformer和CNN的模型在CT图像中实现临床有效的二维和三维骨盆分割

阅读:1

Abstract

OBJECTIVE: Accurate segmentation of pelvic structures and fracture fragments in trauma CT scans remains a technical challenge due to anatomical complexity and image variability. This study provides the first systematic comparison of CNN (U-Net, LinkNet) and transformer-based (UNETR) architectures in 2D and 3D formats for automated pelvic bone segmentation on the PENGWIN MICCAI 2024 challenge dataset, evaluating encoder backbones (VGG19, ResNet50) to identify optimal strategies for clinical trauma imaging. MATERIALS AND METHODS: CT data from 150 patients with pelvic fractures were obtained from the PENGWIN MICCAI 2024 challenge. Preprocessing included normalization, resampling, one-hot encoding of segmentation masks, and format conversion. The segmentation task focused on classifying sacrum, left hipbone, and right hipbone, consolidating original 30-class annotations into 4 classes, providing novel benchmarks for multi-fragment pelvic fracture analysis using off-the-shelf architectures adapted to clinical trauma challenges. 2D models used a 192 × 192 resolution, and 3D models used volumetric inputs of 128 × 128 × 128 or 128 × 128 × 16. Models were trained using composite loss functions (Dice + Focal or BCE), evaluated with 5-fold cross-validation. U-Net and LinkNet were implemented with VGG19 and ResNet50 backbones; UNETR used a transformer-based encoder. Model performance was assessed via Dice coefficient, IoU, accuracy, sensitivity, and specificity. Inference speed and cross-model comparisons were conducted to analyze computational efficiency and segmentation quality. RESULTS: U-Net with ResNet50 achieved the highest 2D performance (Dice 0.991, IoU 0.982), while 3D U-Net with VGG19 led volumetric segmentation (Dice 0.9112). UNETR offered superior specificity (0.993) and inference consistency (< 1 min/case), albeit with lower sensitivity (0.730) in complex fragment localization. Visual analyses confirmed high anatomical fidelity. Comparison with existing literature showed our models exceeded or matched top benchmarks in Dice scores and generalizability. CONCLUSIONS: Our results confirm the efficacy of deep learning-based segmentation for pelvic fracture analysis, with high accuracy and computational efficiency supporting potential clinical integration. The models presented here demonstrate high accuracy and computational efficiency, offering a strong foundation for potential clinical integration into automated trauma imaging systems, pending external validation and workflow assessment.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。