Abstract
The increasing utilization of MRI in radiation therapy planning for head and neck cancer (HNC) highlights the need for precise tumor segmentation to enhance treatment efficacy and reduce side effects. This work presents segmentation models developed for the HNTS-MRG 2024 challenge by the team mic-dkfz, focusing on automated segmentation of HNC tumors from MRI images at two radiotherapy (RT) stages: before (pre-RT) and 2-4 weeks into RT (mid-RT). For Task 1 (pre-RT segmentation), we built upon the nnU-Net framework, enhancing it with the larger Residual Encoder architecture. We incorporated extensive data augmentation and applied transfer learning by pre-training the model on a diverse set of public 3D medical imaging datasets. For Task 2 (mid-RT segmentation), we adopted a longitudinal approach by integrating registered pre-RT images and their segmentations as additional inputs into the nnU-Net framework. On the test set, our models achieved mean aggregated Dice Similarity Coefficient (aggDSC) scores of 81.2 for Task 1 and 72.7 for Task 2. Especially the primary tumor (GTVp) segmentation is challenging and presents potential for further optimization. These results demonstrate the effectiveness of combining advanced architectures, transfer learning, and longitudinal data integration for automated tumor segmentation in MRI-guided adaptive radiation therapy.