Multimodal multi-instance evidence fusion neural networks for cancer survival prediction

用于癌症生存预测的多模态多实例证据融合神经网络

阅读:3

Abstract

Accurate cancer survival prediction plays a crucial role in assisting clinicians in formulating treatment plans. Multimodal data, such as histopathological images, genomic data, and clinical information, provide complementary and comprehensive information, significantly enhancing the accuracy of this task. However, existing methods, despite achieving some promising results, still exhibit two significant limitations: they fail to effectively utilize global context and overlook the uncertainty of different modalities, which may lead to unreliable predictions. In this study, we propose a multimodal multi-instance evidence fusion neural network for cancer survival prediction, called M2EF-NNs. Specifically, to better capture global information from images, we employ a pre-trained vision transformer model to extract patch feature embeddings from histopathological images. Additionally, we are the first to apply the Dempster-Shafer evidence theory to the cancer survival prediction task and introduce subjective logic to estimate the uncertainty of different modalities. We then dynamically adjust the weights of the class probability distribution after multimodal fusion based on the estimated evidence from the fused multimodal data to achieve trusted survival prediction. Finally, the experimental results on three cancer datasets demonstrate that our method significantly improves cancer survival prediction regarding overall C-index and AUC, thereby validating the model's reliability.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。