Out-of-Distribution Detection as a Risk-Control Strategy for Medical Classification Machine Learning Models

分布外检测作为医疗分类机器学习模型的风险控制策略

阅读:1

Abstract

Recent advances in artificial intelligence (AI) and machine learning (ML) algorithms have significantly broadened the scope in which these algorithms are deployed. The use of these algorithms in high-stakes medical contexts requires that the algorithms have appropriate guardrails to ensure their performance. One such guardrail involves using out-of-distribution (OOD) detection algorithms, which detect if observations are unlikely to be sampled from the model's training distribution. Since these observations are not observed/rare in the training data, the model is likely to be unreliable when performing on these observations. In the medical context, identifying which patients are OOD may improve the performance of the model by filtering out patients on which the model has not been properly trained or tested. Here, we assess the performance of state-of-the-art OOD detection algorithms on three medical datasets of image, transcriptomics, and time series modalities, respectively. Using a simulated training-deployment scenario, we find that several OOD detectors consistently identify patients on which the model performs worse. Further, several OOD detectors identified patient subsets that were underrepresented in the training data, prompting further investigation. We present that the use of OOD detection methods could help mitigate model risk when deploying medical AI in the real world.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。