Self-supervised contrastive learning enables robust electrocardiogram-based cardiac classification

自监督对比学习能够实现稳健的基于心电图的心脏分类

阅读:1

Abstract

BACKGROUND: Self-supervised contrastive learning has emerged as a powerful paradigm for learning generalizable representations from unlabeled data. In the context of electrocardiogram (ECG) analysis, such pre-training can significantly enhance classification performance, especially when labeled data is scarce. OBJECTIVE: We aimed to investigate and improve contrastive self-supervised learning techniques for ECGs by systematically combining recent methodological advances in augmentation design, contrastive loss formulation, and encoder architectures. METHODS: We implemented a contrastive pre-training framework combining vectorcardiography (VCG)-based physiologically-inspired augmentations, interlead, intersegment, contrastive loss, and patient-aware positive sampling. In addition, we developed a dual-stream architecture, extending the TemporalNet model by processing grouped ECG leads independently. Pretraining was conducted on a large corpus of approximately 1 million unlabeled ECGs. We evaluated performance on 2 downstream classification tasks-low left ventricular ejection fraction (LVEF) and high serum potassium chloride-using various levels of labeled supervision (1%, 5%, 10%, 50%, and 100%). The pre-trained models were compared with the randomly initialized models under both frozen and finetuned conditions. RESULTS: Contrastive pre-training consistently improved performance across all supervision levels. In low-label settings (1%-10% supervision), the pre-trained model achieved 3%-4% higher area under the receiver operator curve on the LVEF task and 5%-7% higher area under the receiver operator curve on the potassium chloride task compared with the baseline. The performance gap narrowed with increased supervision but remained favorable toward pre-trained models. CONCLUSION: Our findings demonstrate that contrastive pre-training can substantially enhance ECG classification, especially when labeled data is limited. By unifying and extending ideas from recent literature into a scalable framework trained on 1 million ECGs, we provide practical guidance and architectural innovations for building strong ECG foundation models applicable to a broad range of clinical prediction tasks.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。