Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Learning from Time Series for Health

Adversarial Masking for Pretraining ECG Data Improves Downstream Model Generalizability

Jessica Yi Fei Bo · Hen-Wei Huang · Alvin Chan · Giovanni Traverso


Abstract:

Medical datasets often face the problem of data scarcity, as ground truth labels must be generated by medical professionals. One mitigation strategy is to pretrain deep learning models on large, unlabelled datasets with self-supervised learning (SSL), but this introduces the issue of domain shift if the pretraining and task dataset distributions differ. Data augmentations are essential for improving the generalizability of SSL-pretrained models, but they tend to be either handcrafted or randomly applied. We use an adversarial model to generate masks as augmentations for 12-lead electrocardiogram (ECG) data, where masks learn to occlude diagnostically-relevant regions. Compared to random augmentations, adversarial masking reaches better accuracy on a downstream arrhythmia classification task under a domain shift condition and in data-scarce regimes. Adversarial masking is competitive with, and even reaches further improvements when combined with state-of-art ECG augmentation methods, 3KG and random lead masking.

Chat is not available.