Label-Efficient Sleep Staging Using Transformers Pre-trained with Position Prediction
AuthorsSayeri Lala, Hanlin Goh, Chris Sandino
AuthorsSayeri Lala, Hanlin Goh, Chris Sandino
Sleep staging is a clinically important task for diagnosing various sleep disorders but remains challenging to deploy at scale because it requires clinical expertise, among other reasons. Deep learning models can perform the task but at the expense of large labeled datasets, which are unfeasible to procure at scale. While self-supervised learning (SSL) can mitigate this need, recent studies on SSL for sleep staging have shown performance gains saturate after training with labeled data from only tens of subjects, hence are unable to match peak performance attained with larger datasets. We hypothesize that the rapid saturation stems from applying a pretraining scheme that only pretrains a portion of the architecture, i.e., the feature encoder but not the temporal encoder; therefore, we propose adopting an architecture that seamlessly couples the feature and temporal encoding and a suitable pretraining scheme that pretrains the entire model. On a sample sleep staging dataset, we find that the proposed scheme offers performance gains that do not saturate with the labeled training dataset size (e.g., 3-5% improvement in balanced accuracy across low- to high-labeled data settings), which translate into significant reductions in the amount of labeled training data needed for high performance (e.g., by 800 subjects). Based on our findings, we recommend adopting this SSL paradigm for subsequent work on SSL for sleep staging.