Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MAEEG: Masked Auto-encoder for EEG Representation Learning

About

Decoding information from bio-signals such as EEG, using machine learning has been a challenge due to the small data-sets and difficulty to obtain labels. We propose a reconstruction-based self-supervised learning model, the masked auto-encoder for EEG (MAEEG), for learning EEG representations by learning to reconstruct the masked EEG features using a transformer architecture. We found that MAEEG can learn representations that significantly improve sleep stage classification (~5% accuracy increase) when only a small number of labels are given. We also found that input sample lengths and different ways of masking during reconstruction-based SSL pretraining have a huge effect on downstream model performance. Specifically, learning to reconstruct a larger proportion and more concentrated masked signal results in better performance on sleep classification. Our findings provide insight into how reconstruction-based SSL could help representation learning for EEG.

Hsiang-Yun Sherry Chien, Hanlin Goh, Christopher M. Sandino, Joseph Y. Cheng• 2022

Related benchmarks

TaskDatasetResultRank
EEG ClassificationEmotiv Crowdsourced
Accuracy86.75
16
EEG ClassificationEmotiv Attention
Accuracy82.61
8
EEG ClassificationEmotiv STEW
Accuracy72.46
8
EEG ClassificationEmotiv DriverDistraction
Accuracy (ACC)74.58
8
Showing 4 of 4 rows

Other info

Follow for update