The present paper introduces a novel method, named AutoEncoder-Filter Bank Common Spatial Patterns (AE-FBCSP), to decode imagined movements from electroencephalography (EEG). AE-FBCSP is an extension of the well-established FBCSP and… Click to show full abstract
The present paper introduces a novel method, named AutoEncoder-Filter Bank Common Spatial Patterns (AE-FBCSP), to decode imagined movements from electroencephalography (EEG). AE-FBCSP is an extension of the well-established FBCSP and is based on a global (cross-subject) and subsequent transfer learning subject-specific (intra-subject) approach. A multi-way extension of AE-FBCSP is also introduced in this paper. Features are extracted from high-density EEG (64 electrodes), by means of FBCSP, and used to train a custom AE, in an unsupervised way, to project the features into a compressed latent space. Latent features are used to train a supervised classifier (feed forward neural network) to decode the imagined movement. The proposed method was tested using a public dataset of EEGs collected from 109 subjects. The dataset consists of right-hand, left-hand, both hands, both feet motor imagery and resting EEGs. AE-FBCSP was extensively tested in the 3-way classification (right hand vs left hand vs resting) and also in the 2-way, 4-way and 5-way ones, both in cross- and intra-subject analysis. AE-FBCSP outperformed standard FBCSP in a statistically significant way (p > 0.05) and achieved a subject-specific average accuracy of 89.09% in the 3-way classification. The proposed methodology performed subject-specific classification better than other comparable methods in the literature, applied to the same dataset, also in the 2-way, 4-way and 5-way tasks. One of the most interesting outcomes is that AE-FBCSP remarkably increased the number of subjects that responded with a very high accuracy, which is a fundamental requirement for BCI systems to be applied in practice.
               
Click one of the above tabs to view related content.