Subject-independent Classification on Brain-Computer Interface using Autonomous Deep Learning for Finger Movement Recognition
Date
2020-07-20Author
ANAM, Khairul
BUKHORI, Saiful
HANGGARA, Faruq Sandi
PRATAMA, Mahardhika
Metadata
Show full item recordAbstract
The degradation of the subject-independent
classification on a brain-computer interface is a challenging
issue. One method mostly taken to overcome this problem is by
collecting as many subjects as possible and then training the
system across all subjects. This article introduces streaming
online learning called autonomous deep learning (ADL) to
classify five individual fingers based on electroencephalography
(EEG) signals to overcome the issue above. ADL is a deep
learning architecture that can construct its structure by itself
through streaming learning and adapt its structure to the
changes occurring in the input. In this article, the input of ADL
is a common spatial pattern (CSP) extracted from the EEG
signal of healthy subjects. The experimental results on the
subject-dependence classification across four subjects using 5fold
cross-validation show that that ADL achieved the
classification accuracy of around 77%. This performance was
excellent compared to a random forest (RF) and a convolutional
neural network (CNN). They achieved accuracies of about 53%
and 72%, respectively. On the subject-independent
classification, ADL outperforms CNN by resulting stable
accuracies for both training and testing, different from CNN
that experience accuracy degradation to approximately 50%.
These results imply that ADL is a promising machine learning
in dealing with the issue in the subject-independent
classification.
Collections
- LSP-Conference Proceeding [1874]