Advancing BCI with a transformer-based model for motor imagery classification

Abstract Brain-computer interfaces (BCIs) harness electroencephalographic signals for direct neural control of devices, offering significant benefits for individuals with motor impairments. Traditional machine learning methods for EEG-based motor imagery (MI) classification encounter challenges such...

Full description

Saved in:
Bibliographic Details
Main Authors: Wangdan Liao, Hongyun Liu, Weidong Wang
Format: Article
Language:English
Published: Nature Portfolio 2025-07-01
Series:Scientific Reports
Subjects:
Online Access:https://doi.org/10.1038/s41598-025-06364-4
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Abstract Brain-computer interfaces (BCIs) harness electroencephalographic signals for direct neural control of devices, offering significant benefits for individuals with motor impairments. Traditional machine learning methods for EEG-based motor imagery (MI) classification encounter challenges such as manual feature extraction and susceptibility to noise. This paper introduces EEGEncoder, a deep learning framework that employs modified transformers and Temporal Convolutional Networks (TCNs) to surmount these limitations. We propose a novel fusion architecture, named Dual-Stream Temporal-Spatial Block (DSTS), to capture temporal and spatial features, improving the accuracy of Motor Imagery classification task. Additionally, we use multiple parallel structures to enhance the model’s performance. When tested on the BCI Competition IV-2a dataset, our proposed model achieved an average accuracy of 86.46% for subject dependent and average 74.48% for subject independent.
ISSN:2045-2322