A Transformer-based deep neural network model for SSVEP classification_2

2025-04-27 0 0 536.43KB 33 页 10玖币
侵权投诉
arXiv:2210.04172v1 [q-bio.NC] 9 Oct 2022
A Transformer-based deep neural network model for
SSVEP classification
Jianbo Chena, Yangsong Zhanga,
, Yudong Pana, Peng Xub,
, Cuntai Guanc
aLaboratory for Brain Science and Medical Artificial Intelligence, School of Computer
Science and Technology, Southwest University of Science and Technology, Mianyang, China
bMOE Key Laboratory for NeuroInformation, Clinical Hospital of Chengdu Brain Science
Institute, and Center for Information in BioMedicine, School of Life Science and
Technology, University of Electronic Science and Technology of China, Chengdu, China
cSchool of Computer Science and Engineering, Nanyang Technological University, Singapore
Abstract
Steady-state visual evoked potential (SSVEP) is one of the most commonly used
control signal in the brain-computer interface (BCI) systems. However, the con-
ventional spatial filtering methods for SSVEP classification highly depend on
the subject-specific calibration data. The need for the methods that can al-
leviate the demand for the calibration data become urgent. In recent years,
developing the methods that can work in inter-subject classification scenario
has become a promising new direction. As the popular deep learning model
nowadays, Transformer has excellent performance and has been used in EEG
signal classification tasks. Therefore, in this study, we propose a deep learning
model for SSVEP classification based on Transformer structure in inter-subject
classification scenario, termed as SSVEPformer, which is the first application of
the transformer to the classification of SSVEP. Inspired by previous studies, the
model adopts the frequency spectrum of SSVEP data as input, and explores the
spectral and spatial domain information for classification. Furthermore, to fully
utilize the harmonic information, an extended SSVEPformer based on the filter
bank technology (FB-SSVEPformer) is proposed to further improve the clas-
sification performance. Experiments were conducted using two open datasets
Corresponding authors: Yangsong Zhang(zhangysacademy@gmail.com); Peng
Xu(xupeng@uestc.edu.cn)
Preprint submitted to Neural Networks October 11, 2022
(Dataset 1: 10 subjects, 12-class task; Dataset 2: 35 subjects, 40-class task)
in the inter-subject classification scenario. The experimental results show that
the proposed models could achieve better results in terms of classification ac-
curacy and information transfer rate, compared with other baseline methods.
The proposed model validates the feasibility of deep learning models based on
Transformer structure for SSVEP classification task, and could serve as a po-
tential model to alleviate the calibration procedure in the practical application
of SSVEP-based BCI systems.
Keywords: Brain-computer interface, Steady-state visual evoked potential,
Transformer, Deep learning, Filter bank
1. Introduction
Brain-computer interface (BCI) has become a popular research direction
in human-computer interaction and medical rehabilitation, which can directly
connect the brain to external devices without going through the peripheral
nervous system, enabling bidirectional information transmission and feedback
[42, 47]. Electroencephalogram (EEG)-based BCIs obtain the intentions of the
brain through EEG signals, and have attracted attention due to the advantages
of convenience, low cost, and non-invasiveness [1]. Among the various EEG
paradigms, the high signal-to-noise ratio and low training time of steady-state
visual evoked potential (SSVEP) make it one of the most popular paradigms.
SSVEP refers to the EEG in the visual cortex when the subject gazes at a
flickering visual stimulus modulated by a constant frequency [56]. The frequen-
cies of SSVEP are the same as the coding frequency of received visual stimuli
as well as its harmnoics[33]. By virtue of this characteristic of SSVEP, it is
possible to design SSVEP-based BCI system, such as SSVEP-based speller [27],
in which different targets are encoded by different stimulus frequencies. When
the subjects need to select a command, they can gaze at the corresponding flick-
ering target stimulus that coding the command on the interface. The generated
SSVEP can be identified by a specially designed decoder to obtain the intention
2
of the subject.
In the SSVEP-based BCI system, the robust classification of the SSVEPs
is very important [28]. As the SSVEP frequency is the same as the stimulus
frequency, some researches developed the algorithms based on the prior fre-
quency information, such as power spectral density analysis (PSDA) [40] and
canonical correlation analysis (CCA) [22], etc. In addition to the fundamen-
tal frequency component, SSVEP also contains harmonic components whose
frequencies are multiples of the fundamental frequency [25]. Based on this char-
acteristic, filter bank technology was introduced to extend the original CCA
(FBCCA) [5]. FBCCA uses CCA in multiple subbands of SSVEP data, and
finally weights the correlation coefficients calculated from these subbands. The
FBCCA improves the classification performance by distinguishing the funda-
mental frequency and harmonics, demonstrating the effectiveness of the filter
bank technique on SSVEP classification. Nowadays, filter bank technology has
been widely used in various methods [57, 31].
However, due to the complexity of EEG, SSVEP data always contains noises,
such as spontaneous EEG and electromagnetic interference, seriously polluting
the signal [17]. Traditional training-free methods (such as PSDA, CCA) have
better results only when the data length is long. To address the noise inter-
ference in SSVEP, a series of recognition algorithms based on machine learning
have been proposed. Such methods perform under the intra-subject classifi-
cation condition, in which the training and testing data are from the same
subjects, as shown in Fig. 1(a). In this condition, the model can obtain pa-
rameters that are more suitable for a specific subject, thereby reducing noise
interference [50]. For example, individual template based CCA (IT-CCA) cal-
culates the average of the subject’s existing SSVEP signals at each stimulation
frequency and uses it as the reference signal for CCA [4]. This method can add
subject-specific patterns to the reference signal, and is widely used in subsequent
algorithms. Task-related component analysis (TRCA) method obtains spatial
filters by maximizing the reconstitution between SSVEPs of different trials to
reduce the noise of SSVEPs and reference signals [26]. Correlated component
3
Training data
New subjects
Training Testing
Classifier
瀥瀥瀥
Testing data
Existing subjects
瀥瀥瀥
Training data
Training Testing
Classifier
Testing data
(a)
(b)
Figure 1: The diagram of two classification scenarios. (a) intra-subject classification; (b)
inter-subject classification.
analysis (CORCA) learns spatial filters by maximizing the correlation between
data to reduce background noise [55]. Task-discriminant component analysis
(TDCA) uses multi-class linear discriminant analysis to learn spatiotemporal
filters and classify in a discriminant manner [23].
The above method has significant effect in the intra-subject classification
experiment, in which the training data and the testing data belong to the same
subject [32]. However, the collection of SSVEP data is a time-consuming and
laborious work. Hence, a potential and challenging direction is to transfer the
data from existing subjects to new subjects in the inter-subject classification
scenario, under which a classifier can be obtained with the data from already
existing subjects and then used the classifier to test the data from new subjects,
as shown in Fig. 1(b). Although many works have improved traditional state-of-
the-art methods to adapt them to inter-subject scenario, the results may be not
optimal [46]. Because the brain processes natural sensory stimuli in a dynamic,
4
non-fixed, and nonlinear manner, SSVEP is non-stationary and varies widely
among individuals [18]. Even the data collected by the same subject, the data
acquired at different times may also have different distribution. These situa-
tions pose great challenges for inter-subject experiments, and the performance
of traditional machine learning-based algorithms under inter-subject condition
degrades greatly, which is far from its performance under intra-subject condi-
tion.
In recent years, deep learning has been developed significantly and has made
milestone progress in areas such as computer vision and natural language pro-
cessing [19, 8]. Deep learning models have powerful feature extraction capabil-
ities and can directly be applied on the raw data[9, 34]. Deep learning models
have been used on many EEG classification tasks, including convolutional neu-
ral networks (CNN) [60], recurrent neural networks (RNN) [12], graph neural
networks (GNN) [59], etc. Several studies have used deep learning to process
SSVEP data, achieving outstanding performance on classification tasks, espe-
cially inter-subject classification. For instances, EEGNet is a compact convolu-
tional neural network (CNN) that uses CNNs to implement the spatial-temporal
filtering and feature extraction, achieving significantly better results than tra-
ditional methods under inter-subject conditions [41]. The idea of using tem-
poral and spatial convolutions has achieved promising results, which has also
influenced many later algorithms. The subject invariant SSVEP generative ad-
versarial network (SIS-GAN) uses generative adversarial networks to generate
artificial SSVEP data to expand the training dataset [2]. Complex convolu-
tional neural network (CCNN) uses the complex spectrum of SSVEP signal
as the input of CNN for classification, demonstrating the effectiveness of com-
plex spectral features on SSVEP classification [32]. InceptionEEG-Net (IENet)
combines Inception with residual connections and uses multi-scale convolution
kernels to extract features from receptive fields of different sizes [13]. In addi-
tion, filter bank technology is also applied in deep learning models to extend
the existing models, such as FB-EEGNet and FBCNN [48, 58].
Although the deep learning-based SSVEP recognition model has made great
5
摘要:

arXiv:2210.04172v1[q-bio.NC]9Oct2022ATransformer-baseddeepneuralnetworkmodelforSSVEPclassificationJianboChena,YangsongZhanga,∗,YudongPana,PengXub,∗,CuntaiGuancaLaboratoryforBrainScienceandMedicalArtificialIntelligence,SchoolofComputerScienceandTechnology,SouthwestUniversityofScienceandTechnology,Miany...

展开>> 收起<<
A Transformer-based deep neural network model for SSVEP classification_2.pdf

共33页,预览5页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!
分类:图书资源 价格:10玖币 属性:33 页 大小:536.43KB 格式:PDF 时间:2025-04-27

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 33
客服
关注