论文标题
ECOG脑计算机界面的深度学习:端到端与手工制作的功能
Deep learning for ECoG brain-computer interface: end-to-end vs. hand-crafted features
论文作者
论文摘要
在大脑信号处理中,深度学习(DL)模型已被普遍使用。但是,与常规ML方法相比,使用端到端DL模型的性能增长通常很重要,但通常为中等,通常是以增加计算负载和降低解释性为代价。深度学习方法背后的核心思想是通过更大的数据集扩展性能。但是,大脑信号是及时的信噪比,不确定标签和非平稳数据的时间数据。这些因素可能会影响训练过程并减慢模型的性能改善。这些因素的影响对于端到端DL模型可能有所不同,并且使用手工制作的功能有所不同。如前所述,本文比较了使用原始的ECOG信号和时间频率特征进行BCI运动图像解码的模型。我们研究当前数据集大小是否是任何模型的限制。最后,比较获得的过滤器,以确定手工制作的特征之间的差异,并通过反向传播进行了优化。为了比较两种策略的有效性,我们使用了多层感知器以及已经证明在这项任务中有效的卷积和LSTM层的混合物。该分析是对四脑术患者执行3D手翻译的运动成像任务的长期临床试验数据库(记录的近600分钟)进行的。对于给定的数据集,结果表明,端到端训练可能不会比基于手工的功能模型更好。通过更大的数据集降低了性能差距,但是考虑到增加的计算负载,端到端培训对于此应用程序可能无法获利。
In brain signal processing, deep learning (DL) models have become commonly used. However, the performance gain from using end-to-end DL models compared to conventional ML approaches is usually significant but moderate, typically at the cost of increased computational load and deteriorated explainability. The core idea behind deep learning approaches is scaling the performance with bigger datasets. However, brain signals are temporal data with a low signal-to-noise ratio, uncertain labels, and nonstationary data in time. Those factors may influence the training process and slow down the models' performance improvement. These factors' influence may differ for end-to-end DL model and one using hand-crafted features. As not studied before, this paper compares models that use raw ECoG signal and time-frequency features for BCI motor imagery decoding. We investigate whether the current dataset size is a stronger limitation for any models. Finally, obtained filters were compared to identify differences between hand-crafted features and optimized with backpropagation. To compare the effectiveness of both strategies, we used a multilayer perceptron and a mix of convolutional and LSTM layers that were already proved effective in this task. The analysis was performed on the long-term clinical trial database (almost 600 minutes of recordings) of a tetraplegic patient executing motor imagery tasks for 3D hand translation. For a given dataset, the results showed that end-to-end training might not be significantly better than the hand-crafted features-based model. The performance gap is reduced with bigger datasets, but considering the increased computational load, end-to-end training may not be profitable for this application.