Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 70% Match Research Paper Biomedical engineers,Neuroscientists,Machine learning researchers,Medical device developers 2 days ago

ESTformer: Transformer utilising spatiotemporal dependencies for electroencephalogram super-resolution

speech-audio › audio-generation
📄 Abstract

Abstract: Towards practical applications of Electroencephalography (EEG), lightweight acquisition devices garner significant attention. However, EEG channel selection methods are commonly data-sensitive and cannot establish a unified sound paradigm for EEG acquisition devices. Through reverse conceptualisation, we formulated EEG applications in an EEG super-resolution (SR) manner, but suffered from high computation costs, extra interpolation bias, and few insights into spatiotemporal dependency modelling. To this end, we propose ESTformer, an EEG SR framework that utilises spatiotemporal dependencies based on the transformer. ESTformer applies positional encoding methods and a multihead self-attention mechanism to the space and time dimensions, which can learn spatial structural correlations and temporal functional variations. ESTformer, with the fixed mask strategy, adopts a mask token to upsample low-resolution (LR) EEG data in the case of disturbance from mathematical interpolation methods. On this basis, we designed various transformer blocks to construct a spatial interpolation module (SIM) and a temporal reconstruction module (TRM). Finally, ESTformer cascades the SIM and TRM to capture and model the spatiotemporal dependencies for EEG SR with fidelity. Extensive experimental results on two EEG datasets show the effectiveness of ESTformer against previous state-of-the-art methods, demonstrating the versatility of the Transformer for EEG SR tasks. The superiority of the SR data was verified in an EEG-based person identification and emotion recognition task, achieving a 2% to 38% improvement compared with the LR data at different sampling scales.
Authors (4)
Dongdong Li
Zhongliang Zeng
Zhe Wang
Hai Yang
Submitted
December 3, 2023
arXiv Category
eess.SP
Knowledge-Based Systems, 317, 113345 (2025)
arXiv PDF

Key Contributions

Proposes ESTformer, a Transformer-based framework for EEG super-resolution that effectively utilizes spatiotemporal dependencies. It employs positional encoding and multi-head self-attention across space and time, along with a mask token strategy for upsampling, addressing high computation costs and interpolation bias of previous methods.

Business Value

Enables the development of more practical and informative EEG systems, potentially leading to better diagnosis and monitoring of neurological conditions using more accessible hardware.