Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
This paper presents a neuromuscular speech interface that synthesizes speech directly from electromyographic (EMG) signals of orofacial muscles. It demonstrates a strong linear relationship between self-supervised speech (SS) representations and EMG power, enabling direct mapping of EMG to SS features for end-to-end speech synthesis. This approach bypasses explicit articulatory models and vocoders, offering a novel pathway for communication aids.
Enabling individuals who have lost the ability to speak due to motor impairments to communicate effectively through synthesized speech can dramatically improve their quality of life and social integration, opening new markets for assistive communication technologies.