Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Clarifying the neural basis of speech intelligibility is critical for
computational neuroscience and digital speech processing. Recent neuroimaging
studies have shown that intelligibility modulates cortical activity beyond
simple acoustics, primarily in the superior temporal and inferior frontal gyri.
However, previous studies have been largely confined to clean speech, leaving
it unclear whether the brain employs condition-invariant neural codes across
diverse listening environments. To address this gap, we propose a novel
architecture built upon a deep state space model for decoding intelligibility
from fMRI signals, specifically tailored to their high-dimensional temporal
structure. We present the first attempt to decode intelligibility across
acoustically distinct conditions, showing our method significantly outperforms
classical approaches. Furthermore, region-wise analysis highlights
contributions from auditory, frontal, and parietal regions, and cross-condition
transfer indicates the presence of condition-invariant neural codes, thereby
advancing understanding of abstract linguistic representations in the brain.
Key Contributions
This paper proposes a novel deep state space model (DSSM) for decoding speech intelligibility from fMRI signals, specifically designed to capture high-dimensional temporal structure. It demonstrates condition-invariant decoding across diverse listening environments, significantly outperforming classical approaches and highlighting contributions from auditory, frontal, and parietal regions.
Business Value
Advances understanding of speech perception and has potential applications in developing better speech prosthetics, hearing aids, and brain-computer interfaces for communication.