Contributed Talk Sessions | Poster Sessions | All Posters | Search Papers
Poster Session B: Wednesday, August 13, 1:00 – 4:00 pm, de Brug & E‑Hall
Intracranial EEG reveals multiplexed encoding of auditory, speech, and language embeddings in the human temporal lobe
Atlas Kazemian1, Josef Parvizi1, Daniel LK Yamins1, Laura Gwilliams1; 1Stanford University
Presenter: Atlas Kazemian
Speech understanding in the human brain involves several representational transformations: Air pressure fluctuations become a time-frequency representation in the cochlear; auditory cortex extracts speech-relevant units; distributed networks extract amodal representations of meaning and structure. Recent advances in speech and language models have led to a series of studies using text-based large language model (LLM) representations to model these neural transformations. Here we examine the brain’s end-to-end processing of language by extending the investigation to include biophysical models of the cochlear and auditory cortex, as well as performance-optimized models of speech (Whisper) and text (GPT-2, Llama-3). We use these models to predict the activity of spatially precise neural populations recorded via intracranial EEG (iEEG) as participants listened to audiobooks. Our findings are twofold. First, we observe clear differences in encoding performance within both auditory and language model families. Second, each model type captures distinct aspects of the signal in temporal lobe electrodes, suggesting that these regions encode a mixture of intermediate auditory and higher-level semantic features. Together, these results highlight the importance of examining model–brain alignment with fine-grained temporal precision.
Topic Area: Language & Communication
Extended Abstract: Full Text PDF