Decoding age-related changes in the spatiotemporal neural processing of speech using machine learning

2019 
Speech comprehension in noisy environments depends on complex interactions between sensory and cognitive systems. In older adults, such interactions may be affected, especially in those individuals who have more severe age-related hearing loss. Using a data-driven approach, we assessed the temporal (when in time) and spatial (where in the brain) characteristics of the cortex9s speech-evoked response that distinguish older adults with or without mild hearing loss. We used source montage to model scalp-recorded during a phoneme discrimination task conducted under clear and noise-degraded conditions. We applied machine learning analyses (stability selection and control) to choose features of the speech-evoked response that are consistent over a range of model parameters and support vector machine (SVM) classification to investigate the time course and brain regions that segregate groups and speech clarity. Whole-brain data analysis revealed a classification accuracy of 82.03% [area under the curve (AUC)=81.18%; F1-score 82.00%], distinguishing groups within ~50 ms after speech onset (i.e., as early as the P1 wave). We observed lower accuracy of 78.39% [AUC=78.74%; F1-score=79.00%] and delayed classification performance when the speech token were embedded in noise, with group segregation at 60 ms. Separate analysis using left (LH) and right hemisphere (RH) regions showed that LH speech activity was better at distinguishing hearing groups than activity measured over the RH. Moreover, stability selection analysis identified 13 brain regions (among 1428 total spatiotemporal features from 68 regions) where source activity segregated groups with >80% accuracy (clear speech); whereas 15 regions were critical for noise-degraded speech to achieve a comparable level of group segregation (76% accuracy). Our results identify two core neural networks associated with complex speech perception in older adults and confirm a larger number of neural regions, particularly in RH and frontal lobe, are active when processing degraded speech information.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    85
    References
    3
    Citations
    NaN
    KQI
    []