Modularity allows classification of human brain networks during music and speech perception

2020 
We investigate the use of modularity as a quantifier of whole-brain functional networks. Brain networks are constructed from functional magnetic resonance imaging while subjects listened to auditory pieces that varied in emotivity and cultural familiarity. The results of our analysis reveal high and low modularity groups based on the network configuration during a subject's favorite song, and this classification can predict network reconfiguration during the other auditory pieces. In particular, subjects in the low modularity group show significant brain network reconfiguration during both familiar and unfamiliar pieces. In contrast, the high modularity brain networks appear more robust and only exhibit significant changes during the unfamiliar music and speech. We also find differences in the stability of module composition for the two groups during each auditory piece. Our results suggest that the modularity of the whole-brain network plays a significant role in the way the network reconfigures during varying auditory processing demands, and it may therefore contribute to individual differences in neuroplasticity capability during therapeutic music engagement.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    1
    References
    2
    Citations
    NaN
    KQI
    []