What do reinforcement learning models measure? Interpreting model parameters in cognition and neuroscience

2021 
Author(s): Eckstein, MK; Wilbrecht, L; Collins, AG | Abstract: Reinforcement learning (RL) is a concept that has been invaluable to fields including machine learning, neuroscience, and cognitive science. However, what RL entails differs between fields, leading to difficulties when interpreting and translating findings. After laying out these differences, this paper focuses on cognitive (neuro)science to discuss how we as a field might overinterpret RL modeling results. We too often assume — implicitly — that modeling results generalize between tasks, models, and participant populations, despite negative empirical evidence for this assumption. We also often assume that parameters measure specific, unique (neuro)cognitive processes, a concept we call interpretability, when evidence suggests that they capture different functions across studies and tasks. We conclude that future computational research needs to pay increased attention to implicit assumptions when using RL models, and suggest that a more systematic understanding of contextual factors will help address issues and improve the ability of RL to explain brain and behavior.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    87
    References
    6
    Citations
    NaN
    KQI
    []