Back
Tracking the Unknown: Modeling Long-Term Implicit Skill Acquisition as Non-Parametric Bayesian Sequence Learning
{Long perceptuo-motor sequences underlie skills from walking to language learning, and are often learned gradually and unconsciously in the face of noise. We used a non-parametric Bayesian n-gram model (Teh, 2006) to characterize the multi-day evolution of human subjects\textquoteright implicit representation of a serial reaction time task sequence with second-order contingencies. The reaction time for an element in the sequence depended on zero, one and more preceding elements at the same time, predicting frequency, repetition and higher-order learning effects. Our trial-level dynamic model captured these coexistent facilitation effects by seamlessly combining information from shorter and longer windows onto past events. We show how shifting their priors over window lengths allowed subjects to grow and refine their internal sequence representations week by week.}
@misc{item_3331502, title = {{Tracking the Unknown: Modeling Long-Term Implicit Skill Acquisition as Non-Parametric Bayesian Sequence Learning}}, booktitle = {{CogSci 2021 Virtual: Comparative Cognition, Cognitive Animals}}, abstract = {{Long perceptuo-motor sequences underlie skills from walking to language learning, and are often learned gradually and unconsciously in the face of noise. We used a non-parametric Bayesian n-gram model (Teh, 2006) to characterize the multi-day evolution of human subjects\textquoteright implicit representation of a serial reaction time task sequence with second-order contingencies. The reaction time for an element in the sequence depended on zero, one and more preceding elements at the same time, predicting frequency, repetition and higher-order learning effects. Our trial-level dynamic model captured these coexistent facilitation effects by seamlessly combining information from shorter and longer windows onto past events. We show how shifting their priors over window lengths allowed subjects to grow and refine their internal sequence representations week by week.}}, pages = {3307}, year = {2021}, slug = {item_3331502}, author = {Elteto, N and Nemeth, D and Janacsek, K and Dayan, P} }