@article {Silva107698, author = {Carolina Feher da Silva and Yuan-Wei Yao and Todd A. Hare}, title = {Model-free reinforcement learning operates over information stored in working-memory to drive human choices}, elocation-id = {107698}, year = {2017}, doi = {10.1101/107698}, publisher = {Cold Spring Harbor Laboratory}, abstract = {Model-free learning creates stimulus-response associations, but are there limits to the types of stimuli it can operate over? Most experiments on reward-learning have used discrete sensory stimuli, but there is no algorithmic reason to restrict model-free learning to external stimuli, and theories suggest that model-free processes may operate over highly abstract concepts and goals. Our study aimed to determine whether model-free learning can operate over environmental states defined by information held in working memory. We compared the data from human participants in two conditions that presented learning cues either simultaneously or as a temporal sequence that required working memory. There was a significant influence of model-free learning in the working memory condition. Moreover, both groups showed greater model-free effects than simulated model-based agents. Thus, we show that model-free learning processes operate not just in parallel, but also in cooperation with canonical executive functions such as working memory to support behavior.}, URL = {https://www.biorxiv.org/content/early/2017/02/11/107698}, eprint = {https://www.biorxiv.org/content/early/2017/02/11/107698.full.pdf}, journal = {bioRxiv} }