PT - JOURNAL ARTICLE AU - Mehdi Keramati AU - Boris Gutkin TI - Collecting reward to defend homeostasis: A homeostatic reinforcement learning theory AID - 10.1101/005140 DP - 2014 Jan 01 TA - bioRxiv PG - 005140 4099 - http://biorxiv.org/content/early/2014/05/14/005140.short 4100 - http://biorxiv.org/content/early/2014/05/14/005140.full AB - Efficient regulation of internal homeostasis and defending it against perturbations requires complex behavioral strategies. However, the computational principles mediating brain’s homeostatic regulation of reward and associative learning remain undefined. Here we use a definition of primary rewards, as outcomes fulfilling physiological needs, to build a normative theory showing how learning motivated behavior is modulated by the internal state of the animal. The theory proves that seeking rewards is equivalent to the fundamental objective of physiological stability, defining the notion of physiological rationality of behavior. We further give a formal basis for temporal discounting of reward. It also explains how animals learn to act predictively to preclude prospective homeostatic challenges, and attributes a normative computational role to the modulation of midbrain dopaminergic activity by hypothalamic signals.