Hedonic Value: Enhancing Adaptation for Motivated Agents

Ignasi Cos-Aguilera, Lola Cañamero, Gillian M. Hayes, Andrew Gillies

Research output: Contribution to journalArticlepeer-review

15 Citations (Scopus)
190 Downloads (Pure)


Reinforcement learning (RL) in the context of artificial agents is typically used to produce behavioural responses as a function of the reward obtained by interaction with the environment. When the problem consists of learning the shortest path to a goal, it is common to use reward functions yielding a fixed value after each decision, for example a positive value if the target location has been attained and a negative one at each intermediate step. However, this fixed strategy may be overly simplistic for agents to adapt to dynamic environments, in which resources may vary from time to time. By contrast, there is significant evidence that most living beings internally modulate reward value as a function of their context to expand their range of adaptivity. Inspired by the potential of
this operation, we present a review of its underlying processes and we introduce a simplified formalisation for artificial agents. The performance of this formalism is tested by monitoring the adaptation of an agent endowed with a model of motivated actor-critic, embedded with our formalisation of value and constrained by physiological stability, to environments with different resource distribution.
Our main result shows that the manner in which reward is internally processed as a function of the agent’s motivational state, strongly influences adaptivity of the behavioural cycles generated and the agent’s physiological stability.
Original languageEnglish
Pages (from-to)465-483
JournalAdaptive Behavior
Issue number6
Early online date6 Aug 2013
Publication statusPublished - 2013


  • Hedonic Value, Motivation, Reinforcement Learning, Actor-Critic, Grounding


Dive into the research topics of 'Hedonic Value: Enhancing Adaptation for Motivated Agents'. Together they form a unique fingerprint.

Cite this