Skip navigation
Skip navigation

Avoiding wireheading with value reinforcement learning

Everitt, Tom; Hutter, Marcus


How can we design good goals for arbitrarily intelligent agents? Reinforcement learning (RL) may seem like a natural approach. Unfortunately, RL does not work well for generally intelligent agents, as RL agents are incentivised to shortcut the reward sensor for maximum reward – the so-called wireheading problem. In this paper we suggest an alternative to RL called value reinforcement learning (VRL). In VRL, agents use the reward signal to a utility function. The VRL setup allows us to...[Show more]

CollectionsANU Research Publications
Date published: 2016-06-25
Type: Journal article
Source: Lecture Notes in Computer Science
DOI: 10.1007/978-3-319-41649-6_2
Access Rights: Open Access


There are no files associated with this item.

Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.

Updated:  19 May 2020/ Responsible Officer:  University Librarian/ Page Contact:  Library Systems & Web Coordinator