Skip navigation
Skip navigation

Avoiding wireheading with value reinforcement learning

Everitt, Tom; Hutter, Marcus

Description

How can we design good goals for arbitrarily intelligent agents? Reinforcement learning (RL) may seem like a natural approach. Unfortunately, RL does not work well for generally intelligent agents, as RL agents are incentivised to shortcut the reward sensor for maximum reward – the so-called wireheading problem. In this paper we suggest an alternative to RL called value reinforcement learning (VRL). In VRL, agents use the reward signal to a utility function. The VRL setup allows us to...[Show more]

CollectionsANU Research Publications
Date published: 2016-06-25
Type: Journal article
URI: http://hdl.handle.net/1885/111445
Source: Lecture Notes in Computer Science
DOI: 10.1007/978-3-319-41649-6_2

Download

There are no files associated with this item.


Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.

Updated:  20 July 2017/ Responsible Officer:  University Librarian/ Page Contact:  Library Systems & Web Coordinator