Using Localization and Factorization to Reduce the Complexity of Reinforcement Learning
Date
2015
Authors
Sunehag, Peter
Hutter, Marcus
Journal Title
Journal ISSN
Volume Title
Publisher
AGI Network
Abstract
General reinforcement learning is a powerful framework for artificial intelligence that has seen much theoretical progress since introduced fifteen years ago. We have previously provided guarantees for cases with finitely many possible environments. Though the results are the best possible in general, a linear dependence on the size of the hypothesis class renders them impractical. However, we dramatically improved on these by introducing the concept of environments generated by combining laws. The bounds are then linear in the number of laws needed
to generate the environment class. This number is identified as a natural complexity measure for classes of environments. The individual law might only predict some feature (factorization) and only in some contexts (localization). We here extend previous deterministic results to the important stochastic setting.
Description
Keywords
Citation
Collections
Source
Journal of Artificial General Intelligence
Type
Journal article
Book Title
Entity type
Access Statement
License Rights
Restricted until
2037-12-31
Downloads
File
Description