Curriculum Learning for Reinforcement Learning Domains: A Framework and Survey

Sanmit Narvekar,Bei Peng,M. Leonetti,Jivko Sinapov,Matthew E. Taylor,P. Stone

Published 2020 in Journal of machine learning research

ABSTRACT

Reinforcement learning (RL) is a popular paradigm for addressing sequential decision tasks in which the agent has only limited environmental feedback. Despite many advances over the past three decades, learning in many domains still requires a large amount of interaction with the environment, which can be prohibitively expensive in realistic scenarios. To address this problem, transfer learning has been applied to reinforcement learning such that experience gained in one task can be leveraged when starting to learn the next, harder task. More recently, several lines of research have explored how tasks, or data samples themselves, can be sequenced into a curriculum for the purpose of learning a problem that may otherwise be too difficult to learn from scratch. In this article, we present a framework for curriculum learning (CL) in reinforcement learning, and use it to survey and classify existing CL methods in terms of their assumptions, capabilities, and goals. Finally, we use our framework to find open problems and suggest directions for future RL curriculum learning research.

PUBLICATION RECORD

  • Publication year

    2020

  • Venue

    Journal of machine learning research

  • Publication date

    2020-03-10

  • Fields of study

    Mathematics, Computer Science, Education

  • Identifiers
  • External record

    Open on Semantic Scholar

  • Source metadata

    Semantic Scholar

CITATION MAP

EXTRACTION MAP

CLAIMS

  • No claims are published for this paper.

CONCEPTS

  • No concepts are published for this paper.

REFERENCES

Showing 1-100 of 119 references · Page 1 of 2

CITED BY

Showing 1-100 of 644 citing papers · Page 1 of 7