Self-Paced Context Evaluation for Contextual Reinforcement Learning

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

External Research Organisations

  • University of Freiburg
  • Bosch Center for Artificial Intelligence (BCAI)
View graph of relations

Details

Original languageEnglish
Title of host publicationProceedings of the international conference on machine learning (ICML)
Number of pages14
Publication statusE-pub ahead of print - 2021

Abstract

Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

Keywords

    cs.LG

Cite this

Self-Paced Context Evaluation for Contextual Reinforcement Learning. / Eimer, Theresa; Biedenkapp, André; Hutter, Frank et al.
Proceedings of the international conference on machine learning (ICML). 2021.

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Eimer, T, Biedenkapp, A, Hutter, F & Lindauer, M 2021, Self-Paced Context Evaluation for Contextual Reinforcement Learning. in Proceedings of the international conference on machine learning (ICML). <https://www.tnt.uni-hannover.de/papers/data/1454/space.pdf>
Eimer, T., Biedenkapp, A., Hutter, F., & Lindauer, M. (2021). Self-Paced Context Evaluation for Contextual Reinforcement Learning. In Proceedings of the international conference on machine learning (ICML) Advance online publication. https://www.tnt.uni-hannover.de/papers/data/1454/space.pdf
Eimer T, Biedenkapp A, Hutter F, Lindauer M. Self-Paced Context Evaluation for Contextual Reinforcement Learning. In Proceedings of the international conference on machine learning (ICML). 2021 Epub 2021.
Eimer, Theresa ; Biedenkapp, André ; Hutter, Frank et al. / Self-Paced Context Evaluation for Contextual Reinforcement Learning. Proceedings of the international conference on machine learning (ICML). 2021.
Download
@inproceedings{b7f481e4815a453c97f181a48cc71619,
title = "Self-Paced Context Evaluation for Contextual Reinforcement Learning",
abstract = " Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach. ",
keywords = "cs.LG",
author = "Theresa Eimer and Andr{\'e} Biedenkapp and Frank Hutter and Marius Lindauer",
note = "Funding Information: Theresa Eimer and Marius Lindauer acknowledge funding by the German Research Foundation (DFG) under LI 2801/4-1. All authors acknowledge funding by the Robert Bosch GmbH.",
year = "2021",
language = "English",
booktitle = "Proceedings of the international conference on machine learning (ICML)",

}

Download

TY - GEN

T1 - Self-Paced Context Evaluation for Contextual Reinforcement Learning

AU - Eimer, Theresa

AU - Biedenkapp, André

AU - Hutter, Frank

AU - Lindauer, Marius

N1 - Funding Information: Theresa Eimer and Marius Lindauer acknowledge funding by the German Research Foundation (DFG) under LI 2801/4-1. All authors acknowledge funding by the Robert Bosch GmbH.

PY - 2021

Y1 - 2021

N2 - Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

AB - Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

KW - cs.LG

M3 - Conference contribution

BT - Proceedings of the international conference on machine learning (ICML)

ER -

By the same author(s)