Self-Paced Context Evaluation for Contextual Reinforcement Learning

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

Externe Organisationen

  • Albert-Ludwigs-Universität Freiburg
  • Bosch Center for Artificial Intelligence (BCAI)
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksProceedings of the international conference on machine learning (ICML)
Seitenumfang14
PublikationsstatusElektronisch veröffentlicht (E-Pub) - 2021

Abstract

Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

Zitieren

Self-Paced Context Evaluation for Contextual Reinforcement Learning. / Eimer, Theresa; Biedenkapp, André; Hutter, Frank et al.
Proceedings of the international conference on machine learning (ICML). 2021.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Eimer, T, Biedenkapp, A, Hutter, F & Lindauer, M 2021, Self-Paced Context Evaluation for Contextual Reinforcement Learning. in Proceedings of the international conference on machine learning (ICML). <https://www.tnt.uni-hannover.de/papers/data/1454/space.pdf>
Eimer, T., Biedenkapp, A., Hutter, F., & Lindauer, M. (2021). Self-Paced Context Evaluation for Contextual Reinforcement Learning. In Proceedings of the international conference on machine learning (ICML) Vorabveröffentlichung online. https://www.tnt.uni-hannover.de/papers/data/1454/space.pdf
Eimer T, Biedenkapp A, Hutter F, Lindauer M. Self-Paced Context Evaluation for Contextual Reinforcement Learning. in Proceedings of the international conference on machine learning (ICML). 2021 Epub 2021.
Eimer, Theresa ; Biedenkapp, André ; Hutter, Frank et al. / Self-Paced Context Evaluation for Contextual Reinforcement Learning. Proceedings of the international conference on machine learning (ICML). 2021.
Download
@inproceedings{b7f481e4815a453c97f181a48cc71619,
title = "Self-Paced Context Evaluation for Contextual Reinforcement Learning",
abstract = " Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach. ",
keywords = "cs.LG",
author = "Theresa Eimer and Andr{\'e} Biedenkapp and Frank Hutter and Marius Lindauer",
note = "Funding Information: Theresa Eimer and Marius Lindauer acknowledge funding by the German Research Foundation (DFG) under LI 2801/4-1. All authors acknowledge funding by the Robert Bosch GmbH.",
year = "2021",
language = "English",
booktitle = "Proceedings of the international conference on machine learning (ICML)",

}

Download

TY - GEN

T1 - Self-Paced Context Evaluation for Contextual Reinforcement Learning

AU - Eimer, Theresa

AU - Biedenkapp, André

AU - Hutter, Frank

AU - Lindauer, Marius

N1 - Funding Information: Theresa Eimer and Marius Lindauer acknowledge funding by the German Research Foundation (DFG) under LI 2801/4-1. All authors acknowledge funding by the Robert Bosch GmbH.

PY - 2021

Y1 - 2021

N2 - Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

AB - Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

KW - cs.LG

M3 - Conference contribution

BT - Proceedings of the international conference on machine learning (ICML)

ER -

Von denselben Autoren