Self-Paced Context Evaluation for Contextual Reinforcement Learning

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

External Research Organisations

  • University of Freiburg
  • Bosch Center for Artificial Intelligence (BCAI)
View graph of relations

Details

Original languageEnglish
Title of host publicationProceedings of the international conference on machine learning (ICML)
PublisherML Research Press
Number of pages14
ISBN (print)978-171384506-5
Publication statusPublished - 18 Jul 2021

Publication series

NameProceedings of Machine Learning Research
ISSN (Print)2640-3498

Abstract

Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

Keywords

    cs.LG

Cite this

Self-Paced Context Evaluation for Contextual Reinforcement Learning. / Eimer, Theresa; Biedenkapp, André; Hutter, Frank et al.
Proceedings of the international conference on machine learning (ICML). ML Research Press, 2021. (Proceedings of Machine Learning Research).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Eimer, T, Biedenkapp, A, Hutter, F & Lindauer, M 2021, Self-Paced Context Evaluation for Contextual Reinforcement Learning. in Proceedings of the international conference on machine learning (ICML). Proceedings of Machine Learning Research, ML Research Press. <https://www.tnt.uni-hannover.de/papers/data/1454/space.pdf>
Eimer, T., Biedenkapp, A., Hutter, F., & Lindauer, M. (2021). Self-Paced Context Evaluation for Contextual Reinforcement Learning. In Proceedings of the international conference on machine learning (ICML) (Proceedings of Machine Learning Research). ML Research Press. https://www.tnt.uni-hannover.de/papers/data/1454/space.pdf
Eimer T, Biedenkapp A, Hutter F, Lindauer M. Self-Paced Context Evaluation for Contextual Reinforcement Learning. In Proceedings of the international conference on machine learning (ICML). ML Research Press. 2021. (Proceedings of Machine Learning Research).
Eimer, Theresa ; Biedenkapp, André ; Hutter, Frank et al. / Self-Paced Context Evaluation for Contextual Reinforcement Learning. Proceedings of the international conference on machine learning (ICML). ML Research Press, 2021. (Proceedings of Machine Learning Research).
Download
@inproceedings{b7f481e4815a453c97f181a48cc71619,
title = "Self-Paced Context Evaluation for Contextual Reinforcement Learning",
abstract = " Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach. ",
keywords = "cs.LG",
author = "Theresa Eimer and Andr{\'e} Biedenkapp and Frank Hutter and Marius Lindauer",
year = "2021",
month = jul,
day = "18",
language = "English",
isbn = "978-171384506-5",
series = "Proceedings of Machine Learning Research",
publisher = "ML Research Press",
booktitle = "Proceedings of the international conference on machine learning (ICML)",

}

Download

TY - GEN

T1 - Self-Paced Context Evaluation for Contextual Reinforcement Learning

AU - Eimer, Theresa

AU - Biedenkapp, André

AU - Hutter, Frank

AU - Lindauer, Marius

PY - 2021/7/18

Y1 - 2021/7/18

N2 - Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

AB - Reinforcement learning (RL) has made a lot of advances for solving a single problem in a given environment; but learning policies that generalize to unseen variations of a problem remains challenging. To improve sample efficiency for learning on such instances of a problem domain, we present Self-Paced Context Evaluation (SPaCE). Based on self-paced learning, \spc automatically generates \task curricula online with little computational overhead. To this end, SPaCE leverages information contained in state values during training to accelerate and improve training performance as well as generalization capabilities to new instances from the same problem domain. Nevertheless, SPaCE is independent of the problem domain at hand and can be applied on top of any RL agent with state-value function approximation. We demonstrate SPaCE's ability to speed up learning of different value-based RL agents on two environments, showing better generalization capabilities and up to 10x faster learning compared to naive approaches such as round robin or SPDRL, as the closest state-of-the-art approach.

KW - cs.LG

UR - http://www.scopus.com/inward/record.url?scp=85161344151&partnerID=8YFLogxK

M3 - Conference contribution

SN - 978-171384506-5

T3 - Proceedings of Machine Learning Research

BT - Proceedings of the international conference on machine learning (ICML)

PB - ML Research Press

ER -

By the same author(s)