Loading [MathJax]/extensions/tex2jax.js

Extended Abstract: AutoRL Hyperparameter Landscapes

Research output: Contribution to conferenceAbstractResearchpeer review

Details

Original languageEnglish
Number of pages6
Publication statusE-pub ahead of print - 20 Jul 2023
EventEuropean Workshop on Reinforcement Learning 2023 - Brüssel
Duration: 13 Sept 202316 Sept 2023
https://ewrl.wordpress.com/ewrl16-2023/

Workshop

WorkshopEuropean Workshop on Reinforcement Learning 2023
CityBrüssel
Period13 Sept 202316 Sept 2023
Internet address

Abstract

Although Reinforcement Learning (RL) has shown to be capable of producing impressive results, its use is limited by the impact of its hyperparameters on performance. This often makes it difficult to achieve good results in practice. Automated RL (AutoRL) addresses this difficulty, yet little is known about the dynamics of the hyperparameter landscapes that hyperparameter optimization (HPO) methods traverse in search of optimal configurations. In view of existing AutoRL approaches dynamically adjusting hyperparameter configurations, we propose an approach to build and analyze these hyperparameter landscapes not just for one point in time but at multiple points in time throughout training. Addressing an important open question on the legitimacy of such dynamic AutoRL approaches, we provide thorough empirical evidence that the hyperparameter landscapes strongly vary over time across representative algorithms from RL literature (DQN, PPO, and SAC) in different kinds of environments (Cartpole, Bipedal Walker, and Hopper). This supports the theory that hyperparameters should be dynamically adjusted during training and shows the potential for more insights on AutoRL problems that can be gained through landscape analysis.

Cite this

Extended Abstract: AutoRL Hyperparameter Landscapes. / Mohan, Aditya; Benjamins, Carolin; Wienecke, Konrad et al.
2023. Abstract from European Workshop on Reinforcement Learning 2023, Brüssel.

Research output: Contribution to conferenceAbstractResearchpeer review

Mohan, A, Benjamins, C, Wienecke, K, Dockhorn, A & Lindauer, M 2023, 'Extended Abstract: AutoRL Hyperparameter Landscapes', European Workshop on Reinforcement Learning 2023, Brüssel, 13 Sept 2023 - 16 Sept 2023. <https://openreview.net/forum?id=4Zu0l5lBgc>
Mohan, A., Benjamins, C., Wienecke, K., Dockhorn, A., & Lindauer, M. (2023). Extended Abstract: AutoRL Hyperparameter Landscapes. Abstract from European Workshop on Reinforcement Learning 2023, Brüssel. Advance online publication. https://openreview.net/forum?id=4Zu0l5lBgc
Mohan A, Benjamins C, Wienecke K, Dockhorn A, Lindauer M. Extended Abstract: AutoRL Hyperparameter Landscapes. 2023. Abstract from European Workshop on Reinforcement Learning 2023, Brüssel. Epub 2023 Jul 20.
Mohan, Aditya ; Benjamins, Carolin ; Wienecke, Konrad et al. / Extended Abstract : AutoRL Hyperparameter Landscapes. Abstract from European Workshop on Reinforcement Learning 2023, Brüssel.6 p.
Download
@conference{4d63a36a8a73415f95f9a197080dd36f,
title = "Extended Abstract: AutoRL Hyperparameter Landscapes",
abstract = "Although Reinforcement Learning (RL) has shown to be capable of producing impressive results, its use is limited by the impact of its hyperparameters on performance. This often makes it difficult to achieve good results in practice. Automated RL (AutoRL) addresses this difficulty, yet little is known about the dynamics of the hyperparameter landscapes that hyperparameter optimization (HPO) methods traverse in search of optimal configurations. In view of existing AutoRL approaches dynamically adjusting hyperparameter configurations, we propose an approach to build and analyze these hyperparameter landscapes not just for one point in time but at multiple points in time throughout training. Addressing an important open question on the legitimacy of such dynamic AutoRL approaches, we provide thorough empirical evidence that the hyperparameter landscapes strongly vary over time across representative algorithms from RL literature (DQN, PPO, and SAC) in different kinds of environments (Cartpole, Bipedal Walker, and Hopper). This supports the theory that hyperparameters should be dynamically adjusted during training and shows the potential for more insights on AutoRL problems that can be gained through landscape analysis.",
author = "Aditya Mohan and Carolin Benjamins and Konrad Wienecke and Alexander Dockhorn and Marius Lindauer",
year = "2023",
month = jul,
day = "20",
language = "English",
note = "European Workshop on Reinforcement Learning 2023 ; Conference date: 13-09-2023 Through 16-09-2023",
url = "https://ewrl.wordpress.com/ewrl16-2023/",

}

Download

TY - CONF

T1 - Extended Abstract

T2 - European Workshop on Reinforcement Learning 2023

AU - Mohan, Aditya

AU - Benjamins, Carolin

AU - Wienecke, Konrad

AU - Dockhorn, Alexander

AU - Lindauer, Marius

PY - 2023/7/20

Y1 - 2023/7/20

N2 - Although Reinforcement Learning (RL) has shown to be capable of producing impressive results, its use is limited by the impact of its hyperparameters on performance. This often makes it difficult to achieve good results in practice. Automated RL (AutoRL) addresses this difficulty, yet little is known about the dynamics of the hyperparameter landscapes that hyperparameter optimization (HPO) methods traverse in search of optimal configurations. In view of existing AutoRL approaches dynamically adjusting hyperparameter configurations, we propose an approach to build and analyze these hyperparameter landscapes not just for one point in time but at multiple points in time throughout training. Addressing an important open question on the legitimacy of such dynamic AutoRL approaches, we provide thorough empirical evidence that the hyperparameter landscapes strongly vary over time across representative algorithms from RL literature (DQN, PPO, and SAC) in different kinds of environments (Cartpole, Bipedal Walker, and Hopper). This supports the theory that hyperparameters should be dynamically adjusted during training and shows the potential for more insights on AutoRL problems that can be gained through landscape analysis.

AB - Although Reinforcement Learning (RL) has shown to be capable of producing impressive results, its use is limited by the impact of its hyperparameters on performance. This often makes it difficult to achieve good results in practice. Automated RL (AutoRL) addresses this difficulty, yet little is known about the dynamics of the hyperparameter landscapes that hyperparameter optimization (HPO) methods traverse in search of optimal configurations. In view of existing AutoRL approaches dynamically adjusting hyperparameter configurations, we propose an approach to build and analyze these hyperparameter landscapes not just for one point in time but at multiple points in time throughout training. Addressing an important open question on the legitimacy of such dynamic AutoRL approaches, we provide thorough empirical evidence that the hyperparameter landscapes strongly vary over time across representative algorithms from RL literature (DQN, PPO, and SAC) in different kinds of environments (Cartpole, Bipedal Walker, and Hopper). This supports the theory that hyperparameters should be dynamically adjusted during training and shows the potential for more insights on AutoRL problems that can be gained through landscape analysis.

M3 - Abstract

Y2 - 13 September 2023 through 16 September 2023

ER -

By the same author(s)