Graph learning-based generation of abstractions for reinforcement learning

Research output: Contribution to journalArticleResearchpeer review

Authors

  • Yuan Xue
  • Daniel Kudenko
  • Megha Khosla

Research Organisations

External Research Organisations

  • Delft University of Technology
View graph of relations

Details

Original languageEnglish
JournalNeural Computing and Applications
Volume2023
Early online date9 Feb 2023
Publication statusPublished - 2023

Abstract

The application of reinforcement learning (RL) algorithms is often hindered by the combinatorial explosion of the state space. Previous works have leveraged abstractions which condense large state spaces to find tractable solutions. However, they assumed that the abstractions are provided by a domain expert. In this work, we propose a new approach to automatically construct abstract Markov decision processes (AMDPs) for potential-based reward shaping to improve the sample efficiency of RL algorithms. Our approach to constructing abstract states is inspired by graph representation learning methods, it effectively encodes the topological and reward structure of the ground-level MDP. We perform large-scale quantitative experiments on a range of navigation and gathering tasks under both stationary and stochastic settings. Our approach shows improvements of up to 8.5 times in sample efficiency and up to 3 times in run time over the baseline approach. Besides, with our qualitative analyses of the generated AMDPs, we are able to visually demonstrate the capability of our approach to preserve the topological and reward structure of the ground-level MDP.

Keywords

    Abstract MDP, Graph representations, Reinforcement learning, State representations

ASJC Scopus subject areas

Cite this

Graph learning-based generation of abstractions for reinforcement learning. / Xue, Yuan; Kudenko, Daniel; Khosla, Megha.
In: Neural Computing and Applications, Vol. 2023, 2023.

Research output: Contribution to journalArticleResearchpeer review

Xue Y, Kudenko D, Khosla M. Graph learning-based generation of abstractions for reinforcement learning. Neural Computing and Applications. 2023;2023. Epub 2023 Feb 9. doi: 10.1007/s00521-023-08211-x, 10.15488/15413
Xue, Yuan ; Kudenko, Daniel ; Khosla, Megha. / Graph learning-based generation of abstractions for reinforcement learning. In: Neural Computing and Applications. 2023 ; Vol. 2023.
Download
@article{f9f162a825664d2bbaffb624c5ff6ddb,
title = "Graph learning-based generation of abstractions for reinforcement learning",
abstract = "The application of reinforcement learning (RL) algorithms is often hindered by the combinatorial explosion of the state space. Previous works have leveraged abstractions which condense large state spaces to find tractable solutions. However, they assumed that the abstractions are provided by a domain expert. In this work, we propose a new approach to automatically construct abstract Markov decision processes (AMDPs) for potential-based reward shaping to improve the sample efficiency of RL algorithms. Our approach to constructing abstract states is inspired by graph representation learning methods, it effectively encodes the topological and reward structure of the ground-level MDP. We perform large-scale quantitative experiments on a range of navigation and gathering tasks under both stationary and stochastic settings. Our approach shows improvements of up to 8.5 times in sample efficiency and up to 3 times in run time over the baseline approach. Besides, with our qualitative analyses of the generated AMDPs, we are able to visually demonstrate the capability of our approach to preserve the topological and reward structure of the ground-level MDP.",
keywords = "Abstract MDP, Graph representations, Reinforcement learning, State representations",
author = "Yuan Xue and Daniel Kudenko and Megha Khosla",
note = "Funding Information: This work was partially funded by the Federal Ministry of Education and Research (BMBF), Germany, under the project LeibnizKILabor (Grant No. 01DD20003). Open Access funding enabled and organized by Projekt DEAL ",
year = "2023",
doi = "10.1007/s00521-023-08211-x",
language = "English",
volume = "2023",
journal = "Neural Computing and Applications",
issn = "0941-0643",
publisher = "Springer London",

}

Download

TY - JOUR

T1 - Graph learning-based generation of abstractions for reinforcement learning

AU - Xue, Yuan

AU - Kudenko, Daniel

AU - Khosla, Megha

N1 - Funding Information: This work was partially funded by the Federal Ministry of Education and Research (BMBF), Germany, under the project LeibnizKILabor (Grant No. 01DD20003). Open Access funding enabled and organized by Projekt DEAL

PY - 2023

Y1 - 2023

N2 - The application of reinforcement learning (RL) algorithms is often hindered by the combinatorial explosion of the state space. Previous works have leveraged abstractions which condense large state spaces to find tractable solutions. However, they assumed that the abstractions are provided by a domain expert. In this work, we propose a new approach to automatically construct abstract Markov decision processes (AMDPs) for potential-based reward shaping to improve the sample efficiency of RL algorithms. Our approach to constructing abstract states is inspired by graph representation learning methods, it effectively encodes the topological and reward structure of the ground-level MDP. We perform large-scale quantitative experiments on a range of navigation and gathering tasks under both stationary and stochastic settings. Our approach shows improvements of up to 8.5 times in sample efficiency and up to 3 times in run time over the baseline approach. Besides, with our qualitative analyses of the generated AMDPs, we are able to visually demonstrate the capability of our approach to preserve the topological and reward structure of the ground-level MDP.

AB - The application of reinforcement learning (RL) algorithms is often hindered by the combinatorial explosion of the state space. Previous works have leveraged abstractions which condense large state spaces to find tractable solutions. However, they assumed that the abstractions are provided by a domain expert. In this work, we propose a new approach to automatically construct abstract Markov decision processes (AMDPs) for potential-based reward shaping to improve the sample efficiency of RL algorithms. Our approach to constructing abstract states is inspired by graph representation learning methods, it effectively encodes the topological and reward structure of the ground-level MDP. We perform large-scale quantitative experiments on a range of navigation and gathering tasks under both stationary and stochastic settings. Our approach shows improvements of up to 8.5 times in sample efficiency and up to 3 times in run time over the baseline approach. Besides, with our qualitative analyses of the generated AMDPs, we are able to visually demonstrate the capability of our approach to preserve the topological and reward structure of the ground-level MDP.

KW - Abstract MDP

KW - Graph representations

KW - Reinforcement learning

KW - State representations

UR - http://www.scopus.com/inward/record.url?scp=85147750411&partnerID=8YFLogxK

U2 - 10.1007/s00521-023-08211-x

DO - 10.1007/s00521-023-08211-x

M3 - Article

AN - SCOPUS:85147750411

VL - 2023

JO - Neural Computing and Applications

JF - Neural Computing and Applications

SN - 0941-0643

ER -