Latent Property State Abstraction For Reinforcement learning

Publikation: KonferenzbeitragPaperForschungPeer-Review

Autoren

  • John Burden
  • Sajjad Kamali Siahroudi
  • Daniel Kudenko

Organisationseinheiten

Externe Organisationen

  • University of Cambridge
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Seitenumfang8
PublikationsstatusVeröffentlicht - 2021
VeranstaltungAdaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021 - Virtual, Online, Großbritannien / Vereinigtes Königreich
Dauer: 3 Mai 20214 Mai 2021

Konferenz

KonferenzAdaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021
Land/GebietGroßbritannien / Vereinigtes Königreich
OrtVirtual, Online
Zeitraum3 Mai 20214 Mai 2021

Abstract

Potential Based Reward Shaping has proven itself to be an effective method for improving the learning rate for Reinforcement Learning algorithms - especially when the potential function is derived from the solution to an Abstract Markov Decision Process (AMDP) encapsulating an abstraction of the desired task. The provenance of the AMDP is often a domain expert. In this paper we introduce a novel method for the full automation of creating and solving an AMDP to induce a potential function. We then show empirically that the potential function our method creates improves the sample efficiency of DQN in the domain in which we test our approach.

ASJC Scopus Sachgebiete

Zitieren

Latent Property State Abstraction For Reinforcement learning. / Burden, John; Siahroudi, Sajjad Kamali; Kudenko, Daniel.
2021. Beitrag in Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, Großbritannien / Vereinigtes Königreich.

Publikation: KonferenzbeitragPaperForschungPeer-Review

Burden, J, Siahroudi, SK & Kudenko, D 2021, 'Latent Property State Abstraction For Reinforcement learning', Beitrag in Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, Großbritannien / Vereinigtes Königreich, 3 Mai 2021 - 4 Mai 2021. <https://ala2021.vub.ac.be/papers/ALA2021_paper_50.pdf>
Burden, J., Siahroudi, S. K., & Kudenko, D. (2021). Latent Property State Abstraction For Reinforcement learning. Beitrag in Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, Großbritannien / Vereinigtes Königreich. https://ala2021.vub.ac.be/papers/ALA2021_paper_50.pdf
Burden J, Siahroudi SK, Kudenko D. Latent Property State Abstraction For Reinforcement learning. 2021. Beitrag in Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, Großbritannien / Vereinigtes Königreich.
Burden, John ; Siahroudi, Sajjad Kamali ; Kudenko, Daniel. / Latent Property State Abstraction For Reinforcement learning. Beitrag in Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, Großbritannien / Vereinigtes Königreich.8 S.
Download
@conference{5629cd8c858642bb913d0c9e88c83f50,
title = "Latent Property State Abstraction For Reinforcement learning",
abstract = "Potential Based Reward Shaping has proven itself to be an effective method for improving the learning rate for Reinforcement Learning algorithms - especially when the potential function is derived from the solution to an Abstract Markov Decision Process (AMDP) encapsulating an abstraction of the desired task. The provenance of the AMDP is often a domain expert. In this paper we introduce a novel method for the full automation of creating and solving an AMDP to induce a potential function. We then show empirically that the potential function our method creates improves the sample efficiency of DQN in the domain in which we test our approach.",
keywords = "Abstraction, Reinforcement Learning, Reward Shaping",
author = "John Burden and Siahroudi, {Sajjad Kamali} and Daniel Kudenko",
year = "2021",
language = "English",
note = "Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021 ; Conference date: 03-05-2021 Through 04-05-2021",

}

Download

TY - CONF

T1 - Latent Property State Abstraction For Reinforcement learning

AU - Burden, John

AU - Siahroudi, Sajjad Kamali

AU - Kudenko, Daniel

PY - 2021

Y1 - 2021

N2 - Potential Based Reward Shaping has proven itself to be an effective method for improving the learning rate for Reinforcement Learning algorithms - especially when the potential function is derived from the solution to an Abstract Markov Decision Process (AMDP) encapsulating an abstraction of the desired task. The provenance of the AMDP is often a domain expert. In this paper we introduce a novel method for the full automation of creating and solving an AMDP to induce a potential function. We then show empirically that the potential function our method creates improves the sample efficiency of DQN in the domain in which we test our approach.

AB - Potential Based Reward Shaping has proven itself to be an effective method for improving the learning rate for Reinforcement Learning algorithms - especially when the potential function is derived from the solution to an Abstract Markov Decision Process (AMDP) encapsulating an abstraction of the desired task. The provenance of the AMDP is often a domain expert. In this paper we introduce a novel method for the full automation of creating and solving an AMDP to induce a potential function. We then show empirically that the potential function our method creates improves the sample efficiency of DQN in the domain in which we test our approach.

KW - Abstraction

KW - Reinforcement Learning

KW - Reward Shaping

UR - http://www.scopus.com/inward/record.url?scp=85134046472&partnerID=8YFLogxK

M3 - Paper

AN - SCOPUS:85134046472

T2 - Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021

Y2 - 3 May 2021 through 4 May 2021

ER -