Latent Property State Abstraction For Reinforcement learning

Research output: Contribution to conferencePaperResearchpeer review

Authors

  • John Burden
  • Sajjad Kamali Siahroudi
  • Daniel Kudenko

Research Organisations

External Research Organisations

  • University of Cambridge
View graph of relations

Details

Original languageEnglish
Number of pages8
Publication statusPublished - 2021
EventAdaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021 - Virtual, Online, United Kingdom (UK)
Duration: 3 May 20214 May 2021

Conference

ConferenceAdaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021
Country/TerritoryUnited Kingdom (UK)
CityVirtual, Online
Period3 May 20214 May 2021

Abstract

Potential Based Reward Shaping has proven itself to be an effective method for improving the learning rate for Reinforcement Learning algorithms - especially when the potential function is derived from the solution to an Abstract Markov Decision Process (AMDP) encapsulating an abstraction of the desired task. The provenance of the AMDP is often a domain expert. In this paper we introduce a novel method for the full automation of creating and solving an AMDP to induce a potential function. We then show empirically that the potential function our method creates improves the sample efficiency of DQN in the domain in which we test our approach.

Keywords

    Abstraction, Reinforcement Learning, Reward Shaping

ASJC Scopus subject areas

Cite this

Latent Property State Abstraction For Reinforcement learning. / Burden, John; Siahroudi, Sajjad Kamali; Kudenko, Daniel.
2021. Paper presented at Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, United Kingdom (UK).

Research output: Contribution to conferencePaperResearchpeer review

Burden, J, Siahroudi, SK & Kudenko, D 2021, 'Latent Property State Abstraction For Reinforcement learning', Paper presented at Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, United Kingdom (UK), 3 May 2021 - 4 May 2021. <https://ala2021.vub.ac.be/papers/ALA2021_paper_50.pdf>
Burden, J., Siahroudi, S. K., & Kudenko, D. (2021). Latent Property State Abstraction For Reinforcement learning. Paper presented at Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, United Kingdom (UK). https://ala2021.vub.ac.be/papers/ALA2021_paper_50.pdf
Burden J, Siahroudi SK, Kudenko D. Latent Property State Abstraction For Reinforcement learning. 2021. Paper presented at Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, United Kingdom (UK).
Burden, John ; Siahroudi, Sajjad Kamali ; Kudenko, Daniel. / Latent Property State Abstraction For Reinforcement learning. Paper presented at Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021, Virtual, Online, United Kingdom (UK).8 p.
Download
@conference{5629cd8c858642bb913d0c9e88c83f50,
title = "Latent Property State Abstraction For Reinforcement learning",
abstract = "Potential Based Reward Shaping has proven itself to be an effective method for improving the learning rate for Reinforcement Learning algorithms - especially when the potential function is derived from the solution to an Abstract Markov Decision Process (AMDP) encapsulating an abstraction of the desired task. The provenance of the AMDP is often a domain expert. In this paper we introduce a novel method for the full automation of creating and solving an AMDP to induce a potential function. We then show empirically that the potential function our method creates improves the sample efficiency of DQN in the domain in which we test our approach.",
keywords = "Abstraction, Reinforcement Learning, Reward Shaping",
author = "John Burden and Siahroudi, {Sajjad Kamali} and Daniel Kudenko",
year = "2021",
language = "English",
note = "Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021 ; Conference date: 03-05-2021 Through 04-05-2021",

}

Download

TY - CONF

T1 - Latent Property State Abstraction For Reinforcement learning

AU - Burden, John

AU - Siahroudi, Sajjad Kamali

AU - Kudenko, Daniel

PY - 2021

Y1 - 2021

N2 - Potential Based Reward Shaping has proven itself to be an effective method for improving the learning rate for Reinforcement Learning algorithms - especially when the potential function is derived from the solution to an Abstract Markov Decision Process (AMDP) encapsulating an abstraction of the desired task. The provenance of the AMDP is often a domain expert. In this paper we introduce a novel method for the full automation of creating and solving an AMDP to induce a potential function. We then show empirically that the potential function our method creates improves the sample efficiency of DQN in the domain in which we test our approach.

AB - Potential Based Reward Shaping has proven itself to be an effective method for improving the learning rate for Reinforcement Learning algorithms - especially when the potential function is derived from the solution to an Abstract Markov Decision Process (AMDP) encapsulating an abstraction of the desired task. The provenance of the AMDP is often a domain expert. In this paper we introduce a novel method for the full automation of creating and solving an AMDP to induce a potential function. We then show empirically that the potential function our method creates improves the sample efficiency of DQN in the domain in which we test our approach.

KW - Abstraction

KW - Reinforcement Learning

KW - Reward Shaping

UR - http://www.scopus.com/inward/record.url?scp=85134046472&partnerID=8YFLogxK

M3 - Paper

AN - SCOPUS:85134046472

T2 - Adaptive and Learning Agents Workshop, ALA 2021 at AAMAS 2021

Y2 - 3 May 2021 through 4 May 2021

ER -