Utilising Assured Multi-Agent Reinforcement Learning within Safety-Critical Scenarios

Publikation: Beitrag in FachzeitschriftKonferenzaufsatz in FachzeitschriftForschungPeer-Review

Autorschaft

  • Joshua Riley
  • Radu Calinescu
  • Colin Paterson
  • Daniel Kudenko
  • Alec Banks

Organisationseinheiten

Externe Organisationen

  • University of York
  • Defence Science and Technology Laboratory
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Seiten (von - bis)1061-1070
Seitenumfang10
FachzeitschriftProcedia Computer Science
Jahrgang192
Frühes Online-Datum1 Okt. 2021
PublikationsstatusVeröffentlicht - 2021
Veranstaltung25th KES International Conference on Knowledge-Based and Intelligent Information and Engineering Systems, KES 2021 - Szczecin, Polen
Dauer: 8 Sept. 202110 Sept. 2021

Abstract

Multi-agent reinforcement learning allows a team of agents to learn how to work together to solve complex decision-making problems in a shared environment. However, this learning process utilises stochastic mechanisms, meaning that its use in safety-critical domains can be problematic. To overcome this issue, we propose an Assured Multi-Agent Reinforcement Learning (AMARL) approach that uses a model checking technique called quantitative verification to provide formal guarantees of agent compliance with safety, performance, and other non-functional requirements during and after the reinforcement learning process. We demonstrate the applicability of our AMARL approach in three different patrolling navigation domains in which multi-agent systems must learn to visit key areas by using different types of reinforcement learning algorithms (temporal difference learning, game theory, and direct policy search). Furthermore, we compare the effectiveness of these algorithms when used in combination with and without our approach. Our extensive experiments with both homogeneous and heterogeneous multi-agent systems of different sizes show that the use of AMARL leads to safety requirements being consistently satisfied and to better overall results than standard reinforcement learning.

ASJC Scopus Sachgebiete

Zitieren

Utilising Assured Multi-Agent Reinforcement Learning within Safety-Critical Scenarios. / Riley, Joshua; Calinescu, Radu; Paterson, Colin et al.
in: Procedia Computer Science, Jahrgang 192, 2021, S. 1061-1070.

Publikation: Beitrag in FachzeitschriftKonferenzaufsatz in FachzeitschriftForschungPeer-Review

Riley, J, Calinescu, R, Paterson, C, Kudenko, D & Banks, A 2021, 'Utilising Assured Multi-Agent Reinforcement Learning within Safety-Critical Scenarios', Procedia Computer Science, Jg. 192, S. 1061-1070. https://doi.org/10.1016/j.procs.2021.08.109
Riley, J., Calinescu, R., Paterson, C., Kudenko, D., & Banks, A. (2021). Utilising Assured Multi-Agent Reinforcement Learning within Safety-Critical Scenarios. Procedia Computer Science, 192, 1061-1070. https://doi.org/10.1016/j.procs.2021.08.109
Riley J, Calinescu R, Paterson C, Kudenko D, Banks A. Utilising Assured Multi-Agent Reinforcement Learning within Safety-Critical Scenarios. Procedia Computer Science. 2021;192:1061-1070. Epub 2021 Okt 1. doi: 10.1016/j.procs.2021.08.109
Riley, Joshua ; Calinescu, Radu ; Paterson, Colin et al. / Utilising Assured Multi-Agent Reinforcement Learning within Safety-Critical Scenarios. in: Procedia Computer Science. 2021 ; Jahrgang 192. S. 1061-1070.
Download
@article{8c2cd6542b954a09be3b22d689ec1902,
title = "Utilising Assured Multi-Agent Reinforcement Learning within Safety-Critical Scenarios",
abstract = "Multi-agent reinforcement learning allows a team of agents to learn how to work together to solve complex decision-making problems in a shared environment. However, this learning process utilises stochastic mechanisms, meaning that its use in safety-critical domains can be problematic. To overcome this issue, we propose an Assured Multi-Agent Reinforcement Learning (AMARL) approach that uses a model checking technique called quantitative verification to provide formal guarantees of agent compliance with safety, performance, and other non-functional requirements during and after the reinforcement learning process. We demonstrate the applicability of our AMARL approach in three different patrolling navigation domains in which multi-agent systems must learn to visit key areas by using different types of reinforcement learning algorithms (temporal difference learning, game theory, and direct policy search). Furthermore, we compare the effectiveness of these algorithms when used in combination with and without our approach. Our extensive experiments with both homogeneous and heterogeneous multi-agent systems of different sizes show that the use of AMARL leads to safety requirements being consistently satisfied and to better overall results than standard reinforcement learning.",
keywords = "Assurance, Assured multi-agent reinforcement learning, Multi-agent reinforcement learning, Multi-agent systems, Quantitative verification, Reinforcement learning, Safe multi-agent reinforcement learning, Safety-critical scenarios",
author = "Joshua Riley and Radu Calinescu and Colin Paterson and Daniel Kudenko and Alec Banks",
year = "2021",
doi = "10.1016/j.procs.2021.08.109",
language = "English",
volume = "192",
pages = "1061--1070",
note = "25th KES International Conference on Knowledge-Based and Intelligent Information and Engineering Systems, KES 2021 ; Conference date: 08-09-2021 Through 10-09-2021",

}

Download

TY - JOUR

T1 - Utilising Assured Multi-Agent Reinforcement Learning within Safety-Critical Scenarios

AU - Riley, Joshua

AU - Calinescu, Radu

AU - Paterson, Colin

AU - Kudenko, Daniel

AU - Banks, Alec

PY - 2021

Y1 - 2021

N2 - Multi-agent reinforcement learning allows a team of agents to learn how to work together to solve complex decision-making problems in a shared environment. However, this learning process utilises stochastic mechanisms, meaning that its use in safety-critical domains can be problematic. To overcome this issue, we propose an Assured Multi-Agent Reinforcement Learning (AMARL) approach that uses a model checking technique called quantitative verification to provide formal guarantees of agent compliance with safety, performance, and other non-functional requirements during and after the reinforcement learning process. We demonstrate the applicability of our AMARL approach in three different patrolling navigation domains in which multi-agent systems must learn to visit key areas by using different types of reinforcement learning algorithms (temporal difference learning, game theory, and direct policy search). Furthermore, we compare the effectiveness of these algorithms when used in combination with and without our approach. Our extensive experiments with both homogeneous and heterogeneous multi-agent systems of different sizes show that the use of AMARL leads to safety requirements being consistently satisfied and to better overall results than standard reinforcement learning.

AB - Multi-agent reinforcement learning allows a team of agents to learn how to work together to solve complex decision-making problems in a shared environment. However, this learning process utilises stochastic mechanisms, meaning that its use in safety-critical domains can be problematic. To overcome this issue, we propose an Assured Multi-Agent Reinforcement Learning (AMARL) approach that uses a model checking technique called quantitative verification to provide formal guarantees of agent compliance with safety, performance, and other non-functional requirements during and after the reinforcement learning process. We demonstrate the applicability of our AMARL approach in three different patrolling navigation domains in which multi-agent systems must learn to visit key areas by using different types of reinforcement learning algorithms (temporal difference learning, game theory, and direct policy search). Furthermore, we compare the effectiveness of these algorithms when used in combination with and without our approach. Our extensive experiments with both homogeneous and heterogeneous multi-agent systems of different sizes show that the use of AMARL leads to safety requirements being consistently satisfied and to better overall results than standard reinforcement learning.

KW - Assurance

KW - Assured multi-agent reinforcement learning

KW - Multi-agent reinforcement learning

KW - Multi-agent systems

KW - Quantitative verification

KW - Reinforcement learning

KW - Safe multi-agent reinforcement learning

KW - Safety-critical scenarios

UR - http://www.scopus.com/inward/record.url?scp=85116918514&partnerID=8YFLogxK

U2 - 10.1016/j.procs.2021.08.109

DO - 10.1016/j.procs.2021.08.109

M3 - Conference article

AN - SCOPUS:85116918514

VL - 192

SP - 1061

EP - 1070

JO - Procedia Computer Science

JF - Procedia Computer Science

T2 - 25th KES International Conference on Knowledge-Based and Intelligent Information and Engineering Systems, KES 2021

Y2 - 8 September 2021 through 10 September 2021

ER -