Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

  • Joshua Riley
  • Radu Calinescu
  • Colin Paterson
  • Daniel Kudenko
  • Alec Banks

Research Organisations

External Research Organisations

  • Univ. York, Dep. Comput. Sci., Non-Stand. Comput. Group
  • Defence Science and Technology Laboratory
View graph of relations

Details

Original languageEnglish
Title of host publicationProceedings of the 13th International Conference on Agents and Artificial Intelligence
Subtitle of host publicationVolume 2: ICAART
EditorsAna Paula Rocha, Luc Steels, Jaap van den Herik
Pages237-245
Number of pages9
ISBN (electronic)9789897584848
Publication statusPublished - 2021
Event13th International Conference on Agents and Artificial Intelligence, ICAART 2021 - Virtual, Online, Austria
Duration: 4 Feb 20216 Feb 2021

Publication series

NameICAART
ISSN (electronic)2184-433X

Abstract

In multi-agent reinforcement learning, several agents converge together towards optimal policies that solve complex decision-making problems. This convergence process is inherently stochastic, meaning that its use in safety-critical domains can be problematic. To address this issue, we introduce a new approach that combines multi-agent reinforcement learning with a formal verification technique termed quantitative verification. Our assured multi-agent reinforcement learning approach constrains agent behaviours in ways that ensure the satisfaction of requirements associated with the safety, reliability, and other non-functional aspects of the decision-making problem being solved. The approach comprises three stages. First, it models the problem as an abstract Markov decision process, allowing quantitative verification to be applied. Next, this abstract model is used to synthesise a policy which satisfies safety, reliability, and performance constraints. Finally, the synthesised policy is used to constrain agent behaviour within the low-level problem with a greatly lowered risk of constraint violations. We demonstrate our approach using a safety-critical multi-agent patrolling problem.

Keywords

    Assurance, Multi-agent reinforcement learning, Multi-agent system, Quantitative verification, Reinforcement learning

ASJC Scopus subject areas

Cite this

Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies. / Riley, Joshua; Calinescu, Radu; Paterson, Colin et al.
Proceedings of the 13th International Conference on Agents and Artificial Intelligence: Volume 2: ICAART . ed. / Ana Paula Rocha; Luc Steels; Jaap van den Herik. 2021. p. 237-245 (ICAART).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Riley, J, Calinescu, R, Paterson, C, Kudenko, D & Banks, A 2021, Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies. in AP Rocha, L Steels & J van den Herik (eds), Proceedings of the 13th International Conference on Agents and Artificial Intelligence: Volume 2: ICAART . ICAART, pp. 237-245, 13th International Conference on Agents and Artificial Intelligence, ICAART 2021, Online, Austria, 4 Feb 2021. https://doi.org/10.5220/0010258102370245
Riley, J., Calinescu, R., Paterson, C., Kudenko, D., & Banks, A. (2021). Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies. In A. P. Rocha, L. Steels, & J. van den Herik (Eds.), Proceedings of the 13th International Conference on Agents and Artificial Intelligence: Volume 2: ICAART (pp. 237-245). (ICAART). https://doi.org/10.5220/0010258102370245
Riley J, Calinescu R, Paterson C, Kudenko D, Banks A. Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies. In Rocha AP, Steels L, van den Herik J, editors, Proceedings of the 13th International Conference on Agents and Artificial Intelligence: Volume 2: ICAART . 2021. p. 237-245. (ICAART). doi: 10.5220/0010258102370245
Riley, Joshua ; Calinescu, Radu ; Paterson, Colin et al. / Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies. Proceedings of the 13th International Conference on Agents and Artificial Intelligence: Volume 2: ICAART . editor / Ana Paula Rocha ; Luc Steels ; Jaap van den Herik. 2021. pp. 237-245 (ICAART).
Download
@inproceedings{e755ab7bd0bb47ac846f9d0fbb1e693e,
title = "Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies",
abstract = "In multi-agent reinforcement learning, several agents converge together towards optimal policies that solve complex decision-making problems. This convergence process is inherently stochastic, meaning that its use in safety-critical domains can be problematic. To address this issue, we introduce a new approach that combines multi-agent reinforcement learning with a formal verification technique termed quantitative verification. Our assured multi-agent reinforcement learning approach constrains agent behaviours in ways that ensure the satisfaction of requirements associated with the safety, reliability, and other non-functional aspects of the decision-making problem being solved. The approach comprises three stages. First, it models the problem as an abstract Markov decision process, allowing quantitative verification to be applied. Next, this abstract model is used to synthesise a policy which satisfies safety, reliability, and performance constraints. Finally, the synthesised policy is used to constrain agent behaviour within the low-level problem with a greatly lowered risk of constraint violations. We demonstrate our approach using a safety-critical multi-agent patrolling problem.",
keywords = "Assurance, Multi-agent reinforcement learning, Multi-agent system, Quantitative verification, Reinforcement learning",
author = "Joshua Riley and Radu Calinescu and Colin Paterson and Daniel Kudenko and Alec Banks",
note = "Funding Information: This paper presents research sponsored by the UK MOD. The information contained in it should not be interpreted as representing the views of the UK MOD, nor should it be assumed it reflects any current or future UK MOD policy.; 13th International Conference on Agents and Artificial Intelligence, ICAART 2021 ; Conference date: 04-02-2021 Through 06-02-2021",
year = "2021",
doi = "10.5220/0010258102370245",
language = "English",
series = "ICAART",
pages = "237--245",
editor = "Rocha, {Ana Paula} and Luc Steels and {van den Herik}, Jaap",
booktitle = "Proceedings of the 13th International Conference on Agents and Artificial Intelligence",

}

Download

TY - GEN

T1 - Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies

AU - Riley, Joshua

AU - Calinescu, Radu

AU - Paterson, Colin

AU - Kudenko, Daniel

AU - Banks, Alec

N1 - Funding Information: This paper presents research sponsored by the UK MOD. The information contained in it should not be interpreted as representing the views of the UK MOD, nor should it be assumed it reflects any current or future UK MOD policy.

PY - 2021

Y1 - 2021

N2 - In multi-agent reinforcement learning, several agents converge together towards optimal policies that solve complex decision-making problems. This convergence process is inherently stochastic, meaning that its use in safety-critical domains can be problematic. To address this issue, we introduce a new approach that combines multi-agent reinforcement learning with a formal verification technique termed quantitative verification. Our assured multi-agent reinforcement learning approach constrains agent behaviours in ways that ensure the satisfaction of requirements associated with the safety, reliability, and other non-functional aspects of the decision-making problem being solved. The approach comprises three stages. First, it models the problem as an abstract Markov decision process, allowing quantitative verification to be applied. Next, this abstract model is used to synthesise a policy which satisfies safety, reliability, and performance constraints. Finally, the synthesised policy is used to constrain agent behaviour within the low-level problem with a greatly lowered risk of constraint violations. We demonstrate our approach using a safety-critical multi-agent patrolling problem.

AB - In multi-agent reinforcement learning, several agents converge together towards optimal policies that solve complex decision-making problems. This convergence process is inherently stochastic, meaning that its use in safety-critical domains can be problematic. To address this issue, we introduce a new approach that combines multi-agent reinforcement learning with a formal verification technique termed quantitative verification. Our assured multi-agent reinforcement learning approach constrains agent behaviours in ways that ensure the satisfaction of requirements associated with the safety, reliability, and other non-functional aspects of the decision-making problem being solved. The approach comprises three stages. First, it models the problem as an abstract Markov decision process, allowing quantitative verification to be applied. Next, this abstract model is used to synthesise a policy which satisfies safety, reliability, and performance constraints. Finally, the synthesised policy is used to constrain agent behaviour within the low-level problem with a greatly lowered risk of constraint violations. We demonstrate our approach using a safety-critical multi-agent patrolling problem.

KW - Assurance

KW - Multi-agent reinforcement learning

KW - Multi-agent system

KW - Quantitative verification

KW - Reinforcement learning

UR - http://www.scopus.com/inward/record.url?scp=85103859598&partnerID=8YFLogxK

U2 - 10.5220/0010258102370245

DO - 10.5220/0010258102370245

M3 - Conference contribution

AN - SCOPUS:85103859598

T3 - ICAART

SP - 237

EP - 245

BT - Proceedings of the 13th International Conference on Agents and Artificial Intelligence

A2 - Rocha, Ana Paula

A2 - Steels, Luc

A2 - van den Herik, Jaap

T2 - 13th International Conference on Agents and Artificial Intelligence, ICAART 2021

Y2 - 4 February 2021 through 6 February 2021

ER -