Learning Step-Size Adaptation in CMA-ES

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

  • Gresa Shala
  • André Biedenkapp
  • Noor Awad
  • Steven Adriaensen
  • Marius Lindauer
  • Frank Hutter

External Research Organisations

  • University of Freiburg
  • Bosch Center for Artificial Intelligence (BCAI)
View graph of relations

Details

Original languageEnglish
Title of host publicationParallel Problem Solving from Nature – PPSN XVI
Subtitle of host publication16th International Conference, PPSN 2020, Leiden, The Netherlands, September 5-9, 2020, Proceedings, Part I
EditorsThomas Bäck, Mike Preuss, André Deutz, Michael Emmerich, Hao Wang, Carola Doerr, Heike Trautmann
Place of PublicationCham
Pages691-706
Number of pages16
ISBN (electronic)978-3-030-58112-1
Publication statusPublished - 2020
Event16th International Conference on Parallel Problem Solving from Nature, PPSN 2020 - Leiden, Netherlands
Duration: 5 Sept 20209 Sept 2020

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume12269
ISSN (Print)0302-9743
ISSN (electronic)1611-3349

Abstract

An algorithm’s parameter setting often affects its ability to solve a given problem, e.g., population-size, mutation-rate or crossover-rate of an evolutionary algorithm. Furthermore, some parameters have to be adjusted dynamically, such as lowering the mutation-strength over time. While hand-crafted heuristics offer a way to fine-tune and dynamically configure these parameters, their design is tedious, time-consuming and typically involves analyzing the algorithm’s behavior on simple problems that may not be representative for those that arise in practice. In this paper, we show that formulating dynamic algorithm configuration as a reinforcement learning problem allows us to automatically learn policies that can dynamically configure the mutation step-size parameter of Covariance Matrix Adaptation Evolution Strategy (CMA-ES). We evaluate our approach on a wide range of black-box optimization problems, and show that (i) learning step-size policies has the potential to improve the performance of CMA-ES; (ii) learned step-size policies can outperform the default Cumulative Step-Size Adaptation of CMA-ES; and transferring the policies to (iii) different function classes and to (iv) higher dimensions is also possible.

Keywords

    Algorithm configuration, Evolutionary algorithms, Reinforcement learning

ASJC Scopus subject areas

Cite this

Learning Step-Size Adaptation in CMA-ES. / Shala, Gresa; Biedenkapp, André; Awad, Noor et al.
Parallel Problem Solving from Nature – PPSN XVI: 16th International Conference, PPSN 2020, Leiden, The Netherlands, September 5-9, 2020, Proceedings, Part I. ed. / Thomas Bäck; Mike Preuss; André Deutz; Michael Emmerich; Hao Wang; Carola Doerr; Heike Trautmann. Cham, 2020. p. 691-706 (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Vol. 12269).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Shala, G, Biedenkapp, A, Awad, N, Adriaensen, S, Lindauer, M & Hutter, F 2020, Learning Step-Size Adaptation in CMA-ES. in T Bäck, M Preuss, A Deutz, M Emmerich, H Wang, C Doerr & H Trautmann (eds), Parallel Problem Solving from Nature – PPSN XVI: 16th International Conference, PPSN 2020, Leiden, The Netherlands, September 5-9, 2020, Proceedings, Part I. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 12269, Cham, pp. 691-706, 16th International Conference on Parallel Problem Solving from Nature, PPSN 2020, Leiden, Netherlands, 5 Sept 2020. https://doi.org/10.1007/978-3-030-58112-1_48
Shala, G., Biedenkapp, A., Awad, N., Adriaensen, S., Lindauer, M., & Hutter, F. (2020). Learning Step-Size Adaptation in CMA-ES. In T. Bäck, M. Preuss, A. Deutz, M. Emmerich, H. Wang, C. Doerr, & H. Trautmann (Eds.), Parallel Problem Solving from Nature – PPSN XVI: 16th International Conference, PPSN 2020, Leiden, The Netherlands, September 5-9, 2020, Proceedings, Part I (pp. 691-706). (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Vol. 12269).. https://doi.org/10.1007/978-3-030-58112-1_48
Shala G, Biedenkapp A, Awad N, Adriaensen S, Lindauer M, Hutter F. Learning Step-Size Adaptation in CMA-ES. In Bäck T, Preuss M, Deutz A, Emmerich M, Wang H, Doerr C, Trautmann H, editors, Parallel Problem Solving from Nature – PPSN XVI: 16th International Conference, PPSN 2020, Leiden, The Netherlands, September 5-9, 2020, Proceedings, Part I. Cham. 2020. p. 691-706. (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)). Epub 2020 Aug 31. doi: 10.1007/978-3-030-58112-1_48
Shala, Gresa ; Biedenkapp, André ; Awad, Noor et al. / Learning Step-Size Adaptation in CMA-ES. Parallel Problem Solving from Nature – PPSN XVI: 16th International Conference, PPSN 2020, Leiden, The Netherlands, September 5-9, 2020, Proceedings, Part I. editor / Thomas Bäck ; Mike Preuss ; André Deutz ; Michael Emmerich ; Hao Wang ; Carola Doerr ; Heike Trautmann. Cham, 2020. pp. 691-706 (Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)).
Download
@inproceedings{c2ab955cefa549a5a965fe626d5fe316,
title = "Learning Step-Size Adaptation in CMA-ES",
abstract = "An algorithm{\textquoteright}s parameter setting often affects its ability to solve a given problem, e.g., population-size, mutation-rate or crossover-rate of an evolutionary algorithm. Furthermore, some parameters have to be adjusted dynamically, such as lowering the mutation-strength over time. While hand-crafted heuristics offer a way to fine-tune and dynamically configure these parameters, their design is tedious, time-consuming and typically involves analyzing the algorithm{\textquoteright}s behavior on simple problems that may not be representative for those that arise in practice. In this paper, we show that formulating dynamic algorithm configuration as a reinforcement learning problem allows us to automatically learn policies that can dynamically configure the mutation step-size parameter of Covariance Matrix Adaptation Evolution Strategy (CMA-ES). We evaluate our approach on a wide range of black-box optimization problems, and show that (i) learning step-size policies has the potential to improve the performance of CMA-ES; (ii) learned step-size policies can outperform the default Cumulative Step-Size Adaptation of CMA-ES; and transferring the policies to (iii) different function classes and to (iv) higher dimensions is also possible.",
keywords = "Algorithm configuration, Evolutionary algorithms, Reinforcement learning",
author = "Gresa Shala and Andr{\'e} Biedenkapp and Noor Awad and Steven Adriaensen and Marius Lindauer and Frank Hutter",
note = "Funding Information: The authors acknowledge funding by the Robert Bosch GmbH.; 16th International Conference on Parallel Problem Solving from Nature, PPSN 2020 ; Conference date: 05-09-2020 Through 09-09-2020",
year = "2020",
doi = "10.1007/978-3-030-58112-1_48",
language = "English",
isbn = "9783030581114",
series = "Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)",
pages = "691--706",
editor = "Thomas B{\"a}ck and Mike Preuss and Andr{\'e} Deutz and Michael Emmerich and Hao Wang and Carola Doerr and Heike Trautmann",
booktitle = "Parallel Problem Solving from Nature – PPSN XVI",

}

Download

TY - GEN

T1 - Learning Step-Size Adaptation in CMA-ES

AU - Shala, Gresa

AU - Biedenkapp, André

AU - Awad, Noor

AU - Adriaensen, Steven

AU - Lindauer, Marius

AU - Hutter, Frank

N1 - Funding Information: The authors acknowledge funding by the Robert Bosch GmbH.

PY - 2020

Y1 - 2020

N2 - An algorithm’s parameter setting often affects its ability to solve a given problem, e.g., population-size, mutation-rate or crossover-rate of an evolutionary algorithm. Furthermore, some parameters have to be adjusted dynamically, such as lowering the mutation-strength over time. While hand-crafted heuristics offer a way to fine-tune and dynamically configure these parameters, their design is tedious, time-consuming and typically involves analyzing the algorithm’s behavior on simple problems that may not be representative for those that arise in practice. In this paper, we show that formulating dynamic algorithm configuration as a reinforcement learning problem allows us to automatically learn policies that can dynamically configure the mutation step-size parameter of Covariance Matrix Adaptation Evolution Strategy (CMA-ES). We evaluate our approach on a wide range of black-box optimization problems, and show that (i) learning step-size policies has the potential to improve the performance of CMA-ES; (ii) learned step-size policies can outperform the default Cumulative Step-Size Adaptation of CMA-ES; and transferring the policies to (iii) different function classes and to (iv) higher dimensions is also possible.

AB - An algorithm’s parameter setting often affects its ability to solve a given problem, e.g., population-size, mutation-rate or crossover-rate of an evolutionary algorithm. Furthermore, some parameters have to be adjusted dynamically, such as lowering the mutation-strength over time. While hand-crafted heuristics offer a way to fine-tune and dynamically configure these parameters, their design is tedious, time-consuming and typically involves analyzing the algorithm’s behavior on simple problems that may not be representative for those that arise in practice. In this paper, we show that formulating dynamic algorithm configuration as a reinforcement learning problem allows us to automatically learn policies that can dynamically configure the mutation step-size parameter of Covariance Matrix Adaptation Evolution Strategy (CMA-ES). We evaluate our approach on a wide range of black-box optimization problems, and show that (i) learning step-size policies has the potential to improve the performance of CMA-ES; (ii) learned step-size policies can outperform the default Cumulative Step-Size Adaptation of CMA-ES; and transferring the policies to (iii) different function classes and to (iv) higher dimensions is also possible.

KW - Algorithm configuration

KW - Evolutionary algorithms

KW - Reinforcement learning

UR - http://www.scopus.com/inward/record.url?scp=85091295368&partnerID=8YFLogxK

U2 - 10.1007/978-3-030-58112-1_48

DO - 10.1007/978-3-030-58112-1_48

M3 - Conference contribution

AN - SCOPUS:85091295368

SN - 9783030581114

T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

SP - 691

EP - 706

BT - Parallel Problem Solving from Nature – PPSN XVI

A2 - Bäck, Thomas

A2 - Preuss, Mike

A2 - Deutz, André

A2 - Emmerich, Michael

A2 - Wang, Hao

A2 - Doerr, Carola

A2 - Trautmann, Heike

CY - Cham

T2 - 16th International Conference on Parallel Problem Solving from Nature, PPSN 2020

Y2 - 5 September 2020 through 9 September 2020

ER -

By the same author(s)