Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework

Research output: Chapter in book/report/conference proceedingContribution to book/anthologyResearch

Authors

External Research Organisations

  • University of Freiburg
  • Robert Bosch GmbH
View graph of relations

Details

Original languageEnglish
Title of host publicationECAI 2020 - 24th European Conference on Artificial Intelligence
EditorsGiuseppe De Giacomo, Alejandro Catala, Bistra Dilkina, Michela Milano, Senen Barro, Alberto Bugarin, Jerome Lang
Pages427-434
Number of pages8
ISBN (electronic)9781643681009
Publication statusPublished - 24 Aug 2020
Event24th European Conference on Artificial Intelligence - ECAI 2020 - Santiago de Compostela, Spain
Duration: 29 Aug 20208 Sept 2020
http://ecai2020.eu/

Publication series

NameFrontiers in Artificial Intelligence and Applications
Volume325
ISSN (Print)0922-6389
ISSN (electronic)1879-8314

Abstract

The performance of many algorithms in the fields of hard combinatorial problem solving, machine learning or AI in general depends on parameter tuning. Automated methods have been proposed to alleviate users from the tedious and error-prone task of manually searching for performance-optimized configurations across a set of problem instances. However, there is still a lot of untapped potential through adjusting an algorithm's parameters online since different parameter values can be optimal at different stages of the algorithm. Prior work showed that reinforcement learning is an effective approach to learn policies for online adjustments of algorithm parameters in a data-driven way. We extend that approach by formulating the resulting dynamic algorithm configuration as a contextual MDP, such that RL not only learns a policy for a single instance, but across a set of instances. To lay the foundation for studying dynamic algorithm configuration with RL in a controlled setting, we propose white-box benchmarks covering major aspects that make dynamic algorithm configuration a hard problem in practice and study the performance of various types of configuration strategies for them. On these white-box benchmarks, we show that (i) RL is a robust candidate for learning configuration policies, outperforming standard parameter optimization approaches, such as classical algorithm configuration; (ii) based on function approximation, RL agents can learn to generalize to new types of instances; and (iii) self-paced learning can substantially improve the performance by selecting a useful sequence of training instances automatically.

ASJC Scopus subject areas

Cite this

Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. / Biedenkapp, André; Bozkurt, H. Furkan; Eimer, Theresa et al.
ECAI 2020 - 24th European Conference on Artificial Intelligence. ed. / Giuseppe De Giacomo; Alejandro Catala; Bistra Dilkina; Michela Milano; Senen Barro; Alberto Bugarin; Jerome Lang. 2020. p. 427-434 (Frontiers in Artificial Intelligence and Applications; Vol. 325).

Research output: Chapter in book/report/conference proceedingContribution to book/anthologyResearch

Biedenkapp, A, Bozkurt, HF, Eimer, T, Hutter, F & Lindauer, MT 2020, Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. in G De Giacomo, A Catala, B Dilkina, M Milano, S Barro, A Bugarin & J Lang (eds), ECAI 2020 - 24th European Conference on Artificial Intelligence. Frontiers in Artificial Intelligence and Applications, vol. 325, pp. 427-434, 24th European Conference on Artificial Intelligence - ECAI 2020, Santiago de Compostela, Spain, 29 Aug 2020. https://doi.org/10.3233/FAIA200122
Biedenkapp, A., Bozkurt, H. F., Eimer, T., Hutter, F., & Lindauer, M. T. (2020). Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. In G. De Giacomo, A. Catala, B. Dilkina, M. Milano, S. Barro, A. Bugarin, & J. Lang (Eds.), ECAI 2020 - 24th European Conference on Artificial Intelligence (pp. 427-434). (Frontiers in Artificial Intelligence and Applications; Vol. 325). https://doi.org/10.3233/FAIA200122
Biedenkapp A, Bozkurt HF, Eimer T, Hutter F, Lindauer MT. Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. In De Giacomo G, Catala A, Dilkina B, Milano M, Barro S, Bugarin A, Lang J, editors, ECAI 2020 - 24th European Conference on Artificial Intelligence. 2020. p. 427-434. (Frontiers in Artificial Intelligence and Applications). doi: 10.3233/FAIA200122
Biedenkapp, André ; Bozkurt, H. Furkan ; Eimer, Theresa et al. / Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. ECAI 2020 - 24th European Conference on Artificial Intelligence. editor / Giuseppe De Giacomo ; Alejandro Catala ; Bistra Dilkina ; Michela Milano ; Senen Barro ; Alberto Bugarin ; Jerome Lang. 2020. pp. 427-434 (Frontiers in Artificial Intelligence and Applications).
Download
@inbook{ceeb2d01d5dd4d9eab5f993b2b6b7df8,
title = "Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework",
abstract = "The performance of many algorithms in the fields of hard combinatorial problem solving, machine learning or AI in general depends on parameter tuning. Automated methods have been proposed to alleviate users from the tedious and error-prone task of manually searching for performance-optimized configurations across a set of problem instances. However, there is still a lot of untapped potential through adjusting an algorithm's parameters online since different parameter values can be optimal at different stages of the algorithm. Prior work showed that reinforcement learning is an effective approach to learn policies for online adjustments of algorithm parameters in a data-driven way. We extend that approach by formulating the resulting dynamic algorithm configuration as a contextual MDP, such that RL not only learns a policy for a single instance, but across a set of instances. To lay the foundation for studying dynamic algorithm configuration with RL in a controlled setting, we propose white-box benchmarks covering major aspects that make dynamic algorithm configuration a hard problem in practice and study the performance of various types of configuration strategies for them. On these white-box benchmarks, we show that (i) RL is a robust candidate for learning configuration policies, outperforming standard parameter optimization approaches, such as classical algorithm configuration; (ii) based on function approximation, RL agents can learn to generalize to new types of instances; and (iii) self-paced learning can substantially improve the performance by selecting a useful sequence of training instances automatically.",
author = "Andr{\'e} Biedenkapp and Bozkurt, {H. Furkan} and Theresa Eimer and Frank Hutter and Lindauer, {Marius Thomas}",
note = "Funding information: The authors acknowledge funding by the Robert Bosch GmbH, support by the state of Baden-W{\"u}rttemberg through bwHPC and the German Research Foundation through INST 39/963-1 FUGG.; 24th European Conference on Artificial Intelligence - ECAI 2020 ; Conference date: 29-08-2020 Through 08-09-2020",
year = "2020",
month = aug,
day = "24",
doi = "10.3233/FAIA200122",
language = "English",
isbn = " 978-1-64368-100-9",
series = "Frontiers in Artificial Intelligence and Applications",
pages = "427--434",
editor = "{De Giacomo}, Giuseppe and Alejandro Catala and Bistra Dilkina and Michela Milano and Senen Barro and Alberto Bugarin and Jerome Lang",
booktitle = "ECAI 2020 - 24th European Conference on Artificial Intelligence",
url = "http://ecai2020.eu/",

}

Download

TY - CHAP

T1 - Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework

AU - Biedenkapp, André

AU - Bozkurt, H. Furkan

AU - Eimer, Theresa

AU - Hutter, Frank

AU - Lindauer, Marius Thomas

N1 - Funding information: The authors acknowledge funding by the Robert Bosch GmbH, support by the state of Baden-Württemberg through bwHPC and the German Research Foundation through INST 39/963-1 FUGG.

PY - 2020/8/24

Y1 - 2020/8/24

N2 - The performance of many algorithms in the fields of hard combinatorial problem solving, machine learning or AI in general depends on parameter tuning. Automated methods have been proposed to alleviate users from the tedious and error-prone task of manually searching for performance-optimized configurations across a set of problem instances. However, there is still a lot of untapped potential through adjusting an algorithm's parameters online since different parameter values can be optimal at different stages of the algorithm. Prior work showed that reinforcement learning is an effective approach to learn policies for online adjustments of algorithm parameters in a data-driven way. We extend that approach by formulating the resulting dynamic algorithm configuration as a contextual MDP, such that RL not only learns a policy for a single instance, but across a set of instances. To lay the foundation for studying dynamic algorithm configuration with RL in a controlled setting, we propose white-box benchmarks covering major aspects that make dynamic algorithm configuration a hard problem in practice and study the performance of various types of configuration strategies for them. On these white-box benchmarks, we show that (i) RL is a robust candidate for learning configuration policies, outperforming standard parameter optimization approaches, such as classical algorithm configuration; (ii) based on function approximation, RL agents can learn to generalize to new types of instances; and (iii) self-paced learning can substantially improve the performance by selecting a useful sequence of training instances automatically.

AB - The performance of many algorithms in the fields of hard combinatorial problem solving, machine learning or AI in general depends on parameter tuning. Automated methods have been proposed to alleviate users from the tedious and error-prone task of manually searching for performance-optimized configurations across a set of problem instances. However, there is still a lot of untapped potential through adjusting an algorithm's parameters online since different parameter values can be optimal at different stages of the algorithm. Prior work showed that reinforcement learning is an effective approach to learn policies for online adjustments of algorithm parameters in a data-driven way. We extend that approach by formulating the resulting dynamic algorithm configuration as a contextual MDP, such that RL not only learns a policy for a single instance, but across a set of instances. To lay the foundation for studying dynamic algorithm configuration with RL in a controlled setting, we propose white-box benchmarks covering major aspects that make dynamic algorithm configuration a hard problem in practice and study the performance of various types of configuration strategies for them. On these white-box benchmarks, we show that (i) RL is a robust candidate for learning configuration policies, outperforming standard parameter optimization approaches, such as classical algorithm configuration; (ii) based on function approximation, RL agents can learn to generalize to new types of instances; and (iii) self-paced learning can substantially improve the performance by selecting a useful sequence of training instances automatically.

UR - http://www.scopus.com/inward/record.url?scp=85091309770&partnerID=8YFLogxK

U2 - 10.3233/FAIA200122

DO - 10.3233/FAIA200122

M3 - Contribution to book/anthology

SN - 978-1-64368-100-9

T3 - Frontiers in Artificial Intelligence and Applications

SP - 427

EP - 434

BT - ECAI 2020 - 24th European Conference on Artificial Intelligence

A2 - De Giacomo, Giuseppe

A2 - Catala, Alejandro

A2 - Dilkina, Bistra

A2 - Milano, Michela

A2 - Barro, Senen

A2 - Bugarin, Alberto

A2 - Lang, Jerome

T2 - 24th European Conference on Artificial Intelligence - ECAI 2020

Y2 - 29 August 2020 through 8 September 2020

ER -

By the same author(s)