Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

Externe Organisationen

  • Albert-Ludwigs-Universität Freiburg
  • Robert Bosch GmbH
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksECAI 2020 - 24th European Conference on Artificial Intelligence
Herausgeber/-innenGiuseppe De Giacomo, Alejandro Catala, Bistra Dilkina, Michela Milano, Senen Barro, Alberto Bugarin, Jerome Lang
Seiten427-434
Seitenumfang8
ISBN (elektronisch)9781643681009
PublikationsstatusVeröffentlicht - 24 Aug. 2020
Veranstaltung24th European Conference on Artificial Intelligence - ECAI 2020 - Santiago de Compostela, Spanien
Dauer: 29 Aug. 20208 Sept. 2020
http://ecai2020.eu/

Publikationsreihe

NameFrontiers in Artificial Intelligence and Applications
Band325
ISSN (Print)0922-6389
ISSN (elektronisch)1879-8314

Abstract

The performance of many algorithms in the fields of hard combinatorial problem solving, machine learning or AI in general depends on parameter tuning. Automated methods have been proposed to alleviate users from the tedious and error-prone task of manually searching for performance-optimized configurations across a set of problem instances. However, there is still a lot of untapped potential through adjusting an algorithm's parameters online since different parameter values can be optimal at different stages of the algorithm. Prior work showed that reinforcement learning is an effective approach to learn policies for online adjustments of algorithm parameters in a data-driven way. We extend that approach by formulating the resulting dynamic algorithm configuration as a contextual MDP, such that RL not only learns a policy for a single instance, but across a set of instances. To lay the foundation for studying dynamic algorithm configuration with RL in a controlled setting, we propose white-box benchmarks covering major aspects that make dynamic algorithm configuration a hard problem in practice and study the performance of various types of configuration strategies for them. On these white-box benchmarks, we show that (i) RL is a robust candidate for learning configuration policies, outperforming standard parameter optimization approaches, such as classical algorithm configuration; (ii) based on function approximation, RL agents can learn to generalize to new types of instances; and (iii) self-paced learning can substantially improve the performance by selecting a useful sequence of training instances automatically.

ASJC Scopus Sachgebiete

Zitieren

Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. / Biedenkapp, André; Bozkurt, H. Furkan; Eimer, Theresa et al.
ECAI 2020 - 24th European Conference on Artificial Intelligence. Hrsg. / Giuseppe De Giacomo; Alejandro Catala; Bistra Dilkina; Michela Milano; Senen Barro; Alberto Bugarin; Jerome Lang. 2020. S. 427-434 (Frontiers in Artificial Intelligence and Applications; Band 325).

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Biedenkapp, A, Bozkurt, HF, Eimer, T, Hutter, F & Lindauer, MT 2020, Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. in G De Giacomo, A Catala, B Dilkina, M Milano, S Barro, A Bugarin & J Lang (Hrsg.), ECAI 2020 - 24th European Conference on Artificial Intelligence. Frontiers in Artificial Intelligence and Applications, Bd. 325, S. 427-434, 24th European Conference on Artificial Intelligence - ECAI 2020, Santiago de Compostela, Spanien, 29 Aug. 2020. https://doi.org/10.3233/FAIA200122
Biedenkapp, A., Bozkurt, H. F., Eimer, T., Hutter, F., & Lindauer, M. T. (2020). Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. In G. De Giacomo, A. Catala, B. Dilkina, M. Milano, S. Barro, A. Bugarin, & J. Lang (Hrsg.), ECAI 2020 - 24th European Conference on Artificial Intelligence (S. 427-434). (Frontiers in Artificial Intelligence and Applications; Band 325). https://doi.org/10.3233/FAIA200122
Biedenkapp A, Bozkurt HF, Eimer T, Hutter F, Lindauer MT. Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. in De Giacomo G, Catala A, Dilkina B, Milano M, Barro S, Bugarin A, Lang J, Hrsg., ECAI 2020 - 24th European Conference on Artificial Intelligence. 2020. S. 427-434. (Frontiers in Artificial Intelligence and Applications). doi: 10.3233/FAIA200122
Biedenkapp, André ; Bozkurt, H. Furkan ; Eimer, Theresa et al. / Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework. ECAI 2020 - 24th European Conference on Artificial Intelligence. Hrsg. / Giuseppe De Giacomo ; Alejandro Catala ; Bistra Dilkina ; Michela Milano ; Senen Barro ; Alberto Bugarin ; Jerome Lang. 2020. S. 427-434 (Frontiers in Artificial Intelligence and Applications).
Download
@inproceedings{ceeb2d01d5dd4d9eab5f993b2b6b7df8,
title = "Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework",
abstract = "The performance of many algorithms in the fields of hard combinatorial problem solving, machine learning or AI in general depends on parameter tuning. Automated methods have been proposed to alleviate users from the tedious and error-prone task of manually searching for performance-optimized configurations across a set of problem instances. However, there is still a lot of untapped potential through adjusting an algorithm's parameters online since different parameter values can be optimal at different stages of the algorithm. Prior work showed that reinforcement learning is an effective approach to learn policies for online adjustments of algorithm parameters in a data-driven way. We extend that approach by formulating the resulting dynamic algorithm configuration as a contextual MDP, such that RL not only learns a policy for a single instance, but across a set of instances. To lay the foundation for studying dynamic algorithm configuration with RL in a controlled setting, we propose white-box benchmarks covering major aspects that make dynamic algorithm configuration a hard problem in practice and study the performance of various types of configuration strategies for them. On these white-box benchmarks, we show that (i) RL is a robust candidate for learning configuration policies, outperforming standard parameter optimization approaches, such as classical algorithm configuration; (ii) based on function approximation, RL agents can learn to generalize to new types of instances; and (iii) self-paced learning can substantially improve the performance by selecting a useful sequence of training instances automatically.",
author = "Andr{\'e} Biedenkapp and Bozkurt, {H. Furkan} and Theresa Eimer and Frank Hutter and Lindauer, {Marius Thomas}",
note = "Funding information: The authors acknowledge funding by the Robert Bosch GmbH, support by the state of Baden-W{\"u}rttemberg through bwHPC and the German Research Foundation through INST 39/963-1 FUGG.; 24th European Conference on Artificial Intelligence - ECAI 2020 ; Conference date: 29-08-2020 Through 08-09-2020",
year = "2020",
month = aug,
day = "24",
doi = "10.3233/FAIA200122",
language = "English",
isbn = " 978-1-64368-100-9",
series = "Frontiers in Artificial Intelligence and Applications",
pages = "427--434",
editor = "{De Giacomo}, Giuseppe and Alejandro Catala and Bistra Dilkina and Michela Milano and Senen Barro and Alberto Bugarin and Jerome Lang",
booktitle = "ECAI 2020 - 24th European Conference on Artificial Intelligence",
url = "http://ecai2020.eu/",

}

Download

TY - GEN

T1 - Dynamic Algorithm Configuration: Foundation of a New Meta-Algorithmic Framework

AU - Biedenkapp, André

AU - Bozkurt, H. Furkan

AU - Eimer, Theresa

AU - Hutter, Frank

AU - Lindauer, Marius Thomas

N1 - Funding information: The authors acknowledge funding by the Robert Bosch GmbH, support by the state of Baden-Württemberg through bwHPC and the German Research Foundation through INST 39/963-1 FUGG.

PY - 2020/8/24

Y1 - 2020/8/24

N2 - The performance of many algorithms in the fields of hard combinatorial problem solving, machine learning or AI in general depends on parameter tuning. Automated methods have been proposed to alleviate users from the tedious and error-prone task of manually searching for performance-optimized configurations across a set of problem instances. However, there is still a lot of untapped potential through adjusting an algorithm's parameters online since different parameter values can be optimal at different stages of the algorithm. Prior work showed that reinforcement learning is an effective approach to learn policies for online adjustments of algorithm parameters in a data-driven way. We extend that approach by formulating the resulting dynamic algorithm configuration as a contextual MDP, such that RL not only learns a policy for a single instance, but across a set of instances. To lay the foundation for studying dynamic algorithm configuration with RL in a controlled setting, we propose white-box benchmarks covering major aspects that make dynamic algorithm configuration a hard problem in practice and study the performance of various types of configuration strategies for them. On these white-box benchmarks, we show that (i) RL is a robust candidate for learning configuration policies, outperforming standard parameter optimization approaches, such as classical algorithm configuration; (ii) based on function approximation, RL agents can learn to generalize to new types of instances; and (iii) self-paced learning can substantially improve the performance by selecting a useful sequence of training instances automatically.

AB - The performance of many algorithms in the fields of hard combinatorial problem solving, machine learning or AI in general depends on parameter tuning. Automated methods have been proposed to alleviate users from the tedious and error-prone task of manually searching for performance-optimized configurations across a set of problem instances. However, there is still a lot of untapped potential through adjusting an algorithm's parameters online since different parameter values can be optimal at different stages of the algorithm. Prior work showed that reinforcement learning is an effective approach to learn policies for online adjustments of algorithm parameters in a data-driven way. We extend that approach by formulating the resulting dynamic algorithm configuration as a contextual MDP, such that RL not only learns a policy for a single instance, but across a set of instances. To lay the foundation for studying dynamic algorithm configuration with RL in a controlled setting, we propose white-box benchmarks covering major aspects that make dynamic algorithm configuration a hard problem in practice and study the performance of various types of configuration strategies for them. On these white-box benchmarks, we show that (i) RL is a robust candidate for learning configuration policies, outperforming standard parameter optimization approaches, such as classical algorithm configuration; (ii) based on function approximation, RL agents can learn to generalize to new types of instances; and (iii) self-paced learning can substantially improve the performance by selecting a useful sequence of training instances automatically.

UR - http://www.scopus.com/inward/record.url?scp=85091309770&partnerID=8YFLogxK

U2 - 10.3233/FAIA200122

DO - 10.3233/FAIA200122

M3 - Conference contribution

SN - 978-1-64368-100-9

T3 - Frontiers in Artificial Intelligence and Applications

SP - 427

EP - 434

BT - ECAI 2020 - 24th European Conference on Artificial Intelligence

A2 - De Giacomo, Giuseppe

A2 - Catala, Alejandro

A2 - Dilkina, Bistra

A2 - Milano, Michela

A2 - Barro, Senen

A2 - Bugarin, Alberto

A2 - Lang, Jerome

T2 - 24th European Conference on Artificial Intelligence - ECAI 2020

Y2 - 29 August 2020 through 8 September 2020

ER -

Von denselben Autoren