Explaining Hyperparameter Optimization via Partial Dependence Plots

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

  • Julia Moosbauer
  • Julia Herbinger
  • Giuseppe Casalicchio
  • Marius Lindauer
  • Bernd Bischl

Externe Organisationen

  • Ludwig-Maximilians-Universität München (LMU)
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksProceedings of the international conference on Neural Information Processing Systems (NeurIPS)
Seitenumfang21
PublikationsstatusElektronisch veröffentlicht (E-Pub) - 8 Nov. 2021

Abstract

Automated hyperparameter optimization (HPO) can support practitioners to obtain peak performance in machine learning models. However, there is often a lack of valuable insights into the effects of different hyperparameters on the final model performance. This lack of explainability makes it difficult to trust and understand the automated HPO process and its results. We suggest using interpretable machine learning (IML) to gain insights from the experimental data obtained during HPO with Bayesian optimization (BO). BO tends to focus on promising regions with potential high-performance configurations and thus induces a sampling bias. Hence, many IML techniques, such as the partial dependence plot (PDP), carry the risk of generating biased interpretations. By leveraging the posterior uncertainty of the BO surrogate model, we introduce a variant of the PDP with estimated confidence bands. We propose to partition the hyperparameter space to obtain more confident and reliable PDPs in relevant sub-regions. In an experimental study, we provide quantitative evidence for the increased quality of the PDPs within sub-regions.

Zitieren

Explaining Hyperparameter Optimization via Partial Dependence Plots. / Moosbauer, Julia; Herbinger, Julia; Casalicchio, Giuseppe et al.
Proceedings of the international conference on Neural Information Processing Systems (NeurIPS) . 2021.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Moosbauer, J, Herbinger, J, Casalicchio, G, Lindauer, M & Bischl, B 2021, Explaining Hyperparameter Optimization via Partial Dependence Plots. in Proceedings of the international conference on Neural Information Processing Systems (NeurIPS) . <https://arxiv.org/abs/2111.04820>
Moosbauer, J., Herbinger, J., Casalicchio, G., Lindauer, M., & Bischl, B. (2021). Explaining Hyperparameter Optimization via Partial Dependence Plots. In Proceedings of the international conference on Neural Information Processing Systems (NeurIPS) Vorabveröffentlichung online. https://arxiv.org/abs/2111.04820
Moosbauer J, Herbinger J, Casalicchio G, Lindauer M, Bischl B. Explaining Hyperparameter Optimization via Partial Dependence Plots. in Proceedings of the international conference on Neural Information Processing Systems (NeurIPS) . 2021 Epub 2021 Nov 8.
Moosbauer, Julia ; Herbinger, Julia ; Casalicchio, Giuseppe et al. / Explaining Hyperparameter Optimization via Partial Dependence Plots. Proceedings of the international conference on Neural Information Processing Systems (NeurIPS) . 2021.
Download
@inproceedings{273718dcd35c4030a2105718f4c31677,
title = "Explaining Hyperparameter Optimization via Partial Dependence Plots",
abstract = " Automated hyperparameter optimization (HPO) can support practitioners to obtain peak performance in machine learning models. However, there is often a lack of valuable insights into the effects of different hyperparameters on the final model performance. This lack of explainability makes it difficult to trust and understand the automated HPO process and its results. We suggest using interpretable machine learning (IML) to gain insights from the experimental data obtained during HPO with Bayesian optimization (BO). BO tends to focus on promising regions with potential high-performance configurations and thus induces a sampling bias. Hence, many IML techniques, such as the partial dependence plot (PDP), carry the risk of generating biased interpretations. By leveraging the posterior uncertainty of the BO surrogate model, we introduce a variant of the PDP with estimated confidence bands. We propose to partition the hyperparameter space to obtain more confident and reliable PDPs in relevant sub-regions. In an experimental study, we provide quantitative evidence for the increased quality of the PDPs within sub-regions. ",
keywords = "cs.LG, stat.ML",
author = "Julia Moosbauer and Julia Herbinger and Giuseppe Casalicchio and Marius Lindauer and Bernd Bischl",
note = "This work has been partially supported by the German Federal Ministry of Education and Research (BMBF) under Grant No. 01IS18036A. The authors of this work take full responsibilities for its content.",
year = "2021",
month = nov,
day = "8",
language = "English",
booktitle = "Proceedings of the international conference on Neural Information Processing Systems (NeurIPS)",

}

Download

TY - GEN

T1 - Explaining Hyperparameter Optimization via Partial Dependence Plots

AU - Moosbauer, Julia

AU - Herbinger, Julia

AU - Casalicchio, Giuseppe

AU - Lindauer, Marius

AU - Bischl, Bernd

N1 - This work has been partially supported by the German Federal Ministry of Education and Research (BMBF) under Grant No. 01IS18036A. The authors of this work take full responsibilities for its content.

PY - 2021/11/8

Y1 - 2021/11/8

N2 - Automated hyperparameter optimization (HPO) can support practitioners to obtain peak performance in machine learning models. However, there is often a lack of valuable insights into the effects of different hyperparameters on the final model performance. This lack of explainability makes it difficult to trust and understand the automated HPO process and its results. We suggest using interpretable machine learning (IML) to gain insights from the experimental data obtained during HPO with Bayesian optimization (BO). BO tends to focus on promising regions with potential high-performance configurations and thus induces a sampling bias. Hence, many IML techniques, such as the partial dependence plot (PDP), carry the risk of generating biased interpretations. By leveraging the posterior uncertainty of the BO surrogate model, we introduce a variant of the PDP with estimated confidence bands. We propose to partition the hyperparameter space to obtain more confident and reliable PDPs in relevant sub-regions. In an experimental study, we provide quantitative evidence for the increased quality of the PDPs within sub-regions.

AB - Automated hyperparameter optimization (HPO) can support practitioners to obtain peak performance in machine learning models. However, there is often a lack of valuable insights into the effects of different hyperparameters on the final model performance. This lack of explainability makes it difficult to trust and understand the automated HPO process and its results. We suggest using interpretable machine learning (IML) to gain insights from the experimental data obtained during HPO with Bayesian optimization (BO). BO tends to focus on promising regions with potential high-performance configurations and thus induces a sampling bias. Hence, many IML techniques, such as the partial dependence plot (PDP), carry the risk of generating biased interpretations. By leveraging the posterior uncertainty of the BO surrogate model, we introduce a variant of the PDP with estimated confidence bands. We propose to partition the hyperparameter space to obtain more confident and reliable PDPs in relevant sub-regions. In an experimental study, we provide quantitative evidence for the increased quality of the PDPs within sub-regions.

KW - cs.LG

KW - stat.ML

M3 - Conference contribution

BT - Proceedings of the international conference on Neural Information Processing Systems (NeurIPS)

ER -

Von denselben Autoren