Towards Benchmarking the Utility of Explanations for Model Debugging

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autorschaft

  • Maximilian Idahl
  • Lijun Lyu
  • Ujwal Gadiraju
  • Avishek Anand

Organisationseinheiten

Externe Organisationen

  • Delft University of Technology
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksTrustNL
UntertitelFirst Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop
Herausgeber/-innenYada Pruksachatkun, Anil Ramakrishna, Kai-Wei Chang, Satyapriya Krishna, Jwala Dhamala, Tanaya Guha, Xiang Ren
Seiten68-73
Seitenumfang6
ISBN (elektronisch)9781954085336
PublikationsstatusVeröffentlicht - 2021
Veranstaltung1st Workshop on Trustworthy Natural Language Processing, TrustNLP 2021 - Virtual, Online
Dauer: 10 Juni 20215 Juli 2021

Abstract

Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the need for a benchmark to facilitate evaluations of the utility of post-hoc explanation methods. As a first step to this end, we enumerate desirable properties that such a benchmark should possess for the task of debugging text classifiers. Additionally, we highlight that such a benchmark facilitates not only assessing the effectiveness of explanations but also their efficiency.

ASJC Scopus Sachgebiete

Zitieren

Towards Benchmarking the Utility of Explanations for Model Debugging. / Idahl, Maximilian; Lyu, Lijun; Gadiraju, Ujwal et al.
TrustNL: First Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop. Hrsg. / Yada Pruksachatkun; Anil Ramakrishna; Kai-Wei Chang; Satyapriya Krishna; Jwala Dhamala; Tanaya Guha; Xiang Ren. 2021. S. 68-73.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Idahl, M, Lyu, L, Gadiraju, U & Anand, A 2021, Towards Benchmarking the Utility of Explanations for Model Debugging. in Y Pruksachatkun, A Ramakrishna, K-W Chang, S Krishna, J Dhamala, T Guha & X Ren (Hrsg.), TrustNL: First Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop. S. 68-73, 1st Workshop on Trustworthy Natural Language Processing, TrustNLP 2021, Virtual, Online, 10 Juni 2021. https://doi.org/10.48550/arXiv.2105.04505, https://doi.org/10.18653/v1/2021.trustnlp-1.8
Idahl, M., Lyu, L., Gadiraju, U., & Anand, A. (2021). Towards Benchmarking the Utility of Explanations for Model Debugging. In Y. Pruksachatkun, A. Ramakrishna, K.-W. Chang, S. Krishna, J. Dhamala, T. Guha, & X. Ren (Hrsg.), TrustNL: First Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop (S. 68-73) https://doi.org/10.48550/arXiv.2105.04505, https://doi.org/10.18653/v1/2021.trustnlp-1.8
Idahl M, Lyu L, Gadiraju U, Anand A. Towards Benchmarking the Utility of Explanations for Model Debugging. in Pruksachatkun Y, Ramakrishna A, Chang KW, Krishna S, Dhamala J, Guha T, Ren X, Hrsg., TrustNL: First Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop. 2021. S. 68-73 doi: 10.48550/arXiv.2105.04505, 10.18653/v1/2021.trustnlp-1.8
Idahl, Maximilian ; Lyu, Lijun ; Gadiraju, Ujwal et al. / Towards Benchmarking the Utility of Explanations for Model Debugging. TrustNL: First Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop. Hrsg. / Yada Pruksachatkun ; Anil Ramakrishna ; Kai-Wei Chang ; Satyapriya Krishna ; Jwala Dhamala ; Tanaya Guha ; Xiang Ren. 2021. S. 68-73
Download
@inproceedings{b406c2a31a534a158f72923de3e90c90,
title = "Towards Benchmarking the Utility of Explanations for Model Debugging",
abstract = "Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the need for a benchmark to facilitate evaluations of the utility of post-hoc explanation methods. As a first step to this end, we enumerate desirable properties that such a benchmark should possess for the task of debugging text classifiers. Additionally, we highlight that such a benchmark facilitates not only assessing the effectiveness of explanations but also their efficiency.",
author = "Maximilian Idahl and Lijun Lyu and Ujwal Gadiraju and Avishek Anand",
year = "2021",
doi = "10.48550/arXiv.2105.04505",
language = "English",
pages = "68--73",
editor = "Yada Pruksachatkun and Anil Ramakrishna and Kai-Wei Chang and Satyapriya Krishna and Jwala Dhamala and Tanaya Guha and Xiang Ren",
booktitle = "TrustNL",
note = "1st Workshop on Trustworthy Natural Language Processing, TrustNLP 2021 ; Conference date: 10-06-2021 Through 05-07-2021",

}

Download

TY - GEN

T1 - Towards Benchmarking the Utility of Explanations for Model Debugging

AU - Idahl, Maximilian

AU - Lyu, Lijun

AU - Gadiraju, Ujwal

AU - Anand, Avishek

PY - 2021

Y1 - 2021

N2 - Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the need for a benchmark to facilitate evaluations of the utility of post-hoc explanation methods. As a first step to this end, we enumerate desirable properties that such a benchmark should possess for the task of debugging text classifiers. Additionally, we highlight that such a benchmark facilitates not only assessing the effectiveness of explanations but also their efficiency.

AB - Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the need for a benchmark to facilitate evaluations of the utility of post-hoc explanation methods. As a first step to this end, we enumerate desirable properties that such a benchmark should possess for the task of debugging text classifiers. Additionally, we highlight that such a benchmark facilitates not only assessing the effectiveness of explanations but also their efficiency.

UR - http://www.scopus.com/inward/record.url?scp=85119843243&partnerID=8YFLogxK

U2 - 10.48550/arXiv.2105.04505

DO - 10.48550/arXiv.2105.04505

M3 - Conference contribution

AN - SCOPUS:85119843243

SP - 68

EP - 73

BT - TrustNL

A2 - Pruksachatkun, Yada

A2 - Ramakrishna, Anil

A2 - Chang, Kai-Wei

A2 - Krishna, Satyapriya

A2 - Dhamala, Jwala

A2 - Guha, Tanaya

A2 - Ren, Xiang

T2 - 1st Workshop on Trustworthy Natural Language Processing, TrustNLP 2021

Y2 - 10 June 2021 through 5 July 2021

ER -