Details
Originalsprache | Englisch |
---|---|
Titel des Sammelwerks | TrustNL |
Untertitel | First Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop |
Herausgeber/-innen | Yada Pruksachatkun, Anil Ramakrishna, Kai-Wei Chang, Satyapriya Krishna, Jwala Dhamala, Tanaya Guha, Xiang Ren |
Seiten | 68-73 |
Seitenumfang | 6 |
ISBN (elektronisch) | 9781954085336 |
Publikationsstatus | Veröffentlicht - 2021 |
Veranstaltung | 1st Workshop on Trustworthy Natural Language Processing, TrustNLP 2021 - Virtual, Online Dauer: 10 Juni 2021 → 5 Juli 2021 |
Abstract
Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the need for a benchmark to facilitate evaluations of the utility of post-hoc explanation methods. As a first step to this end, we enumerate desirable properties that such a benchmark should possess for the task of debugging text classifiers. Additionally, we highlight that such a benchmark facilitates not only assessing the effectiveness of explanations but also their efficiency.
ASJC Scopus Sachgebiete
- Informatik (insg.)
- Software
- Geisteswissenschaftliche Fächer (insg.)
- Sprache und Linguistik
- Informatik (insg.)
- Theoretische Informatik und Mathematik
- Sozialwissenschaften (insg.)
- Linguistik und Sprache
Zitieren
- Standard
- Harvard
- Apa
- Vancouver
- BibTex
- RIS
TrustNL: First Workshop on Trustworthy Natural Language Processing, Proceedings of the Workshop. Hrsg. / Yada Pruksachatkun; Anil Ramakrishna; Kai-Wei Chang; Satyapriya Krishna; Jwala Dhamala; Tanaya Guha; Xiang Ren. 2021. S. 68-73.
Publikation: Beitrag in Buch/Bericht/Sammelwerk/Konferenzband › Aufsatz in Konferenzband › Forschung › Peer-Review
}
TY - GEN
T1 - Towards Benchmarking the Utility of Explanations for Model Debugging
AU - Idahl, Maximilian
AU - Lyu, Lijun
AU - Gadiraju, Ujwal
AU - Anand, Avishek
PY - 2021
Y1 - 2021
N2 - Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the need for a benchmark to facilitate evaluations of the utility of post-hoc explanation methods. As a first step to this end, we enumerate desirable properties that such a benchmark should possess for the task of debugging text classifiers. Additionally, we highlight that such a benchmark facilitates not only assessing the effectiveness of explanations but also their efficiency.
AB - Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the need for a benchmark to facilitate evaluations of the utility of post-hoc explanation methods. As a first step to this end, we enumerate desirable properties that such a benchmark should possess for the task of debugging text classifiers. Additionally, we highlight that such a benchmark facilitates not only assessing the effectiveness of explanations but also their efficiency.
UR - http://www.scopus.com/inward/record.url?scp=85119843243&partnerID=8YFLogxK
U2 - 10.48550/arXiv.2105.04505
DO - 10.48550/arXiv.2105.04505
M3 - Conference contribution
AN - SCOPUS:85119843243
SP - 68
EP - 73
BT - TrustNL
A2 - Pruksachatkun, Yada
A2 - Ramakrishna, Anil
A2 - Chang, Kai-Wei
A2 - Krishna, Satyapriya
A2 - Dhamala, Jwala
A2 - Guha, Tanaya
A2 - Ren, Xiang
T2 - 1st Workshop on Trustworthy Natural Language Processing, TrustNLP 2021
Y2 - 10 June 2021 through 5 July 2021
ER -