Details
Originalsprache | Englisch |
---|---|
Titel des Sammelwerks | SIGIR 2019 |
Untertitel | Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval |
Herausgeber (Verlag) | Association for Computing Machinery (ACM) |
Seiten | 1005-1008 |
Seitenumfang | 4 |
ISBN (elektronisch) | 9781450361729 |
Publikationsstatus | Veröffentlicht - 18 Juli 2019 |
Veranstaltung | 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2019 - Paris, Frankreich Dauer: 21 Juli 2019 → 25 Juli 2019 |
Abstract
A recent trend in IR has been the usage of neural networks to learn retrieval models for text based adhoc search. While various approaches and architectures have yielded significantly better performance than traditional retrieval models such as BM25, it is still difficult to understand exactly why a document is relevant to a query. In the ML community several approaches for explaining decisions made by deep neural networks have been proposed - including DeepSHAP which modifies the DeepLift algorithm to estimate the relative importance (shapley values) of input features for a given decision by comparing the activations in the network for a given image against the activations caused by a reference input. In image classification, the reference input tends to be a plain black image. While DeepSHAP has been well studied for image classification tasks, it remains to be seen how we can adapt it to explain the output of Neural Retrieval Models (NRMs). In particular, what is a good “black” image in the context of IR? In this paper we explored various reference input document construction techniques. Additionally, we compared the explanations generated by DeepSHAP to LIME (a model agnostic approach) and found that the explanations differ considerably. Our study raises concerns regarding the robustness and accuracy of explanations produced for NRMs. With this paper we aim to shed light on interesting problems surrounding interpretability in NRMs and highlight areas of future work.
ASJC Scopus Sachgebiete
- Informatik (insg.)
- Information systems
- Mathematik (insg.)
- Angewandte Mathematik
- Informatik (insg.)
- Software
Zitieren
- Standard
- Harvard
- Apa
- Vancouver
- BibTex
- RIS
SIGIR 2019: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval. Association for Computing Machinery (ACM), 2019. S. 1005-1008.
Publikation: Beitrag in Buch/Bericht/Sammelwerk/Konferenzband › Aufsatz in Konferenzband › Forschung › Peer-Review
}
TY - GEN
T1 - A study on the Interpretability of Neural Retrieval Models using DeepSHAP
AU - Fernando, Zeon Trevor
AU - Singh, Jaspreet
AU - Anand, Avishek
N1 - Funding Information: This work was supported by the Amazon research award on ‘Interpretability of Neural Rankers’.
PY - 2019/7/18
Y1 - 2019/7/18
N2 - A recent trend in IR has been the usage of neural networks to learn retrieval models for text based adhoc search. While various approaches and architectures have yielded significantly better performance than traditional retrieval models such as BM25, it is still difficult to understand exactly why a document is relevant to a query. In the ML community several approaches for explaining decisions made by deep neural networks have been proposed - including DeepSHAP which modifies the DeepLift algorithm to estimate the relative importance (shapley values) of input features for a given decision by comparing the activations in the network for a given image against the activations caused by a reference input. In image classification, the reference input tends to be a plain black image. While DeepSHAP has been well studied for image classification tasks, it remains to be seen how we can adapt it to explain the output of Neural Retrieval Models (NRMs). In particular, what is a good “black” image in the context of IR? In this paper we explored various reference input document construction techniques. Additionally, we compared the explanations generated by DeepSHAP to LIME (a model agnostic approach) and found that the explanations differ considerably. Our study raises concerns regarding the robustness and accuracy of explanations produced for NRMs. With this paper we aim to shed light on interesting problems surrounding interpretability in NRMs and highlight areas of future work.
AB - A recent trend in IR has been the usage of neural networks to learn retrieval models for text based adhoc search. While various approaches and architectures have yielded significantly better performance than traditional retrieval models such as BM25, it is still difficult to understand exactly why a document is relevant to a query. In the ML community several approaches for explaining decisions made by deep neural networks have been proposed - including DeepSHAP which modifies the DeepLift algorithm to estimate the relative importance (shapley values) of input features for a given decision by comparing the activations in the network for a given image against the activations caused by a reference input. In image classification, the reference input tends to be a plain black image. While DeepSHAP has been well studied for image classification tasks, it remains to be seen how we can adapt it to explain the output of Neural Retrieval Models (NRMs). In particular, what is a good “black” image in the context of IR? In this paper we explored various reference input document construction techniques. Additionally, we compared the explanations generated by DeepSHAP to LIME (a model agnostic approach) and found that the explanations differ considerably. Our study raises concerns regarding the robustness and accuracy of explanations produced for NRMs. With this paper we aim to shed light on interesting problems surrounding interpretability in NRMs and highlight areas of future work.
UR - http://www.scopus.com/inward/record.url?scp=85073786690&partnerID=8YFLogxK
U2 - 10.48550/arXiv.1907.06484
DO - 10.48550/arXiv.1907.06484
M3 - Conference contribution
AN - SCOPUS:85073786690
SP - 1005
EP - 1008
BT - SIGIR 2019
PB - Association for Computing Machinery (ACM)
T2 - 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2019
Y2 - 21 July 2019 through 25 July 2019
ER -