Multihop Attention Networks for Question Answer Matching

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

  • Nam Khanh Tran
  • Claudia Niederee

Research Organisations

View graph of relations

Details

Original languageEnglish
Title of host publication41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2018
Subtitle of host publicationThe 41st International ACM SIGIR Conference on Research & Development in Information Retrieval
Pages325-334
Number of pages10
ISBN (electronic)978-1-4503-5657-2
Publication statusPublished - 27 Jun 2018

Abstract

Attention based neural network models have been successfully applied in answer selection, which is an important subtask of question answering (QA). These models often represent a question by a single vector and find its corresponding matches by attending to candidate answers. However, questions and answers might be related to each other in complicated ways which cannot be captured by single-vector representations. In this paper, we propose Multihop Attention Networks (MAN) which aim to uncover these complex relations for ranking question and answer pairs. Unlike previous models, we do not collapse the question into a single vector, instead we use multiple vectors which focus on different parts of the question for its overall semantic representation and apply multiple steps of attention to learn representations for the candidate answers. For each attention step, in addition to common attention mechanisms, we adopt sequential attention which utilizes context information for computing context-aware attention weights. Via extensive experiments, we show that MAN outperforms state-of-the-art approaches on popular benchmark QA datasets. Empirical studies confirm the effectiveness of sequential attention over other attention mechanisms.

Keywords

    Answer selection, non-factoid QA, representation learning, attention mechanism, Representation learning, Attention mechanism, Non-factoid qa

ASJC Scopus subject areas

Cite this

Multihop Attention Networks for Question Answer Matching. / Nam Khanh Tran; Niederee, Claudia.
41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2018: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. 2018. p. 325-334.

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Nam Khanh Tran & Niederee, C 2018, Multihop Attention Networks for Question Answer Matching. in 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2018: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. pp. 325-334. https://doi.org/10.1145/3209978.3210009
Nam Khanh Tran, & Niederee, C. (2018). Multihop Attention Networks for Question Answer Matching. In 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2018: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval (pp. 325-334) https://doi.org/10.1145/3209978.3210009
Nam Khanh Tran, Niederee C. Multihop Attention Networks for Question Answer Matching. In 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2018: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. 2018. p. 325-334 doi: 10.1145/3209978.3210009
Nam Khanh Tran ; Niederee, Claudia. / Multihop Attention Networks for Question Answer Matching. 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2018: The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. 2018. pp. 325-334
Download
@inproceedings{ecb34283dd864df5b4e4c42aec95b735,
title = "Multihop Attention Networks for Question Answer Matching",
abstract = "Attention based neural network models have been successfully applied in answer selection, which is an important subtask of question answering (QA). These models often represent a question by a single vector and find its corresponding matches by attending to candidate answers. However, questions and answers might be related to each other in complicated ways which cannot be captured by single-vector representations. In this paper, we propose Multihop Attention Networks (MAN) which aim to uncover these complex relations for ranking question and answer pairs. Unlike previous models, we do not collapse the question into a single vector, instead we use multiple vectors which focus on different parts of the question for its overall semantic representation and apply multiple steps of attention to learn representations for the candidate answers. For each attention step, in addition to common attention mechanisms, we adopt sequential attention which utilizes context information for computing context-aware attention weights. Via extensive experiments, we show that MAN outperforms state-of-the-art approaches on popular benchmark QA datasets. Empirical studies confirm the effectiveness of sequential attention over other attention mechanisms.",
keywords = "Answer selection, non-factoid QA, representation learning, attention mechanism, Representation learning, Attention mechanism, Non-factoid qa",
author = "{Nam Khanh Tran} and Claudia Niederee",
note = "Publisher Copyright: {\textcopyright} 2018 ACM.",
year = "2018",
month = jun,
day = "27",
doi = "10.1145/3209978.3210009",
language = "English",
pages = "325--334",
booktitle = "41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2018",

}

Download

TY - GEN

T1 - Multihop Attention Networks for Question Answer Matching

AU - Nam Khanh Tran,

AU - Niederee, Claudia

N1 - Publisher Copyright: © 2018 ACM.

PY - 2018/6/27

Y1 - 2018/6/27

N2 - Attention based neural network models have been successfully applied in answer selection, which is an important subtask of question answering (QA). These models often represent a question by a single vector and find its corresponding matches by attending to candidate answers. However, questions and answers might be related to each other in complicated ways which cannot be captured by single-vector representations. In this paper, we propose Multihop Attention Networks (MAN) which aim to uncover these complex relations for ranking question and answer pairs. Unlike previous models, we do not collapse the question into a single vector, instead we use multiple vectors which focus on different parts of the question for its overall semantic representation and apply multiple steps of attention to learn representations for the candidate answers. For each attention step, in addition to common attention mechanisms, we adopt sequential attention which utilizes context information for computing context-aware attention weights. Via extensive experiments, we show that MAN outperforms state-of-the-art approaches on popular benchmark QA datasets. Empirical studies confirm the effectiveness of sequential attention over other attention mechanisms.

AB - Attention based neural network models have been successfully applied in answer selection, which is an important subtask of question answering (QA). These models often represent a question by a single vector and find its corresponding matches by attending to candidate answers. However, questions and answers might be related to each other in complicated ways which cannot be captured by single-vector representations. In this paper, we propose Multihop Attention Networks (MAN) which aim to uncover these complex relations for ranking question and answer pairs. Unlike previous models, we do not collapse the question into a single vector, instead we use multiple vectors which focus on different parts of the question for its overall semantic representation and apply multiple steps of attention to learn representations for the candidate answers. For each attention step, in addition to common attention mechanisms, we adopt sequential attention which utilizes context information for computing context-aware attention weights. Via extensive experiments, we show that MAN outperforms state-of-the-art approaches on popular benchmark QA datasets. Empirical studies confirm the effectiveness of sequential attention over other attention mechanisms.

KW - Answer selection

KW - non-factoid QA

KW - representation learning

KW - attention mechanism

KW - Representation learning

KW - Attention mechanism

KW - Non-factoid qa

UR - http://www.scopus.com/inward/record.url?scp=85051505288&partnerID=8YFLogxK

U2 - 10.1145/3209978.3210009

DO - 10.1145/3209978.3210009

M3 - Conference contribution

SP - 325

EP - 334

BT - 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2018

ER -