Efficient Neural Ranking using Forward Indexes

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

  • Jurek Leonhardt
  • Koustav Rudra
  • Megha Khosla
  • Abhijit Anand
  • Avishek Anand

Research Organisations

External Research Organisations

  • Indian School of Mines University
View graph of relations

Details

Original languageEnglish
Title of host publicationWWW '22
Subtitle of host publicationProceedings of the ACM Web Conference 2022
Pages266-276
Number of pages11
ISBN (electronic)9781450390965
Publication statusPublished - 25 Apr 2022
Event31st ACM World Wide Web Conference, WWW 2022 - Virtual, Online, France
Duration: 25 Apr 202229 Apr 2022

Abstract

Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.

Keywords

    dense, interpolation, ranking, retrieval, sparse

ASJC Scopus subject areas

Cite this

Efficient Neural Ranking using Forward Indexes. / Leonhardt, Jurek; Rudra, Koustav; Khosla, Megha et al.
WWW '22: Proceedings of the ACM Web Conference 2022. 2022. p. 266-276.

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Leonhardt, J, Rudra, K, Khosla, M, Anand, A & Anand, A 2022, Efficient Neural Ranking using Forward Indexes. in WWW '22: Proceedings of the ACM Web Conference 2022. pp. 266-276, 31st ACM World Wide Web Conference, WWW 2022, Virtual, Online, France, 25 Apr 2022. https://doi.org/10.48550/arXiv.2110.06051, https://doi.org/10.1145/3485447.3511955
Leonhardt, J., Rudra, K., Khosla, M., Anand, A., & Anand, A. (2022). Efficient Neural Ranking using Forward Indexes. In WWW '22: Proceedings of the ACM Web Conference 2022 (pp. 266-276) https://doi.org/10.48550/arXiv.2110.06051, https://doi.org/10.1145/3485447.3511955
Leonhardt J, Rudra K, Khosla M, Anand A, Anand A. Efficient Neural Ranking using Forward Indexes. In WWW '22: Proceedings of the ACM Web Conference 2022. 2022. p. 266-276 doi: 10.48550/arXiv.2110.06051, 10.1145/3485447.3511955
Leonhardt, Jurek ; Rudra, Koustav ; Khosla, Megha et al. / Efficient Neural Ranking using Forward Indexes. WWW '22: Proceedings of the ACM Web Conference 2022. 2022. pp. 266-276
Download
@inproceedings{d5502f5654d44699ac6fe1a3250fa27b,
title = "Efficient Neural Ranking using Forward Indexes",
abstract = "Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.",
keywords = "dense, interpolation, ranking, retrieval, sparse",
author = "Jurek Leonhardt and Koustav Rudra and Megha Khosla and Abhijit Anand and Avishek Anand",
note = "Funding Information: This work is supported by the National Natural Science Foundation of China (62102382,U19A2079), the USTC Research Funds of the Double First-Class Initiative (WK2100000019) and the Alibaba Innovative Research project (ATT50DHZ420003).; 31st ACM World Wide Web Conference, WWW 2022 ; Conference date: 25-04-2022 Through 29-04-2022",
year = "2022",
month = apr,
day = "25",
doi = "10.48550/arXiv.2110.06051",
language = "English",
pages = "266--276",
booktitle = "WWW '22",

}

Download

TY - GEN

T1 - Efficient Neural Ranking using Forward Indexes

AU - Leonhardt, Jurek

AU - Rudra, Koustav

AU - Khosla, Megha

AU - Anand, Abhijit

AU - Anand, Avishek

N1 - Funding Information: This work is supported by the National Natural Science Foundation of China (62102382,U19A2079), the USTC Research Funds of the Double First-Class Initiative (WK2100000019) and the Alibaba Innovative Research project (ATT50DHZ420003).

PY - 2022/4/25

Y1 - 2022/4/25

N2 - Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.

AB - Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.

KW - dense

KW - interpolation

KW - ranking

KW - retrieval

KW - sparse

UR - http://www.scopus.com/inward/record.url?scp=85129831935&partnerID=8YFLogxK

U2 - 10.48550/arXiv.2110.06051

DO - 10.48550/arXiv.2110.06051

M3 - Conference contribution

AN - SCOPUS:85129831935

SP - 266

EP - 276

BT - WWW '22

T2 - 31st ACM World Wide Web Conference, WWW 2022

Y2 - 25 April 2022 through 29 April 2022

ER -