Efficient Neural Ranking using Forward Indexes

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

  • Jurek Leonhardt
  • Koustav Rudra
  • Megha Khosla
  • Abhijit Anand
  • Avishek Anand

Organisationseinheiten

Externe Organisationen

  • Indian School of Mines University
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksWWW '22
UntertitelProceedings of the ACM Web Conference 2022
Seiten266-276
Seitenumfang11
ISBN (elektronisch)9781450390965
PublikationsstatusVeröffentlicht - 25 Apr. 2022
Veranstaltung31st ACM World Wide Web Conference, WWW 2022 - Virtual, Online, Frankreich
Dauer: 25 Apr. 202229 Apr. 2022

Abstract

Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.

ASJC Scopus Sachgebiete

Zitieren

Efficient Neural Ranking using Forward Indexes. / Leonhardt, Jurek; Rudra, Koustav; Khosla, Megha et al.
WWW '22: Proceedings of the ACM Web Conference 2022. 2022. S. 266-276.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Leonhardt, J, Rudra, K, Khosla, M, Anand, A & Anand, A 2022, Efficient Neural Ranking using Forward Indexes. in WWW '22: Proceedings of the ACM Web Conference 2022. S. 266-276, 31st ACM World Wide Web Conference, WWW 2022, Virtual, Online, Frankreich, 25 Apr. 2022. https://doi.org/10.48550/arXiv.2110.06051, https://doi.org/10.1145/3485447.3511955
Leonhardt, J., Rudra, K., Khosla, M., Anand, A., & Anand, A. (2022). Efficient Neural Ranking using Forward Indexes. In WWW '22: Proceedings of the ACM Web Conference 2022 (S. 266-276) https://doi.org/10.48550/arXiv.2110.06051, https://doi.org/10.1145/3485447.3511955
Leonhardt J, Rudra K, Khosla M, Anand A, Anand A. Efficient Neural Ranking using Forward Indexes. in WWW '22: Proceedings of the ACM Web Conference 2022. 2022. S. 266-276 doi: 10.48550/arXiv.2110.06051, 10.1145/3485447.3511955
Leonhardt, Jurek ; Rudra, Koustav ; Khosla, Megha et al. / Efficient Neural Ranking using Forward Indexes. WWW '22: Proceedings of the ACM Web Conference 2022. 2022. S. 266-276
Download
@inproceedings{d5502f5654d44699ac6fe1a3250fa27b,
title = "Efficient Neural Ranking using Forward Indexes",
abstract = "Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.",
keywords = "dense, interpolation, ranking, retrieval, sparse",
author = "Jurek Leonhardt and Koustav Rudra and Megha Khosla and Abhijit Anand and Avishek Anand",
note = "Funding Information: This work is supported by the National Natural Science Foundation of China (62102382,U19A2079), the USTC Research Funds of the Double First-Class Initiative (WK2100000019) and the Alibaba Innovative Research project (ATT50DHZ420003).; 31st ACM World Wide Web Conference, WWW 2022 ; Conference date: 25-04-2022 Through 29-04-2022",
year = "2022",
month = apr,
day = "25",
doi = "10.48550/arXiv.2110.06051",
language = "English",
pages = "266--276",
booktitle = "WWW '22",

}

Download

TY - GEN

T1 - Efficient Neural Ranking using Forward Indexes

AU - Leonhardt, Jurek

AU - Rudra, Koustav

AU - Khosla, Megha

AU - Anand, Abhijit

AU - Anand, Avishek

N1 - Funding Information: This work is supported by the National Natural Science Foundation of China (62102382,U19A2079), the USTC Research Funds of the Double First-Class Initiative (WK2100000019) and the Alibaba Innovative Research project (ATT50DHZ420003).

PY - 2022/4/25

Y1 - 2022/4/25

N2 - Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.

AB - Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.

KW - dense

KW - interpolation

KW - ranking

KW - retrieval

KW - sparse

UR - http://www.scopus.com/inward/record.url?scp=85129831935&partnerID=8YFLogxK

U2 - 10.48550/arXiv.2110.06051

DO - 10.48550/arXiv.2110.06051

M3 - Conference contribution

AN - SCOPUS:85129831935

SP - 266

EP - 276

BT - WWW '22

T2 - 31st ACM World Wide Web Conference, WWW 2022

Y2 - 25 April 2022 through 29 April 2022

ER -