Details
Original language | English |
---|---|
Title of host publication | WWW '22 |
Subtitle of host publication | Proceedings of the ACM Web Conference 2022 |
Pages | 266-276 |
Number of pages | 11 |
ISBN (electronic) | 9781450390965 |
Publication status | Published - 25 Apr 2022 |
Event | 31st ACM World Wide Web Conference, WWW 2022 - Virtual, Online, France Duration: 25 Apr 2022 → 29 Apr 2022 |
Abstract
Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.
Keywords
- dense, interpolation, ranking, retrieval, sparse
ASJC Scopus subject areas
- Computer Science(all)
- Computer Networks and Communications
- Computer Science(all)
- Software
Cite this
- Standard
- Harvard
- Apa
- Vancouver
- BibTeX
- RIS
WWW '22: Proceedings of the ACM Web Conference 2022. 2022. p. 266-276.
Research output: Chapter in book/report/conference proceeding › Conference contribution › Research › peer review
}
TY - GEN
T1 - Efficient Neural Ranking using Forward Indexes
AU - Leonhardt, Jurek
AU - Rudra, Koustav
AU - Khosla, Megha
AU - Anand, Abhijit
AU - Anand, Avishek
N1 - Funding Information: This work is supported by the National Natural Science Foundation of China (62102382,U19A2079), the USTC Research Funds of the Double First-Class Initiative (WK2100000019) and the Alibaba Innovative Research project (ATT50DHZ420003).
PY - 2022/4/25
Y1 - 2022/4/25
N2 - Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.
AB - Neural document ranking approaches, specifically transformer models, have achieved impressive gains in ranking performance. However, query processing using such over-parameterized models is both resource and time intensive. In this paper, we propose the Fast-Forward index - a simple vector forward index that facilitates ranking documents using interpolation of lexical and semantic scores - as a replacement for contextual re-rankers and dense indexes based on nearest neighbor search. Fast-Forward indexes rely on efficient sparse models for retrieval and merely look up pre-computed dense transformer-based vector representations of documents and passages in constant time for fast CPU-based semantic similarity computation during query processing. We propose index pruning and theoretically grounded early stopping techniques to improve the query processing throughput. We conduct extensive large-scale experiments on TREC-DL datasets and show improvements over hybrid indexes in performance and query processing efficiency using only CPUs. Fast-Forward indexes can provide superior ranking performance using interpolation due to the complementary benefits of lexical and semantic similarities.
KW - dense
KW - interpolation
KW - ranking
KW - retrieval
KW - sparse
UR - http://www.scopus.com/inward/record.url?scp=85129831935&partnerID=8YFLogxK
U2 - 10.48550/arXiv.2110.06051
DO - 10.48550/arXiv.2110.06051
M3 - Conference contribution
AN - SCOPUS:85129831935
SP - 266
EP - 276
BT - WWW '22
T2 - 31st ACM World Wide Web Conference, WWW 2022
Y2 - 25 April 2022 through 29 April 2022
ER -