Efficiency Optimizations for Superblock-based Sparse Retrieval
About
Learned sparse retrieval (LSR) is a popular method for first-stage retrieval because it combines the semantic matching of language models with efficient CPU-friendly algorithms. Previous work aggregates blocks into "superblocks" to quickly skip the visitation of blocks during query processing by using an advanced pruning heuristic. This paper proposes a simple and effective superblock pruning scheme that reduces the overhead of superblock score computation while preserving competitive relevance. It combines this scheme with a compact index structure and a robust zero-shot configuration that is effective across LSR models and multiple datasets. This paper provides an analytical justification and evaluation on the MS MARCO and BEIR datasets, demonstrating that the proposed scheme can be a strong alternative for efficient sparse retrieval.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Information Retrieval | SCIDOCS | nDCG@1016.2 | 24 | |
| Information Retrieval | NQ BEIR | -- | 20 | |
| Information Retrieval | DBPedia BEIR (test) | nDCG@1042.9 | 18 | |
| Information Retrieval | FIQA BEIR (test) | nDCG@1035.1 | 15 | |
| Information Retrieval | Quora BEIR | nDCG83.5 | 11 | |
| Information Retrieval | Arguana BEIR | nDCG50.3 | 11 | |
| Information Retrieval | T-COVID BEIR | nDCG72.9 | 5 | |
| Information Retrieval | SciFact BEIR | nDCG0.708 | 5 | |
| Information Retrieval | NFCorpus BEIR | nDCG35.2 | 5 | |
| Information Retrieval | Touche BEIR | nDCG0.273 | 5 |