Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 171 tok/s
Gemini 2.5 Pro 52 tok/s Pro
GPT-5 Medium 38 tok/s Pro
GPT-5 High 43 tok/s Pro
GPT-4o 108 tok/s Pro
Kimi K2 173 tok/s Pro
GPT OSS 120B 442 tok/s Pro
Claude Sonnet 4.5 34 tok/s Pro
2000 character limit reached

SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking (2107.05720v1)

Published 12 Jul 2021 in cs.IR

Abstract: In neural Information Retrieval, ongoing research is directed towards improving the first retriever in ranking pipelines. Learning dense embeddings to conduct retrieval using efficient approximate nearest neighbors methods has proven to work well. Meanwhile, there has been a growing interest in learning sparse representations for documents and queries, that could inherit from the desirable properties of bag-of-words models such as the exact matching of terms and the efficiency of inverted indexes. In this work, we present a new first-stage ranker based on explicit sparsity regularization and a log-saturation effect on term weights, leading to highly sparse representations and competitive results with respect to state-of-the-art dense and sparse methods. Our approach is simple, trained end-to-end in a single stage. We also explore the trade-off between effectiveness and efficiency, by controlling the contribution of the sparsity regularization.

Citations (263)

Summary

  • The paper introduces SPLADE, a sparse ranking model that leverages logarithmic activations to balance sparsity with document-query expansion.
  • It employs a combined ranking and regularization loss, significantly reducing FLOPS while matching performance of dense models on benchmarks.
  • Experiments on MS MARCO demonstrate SPLADE's competitive retrieval effectiveness and efficiency, setting a new standard for sparse IR methods.

SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking

Introduction

The development of large pre-trained LLMs such as BERT has significantly transformed NLP and IR domains. In the context of neural information retrieval, the primary challenge is to enhance the effectiveness of the initial retrieval stage. While dense retrieval using methods like approximate nearest neighbors has yielded impressive outcomes, it lacks the mechanism to precisely match term-level representations. This limitation has driven recent interest in sparse representation learning as a potential solution. The paper introduces SPLADE, a first-stage ranking model that focuses on balancing sparsity and document-query expansion while leveraging properties of BOW models, including efficiency and interpretability.

SPLADE Model Description

SPLADE utilizes a strategic combination of explicit sparsity regularization and a logarithmic activation function. By doing so, the model achieves highly sparse representations while maintaining competitive performance with state-of-the-art sparse and dense models. SPLADE is designed to be trained end-to-end in a single stage, thus simplifying the training complexity compared to other models that require multi-stage processes.

The SPLADE model introduces a logarithmic activation to control the saturation of term weights, preventing certain terms from dominating and naturally ensuring sparsity:

wj=itlog(1+ReLU(wij))w_{j} = \sum_{i \in t} \log \left(1 + \text{ReLU}(w_{ij}) \right)

This approach contrasts with methods that apply a traditional ReLU activation, providing benefits in both effectiveness and efficiency.

Ranking and Regularization

SPLADE's ranking is guided by a loss function integrating both ranking objectives and regularization to ensure sparse embeddings:

L=LrankIBN+λqLregq+λdLregd\mathcal{L} = \mathcal{L}_{rank-IBN} + \lambda_q \mathcal{L}_{reg}^{q} + \lambda_d \mathcal{L}_{reg}^{d}

Where Lreg\mathcal{L}_{reg} can be either 1\ell_1 regularization or the more balanced FLOPS\ell_{FLOPS} regularization. The latter enforces a more even distribution of active terms, which results in efficient index structures (Figure 1). Figure 1

Figure 1: Performance vs FLOPS for SPLADE models trained with different regularization strength λ\lambda on MS MARCO.

Experimental Evaluation

Experiments have been performed utilizing the MS MARCO passage ranking dataset, yielding promising results. SPLADE surpassed other sparse retrieval methods—like DeepCT and doc2query-T5—and echoed the efficacy of dense models such as ANCE. Notably, SPLADE demonstrated a significant reduction in the average floating-point operations (FLOPS) required for a query-document pair, while maintaining competitive retrieval performance.

Comparison Between Dense and Sparse Retrieval

The results highlighted that SPLADE models offer a compelling trade-off between computational efficiency and retrieval effectiveness, overcoming the typical divisiveness observed between dense and sparse methods. The introduction of FLOPS regularization further facilitates the creation of indices that are not just efficient but also aligned with the sparsity of traditional lexical indices.

Future Implications and Conclusion

The approach SPLADE brings to sparse representation learning emphasizes on efficiency and straightforward training pipelines without sacrificing effectiveness. SPLADE stands as a promising candidate for initial retrieval tasks, holding potential for further refinement and application in a range of information retrieval contexts.

Future work can explore enhancements through hybrid indexing solutions or adaptations into more complex ranking architectures. SPLADE not only sets a precedent for sparse retrieval models but also opens avenues for continued research into efficient neural ranking mechanisms.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Github Logo Streamline Icon: https://streamlinehq.com
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 2 tweets and received 1 like.

Upgrade to Pro to view all of the tweets about this paper:

Youtube Logo Streamline Icon: https://streamlinehq.com

HackerNews

  1. Splade: Sparse Neural Search (1 point, 0 comments)