Exploring the Best Practices of Query Expansion with Large Language Models (2401.06311v3)
Abstract: LLMs are foundational in language technologies, particularly in information retrieval (IR). Previous studies have utilized LLMs for query expansion, achieving notable improvements in IR. In this paper, we thoroughly explore the best practice of leveraging LLMs for query expansion. To this end, we introduce a training-free, straightforward yet effective framework called Multi-Text Generation Integration (\textsc{MuGI}). It leverages LLMs to generate multiple pseudo-references, integrating them with queries to enhance both sparse and dense retrievers. Our empirical findings reveal that: (1) Increasing the number of samples from LLMs benefits IR systems; (2) A balance between the query and pseudo-documents, and an effective integration strategy, is critical for high performance; (3) Contextual information from LLMs is essential, even boost a 23M model to outperform a 7B baseline model; (4) Pseudo relevance feedback can further calibrate queries for improved performance; and (5) Query expansion is widely applicable and versatile, consistently enhancing models ranging from 23M to 7B parameters. Our code and all generated references are made available at \url{https://github.com/lezhang7/Retrieval_MuGI}
- Inpars: Data augmentation for information retrieval using large language models. arXiv preprint arXiv:2202.05144.
- Language models are few-shot learners. ArXiv, abs/2005.14165.
- Overview of the trec 2020 deep learning track.
- Overview of the trec 2019 deep learning track. arXiv preprint arXiv:2003.07820.
- Recommender systems in the era of large language models (llms). ArXiv, abs/2307.02046.
- Precise zero-shot dense retrieval without relevance labels.
- Unsupervised dense information retrieval with contrastive learning. arXiv preprint arXiv:2112.09118.
- Gautier Izacard and Edouard Grave. 2020. Leveraging passage retrieval with generative models for open domain question answering. arXiv preprint arXiv:2007.01282.
- Inpars-v2: Large language models as efficient dataset generators for information retrieval. arXiv preprint arXiv:2301.01820.
- Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, 7(3):535–547.
- Dense passage retrieval for open-domain question answering.
- Retrieval-augmented generation for knowledge-intensive nlp tasks.
- Making large language models a better foundation for dense retrieval. arXiv preprint arXiv:2312.15503.
- Pyserini: A Python toolkit for reproducible information retrieval research with sparse and dense representations. In Proceedings of the 44th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2021), pages 2356–2362.
- Zero-shot listwise document reranking with a large language model. arXiv preprint arXiv:2305.02156.
- Mteb: Massive text embedding benchmark. arXiv preprint arXiv:2210.07316.
- Document ranking with a pretrained sequence-to-sequence model. arXiv preprint arXiv:2003.06713.
- Learning implicit user profile for personalized retrieval-based chatbot. Proceedings of the 30th ACM International Conference on Information & Knowledge Management.
- Rocketqa: An optimized training approach to dense passage retrieval for open-domain question answering. In North American Chapter of the Association for Computational Linguistics.
- Okapi at trec-3. In Text Retrieval Conference.
- Improving passage retrieval with zero-shot question generation. arXiv preprint arXiv:2204.07496.
- Large language models are strong zero-shot retriever. arXiv preprint arXiv:2304.14233.
- Is chatgpt good at search? investigating large language models as re-ranking agent. arXiv preprint arXiv:2304.09542.
- Beir: A heterogenous benchmark for zero-shot evaluation of information retrieval models.
- Llama: Open and efficient foundation language models. ArXiv, abs/2302.13971.
- Text embeddings by weakly-supervised contrastive pre-training.
- Improving text embeddings with large language models. arXiv preprint arXiv:2401.00368.
- Query2doc: Query expansion with large language models. arXiv preprint arXiv:2303.07678.
- Approximate nearest neighbor negative contrastive learning for dense text retrieval. arXiv preprint arXiv:2007.00808.
- Multi-hop selector network for multi-turn response selection in retrieval-based chatbots. In Conference on Empirical Methods in Natural Language Processing.
- Recommendation as instruction following: A large language model empowered recommendation approach. ArXiv, abs/2305.07001.
- Retgen: A joint framework for retrieval and grounded text generation modeling. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 11739–11747.
- Large language models for information retrieval: A survey. arXiv preprint arXiv:2308.07107.