Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 136 tok/s
Gemini 2.5 Pro 45 tok/s Pro
GPT-5 Medium 29 tok/s Pro
GPT-5 High 27 tok/s Pro
GPT-4o 88 tok/s Pro
Kimi K2 189 tok/s Pro
GPT OSS 120B 427 tok/s Pro
Claude Sonnet 4.5 38 tok/s Pro
2000 character limit reached

Utilizing BERT for Information Retrieval: Survey, Applications, Resources, and Challenges (2403.00784v1)

Published 18 Feb 2024 in cs.IR, cs.AI, and cs.CL

Abstract: Recent years have witnessed a substantial increase in the use of deep learning to solve various NLP problems. Early deep learning models were constrained by their sequential or unidirectional nature, such that they struggled to capture the contextual relationships across text inputs. The introduction of bidirectional encoder representations from transformers (BERT) leads to a robust encoder for the transformer model that can understand the broader context and deliver state-of-the-art performance across various NLP tasks. This has inspired researchers and practitioners to apply BERT to practical problems, such as information retrieval (IR). A survey that focuses on a comprehensive analysis of prevalent approaches that apply pretrained transformer encoders like BERT to IR can thus be useful for academia and the industry. In light of this, we revisit a variety of BERT-based methods in this survey, cover a wide range of techniques of IR, and group them into six high-level categories: (i) handling long documents, (ii) integrating semantic information, (iii) balancing effectiveness and efficiency, (iv) predicting the weights of terms, (v) query expansion, and (vi) document expansion. We also provide links to resources, including datasets and toolkits, for BERT-based IR systems. A key highlight of our survey is the comparison between BERT's encoder-based models and the latest generative LLMs, such as ChatGPT, which rely on decoders. Despite the popularity of LLMs, we find that for specific tasks, finely tuned BERT encoders still outperform, and at a lower deployment cost. Finally, we summarize the comprehensive outcomes of the survey and suggest directions for future research in the area.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (6)
  1. B Mitra, and N Craswell. Neural models for information retrieval. arXiv preprint arXiv:1705.01509. 2017.
  2. D Hiemstra. Using language models for information retrieval. Taaluitgeverij Neslia Paniculata, 2001.
  3. J X Huang, and Q Hu. A bayesian learning approach to promoting diversity in ranking for biomedical information retrieval. Proceedings of the 32nd international ACM SIGIR conference on Research and development in information retrieval. 2009.
  4. M Li and E Gaussier. BERT-based dense intra-ranking and contextualized late interaction via multi-task learning for long document retrieval. Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2022.
  5. OpenAI, R. ”GPT-4 technical report.” arXiv (2023): 2303-08774.
  6. R Sennrich, H Barry, and B Alexandra. “Neural machine translation of rare words with subword units.” arXiv preprint arXiv:1508.07909, 2015.
Citations (11)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 20 likes.

Upgrade to Pro to view all of the tweets about this paper: