Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 133 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 30 tok/s Pro
GPT-4o 125 tok/s Pro
Kimi K2 188 tok/s Pro
GPT OSS 120B 448 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Corrective Retrieval Augmented Generation (2401.15884v3)

Published 29 Jan 2024 in cs.CL

Abstract: LLMs inevitably exhibit hallucinations since the accuracy of generated texts cannot be secured solely by the parametric knowledge they encapsulate. Although retrieval-augmented generation (RAG) is a practicable complement to LLMs, it relies heavily on the relevance of retrieved documents, raising concerns about how the model behaves if retrieval goes wrong. To this end, we propose the Corrective Retrieval Augmented Generation (CRAG) to improve the robustness of generation. Specifically, a lightweight retrieval evaluator is designed to assess the overall quality of retrieved documents for a query, returning a confidence degree based on which different knowledge retrieval actions can be triggered. Since retrieval from static and limited corpora can only return sub-optimal documents, large-scale web searches are utilized as an extension for augmenting the retrieval results. Besides, a decompose-then-recompose algorithm is designed for retrieved documents to selectively focus on key information and filter out irrelevant information in them. CRAG is plug-and-play and can be seamlessly coupled with various RAG-based approaches. Experiments on four datasets covering short- and long-form generation tasks show that CRAG can significantly improve the performance of RAG-based approaches.

Citations (43)

Summary

  • The paper introduces CRAG, which deploys a lightweight retrieval evaluator to classify document relevance and adjust retrieval actions accordingly.
  • It employs a decomposition-recomposition algorithm combined with dynamic web search integration to selectively refine and augment corpus information.
  • Experiments across four datasets demonstrate CRAG’s superior performance and robustness compared to traditional RAG and self-correcting methods.

Introduction

LLMs have gained significant traction for their impressive language generation abilities. However, they often grapple with factual errors and hallucinations, highlighting inherent limitations in their parametric knowledge. While Retrieval-Augmented Generation (RAG) systems offer a practical supplement to LLMs by augmenting generation with external documents, their efficacy is critically dependent on the relevance and accuracy of these documents. Current approaches, however, may indiscriminately integrate irrelevant information, undermining the robustness of generation.

Corrective Strategies for Robust Generation

The paper introduces an innovative Corrective Retrieval Augmented Generation (CRAG) methodology designed to enhance the resilience of RAG systems. CRAG incorporates a lightweight retrieval evaluator alongside a varying approach to document use depending on the evaluator's output. The evaluator assigns actions based on the estimated relevance, falling under Correct, Incorrect, or Ambiguous categories. This approach pivots on a decomposition-recomposition algorithm which addresses static corpuses' limitations by selectively focusing on essential information and discarding irrelevant content.

Web Search Integration and Experimentation

A novel aspect of CRAG is its integration with large-scale web searches for cases when knowledge retrieval is deemed Incorrect. Leveraging the dynamic nature of the web, CRAG broadens the scope and variety of information at the model’s disposal. This ensures a rich set of external knowledge to amend the initial corpus results. CRAG is demonstrated to be a plug-and-play model and shows compatibility with existing RAG-based approaches. Validation across four datasets emphasizes its performance uplift and generalizability over tasks that demand both short and long-form generation.

Conclusion and Contribution

CRAG's design for self-correcting and enhancing the utilization of retrieved documents marks a notable step in addressing RAG's existing pitfalls. The retrieval evaluator is central, helping to avoid the inclusion of misleading information and prompting robust actions based on document assessment. The refined approach shows significant improvements over the standard RAG and Self-RAG, indicating wide applicability in scenarios where the integrity of retrieved information is questionable. The experiments encapsulate CRAG's adaptability to RAG-based methods and its capability to transcend challenges across various generation categories.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 43 tweets and received 600 likes.

Upgrade to Pro to view all of the tweets about this paper:

Youtube Logo Streamline Icon: https://streamlinehq.com