Emergent Mind

Abstract

Large language models are powerful text processors and reasoners, but are still subject to limitations including outdated knowledge and hallucinations, which necessitates connecting them to the world. Retrieval-augmented LLMs have raised extensive attention for grounding model generation on external knowledge. However, retrievers struggle to capture relevance, especially for queries with complex information needs. Recent work has proposed to improve relevance modeling by having LLMs actively involved in retrieval, i.e., to improve retrieval with generation. In this paper, we show that strong performance can be achieved by a method we call Iter-RetGen, which synergizes retrieval and generation in an iterative manner. A model output shows what might be needed to finish a task, and thus provides an informative context for retrieving more relevant knowledge which in turn helps generate a better output in the next iteration. Compared with recent work which interleaves retrieval with generation when producing an output, Iter-RetGen processes all retrieved knowledge as a whole and largely preserves the flexibility in generation without structural constraints. We evaluate Iter-RetGen on multi-hop question answering, fact verification, and commonsense reasoning, and show that it can flexibly leverage parametric knowledge and non-parametric knowledge, and is superior to or competitive with state-of-the-art retrieval-augmented baselines while causing fewer overheads of retrieval and generation. We can further improve performance via generation-augmented retrieval adaptation.

Overview

  • TER-RETGEN introduces an iterative process that improves retrieval-augmented LLMs by using initial responses to guide subsequent information retrieval.

  • The methodology allows a seamless fusion of parametric and non-parametric knowledge, enhancing content generation for complex queries.

  • Experimental results show TER-RETGEN’s superior accuracy in multi-hop question answering, fact verification, and commonsense reasoning tasks.

  • TER-RETGEN integrates generation into retrieval adaptation, improving performance and reducing iterative and retrieval load.

  • TER-RETGEN is presented as a viable solution for complex retrieval-augmented settings with an emphasis on simplicity and robustness.

Introduction

A new method, termed TER-RETGEN (Iterative Retrieval-Generation Synergy), has been introduced, offering a significant advance in retrieval-augmented LLMs. TER-RETGEN operates under the premise that a model's initial response can inform the retrieval of relevant information for the next iteration, thus enhancing the quality of the generated content. This iterative process allows for a dynamic interplay between retrieval and generation, potentially resolving the challenges faced by LLMs in terms of staying current and avoiding hallucinations, particularly in complex query scenarios.

Retrieval-Augmented Generation

Traditional retrieval-augmented language models typically rely on a single retrieval step, falling short in situations that demand nuanced and elaborate information synthesis. TER-RETGEN distinguishes itself by handling all retrieved knowledge collectively without imposing structural constraints within the generation process. This facilitates a more fluid integration of parametric knowledge (learned from training data) and non-parametric knowledge (external information), even as the complexity of information needs escalates.

Experimental Validation

TER-RETGEN has been benchmarked against several retrieval-augmented models across tasks including multi-hop question answering, fact verification, and commonsense reasoning. The results exhibited notable accuracy improvements, demonstrating TER-RETGEN's adaptability in leveraging both parametric and non-parametric knowledge sources. Furthermore, TER-RETGEN achieves these results with minimal overhead, refining the interaction between retrieval and generation components without necessitating complex processing pipelines.

Further Insights and Adaptations

One of the remarkable aspects of TER-RETGEN is its ability to use generative outputs from previous iterations to guide retrieval adaptation. By distilling knowledge from model generations into the dense retriever, TER-RETGEN not only enhances performance but also reduces the iterative load. Specifics illustrated in the paper show that integrating generation into retrieval adaptation brings considerable improvements while trimming down both the retrieval and iteration overheads, thereby offering a more streamlined approach.

In summary, TER-RETGEN emerges as a strong contender for handling complex informational needs in retrieval-augmented settings, balancing robustness with simplicity of implementation.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.