ActiveRAG: Autonomously Knowledge Assimilation and Accommodation through Retrieval-Augmented Agents (2402.13547v2)
Abstract: Retrieval-Augmented Generation (RAG) enables LLMs to leverage external knowledge, enhancing their performance on knowledge-intensive tasks. However, existing RAG models often treat LLMs as passive recipients of information, which can lead to interference from noisy retrieved content. In this paper, we introduce ActiveRAG, a multi-agent framework that mimics human learning behavior to help LLMs actively engage with and learn from retrieved evidence. ActiveRAG designs a knowledge assimilation agent to form the knowledge understanding by associating external knowledge with the parametric memory of LLMs. Then our model employs the thought accommodation agent to calibrate the internal thought of LLMs for response refinement. Our experiments show that ActiveRAG achieves a 10\% improvement over vanilla RAG on various question-answering benchmarks. Further analysis reveals that ActiveRAG mitigates the impact of noisy retrievals, alleviates conflicts between external knowledge and parametric memory and improves the self-consistency of LLMs in answering the question. All data and codes are available at https://github.com/OpenMatch/ActiveRAG.
- Self-rag: Learning to retrieve, generate, and critique through self-reflection. ArXiv preprint.
- Richard C Atkinson and Richard M Shiffrin. 1968. Human memory: A proposed system and its control processes. In Psychology of learning and motivation, pages 89–195.
- A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity. ArXiv preprint.
- Semantic parsing on Freebase from question-answer pairs. In Proceedings of EMNLP, pages 1533–1544.
- Improving language models by retrieving from trillions of tokens. In Proceedings of ICML, pages 2206–2240.
- Benchmarking large language models in retrieval-augmented generation. ArXiv preprint.
- Augmenting zero-shot dense retrievers with plug-in mixture-of-memories. pages 1796–1812.
- Retrieval augmented language model pre-training. In Proceedings of ICML, pages 3929–3938.
- Rethinking with retrieval: Faithful large language model inference. ArXiv preprint.
- Efficient nearest neighbor language models. In Proceedings of EMNLP, pages 5703–5714.
- Gautier Izacard and Edouard Grave. 2020. Distilling knowledge from reader to retriever for question answering. In Proceedings of ICLR.
- Gautier Izacard and Edouard Grave. 2021. Leveraging passage retrieval with generative models for open domain question answering. In Proceedings of EACL, pages 874–880.
- Atlas: Few-shot learning with retrieval augmented language models. J. Mach. Learn. Res., pages 251:1–251:43.
- Survey of hallucination in natural language generation. ACM Comput. Surv., (12):248:1–248:38.
- Active retrieval augmented generation. pages 7969–7992.
- TriviaQA: A large scale distantly supervised challenge dataset for reading comprehension. In Proceedings of ACL, pages 1601–1611.
- Large language models struggle to learn long-tail knowledge. In Proceedings of ICML, pages 15696–15707. PMLR.
- Dense passage retrieval for open-domain question answering. In Proceedings of EMNLP, pages 6769–6781.
- Natural questions: A benchmark for question answering research. Proceedings of TACL, pages 452–466.
- Retrieval-augmented generation for knowledge-intensive NLP tasks. In Proceedings of NeurIPS.
- Truthfulqa: Measuring how models mimic human falsehoods. In Proceedings of ACL, pages 3214–3252.
- Openmatch: An open source library for neu-ir research. In Proceedings of SIGIR, pages 2531–2535.
- When not to trust language models: Investigating effectiveness of parametric and non-parametric memories. In Proceedings of ACL, pages 9802–9822.
- R OpenAI. 2023. Gpt-4 technical report. arXiv, pages 2303–08774.
- KILT: a benchmark for knowledge intensive language tasks. In Proceedings of NAACL-HLT, pages 2523–2544.
- Jean Piaget. 1970. Genetic epistemology.
- In-context retrieval-augmented language models. ArXiv preprint.
- In-context retrieval-augmented language models. Proceedings of TACL, pages 1316–1331.
- Replug: Retrieval-augmented black-box language models. ArXiv preprint.
- Retrieval augmentation reduces hallucination in conversation. In Proceedings of EMNLP Findings, pages 3784–3803.
- Blenderbot 3: a deployed conversational agent that continually learns to responsibly engage. ArXiv preprint.
- Prompting gpt-3 to be reliable. In Proceedings of ICLR.
- End-to-end training of multi-document reader and retriever for open-domain question answering. Proceedings of NeurIPS, pages 25968–25981.
- Leslie P Steffe and Jerry Edward Gale. 1995. Constructivism in education.
- Llama: Open and efficient foundation language models. ArXiv preprint.
- Interleaving retrieval with chain-of-thought reasoning for knowledge-intensive multi-step questions. In Proceedings of ACL, pages 10014–10037.
- Learning to filter context for retrieval-augmented generation. ArXiv preprint.
- Emergent abilities of large language models. Transactions on Machine Learning Research.
- Chain-of-thought prompting elicits reasoning in large language models. Proceedings of NeurIPS, pages 24824–24837.
- Approximate nearest neighbor negative contrastive learning for dense text retrieval. In Proceedings of ICLR.
- Recomp: Improving retrieval-augmented lms with compression and selective augmentation. ArXiv preprint.
- Answering questions by meta-reasoning over multiple chains of thought. In Proceedings of EMNLP, pages 5942–5966.
- Making retrieval-augmented language models robust to irrelevant context. ArXiv preprint.
- Openmatch-v2: An all-in-one multi-modality plm-based information retrieval toolkit. In Proceedings of SIGIR, pages 3160–3164.
- Generate rather than retrieve: Large language models are strong context generators. In Proceedings of ICLR.
- Chain-of-note: Enhancing robustness in retrieval-augmented language models. ArXiv preprint.
- Augmentation-adapted retriever improves generalization of language models as generic plug-in. pages 2421–2436.
- A survey of large language models. ArXiv preprint.
- Docprompting: Generating code by retrieving the docs. In Proceedings of ICLR.