A Comprehensive Survey of Hallucination Mitigation Techniques in Large Language Models (2401.01313v3)
Abstract: As LLMs continue to advance in their ability to write human-like text, a key challenge remains around their tendency to hallucinate generating content that appears factual but is ungrounded. This issue of hallucination is arguably the biggest hindrance to safely deploying these powerful LLMs into real-world production systems that impact people's lives. The journey toward widespread adoption of LLMs in practical settings heavily relies on addressing and mitigating hallucinations. Unlike traditional AI systems focused on limited tasks, LLMs have been exposed to vast amounts of online text data during training. While this allows them to display impressive language fluency, it also means they are capable of extrapolating information from the biases in training data, misinterpreting ambiguous prompts, or modifying the information to align superficially with the input. This becomes hugely alarming when we rely on language generation capabilities for sensitive applications, such as summarizing medical records, financial analysis reports, etc. This paper presents a comprehensive survey of over 32 techniques developed to mitigate hallucination in LLMs. Notable among these are Retrieval Augmented Generation (Lewis et al, 2021), Knowledge Retrieval (Varshney et al,2023), CoNLI (Lei et al, 2023), and CoVe (Dhuliawala et al, 2023). Furthermore, we introduce a detailed taxonomy categorizing these methods based on various parameters, such as dataset utilization, common tasks, feedback mechanisms, and retriever types. This classification helps distinguish the diverse approaches specifically designed to tackle hallucination issues in LLMs. Additionally, we analyze the challenges and limitations inherent in these techniques, providing a solid foundation for future research in addressing hallucinations and related phenomena within the realm of LLMs.
- Fleek: Factual error detection and correction with evidence retrieved from external knowledge.
- A step closer to comprehensive answers: Constrained multi-stage question decomposition with large language models.
- Dress: Instructing large vision-language models to align and interact with humans via natural language feedback. arXiv preprint arXiv:2311.10081.
- Uprise: Universal prompt retrieval for improving zero-shot evaluation.
- George Chrysostomou and Nikolaos Aletras. 2021. Enjoy the salience: Towards better transformer-based faithful explanations with word salience. arXiv preprint arXiv:2108.13759.
- Dola: Decoding by contrasting layers improves factuality in large language models.
- Scaling instruction-finetuned language models.
- Chain-of-verification reduces hallucination in large language models.
- Halo: Estimation and reduction of hallucinations in open-source weak large language models.
- Trapping llm hallucinations using tagged context prompts.
- Rarr: Researching and revising what language models say, using language models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 16477–16508.
- Opt-iml: Scaling language model instruction meta learning through the lens of generalization.
- RHO: Reducing hallucination in open-domain dialogues with knowledge grounding. In Findings of the Association for Computational Linguistics: ACL 2023, pages 4504–4522, Toronto, Canada. Association for Computational Linguistics.
- Towards mitigating hallucination in large language models via self-reflection.
- Teaching language models to hallucinate less with synthetic tasks.
- Ever: Mitigating hallucination in large language models through real-time verification and rectification.
- Hallucination augmented recitations for language models.
- Albert: A lite bert for self-supervised learning of language representations. In International Conference on Learning Representations.
- Mateusz Lango and Ondrej Dusek. 2023. Critic-driven decoding for mitigating hallucinations in data-to-text generation. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 2853–2862, Singapore. Association for Computational Linguistics.
- Chain of natural language inference for reducing large language model ungrounded hallucinations.
- The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Retrieval-augmented generation for knowledge-intensive nlp tasks. arXiv preprint arXiv:2005.11401v4.
- Inference-time intervention: Eliciting truthful answers from a language model. arXiv preprint arXiv:2306.03341.
- Self-checker: Plug-and-play modules for fact-checking with large language models.
- Mind’s mirror: Distilling self-evaluation capability and comprehensive thinking from large language models.
- Self-refine: Iterative refinement with self-feedback.
- Self-contradictory hallucinations of large language models: Evaluation, detection and mitigation. arXiv preprint arXiv:2305.15852.
- Self-contradictory hallucinations of large language models: Evaluation, detection and mitigation.
- Check your facts and try again: Improving large language models with external knowledge and automated feedback.
- Think while you write: Hypothesis verification promotes faithful knowledge-to-text generation. arXiv preprint arXiv:2311.09467.
- Detecting and mitigating hallucinations in multilingual summarisation.
- The troubling emergence of hallucination in large language models – an extensive definition, quantification, and prescriptive remediations.
- Partha Pratim Ray. 2023. Chatgpt: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope. Internet of Things and Cyber-Physical Systems, 3:121–154.
- Dial BeInfo for Faithfulness: Improving factuality of information-seeking dialogue via behavioural fine-tuning.
- Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108.
- Trusting your evidence: Hallucinate less with context-aware decoding.
- Prompting gpt-3 to be reliable. arXiv preprint arXiv:2210.09150.
- Head-to-tail: How knowledgeable are large language models (llm)? aka will llms replace knowledge graphs? arXiv preprint arXiv:2308.10168.
- Principle-driven self-alignment of language models from scratch with minimal human supervision.
- Fine-tuning language models for factuality.
- Llama 2: Open foundation and fine-tuned chat models.
- A stitch in time saves nine: Detecting and mitigating hallucinations of llms by validating low-confidence generation.
- Freshllms: Refreshing large language models with search engine augmentation.
- Self-instruct: Aligning language models with self-generated instructions.
- A prompt pattern catalog to enhance prompt engineering with chatgpt.
- Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100.
- Wizardlm: Empowering large language models to follow complex instructions.
- On what basis? predicting text preference via structured comparative reasoning.
- Information-theoretic text hallucination reduction for video-grounded dialogue. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 4182–4193, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- R-tuning: Teaching large language models to refuse unknown questions. arXiv preprint arXiv:2311.09677.
- The knowledge alignment problem: Bridging human and external knowledge for large language models.