Emergent Mind

Abstract

We present chain-of-knowledge (CoK), a novel framework that augments LLMs by dynamically incorporating grounding information from heterogeneous sources. It results in more factual rationales and reduced hallucination in generation. Specifically, CoK consists of three stages: reasoning preparation, dynamic knowledge adapting, and answer consolidation. Given a knowledge-intensive question, CoK first prepares several preliminary rationales and answers while identifying the relevant knowledge domains. If there is no majority consensus among the answers from samples, CoK corrects the rationales step by step by adapting knowledge from the identified domains. These corrected rationales can plausibly serve as a better foundation for the final answer consolidation. Unlike prior studies that primarily use unstructured data, CoK also leverages structured knowledge sources such as Wikidata and tables that provide more reliable factual information. To access both unstructured and structured knowledge sources in the dynamic knowledge adapting stage, we propose an adaptive query generator that allows the generation of queries for various types of query languages, including SPARQL, SQL, and natural sentences. Moreover, to minimize error propagation between rationales, CoK corrects the rationales progressively using preceding corrected rationales to generate and correct subsequent rationales. Extensive experiments show that CoK consistently improves the performance of LLMs on knowledge-intensive tasks across different domains.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a detailed summary of this paper with a premium account.

We ran into a problem analyzing this paper.

Subscribe by Email

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.

YouTube
References
  1. Autoregressive entity retrieval. In Proceedings of ICLR
  2. KQA pro: A dataset with explicit compositional programs for complex question answering over knowledge base. In Proceedings of ACL
  3. Reading Wikipedia to answer open-domain questions. In Proceedings of ACL
  4. Is GPT-4 a Good Data Analyst?
  5. PaLM: Scaling Language Modeling with Pathways
  6. Scaling Instruction-Finetuned Language Models
  7. Is gpt-3 a good data annotator? In Proceedings of ACL
  8. Can machine translation systems be evaluated by the crowd alone. Natural Language Engineering
  9. Retrieval augmented language model pre-training. In Proceedings of ICML
  10. MedAlpaca -- An Open-Source Collection of Medical Conversational AI Models and Training Data
  11. Measuring massive multitask language understanding. In Proceedings of ICLR
  12. LoRA: Low-Rank Adaptation of Large Language Models
  13. Survey of hallucination in natural language generation. ACM Computing Surveys
  14. Findings of the 2022 conference on machine translation (WMT22). In Proceedings of WMT
  15. The measurement of observer agreement for categorical data. biometrics
  16. Retrieval-augmented generation for knowledge-intensive nlp tasks. In Proceedings of NIPS
  17. Chain of Hindsight Aligns Language Models with Feedback
  18. Learn to explain: Multimodal reasoning via thought chains for science question answering. In Proceedings of NIPS
  19. Augmented Large Language Models with Parametric Knowledge Guiding
  20. Augmented Language Models: a Survey
  21. FeTaQA: Free-form table question answering. Transactions of the Association for Computational Linguistics
  22. GPT-4 Technical Report
  23. Training language models to follow instructions with human feedback
  24. MedMCQA : A Large-scale Multi-Subject Multi-Choice Dataset for Medical domain Question Answering
  25. Fact-checking complex claims with program-guided reasoning. In Proceedings of ACL
  26. KILT: a benchmark for knowledge intensive language tasks. In Proceedings of NAACL
  27. Toolformer: Language Models Can Teach Themselves to Use Tools
  28. HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face
  29. Replug: Retrieval-augmented black-box language models
  30. FEVER: a large-scale dataset for fact extraction and VERification. In Proceedings of NAACL
  31. Llama 2: Open Foundation and Fine-Tuned Chat Models
  32. Lc-quad: A corpus for complex question answering over knowledge graphs. In Proceedings of ISWC
  33. Self-consistency improves chain of thought reasoning in language models. In Proceedings of ICLR
  34. Chain-of-thought prompting elicits reasoning in large language models. In Proceedings of NIPS
  35. UnifiedSKG: Unifying and multi-tasking structured knowledge grounding with text-to-text language models. In Proceedings of EMNLP
  36. HotpotQA: A dataset for diverse, explainable multi-hop question answering. In Proceedings of EMNLP
  37. React: Synergizing reasoning and acting in language models. In Proceedings of ICLR
  38. Retrieving Multimodal Information for Augmented Generation: A Survey
  39. Verify-and-edit: A knowledge-enhanced chain-of-thought framework. In Proceedings of ACL, 2023b.

Show All 39