Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 47 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 28 tok/s Pro
GPT-5 High 25 tok/s Pro
GPT-4o 104 tok/s Pro
Kimi K2 156 tok/s Pro
GPT OSS 120B 474 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

Towards Logically Consistent Language Models via Probabilistic Reasoning (2404.12843v1)

Published 19 Apr 2024 in cs.LG and cs.CL

Abstract: LLMs are a promising venue for natural language understanding and generation tasks. However, current LLMs are far from reliable: they are prone to generate non-factual information and, more crucially, to contradict themselves when prompted to reason about beliefs of the world. These problems are currently addressed with large scale fine-tuning or by delegating consistent reasoning to external tools. In this work, we strive for a middle ground and introduce a training objective based on principled probabilistic reasoning that teaches a LLM to be consistent with external knowledge in the form of a set of facts and rules. Fine-tuning with our loss on a limited set of facts enables our LLMs to be more logically consistent than previous baselines and allows them to extrapolate to unseen but semantically similar factual knowledge more systematically.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. Semantic Probabilistic Layers for Neuro-Symbolic Learning. In NeurIPS, 2022.
  2. Semantic sensitivities and inconsistent predictions: Measuring the fragility of nli models, 2024.
  3. Roberto Battiti. Maximum satisfiability problemMaximum Satisfiability Problem, pp.  2035–2041. Springer US, Boston, MA, 2009. ISBN 978-0-387-74759-0. doi: 10.1007/978-0-387-74759-0˙364.
  4. Discovering latent knowledge in language models without supervision, 2022.
  5. From statistical relational to neural-symbolic artificial intelligence. In Proceedings of the Twenty-Ninth International Conference on International Joint Conferences on Artificial Intelligence, pp. 4943–4950, 2021.
  6. Truthful ai: Developing and governing ai that does not lie, 2021.
  7. Maieutic prompting: Logically consistent reasoning with recursive explanations, 2022.
  8. Beliefbank: Adding memory to a pre-trained language model for a systematic notion of belief, 2021.
  9. Language models with rationality, 2023.
  10. A logic-driven framework for consistency of neural models, 2019.
  11. Truthfulqa: Measuring how models mimic human falsehoods, 2022.
  12. Vera: A general-purpose plausibility estimation model for commonsense statements, 2023.
  13. Roberta: A robustly optimized bert pretraining approach, 2019.
  14. Sgdr: Stochastic gradient descent with warm restarts. In International Conference on Learning Representations, 2016.
  15. Enhancing self-consistency and performance of pre-trained language models through natural language inference, 2022.
  16. Language models as knowledge bases?, 2019.
  17. Differentiation of blackbox combinatorial solvers. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020.
  18. General-purpose question-answering with macaw, 2021.
  19. Entailer: Answering questions with faithful and truthful chains of reasoning, 2022.
  20. A compositional atlas of tractable circuit operations for probabilistic inference. Advances in Neural Information Processing Systems, 34:13189–13201, 2021.
  21. Fact or fiction: Verifying scientific claims, 2020.
  22. A semantic loss function for deep learning with symbolic knowledge, 2018.
  23. Improved logical reasoning of language models via differentiable symbolic programming, 2023.
Citations (1)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com