Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 65 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 39 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 97 tok/s Pro
Kimi K2 164 tok/s Pro
GPT OSS 120B 466 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

CLARINET: Augmenting Language Models to Ask Clarification Questions for Retrieval (2405.15784v1)

Published 28 Apr 2024 in cs.IR, cs.AI, and cs.CL

Abstract: Users often make ambiguous requests that require clarification. We study the problem of asking clarification questions in an information retrieval setting, where systems often face ambiguous search queries and it is challenging to turn the uncertainty in the retrieval model into a natural language question. We present CLARINET, a system that asks informative clarification questions by choosing questions whose answers would maximize certainty in the correct candidate. Our approach works by augmenting a LLM to condition on a retrieval distribution, finetuning end-to-end to generate the question that would have maximized the rank of the true candidate at each turn. When evaluated on a real-world retrieval dataset of users searching for books, our system outperforms traditional heuristics such as information gain on retrieval success by 17% and vanilla-prompted LLMs by 39% relative.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (17)
  1. ’it’s on the tip of my tongue’ a new dataset for known-item retrieval. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, pages 48–56.
  2. Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416.
  3. Robert Hawkins and Noah Goodman. 2017. Why do you ask? the informational dynamics of questions and answers.
  4. Gautier Izacard and Edouard Grave. 2020. Leveraging passage retrieval with generative models for open domain question answering. arXiv preprint arXiv:2007.01282.
  5. Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906.
  6. Kimiya Keyvan and Jimmy Xiangji Huang. 2022. How to approach ambiguous queries in conversational search: A survey of techniques, approaches, tools, and challenges. ACM Computing Surveys, 55(6):1–40.
  7. Decomposing complex queries for tip-of-the-tongue retrieval. arXiv preprint arXiv:2305.15053.
  8. Experience matters: Information acquisition optimizes probability gain. Psychological science, 21(7):960–969.
  9. Mike Oaksford and Nick Chater. 1994. A rational analysis of the selection task as optimal data selection. Psychological review, 101(4):608.
  10. Reinforced clarification question generation with defeasibility rewards for disambiguating social and moral situations. arXiv preprint arXiv:2212.10409.
  11. Sudha Rao and Hal Daumé III. 2018. Learning to ask good questions: Ranking clarification questions using neural expected value of perfect information. arXiv preprint arXiv:1805.04655.
  12. Sudha Rao and Hal Daumé III. 2019. Answer-based adversarial training for generating clarification questions. arXiv preprint arXiv:1904.02281.
  13. Question asking as program generation. Advances in neural information processing systems, 30.
  14. Robert Van Rooy. 2004. Utility, informativity and protocols. Journal of philosophical logic, 33:389–419.
  15. Open-domain clarification question generation without question examples. arXiv preprint arXiv:2110.09779.
  16. Interactive classification by asking informative questions. arXiv preprint arXiv:1911.03598.
  17. Generating clarifying questions for information retrieval. In Proceedings of the web conference 2020, pages 418–428.
Citations (1)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube