Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
124 tokens/sec
GPT-4o
8 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Play the Shannon Game With Language Models: A Human-Free Approach to Summary Evaluation (2103.10918v2)

Published 19 Mar 2021 in cs.CL and cs.LG

Abstract: The goal of a summary is to concisely state the most important information in a document. With this principle in mind, we introduce new reference-free summary evaluation metrics that use a pretrained LLM to estimate the information content shared between a document and its summary. These metrics are a modern take on the Shannon Game, a method for summary quality scoring proposed decades ago, where we replace human annotators with LLMs. We also view these metrics as an extension of BLANC, a recently proposed approach to summary quality measurement based on the performance of a LLM with and without the help of a summary. Using transformer based LLMs, we empirically verify that our metrics achieve state-of-the-art correlation with human judgement of the summary quality dimensions of both coherence and relevance, as well as competitive correlation with human judgement of consistency and fluency.

Citations (15)

Summary

We haven't generated a summary for this paper yet.