Reading Subtext: Evaluating Large Language Models on Short Story Summarization with Writers (2403.01061v3)
Abstract: We evaluate recent LLMs on the challenging task of summarizing short stories, which can be lengthy, and include nuanced subtext or scrambled timelines. Importantly, we work directly with authors to ensure that the stories have not been shared online (and therefore are unseen by the models), and to obtain informed evaluations of summary quality using judgments from the authors themselves. Through quantitative and qualitative analysis grounded in narrative theory, we compare GPT-4, Claude-2.1, and LLama-2-70B. We find that all three models make faithfulness mistakes in over 50% of summaries and struggle with specificity and interpretation of difficult subtext. We additionally demonstrate that LLM ratings and other automatic metrics for summary quality do not correlate well with the quality ratings from the writers.
- Nina Begus. 2023. Experimental narratives: A comparison of human crowdsourced storytelling and ai storytelling. arXiv preprint arXiv:2310.12902.
- Wayne C Booth. 1983. The rhetoric of fiction. University of Chicago Press.
- Art or artifice? large language models and the false promise of creativity. arXiv preprint arXiv:2309.14556.
- Creativity support in the age of large language models: An empirical study involving emerging writers. arXiv preprint arXiv:2309.12570.
- Help me write a poem: Instruction tuning as a vehicle for collaborative poetry writing. arXiv preprint arXiv:2210.13669.
- Booookscore: A systematic exploration of book-length summarization in the era of llms. arXiv preprint arXiv:2310.00785.
- Summscreen: A dataset for abstractive screenplay summarization. arXiv preprint arXiv:2104.07091.
- Evaluation of african american language bias in natural language generation. arXiv preprint arXiv:2305.14291.
- Summeval: Re-evaluating summarization evaluation. Transactions of the Association for Computational Linguistics, 9:391–409.
- Qafacteval: Improved qa-based factual consistency evaluation for summarization. arXiv preprint arXiv:2112.08542.
- Gérard Genette. 1980. Narrative discourse: An essay in method, volume 3. Cornell University Press.
- FALTE: A toolkit for fine-grained annotation for long text evaluation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 351–358, Abu Dhabi, UAE. Association for Computational Linguistics.
- Snac: Coherence error detection for narrative summarization. arXiv preprint arXiv:2205.09641.
- News summarization and evaluation in the era of gpt-3.
- Jessica A Grieser. 2022. The Black side of the river: Race, language, and belonging in Washington, DC. Georgetown University Press.
- Newsroom: A dataset of 1.3 million summaries with diverse extractive strategies. arXiv preprint arXiv:1804.11283.
- David Herman. 2009. Basic elements of narrative. John Wiley & Sons.
- Teaching machines to read and comprehend. Advances in neural information processing systems, 28.
- Inspo: Writing stories with a flock of ais and humans. arXiv preprint arXiv:2311.16521.
- Creative writing with an ai-powered writing assistant: Perspectives from professional writers.
- A comprehensive evaluation of large language models on benchmark biomedical text processing tasks. Computers in Biology and Medicine, page 108189.
- Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel.
- Longeval: Guidelines for human evaluation of faithfulness in long-form summarization. arXiv preprint arXiv:2301.13298.
- Booksum: A collection of datasets for long-form narrative summarization. arXiv preprint arXiv:2105.08209.
- Exploring content selection in summarization of novel chapters. arXiv preprint arXiv:2005.01840.
- Benchmarking generation and evaluation capabilities of large language models for instruction controllable summarization. arXiv preprint arXiv:2311.09184.
- Unveiling the essence of poetry: Introducing a comprehensive dataset and benchmark for poem summarization. In The 2023 Conference on Empirical Methods in Natural Language Processing.
- Jean M Mandler and Nancy S Johnson. 1977. Remembrance of things parsed: Story structure and recall. Cognitive psychology, 9(1):111–151.
- Factscore: Fine-grained atomic evaluation of factual precision in long form text generation. arXiv preprint arXiv:2305.14251.
- Toni Morrison. 2004. Beloved. 1987. New York: Vintage.
- OpenAI. 2023. Gpt-4 technical report.
- Does writing with language models reduce content diversity? arXiv preprint arXiv:2309.05196.
- Alison H Paris and Scott G Paris. 2003. Assessing narrative comprehension in young children. Reading Research Quarterly, 38(1):36–76.
- Narrative theory for computational narrative understanding. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 298–311.
- Summarization is (almost) dead. arXiv preprint arXiv:2309.09558.
- Understanding factual errors in summarization: Errors, summarizers, datasets, error detectors. arXiv preprint arXiv:2205.12854.
- Tofueval: Evaluating hallucinations of llms on topic-focused dialogue summarization.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
- Squality: Building a long-document summarization dataset the hard way. arXiv preprint arXiv:2205.11465.
- Recursively summarizing books with human feedback. arXiv preprint arXiv:2109.10862.
- Opentom: A comprehensive benchmark for evaluating theory-of-mind reasoning capabilities of large language models. arXiv preprint arXiv:2402.06044.
- Fantastic questions and where to find them: Fairytaleqa–an authentic dataset for narrative comprehension. arXiv preprint arXiv:2203.13947.
- Ghostwriter: Augmenting collaborative human-ai writing experiences through personalization and agency. arXiv preprint arXiv:2402.08855.
- Wordcraft: story writing with large language models. In 27th International Conference on Intelligent User Interfaces, pages 841–852.
- Mug: A general meeting understanding and generation benchmark. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE.
- Benchmarking large language models for news summarization. arXiv preprint arXiv:2301.13848.
- Fiction-writing mode: An effective control for human-machine collaborative writing. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, pages 1744–1757.