Emergent Mind

Correcting misinformation on social media with a large language model

(2403.11169)
Published Mar 17, 2024 in cs.CL and cs.AI

Abstract

Real-world misinformation can be partially correct and even factual but misleading. It undermines public trust in science and democracy, particularly on social media, where it can spread rapidly. High-quality and timely correction of misinformation that identifies and explains its (in)accuracies has been shown to effectively reduce false beliefs. Despite the wide acceptance of manual correction, it is difficult to be timely and scalable, a concern as technologies like LLMs make misinformation easier to produce. LLMs also have versatile capabilities that could accelerate misinformation correction-however, they struggle due to a lack of recent information, a tendency to produce false content, and limitations in addressing multimodal information. We propose MUSE, an LLM augmented with access to and credibility evaluation of up-to-date information. By retrieving evidence as refutations or contexts, MUSE identifies and explains (in)accuracies in a piece of content-not presupposed to be misinformation-with references. It also describes images and conducts multimodal searches to verify and correct multimodal content. Fact-checking experts evaluate responses to social media content that are not presupposed to be (non-)misinformation but broadly include incorrect, partially correct, and correct posts, that may or may not be misleading. We propose and evaluate 13 dimensions of misinformation correction quality, ranging from the accuracy of identifications and factuality of explanations to the relevance and credibility of references. The results demonstrate MUSE's ability to promptly write high-quality responses to potential misinformation on social media-overall, MUSE outperforms GPT-4 by 37% and even high-quality responses from laypeople by 29%. This work reveals LLMs' potential to help combat real-world misinformation effectively and efficiently.

Muse uses LLMs, image recognition, and web search to generate corrections for misinformation with credible references.

Overview

  • This paper introduces Muse, a novel approach that employs a Large Language Model (LLM) for scalable correction of misinformation on social media, augmented with up-to-date, credible web knowledge.

  • Muse can correct both textual and visual misinformation by integrating visual content descriptions and retrieving relevant, credible web pages for evidence-backed corrections.

  • In an evaluation involving experts, Muse's corrections were found to be 37% more effective than those generated by GPT-4 and 29% more than high-quality laypeople's corrections, in terms of factuality and credibility.

  • The development of Muse demonstrates a significant advancement in automated misinformation correction, promising a scalable, efficient solution to the misinformation challenge on social media platforms.

Leveraging LLMs for Scalable Misinformation Correction on Social Media

Introduction

The proliferation of misinformation on social media platforms poses a significant challenge to public perception, often undermining trust in science and democracy. Traditional methods of misinformation correction involve expert and layperson intervention, which although effective, cannot scale to address the volume of misinformation generated daily. This limitation is exacerbated by the evolution of LLMs that, while facilitating misinformation creation, also bear potential for scalable misinformation correction. The paper introduces Muse, a novel approach utilizing an LLM augmented with access to and credibility evaluation of up-to-date information for multimodal misinformation correction on social media. Muse demonstrates superior performance in correcting misinformation compared to GPT-4 and high-quality corrections from laypeople.

Approach

Muse's design enables it to address not only textual but also visual misinformation through the integration of visuals, up-to-date factual, and credible web knowledge retrieval. The process begins with image descriptions utilizing image captioning models enhanced by celebrity recognition and Optical Character Recognition (OCR) for more informative interpretation of visual content. For textual misinformation, relevant web pages are retrieved using generated queries and filtered based on direct relevance and credibility evaluation. Muse generates corrections leveraging extracted evidence from web pages, ensuring accurate, trustworthy references and explanations.

Evaluation

An extensive evaluation involving experts in fact-checking and journalism assessed corrections generated by Muse across 13 dimensions. These included the factuality of explanation, relevance and credibility of references, and overall quality of corrections among others. Muse's corrections outperformed those generated by GPT-4 by 37% and laypeople's high-quality corrections by 29%, showcasing its capability to promptly correct misinformation after it appears on social media. Particularly, Muse excelled in identifying inaccuracies, generating relevant and factual text, and providing credible references.

Implications and Future Directions

The findings underscore the potential of Muse and similar technologies in combating misinformation on social media platforms both effectively and efficiently. The approach outlined not only addresses the scalability issue faced by manual corrections but also introduces a method to enhance the accuracy and trustworthiness of generated corrections. Future research could explore the integration of video inputs, application across multiple languages, and extend the evaluation to other platforms beyond X Community Notes. Moreover, further developments might aim to reduce the correction generation cost and time, already estimated at $0.5 per social media post, and examine the impact of correction immediacy on Muse's performance.

Conclusion

The development of Muse represents a significant advancement in the use of LLMs for misinformation correction on social media. By integrating capabilities for handling multimodal misinformation, accessing up-to-date information, and generating corrections with accurate references, Muse sets a new standard for automated misinformation correction technologies. Its superior performance, demonstrated through a comprehensive expert evaluation, highlights its potential as a scalable solution to the misinformation problem that plagues social media platforms.

Create an account to read this summary for free:

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.