Emergent Mind

Abstract

Generative, multimodal artificial intelligence (GenAI) offers transformative potential across industries, but its misuse poses significant risks. Prior research has shed light on the potential of advanced AI systems to be exploited for malicious purposes. However, we still lack a concrete understanding of how GenAI models are specifically exploited or abused in practice, including the tactics employed to inflict harm. In this paper, we present a taxonomy of GenAI misuse tactics, informed by existing academic literature and a qualitative analysis of approximately 200 observed incidents of misuse reported between January 2023 and March 2024. Through this analysis, we illuminate key and novel patterns in misuse during this time period, including potential motivations, strategies, and how attackers leverage and abuse system capabilities across modalities (e.g. image, text, audio, video) in the wild.

Top strategies linked to each misuse goal.

Overview

  • The paper identifies and categorizes misuse tactics of generative multimodal AI (GenAI) through an analysis of around 200 incidents reported between January 2023 and March 2024.

  • It emphasizes two primary misuse areas: exploitation of GenAI capabilities to create hyper-realistic outputs and attacks on GenAI systems to exploit their vulnerabilities.

  • Key findings reveal that most misuse incidents involved simple, accessible GenAI features like impersonation and sockpuppeting, and outlined the ethical and policy challenges posed by these activities.

Generative AI Misuse: A Taxonomy of Tactics and Insights from Real-World Data

The paper "Generative AI Misuse: A Taxonomy of Tactics and Insights from Real-World Data" focuses on the exploitation of generative multimodal artificial intelligence (GenAI) capabilities, presenting a carefully constructed taxonomy of misuse tactics based on a qualitative analysis of approximately 200 reported incidents between January 2023 and March 2024. This research strives to bridge the gap in understanding how GenAI models are being exploited in practice, identifying the specific tactics and strategies applied by malicious actors.

Taxonomy of GenAI Misuse Tactics

The authors categorize the misuse tactics into two core areas: (1) exploitation of GenAI capabilities and (2) compromising GenAI systems.

Exploitation of GenAI Capabilities: The paper identifies ten distinct tactics leveraging GenAI capabilities to create hyper-realistic outputs across modalities such as text, image, audio, and video.

  1. Impersonation: Generating audio or video clips to replicate individuals, usually public figures, for real-time deception.
  2. Appropriated Likeness: Modifying static depictions of real people to fabricate actions or characteristics.
  3. Sockpuppeting: Creating entirely synthetic personas to simulate human interaction.
  4. NCII and CSAM: Generating non-consensual intimate imagery, infringing on individuals' privacy, often with severe ethical and legal ramifications.
  5. IP Infringement: Replicating intellectual property without authorization.
  6. Counterfeit: Mimicking original works or styles and falsely representing them as authentic.
  7. Falsification: Producing synthetic content that fabricates events, places, or objects.
  8. Scaling: Deploying large networks of fake profiles to generate and distribute content at scale.
  9. Amplification: Enhancing the reach and engagement of content through automated interactions.
  10. Targeting: Using GenAI outputs to create personalized and targeted messaging for specific demographics.

Compromising GenAI Systems: These tactics focus on attacking GenAI models to exploit vulnerabilities within their architecture or data.

  1. Adversarial Inputs: Modifying input data to induce model errors.
  2. Prompt Injections: Manipulating text instructions to bypass security filters.
  3. Jailbreaking: Removing model restrictions entirely to generate harmful outputs.
  4. Model Diversion: Repurposing open-source models for unintended and often malicious activities.
  5. Steganography: Hiding covert messages within GenAI outputs.
  6. Data Poisoning: Corrupting training datasets to cause systematic model errors.
  7. Privacy Compromise: Revealing sensitive or private information from the training data.
  8. Data and Model Extraction: Illicitly obtaining model parameters, architecture, or training data.

Findings and Implications

The findings highlight the predominance of misuse tactics exploiting GenAI capabilities over direct attacks on systems. Approximately 90% of reported misuse incidents involved leveraging GenAI for manipulation of human likeness, including impersonation and sockpuppeting. Recognizable misuse tactics encompassed falsification of content, non-consensual creation of intimate imagery, and fraudulent schemes, often with minimal technical sophistication.

The implications of these findings suggest that while concerns about highly sophisticated state-sponsored attacks are prevalent, most observed misuse involves simple, accessible GenAI capabilities. Consequently, this democratization of technology has widened participation in misuse activities, including by those without significant technical expertise.

Significant misuse cases focused on attempts to shape public opinion and manipulate political perceptions, leveraging GenAI for highly personalized and emotionally charged outputs. Additionally, monetization-oriented misuse highlighted the financial motivations behind content farming and the creation of non-consensual intimate imagery for profit.

The paper also underscores novel lower-level forms of misuse that blur ethical standards and present new challenges for trust and safety teams. These include the use of GenAI for political image cultivation and subtle deceptive practices that do not obviously violate content policies but still prompt significant ethical concerns.

Conclusion

This paper’s rigorous analysis and resulting taxonomy provide a vital reference for policymakers, researchers, and industry practitioners seeking to understand and counteract the misuse of GenAI. By detailing specific tactics and strategies, it guides the development of more effective mitigations and governance frameworks tailored to the rapidly evolving threat landscape posed by generative AI technologies.

While technical advancements can address certain vulnerabilities, the inherently social nature of many misuse tactics calls for broad, user-facing interventions, including psychological strategies like prebunking. The dynamic nature of GenAI capabilities necessitates continuous monitoring and updating of strategies to remain effective against emerging misuse patterns.

The paper lays the groundwork for further longitudinal research to capture the evolution of GenAI misuse. It stresses the importance of ethical considerations in the deployment and regulation of GenAI while offering actionable insights into protecting the integrity of information ecosystems in the digital age.

Create an account to read this summary for free:

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.

YouTube