GPT detectors are biased against non-native English writers
Abstract: The rapid adoption of generative LLMs has brought about substantial advancements in digital communication, while simultaneously raising concerns regarding the potential misuse of AI-generated content. Although numerous detection methods have been proposed to differentiate between AI and human-generated content, the fairness and robustness of these detectors remain underexplored. In this study, we evaluate the performance of several widely-used GPT detectors using writing samples from native and non-native English writers. Our findings reveal that these detectors consistently misclassify non-native English writing samples as AI-generated, whereas native writing samples are accurately identified. Furthermore, we demonstrate that simple prompting strategies can not only mitigate this bias but also effectively bypass GPT detectors, suggesting that GPT detectors may unintentionally penalize writers with constrained linguistic expressions. Our results call for a broader conversation about the ethical implications of deploying ChatGPT content detectors and caution against their use in evaluative or educational settings, particularly when they may inadvertently penalize or exclude non-native English speakers from the global discourse. The published version of this study can be accessed at: www.cell.com/patterns/fulltext/S2666-3899(23)00130-7
- OpenAI. ChatGPT. https://chat.openai.com/ (2022). Accessed: 2022-12-31.
- Hu, K. Chatgpt sets record for fastest-growing user base - analyst note. \JournalTitleReuters (2023).
- Paris, M. Chatgpt hits 100 million users, google invests in ai bot and catgpt goes viral. \JournalTitleForbes (2023).
- Lee, M. et al. Evaluating human-language model interaction. \JournalTitlearXiv preprint arXiv:2212.09746 (2022).
- Kung, T. H. et al. Performance of chatgpt on usmle: Potential for ai-assisted medical education using large language models. \JournalTitlePLoS digital health 2, e0000198 (2023).
- Terwiesch, C. Would chat gpt3 get a wharton mba? a prediction based on its performance in the operations management course. \JournalTitleMack Institute for Innovation Management at the Wharton School, University of Pennsylvania (2023).
- Else, H. Abstracts written by chatgpt fool scientists. \JournalTitleNature (2023).
- Gao, C. A. et al. Comparing scientific abstracts generated by chatgpt to original abstracts using an artificial intelligence output detector, plagiarism detector, and blinded human reviewers. \JournalTitlebioRxiv 2022–12 (2022).
- All the news that’s fit to fabricate: Ai-generated text as a tool of media misinformation. \JournalTitleJournal of Experimental Political Science 9, 104–117, DOI: 10.1017/XPS.2020.37 (2022).
- Editorial, N. Tools such as chatgpt threaten transparent science; here are our ground rules for their use. \JournalTitleNature 613, 612–612 (2023).
- ICML. Clarification on large language model policy LLM. https://icml.cc/Conferences/2023/llm-policy (2023).
- Clark, E. et al. All that’s ‘human’is not gold: Evaluating human evaluation of generated text. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 7282–7296 (2021).
- OpenAI. GPT-2: 1.5B release. https://openai.com/research/gpt-2-1-5b-release (2019). Accessed: 2019-11-05.
- Automatic detection of machine generated text: A critical survey. \JournalTitlearXiv preprint arXiv:2011.01314 (2020).
- Tweepfake: About detecting deepfake tweets. \JournalTitlePlos one 16, e0251415 (2021).
- Automatic detection of generated text is easiest when humans are fooled. \JournalTitlearXiv preprint arXiv:1911.00650 (2019).
- DetectGPT: Zero-shot machine-generated text detection using probability curvature. \JournalTitlearXiv preprint arXiv:2301.11305 (2023).
- Solaiman, I. et al. Release strategies and the social impacts of language models. \JournalTitlearXiv preprint arXiv:1908.09203 (2019).
- Gltr: Statistical detection and visualization of generated text. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, 111–116 (2019).
- Heikkil"a, M. How to spot ai-generated text. \JournalTitleMIT Technology Review (2022).
- Machine generated text: A comprehensive survey of threat models and detection methods. \JournalTitlearXiv preprint arXiv:2210.07321 (2022).
- Rosenblatt, K. Chatgpt banned from new york city public schools’ devices and networks. \JournalTitleNBC News (2023). Accessed: 22.01.2023.
- Kasneci, E. et al. Chatgpt for good? on opportunities and challenges of large language models for education. \JournalTitleLearning and Individual Differences 103, 102274 (2023).
- Kaggle. The hewlett foundation: Automated essay scoring. https://www.kaggle.com/c/asap-aes (2012). Accessed: 2023-03-15.
- Vocabulary size and use: Lexical richness in l2 written production. \JournalTitleApplied linguistics 16, 307–322 (1995).
- Jarvis, S. Short texts, best-fitting curves and new measures of lexical diversity. \JournalTitleLanguage Testing 19, 57–84 (2002).
- Lexical richness in the spontaneous speech of bilinguals. \JournalTitleApplied linguistics 24, 197–222 (2003).
- Lu, X. A corpus-based evaluation of syntactic complexity measures as indices of college-level esl writers’ language development. \JournalTitleTESOL quarterly 45, 36–62 (2011).
- Does writing development equal writing quality? a computational investigation of syntactic complexity in l2 learners. \JournalTitleJournal of Second Language Writing 26, 66–79 (2014).
- Ortega, L. Syntactic complexity measures and their relationship to l2 proficiency: A research synthesis of college-level l2 writing. \JournalTitleApplied linguistics 24, 492–518 (2003).
- Should we use characteristics of conversation to measure grammatical complexity in l2 writing development? \JournalTitleTesol Quarterly 45, 5–35 (2011).
- Paraphrasing evades detectors of ai-generated text, but retrieval is an effective defense. \JournalTitlearXiv preprint arXiv:2303.13408 (2023).
- Can ai-generated text be reliably detected? \JournalTitlearXiv preprint arXiv:2303.11156 (2023).
- Kirchenbauer, J. et al. A watermark for large language models. \JournalTitlearXiv preprint arXiv:2301.10226 (2023).
- Watermarking pre-trained language models with backdooring. \JournalTitlearXiv preprint arXiv:2210.07543 (2022).
- ChatGPT-Detector-Bias: v1.0.0, DOI: 10.5281/zenodo.7893958 (2023).
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.