Analyzing Toxicity in Deep Conversations: A Reddit Case Study (2404.07879v1)
Abstract: Online social media has become increasingly popular in recent years due to its ease of access and ability to connect with others. One of social media's main draws is its anonymity, allowing users to share their thoughts and opinions without fear of judgment or retribution. This anonymity has also made social media prone to harmful content, which requires moderation to ensure responsible and productive use. Several methods using artificial intelligence have been employed to detect harmful content. However, conversation and contextual analysis of hate speech are still understudied. Most promising works only analyze a single text at a time rather than the conversation supporting it. In this work, we employ a tree-based approach to understand how users behave concerning toxicity in public conversation settings. To this end, we collect both the posts and the comment sections of the top 100 posts from 8 Reddit communities that allow profanity, totaling over 1 million responses. We find that toxic comments increase the likelihood of subsequent toxic comments being produced in online conversations. Our analysis also shows that immediate context plays a vital role in shaping a response rather than the original post. We also study the effect of consensual profanity and observe overlapping similarities with non-consensual profanity in terms of user behavior and patterns.
- Twitter users’ behavioral response to toxic replies. arXiv preprint arXiv:2210.13420.
- Reconstruction of threaded conversations in online discussion forums. In Proceedings of the International AAAI Conference on Web and Social Media, volume 5, pages 26–33.
- The role of social networks in information diffusion. In Proceedings of the 21st international conference on World Wide Web, pages 519–528.
- RedditBias: A real-world resource for bias evaluation and debiasing of conversational language models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1941–1955, Online. Association for Computational Linguistics.
- Nuanced metrics for measuring unintended bias with real data for text classification. In Companion proceedings of the 2019 world wide web conference, pages 491–500.
- Racial bias in hate speech and abusive language detection datasets. In Proceedings of the Third Workshop on Abusive Language Online, pages 25–35, Florence, Italy. Association for Computational Linguistics.
- Automated hate speech detection and the problem of offensive language. In Proceedings of the international AAAI conference on web and social media, volume 11, pages 512–515.
- Measuring and mitigating unintended bias in text classification. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, pages 67–73.
- Latent hatred: A benchmark for understanding implicit hate speech. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 345–363, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Large scale crowdsourcing and characterization of twitter abusive behavior. In Twelfth International AAAI Conference on Web and Social Media.
- Sarcasm analysis using conversation context. Computational Linguistics, 44(4):755–792.
- Cosyn: Detecting implicit hate speech in online conversations using a context synergized hyperbolic network. Preprint, arXiv:2303.03387.
- Public wisdom matters! discourse-aware hyperbolic fourier co-attention for social text classification. Advances in Neural Information Processing Systems, 35:9417–9431.
- Fueling toxicity? studying deceitful opinion leaders and behavioral changes of their followers. Politics and Governance, 10(4):336–348.
- ToxiGen: A large-scale machine-generated dataset for adversarial and implicit hate speech detection. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3309–3326, Dublin, Ireland. Association for Computational Linguistics.
- Harold Stanley Heaps. 1978. Information retrieval, computational and theoretical aspects. Academic Press.
- John W Jordan. 2020. Profanity from the heart as exceptional civic rhetoric. Quarterly Journal of Speech, 106(2):111–132.
- Strangers on your phone: Why people use anonymous communication applications. In Proceedings of the 19th ACM conference on computer-supported cooperative work & social computing, pages 359–370.
- Predicting continuity of online conversations on reddit. Telematics and Informatics, 79:101965.
- Dynamics of conversations. In Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 553–562.
- Alex Leavitt and Joshua A Clark. 2014. Upvoting hurricane sandy: event-based news production processes on a social news site. In Proceedings of the SIGCHI conference on human factors in computing systems, pages 1495–1504.
- Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692.
- A measurement study of hate speech in social media. In Proceedings of the 28th ACM conference on hypertext and social media, pages 85–94.
- Abusive language detection in online user content. In Proceedings of the 25th international conference on world wide web, pages 145–153.
- Comparing toxicity across social media platforms for covid-19 discourse. ArXiv, abs/2302.14270.
- Quick, community-specific learning: How distinctive toxicity norms are maintained in political subreddits. In Proceedings of the International AAAI Conference on Web and Social Media, volume 14, pages 557–568.
- HateCheck: Functional tests for hate speech detection models. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 41–58, Online. Association for Computational Linguistics.
- The risk of racial bias in hate speech detection. In Proceedings of the 57th annual meeting of the association for computational linguistics, pages 1668–1678.
- How social media users perceive different forms of online hate speech: A qualitative multi-method study. New Media & Society, page 14614448221091185.
- Namespotting: Username toxicity and actual toxic behavior on reddit. Computers in Human Behavior, 136:107371.
- Empirical analysis of multi-task learning for reducing identity bias in toxic comment detection. In Proceedings of the International AAAI Conference on Web and Social Media, volume 14, pages 683–693.
- Learning from the worst: Dynamically generated datasets to improve online hate detection. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1667–1682, Online. Association for Computational Linguistics.
- Successful new-entry prediction for multi-party online conversations via latent topics and discourse modeling. In Proceedings of the ACM Web Conference 2022, pages 1663–1672.
- Zeerak Waseem. 2016. Are you a racist or am i seeing things? annotator influence on hate speech detection on twitter. In Proceedings of the first workshop on NLP and computational social science, pages 138–142.
- Zeerak Waseem and Dirk Hovy. 2016. Hateful symbols or hateful people? predictive features for hate speech detection on Twitter. In Proceedings of the NAACL Student Research Workshop, pages 88–93, San Diego, California. Association for Computational Linguistics.
- Ex machina: Personal attacks seen at scale. In Proceedings of the 26th international conference on world wide web, pages 1391–1399.
- Demoting racial bias in hate speech detection. In Proceedings of the Eighth International Workshop on Natural Language Processing for Social Media, pages 7–14, Online. Association for Computational Linguistics.
- Detection of harassment on web 2.0. Proceedings of the Content Analysis in the WEB, 2:1–7.
- Towards developing a measure to assess contagiousness of toxic tweets.