NormAd: A Framework for Measuring the Cultural Adaptability of Large Language Models (2404.12464v7)
Abstract: To be effectively and safely deployed to global user populations, LLMs must adapt outputs to user values and culture, not just know about them. We introduce NormAd, an evaluation framework to assess LLMs' cultural adaptability, specifically measuring their ability to judge social acceptability across different levels of cultural norm specificity, from abstract values to explicit social norms. As an instantiation of our framework, we create NormAd-Eti, a benchmark of 2.6k situational descriptions representing social-etiquette related cultural norms from 75 countries. Through comprehensive experiments on NormAd-Eti, we find that LLMs struggle to accurately judge social acceptability across these varying degrees of cultural contexts and show stronger adaptability to English-centric cultures over those from the Global South. Even in the simplest setting where the relevant social norms are provided, our best models' performance (<82%) lags behind humans (>95%). In settings with abstract values and country information, model performance drops substantially (<60%), while human accuracy remains high (>90%). Furthermore, we find that models are better at recognizing socially acceptable versus unacceptable situations. Our findings showcase the current pitfalls in socio-cultural reasoning of LLMs which hinder their adaptability for global audiences.
- Investigating cultural alignment of large language models, 2024.
- Which humans?, Sep 2023. URL osf.io/preprints/psyarxiv/5b26t.
- Training a helpful and harmless assistant with reinforcement learning from human feedback, 2022.
- On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAccT ’21, pp. 610–623, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 9781450383097. doi: 10.1145/3442188.3445922. URL https://doi.org/10.1145/3442188.3445922.
- Assessing llms for moral value pluralism. arXiv preprint arXiv:2312.10075, 2023.
- Culturalteaming: Ai-assisted interactive red-teaming for challenging llms’ (lack of) multicultural knowledge. 2024. URL https://api.semanticscholar.org/CorpusID:269032901.
- Eticor: Corpus for analyzing llms for etiquettes, 2023.
- Kto: Model alignment as prospect theoretic optimization. arXiv preprint arXiv:2402.01306, 2024.
- The cultural atlas. https://culturalatlas.sbs.com.au/, 2024.
- Directions for NLP practices applied to online hate speech detection. In EMNLP, pp. 11794–11805, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics. URL https://aclanthology.org/2022.emnlp-main.809.
- Improving alignment of dialogue agents via targeted human judgements. arXiv preprint arXiv:2209.14375, 2022.
- Reporting bias and knowledge acquisition. In Proceedings of the 2013 workshop on Automated knowledge base construction, pp. 25–30, 2013.
- Cultural Aspects of Gift Giving: A Comparative Analysis of the Significance of Gift Giving in the U.S. and Japan. In Proceedings of the 1997 World Marketing Congress, pp. 283–287. Springer, Cham, Switzerland, 2015. ISBN 978-3-319-17320-7. doi: 10.1007/978-3-319-17320-7˙78.
- Trust, comfort and relatability: Understanding black older adults’ perceptions of chatbot design for health information seeking. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, CHI ’23, New York, NY, USA, 2023. Association for Computing Machinery. ISBN 9781450394215. doi: 10.1145/3544548.3580719. URL https://doi.org/10.1145/3544548.3580719.
- Aligning ai with shared human values, 2023.
- Geert Hofstede. Culture’s Consequences: Comparing Values, Behaviors, Institutions and Organizations Across Nations. SAGE Publications, Inc, 1980.
- The cultural map of the world. https://www.worldvaluessurvey.org/WVSContents.jsp?CMSID=Findings, 2023. [Accessed 03-29-2024].
- The ghost in the machine has an american accent: value conflict in gpt-3. arXiv preprint arXiv:2203.07785, 2022.
- Personalisation within bounds: A risk taxonomy and policy framework for the alignment of large language models with personalised feedback, 2023.
- The Current Formulation of Kohlberg’s Theory and a Response to Critics. Human Development, 28(2):94–100, 12 2009. ISSN 0018-716X. doi: 10.1159/000272945. URL https://doi.org/10.1159/000272945.
- Culturellm: Incorporating cultural differences into large language models. arXiv preprint arXiv:2402.10946, 2024.
- Cultural alignment in large language models: An explanatory analysis based on hofstede’s cultural dimensions. ArXiv, abs/2309.12342, 2023. URL https://api.semanticscholar.org/CorpusID:262216989.
- Andrew Molinsky. Cross-cultural code-switching: The psychological challenges of adapting behavior in foreign cultural interactions. Academy of management review, 32(2):622–640, 2007.
- Having beer after prayer? measuring cultural bias in large language models. ArXiv, abs/2305.14456, 2023. URL https://api.semanticscholar.org/CorpusID:258865272.
- Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35:27730–27744, 2022.
- Discovering language model behaviors with model-written evaluations, 2022.
- Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36, 2024.
- Knowledge of cultural moral norms in large language models. In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 428–446, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.acl-long.26. URL https://aclanthology.org/2023.acl-long.26.
- Ethical reasoning over moral alignment: A case and framework for in-context ethical policies in LLMs. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Findings of the Association for Computational Linguistics: EMNLP 2023, pp. 13370–13388, Singapore, December 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.findings-emnlp.892. URL https://aclanthology.org/2023.findings-emnlp.892.
- Sharon Richardson. Against generalisation: Data-driven decisions need context to be human-compatible. Business Information Review, 38(4):162–169, December 2021. URL https://doi.org/10.1177/02663821211061986.
- Unintended impacts of llm alignment on global representation. arXiv preprint arXiv:2402.15018, 2024.
- The risk of racial bias in hate speech detection. In Proceedings of the 57th annual meeting of the association for computational linguistics, pp. 1668–1678, 2019.
- Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
- Shalom H. Schwartz. An overview of the schwartz theory of basic values. Online Readings in Psychology and Culture, 2, 12 2012. doi: 10.9707/2307-0919.1116.
- Bernd Stauss. Gifts and Culture: What Applies Globally and What Regionally? In Psychology of Gift-Giving, pp. 161–176. Springer, Berlin, Germany, January 2023. ISBN 978-3-662-66393-6. doi: 10.1007/978-3-662-66393-6˙13.
- Learning to summarize with human feedback. Advances in Neural Information Processing Systems, 33:3008–3021, 2020.
- Trustllm: Trustworthiness in large language models, 2024.
- Phantom: Personality has an effect on theory-of-mind reasoning in large language models, 2024.
- Probing the moral development of large language models through defining issues test, 2023.
- Designing for harm reduction: Communication repair for multicultural users’ voice interactions, 2024.
- WVS. WVS Database — worldvaluessurvey.org. https://www.worldvaluessurvey.org/wvs.jsp, 1981. [Accessed 07-12-2023].
- Value fulcra: Mapping large language models to the multidimensional spectrum of basic human values, 2023.
- “don’t take this out of context!” on the need for contextual models and evaluations for stylistic rewriting. In EMNLP, 2023. URL https://arxiv.org/abs/2305.14755.
- Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593, 2019.
- Abhinav Rao (8 papers)
- Akhila Yerukola (14 papers)
- Vishwa Shah (6 papers)
- Katharina Reinecke (16 papers)
- Maarten Sap (87 papers)