Abstract
Language embeds information about social, cultural, and political values people hold. Prior work has explored potentially harmful social biases encoded in Pre-trained Language Models (PLMs). However, there has been no systematic study investigating how values embedded in these models vary across cultures. In this paper, we introduce probes to study which cross-cultural values are embedded in these models, and whether they align with existing theories and cross-cultural values surveys. We find that PLMs capture differences in values across cultures, but those only weakly align with established values surveys. We discuss implications of using mis-aligned models in cross-cultural settings, as well as ways of aligning PLMs with values surveys.
Original language | English |
---|---|
Title of host publication | Proceedings of the First Workshop on Cross-Cultural Considerations in NLP (C3NLP) |
Publisher | Association for Computational Linguistics (ACL) |
Publication date | 2023 |
Pages | 114-130 |
DOIs | |
Publication status | Published - 2023 |
Event | 1st Workshop on Cross-Cultural Considerations in NLP, C3NLP 2023 - Dubrovnik, Croatia Duration: 5 May 2023 → … |
Conference
Conference | 1st Workshop on Cross-Cultural Considerations in NLP, C3NLP 2023 |
---|---|
Country/Territory | Croatia |
City | Dubrovnik |
Period | 05/05/2023 → … |