The State of Multilingual LLM Safety Research: From Measuring the Language Gap to Mitigating It
Abstract
The analysis reveals a significant language gap in LLM safety research, focusing mainly on English, and suggests recommendations for multilingual safety evaluation and crosslingual generalization to enhance global AI safety.
This paper presents a comprehensive analysis of the linguistic diversity of LLM safety research, highlighting the English-centric nature of the field. Through a systematic review of nearly 300 publications from 2020--2024 across major NLP conferences and workshops at *ACL, we identify a significant and growing language gap in LLM safety research, with even high-resource non-English languages receiving minimal attention. We further observe that non-English languages are rarely studied as a standalone language and that English safety research exhibits poor language documentation practice. To motivate future research into multilingual safety, we make several recommendations based on our survey, and we then pose three concrete future directions on safety evaluation, training data generation, and crosslingual safety generalization. Based on our survey and proposed directions, the field can develop more robust, inclusive AI safety practices for diverse global populations.
Community
Multilingual safety training/eval is now standard practice, but a critical question remains: Is multilingual safety actually solved?
Our new survey answers this and dives deep into:
- Language gap in safety research
- Future priority areas
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Multilingual Prompt Engineering in Large Language Models: A Survey Across NLP Tasks (2025)
- The Multilingual Divide and Its Impact on Global AI Safety (2025)
- Refusal Direction is Universal Across Safety-Aligned Languages (2025)
- PolyGuard: A Multilingual Safety Moderation Tool for 17 Languages (2025)
- M-Prometheus: A Suite of Open Multilingual LLM Judges (2025)
- Multilingual Collaborative Defense for Large Language Models (2025)
- Kaleidoscope: In-language Exams for Massively Multilingual Vision Evaluation (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper