Papers
arxiv:2505.24119

The State of Multilingual LLM Safety Research: From Measuring the Language Gap to Mitigating It

Published on May 30
· Submitted by yongzx on Jun 2
Authors:
,
,
,
,

Abstract

The analysis reveals a significant language gap in LLM safety research, focusing mainly on English, and suggests recommendations for multilingual safety evaluation and crosslingual generalization to enhance global AI safety.

AI-generated summary

This paper presents a comprehensive analysis of the linguistic diversity of LLM safety research, highlighting the English-centric nature of the field. Through a systematic review of nearly 300 publications from 2020--2024 across major NLP conferences and workshops at *ACL, we identify a significant and growing language gap in LLM safety research, with even high-resource non-English languages receiving minimal attention. We further observe that non-English languages are rarely studied as a standalone language and that English safety research exhibits poor language documentation practice. To motivate future research into multilingual safety, we make several recommendations based on our survey, and we then pose three concrete future directions on safety evaluation, training data generation, and crosslingual safety generalization. Based on our survey and proposed directions, the field can develop more robust, inclusive AI safety practices for diverse global populations.

Community

Paper submitter

Multilingual safety training/eval is now standard practice, but a critical question remains: Is multilingual safety actually solved?

Our new survey answers this and dives deep into:

  • Language gap in safety research
  • Future priority areas

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2505.24119 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2505.24119 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.