RESEARCH PAPER

Global Security Blind Spots: LLM Safety Failures in Low-Resource Languages

In Summary

By evaluating harmful prompts across 79 languages under controlled conditions, the study shows that models which appear well aligned in English frequently produce unsafe responses in low-resource languages. In many cases, simple translation alone is enough to bypass existing safety guardrails. 

What you’ll learn

  • How unsafe response rates change when prompts move from English to low-resource languages 
  • Why translation can function as a practical jailbreak for modern LLMs 
  • Which language families exhibit the highest safety risk 
  • Which harm categories show the largest cross-lingual safety gaps 
  • How English-only safety benchmarks can misrepresent real-world deployment risk 

This research is intended to support responsible AI deployment and is shared with appropriate safeguards.