top of page


AI Astrolabe Journal
Reflections, Research, and Roadmaps


Testing Arabic LLM Safety with Saudi Dialect: A Follow-up Study on ASAS
Manar Alharbi, Asma Alsakkaf, Reem Bazarah and AI Astrolabe team (work done as interns at AI Astrolabe) Introduction In our previous research , we introduced ASAS (أساس) - the first human-rated Arabic safety benchmark for evaluating LLMs in Modern Standard Arabic (MSA). The results revealed significant safety gaps across frontier models, with the best-performing model achieving only 68% safe responses. Building on that work, we now present a critical follow-up study: evaluati
Jan 236 min read


Redteaming Frontier LLMs with AI Astrolabe Arabic Safety Index (ASAS - أساس)
Arabic is spoken by hundreds of millions, yet most models fall short; not just in fluency, but in safety, cultural grounding, and contextual understanding.
Apr 25, 202512 min read
bottom of page