You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this dataset content.

SAHARA Benchmark



ACL Paper Website HuggingFace Dataset HuggingFace Leaderboard GitHub

Sahara is a comprehensive benchmark for African NLP, part of our ACL 2025 paper, "Where Are We? Evaluating LLM Performance on African Languages". Africa's rich linguistic heritage remains underrepresented in NLP, largely due to historical policies that favor foreign languages and create significant data inequities. In the paper, we integrate theoretical insights on Africa's language landscape with an empirical evaluation using Sahara. Sahara is curated from large-scale, publicly accessible datasets capturing the continent's linguistic diversity. By systematically assessing the performance of leading large language models (LLMs) on Sahara, we demonstrate how policy-induced data variations directly impact model effectiveness across African languages. Our findings reveal that while a few languages perform reasonably well, many Indigenous languages remain marginalized due to sparse data. Sahara supports an impressive 517 languages and varieties, across 16 tasks, making it the most extensive and representative benchmark for African NLP.

Official Website Sahara Official Website
Paper: Where Are We? Evaluating LLM Performance on African Languages
Leaderboards Sahara Leaderboards
GITHUB: https://github.com/UBC-NLP/sahara\

How to Use the Dataset

You can easily load and explore the SAHARA benchmark using the datasets library from Hugging Face.

All information about usage, the evaluation, and the scoring system is available on the official website.

Citation

If you use the Sahara benchmark for your scientific publication, or if you find the resources in this website useful, please cite our paper.


@inproceedings{adebara-etal-2025-evaluating,
    title = "Where Are We? Evaluating {LLM} Performance on {A}frican Languages",
    author = "Adebara, Ife  and
      Toyin, Hawau Olamide  and
      Ghebremichael, Nahom Tesfu  and
      Elmadany, AbdelRahim A.  and
      Abdul-Mageed, Muhammad",
    editor = "Che, Wanxiang  and
      Nabende, Joyce  and
      Shutova, Ekaterina  and
      Pilehvar, Mohammad Taher",
    booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
    month = jul,
    year = "2025",
    address = "Vienna, Austria",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2025.acl-long.1572/",
    pages = "32704--32731",
    ISBN = "979-8-89176-251-0",
}
Downloads last month
12

Space using UBC-NLP/sahara_benchmark 1