Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

InkubaLM: A small language model for low-resource African languages

About

High-resource language models often fall short in the African context, where there is a critical need for models that are efficient, accessible, and locally relevant, even amidst significant computing and data constraints. This paper introduces InkubaLM, a small language model with 0.4 billion parameters, which achieves performance comparable to models with significantly larger parameter counts and more extensive training data on tasks such as machine translation, question-answering, AfriMMLU, and the AfriXnli task. Notably, InkubaLM outperforms many larger models in sentiment analysis and demonstrates remarkable consistency across multiple languages. This work represents a pivotal advancement in challenging the conventional paradigm that effective language models must rely on substantial resources. Our model and datasets are publicly available at https://huggingface.co/lelapa to encourage research and development on low-resource languages.

Atnafu Lambebo Tonja, Bonaventure F. P. Dossou, Jessica Ojo, Jenalea Rajab, Fadel Thior, Eric Peter Wairagala, Anuoluwapo Aremu, Pelonomi Moiloa, Jade Abbott, Vukosi Marivate, Benjamin Rosman• 2024

Related benchmarks

TaskDatasetResultRank
Part-of-Speech TaggingMasakhaPOS isiXhosa
Token Accuracy0.00e+0
12
Part-of-Speech TaggingMasakhaPOS isiZulu
Token Accuracy0.00e+0
12
Part-of-Speech TaggingMasakhaPOS Setswana
Token Accuracy0.00e+0
12
Named Entity RecognitionMasakhaNER isiXhosa 2.0
Macro F10.1
11
Named Entity RecognitionMasakhaNER 2.0
Macro-F1 Score0.00e+0
11
Named Entity RecognitionMasakhaNER Setswana 2.0
Macro-F1 Score0.00e+0
11
Topic ClassificationSIB-200
Accuracy (Xho)8.4
11
Intent ClassificationINJONGO Intent
Accuracy (Eng)0.4
11
Topic ClassificationMasakhaNEWS English
Macro-F120.3
11
Topic ClassificationMasakhaNEWS isiXhosa
Macro F17.4
11
Showing 10 of 14 rows

Other info

Follow for update