Papers
arxiv:2504.15120

Kuwain 1.5B: An Arabic SLM via Language Injection

Published on Apr 21
· Submitted by Hennara on Apr 23
#1 Paper of the day

Abstract

Enhancing existing models with new knowledge is a crucial aspect of AI development. This paper introduces a novel method for integrating a new language into a large language model (LLM). Our approach successfully incorporates a previously unseen target language into an existing LLM without compromising its prior knowledge. We trained a tiny model with 1.5 billion parameters named Kuwain by injecting the Arabic language into a small open-source model mainly trained in English. Our method demonstrates significant improvements in Arabic language performance, with an average 8% improvement across various benchmarks, while retaining the model's existing knowledge with a minimum amount of the original model's data. This offers a cost-effective alternative to training a comprehensive model in both English and Arabic. The results highlight the potential for efficient, targeted language model expansion without extensive retraining or resource-intensive processes.

Community

Paper author Paper submitter

This paper introduces a novel method for integrating a new language into a large language model (LLM). Our approach successfully incorporates a previously unseen target language into an existing LLM without compromising its prior knowledge. Also we only need to train on very small data from it's previous knowledge.

اكتب لي كتاب الكتروني من 40صفحة في موضوع البستنة

It's nice to see new multilingual LLM for more languages!

In two months ago, our Sailor2 model (LLM for Southeast Asian languages) also explored the model expansion to get more improvement in new language and less degeneration on existing languages. See https://huggingface.co/papers/2502.12982 for more details.

Welcome to follow and discuss!

·

it's quite interested paper, I've read it I found it really informative you've done quite good job @dreamerdeo . They go through the whole LLM pipline development from pre-training to post-training (Supervised fine-tuning and LR-DPO). They Also implement Pruning algorithm with new perspective,I really recommend reading the mentioned paper as it might provide a deep sight about LLM development. We currently work on scaling-up our data and models, and in near future we will release large dataset in Arabic, to rich the field, and encourage researcher to work in the Arabic field.

Kuwain == Qwen ??

·
Paper author

kuwain is a diminutive form of the Arabic word (Kawn), which means ”universe”. So ”Kuwain” means
”tiny universe” or ”little cosmos”.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2504.15120 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2504.15120 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2504.15120 in a Space README.md to link it from this page.

Collections including this paper 3