Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

ArabianGPT: Native Arabic GPT-based Large Language Model

Version 1 : Received: 24 February 2024 / Approved: 26 February 2024 / Online: 26 February 2024 (09:29:02 CET)

How to cite: Koubaa, A.; Ammar, A.; Ghouti, L.; Necar, O.; Sibaee, S. ArabianGPT: Native Arabic GPT-based Large Language Model. Preprints 2024, 2024021409. https://doi.org/10.20944/preprints202402.1409.v1 Koubaa, A.; Ammar, A.; Ghouti, L.; Necar, O.; Sibaee, S. ArabianGPT: Native Arabic GPT-based Large Language Model. Preprints 2024, 2024021409. https://doi.org/10.20944/preprints202402.1409.v1

Abstract

The predominance of English and Latin-based large language models (LLMs) has led to a notable deficit in native Arabic LLMs. This discrepancy is accentuated by the prevalent inclusion of English tokens in existing Arabic models, detracting from their efficacy in processing native Arabic's intricate morphology and syntax. Consequently, there is a theoretical and practical imperative for developing LLMs predominantly focused on Arabic linguistic elements. To address this gap, this paper proposes ArabianGPT, a series of transformer-based models within the ArabianLLM suite designed explicitly for Arabic. These models, including ArabianGPT-0.1B and ArabianGPT-0.3B, vary in size and complexity, aligning with the nuanced linguistic characteristics of Arabic. The AraNizer tokenizer, integral to these models, addresses the unique morphological aspects of Arabic script, ensuring more accurate text processing. Empirical results from fine-tuning the models on tasks like sentiment analysis and summarization demonstrate significant improvements. For sentiment analysis, the fine-tuned ArabianGPT-0.1B model achieved a remarkable accuracy of 95%, a substantial increase from the base model's 56%. Similarly, in summarization tasks, fine-tuned models showed enhanced F1 scores, indicating improved precision and recall in generating concise summaries. Comparative analysis of fine-tuned ArabianGPT models against their base versions across various benchmarks reveals nuanced differences in performance, with fine-tuning positively impacting specific tasks like question answering and summarization. These findings underscore the efficacy of fine-tuning in aligning ArabianGPT models more closely with specific NLP tasks, highlighting the potential of tailored transformer architectures in advancing Arabic NLP.

Keywords

Large Language Models, Natural Language Processing, Transformers, Arabic Language, Deep Learning

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.