Abstract
The widespread interest in large language models (LLMs) is rooted in their remarkable capacity to generate human-like and contextually relevant responses. However, the precision of LLMs within specific domains or intricate tasks, such as Arabic dialect identification, remains largely unexplored. This task presents a substantial challenge in Arabic natural language processing, given its language-dependent nature. This paper provides a framework for evaluating LLMs for Arabic dialect identification and conducts a comprehensive evaluation of LLMs, employing both tuning-free and fine-tuning-based learning paradigms. The evaluation encompasses GPT-3.5, chatGPT, GPT-4, and Google BARD for Arabic dialect identification under the tuning-free learning paradigm. Furthermore, it assesses the performance of GPT-3.5 along with AraBERT and MARBERT using the fine-tuning learning paradigm. In the tuning-free approach, GPT-4 achieves the most favorable results, reporting an F1MAC of 45.60%. Under the fine-tuning learning paradigm, both AraBERT and MARBERT exhibit comparable performance (around 50% F 1MAC) to GPT-3.5, without incurring any financial costs, in contrast to the expenses associated with GPT-3.5.
Original language | English |
---|---|
Title of host publication | 2024 International Joint Conference on Neural Networks, IJCNN 2024 - Proceedings |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
ISBN (Electronic) | 9798350359312 |
DOIs | |
State | Published - 2024 |
Event | 2024 International Joint Conference on Neural Networks, IJCNN 2024 - Yokohama, Japan Duration: 30 Jun 2024 → 5 Jul 2024 |
Publication series
Name | Proceedings of the International Joint Conference on Neural Networks |
---|
Conference
Conference | 2024 International Joint Conference on Neural Networks, IJCNN 2024 |
---|---|
Country/Territory | Japan |
City | Yokohama |
Period | 30/06/24 → 5/07/24 |
Bibliographical note
Publisher Copyright:© 2024 IEEE.
Keywords
- Arabic dialect identification
- Evaluation LLMs
- Instruction fine-tuning
- Large language models
- Prompt engineering
ASJC Scopus subject areas
- Software
- Artificial Intelligence