Advancing arabic dialect detection with hybrid stacked transformer models
The rapid expansion of dialectally unique Arabic material on social media and the internet highlights how important it is to categorize dialects accurately to maximize a variety of Natural Language Processing (NLP) applications. The improvement in classification performance highlights the wider vari...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Frontiers Media S.A.
2025-02-01
|
Series: | Frontiers in Human Neuroscience |
Subjects: | |
Online Access: | https://www.frontiersin.org/articles/10.3389/fnhum.2025.1498297/full |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1823859189135441920 |
---|---|
author | Hager Saleh Hager Saleh Hager Saleh Abdulaziz AlMohimeed Rasha Hassan Mandour M. Ibrahim Saeed Hamood Alsamhi Moatamad Refaat Hassan Sherif Mostafa |
author_facet | Hager Saleh Hager Saleh Hager Saleh Abdulaziz AlMohimeed Rasha Hassan Mandour M. Ibrahim Saeed Hamood Alsamhi Moatamad Refaat Hassan Sherif Mostafa |
author_sort | Hager Saleh |
collection | DOAJ |
description | The rapid expansion of dialectally unique Arabic material on social media and the internet highlights how important it is to categorize dialects accurately to maximize a variety of Natural Language Processing (NLP) applications. The improvement in classification performance highlights the wider variety of linguistic variables that the model can capture, providing a reliable solution for precise Arabic dialect recognition and improving the efficacy of NLP applications. Recent advances in deep learning (DL) models have shown promise in overcoming potential challenges in identifying Arabic dialects. In this paper, we propose a novel stacking model based on two transformer models, i.e., Bert-Base-Arabertv02 and Dialectal-Arabic-XLM-R-Base, to enhance the classification of dialectal Arabic. The proposed model consists of two levels, including base models and meta-learners. In the proposed model, Level 1 generates class probabilities from two transformer models for training and testing sets, which are then used in Level 2 to train and evaluate a meta-learner. The stacking model compares various models, including long-short-term memory (LSTM), gated recurrent units (GRU), convolutional neural network (CNN), and two transformer models using different word embedding. The results show that the stacking model combination of two models archives outperformance over single-model approaches due to capturing a broader range of linguistic features, which leads to better generalization across different forms of Arabic. The proposed model is evaluated based on the performance of IADD and Shami. For Shami, the Stacking-Transformer achieves the highest performance in all rates compared to other models with 89.73 accuracy, 89.596 precision, 89.73 recall, and 89.574 F1-score. For IADD, the Stacking-Transformer achieves the highest performance in all rates compared to other models with 93.062 accuracy, 93.368 precision, 93.062 recall, and 93.184 F1 score. The improvement in classification performance highlights the wider variety of linguistic variables that the model can capture, providing a reliable solution for precise Arabic dialect recognition and improving the efficacy of NLP applications. |
format | Article |
id | doaj-art-88ddf5771905431abda778023cf47a3c |
institution | Kabale University |
issn | 1662-5161 |
language | English |
publishDate | 2025-02-01 |
publisher | Frontiers Media S.A. |
record_format | Article |
series | Frontiers in Human Neuroscience |
spelling | doaj-art-88ddf5771905431abda778023cf47a3c2025-02-11T06:59:47ZengFrontiers Media S.A.Frontiers in Human Neuroscience1662-51612025-02-011910.3389/fnhum.2025.14982971498297Advancing arabic dialect detection with hybrid stacked transformer modelsHager Saleh0Hager Saleh1Hager Saleh2Abdulaziz AlMohimeed3Rasha Hassan4Mandour M. Ibrahim5Saeed Hamood Alsamhi6Moatamad Refaat Hassan7Sherif Mostafa8Faculty of Computers and Artificial Intelligence, Hurghada University, Hurghada, EgyptInsight SFI Research Centre for Data Analytics, School of Engineering, University of Galway, Galway, IrelandAtlantic Technological University, Letterkenny, IrelandComputer Science Department College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi ArabiaDepartment of Computer Science, Faculty of Science, Aswan University, Aswan, EgyptInformation Technology Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi ArabiaDepartment of Computer Science and Engineering, College of Informatics, Korea University, Seoul, Republic of KoreaDepartment of Computer Science, Faculty of Science, Aswan University, Aswan, EgyptFaculty of Computers and Artificial Intelligence, Hurghada University, Hurghada, EgyptThe rapid expansion of dialectally unique Arabic material on social media and the internet highlights how important it is to categorize dialects accurately to maximize a variety of Natural Language Processing (NLP) applications. The improvement in classification performance highlights the wider variety of linguistic variables that the model can capture, providing a reliable solution for precise Arabic dialect recognition and improving the efficacy of NLP applications. Recent advances in deep learning (DL) models have shown promise in overcoming potential challenges in identifying Arabic dialects. In this paper, we propose a novel stacking model based on two transformer models, i.e., Bert-Base-Arabertv02 and Dialectal-Arabic-XLM-R-Base, to enhance the classification of dialectal Arabic. The proposed model consists of two levels, including base models and meta-learners. In the proposed model, Level 1 generates class probabilities from two transformer models for training and testing sets, which are then used in Level 2 to train and evaluate a meta-learner. The stacking model compares various models, including long-short-term memory (LSTM), gated recurrent units (GRU), convolutional neural network (CNN), and two transformer models using different word embedding. The results show that the stacking model combination of two models archives outperformance over single-model approaches due to capturing a broader range of linguistic features, which leads to better generalization across different forms of Arabic. The proposed model is evaluated based on the performance of IADD and Shami. For Shami, the Stacking-Transformer achieves the highest performance in all rates compared to other models with 89.73 accuracy, 89.596 precision, 89.73 recall, and 89.574 F1-score. For IADD, the Stacking-Transformer achieves the highest performance in all rates compared to other models with 93.062 accuracy, 93.368 precision, 93.062 recall, and 93.184 F1 score. The improvement in classification performance highlights the wider variety of linguistic variables that the model can capture, providing a reliable solution for precise Arabic dialect recognition and improving the efficacy of NLP applications.https://www.frontiersin.org/articles/10.3389/fnhum.2025.1498297/fullArabic dialectsBert-Base-Arabertv02Dialectal-Arabic-XLM-R-BasetransformerKnowledge representationNLP |
spellingShingle | Hager Saleh Hager Saleh Hager Saleh Abdulaziz AlMohimeed Rasha Hassan Mandour M. Ibrahim Saeed Hamood Alsamhi Moatamad Refaat Hassan Sherif Mostafa Advancing arabic dialect detection with hybrid stacked transformer models Frontiers in Human Neuroscience Arabic dialects Bert-Base-Arabertv02 Dialectal-Arabic-XLM-R-Base transformer Knowledge representation NLP |
title | Advancing arabic dialect detection with hybrid stacked transformer models |
title_full | Advancing arabic dialect detection with hybrid stacked transformer models |
title_fullStr | Advancing arabic dialect detection with hybrid stacked transformer models |
title_full_unstemmed | Advancing arabic dialect detection with hybrid stacked transformer models |
title_short | Advancing arabic dialect detection with hybrid stacked transformer models |
title_sort | advancing arabic dialect detection with hybrid stacked transformer models |
topic | Arabic dialects Bert-Base-Arabertv02 Dialectal-Arabic-XLM-R-Base transformer Knowledge representation NLP |
url | https://www.frontiersin.org/articles/10.3389/fnhum.2025.1498297/full |
work_keys_str_mv | AT hagersaleh advancingarabicdialectdetectionwithhybridstackedtransformermodels AT hagersaleh advancingarabicdialectdetectionwithhybridstackedtransformermodels AT hagersaleh advancingarabicdialectdetectionwithhybridstackedtransformermodels AT abdulazizalmohimeed advancingarabicdialectdetectionwithhybridstackedtransformermodels AT rashahassan advancingarabicdialectdetectionwithhybridstackedtransformermodels AT mandourmibrahim advancingarabicdialectdetectionwithhybridstackedtransformermodels AT saeedhamoodalsamhi advancingarabicdialectdetectionwithhybridstackedtransformermodels AT moatamadrefaathassan advancingarabicdialectdetectionwithhybridstackedtransformermodels AT sherifmostafa advancingarabicdialectdetectionwithhybridstackedtransformermodels |