Zero-BertXGB: An Empirical Technique for Abstract Classification in Systematic Reviews

Abstract classification in systematic reviews (SRs) is a crucial step in evidence synthesis but is often time-consuming and labour-intensive. This study evaluates the effectiveness of various Machine Learning (ML) models and embedding techniques in automating this process. Five diverse datasets are...

Full description

Saved in:
Bibliographic Details
Main Authors: Mohammad Shariful Islam, Mohammad Abu Tareq Rony, Md Rasel Hossain, Samah Alshathri, Walid El-Shafai
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10845770/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832576801265680384
author Mohammad Shariful Islam
Mohammad Abu Tareq Rony
Md Rasel Hossain
Samah Alshathri
Walid El-Shafai
author_facet Mohammad Shariful Islam
Mohammad Abu Tareq Rony
Md Rasel Hossain
Samah Alshathri
Walid El-Shafai
author_sort Mohammad Shariful Islam
collection DOAJ
description Abstract classification in systematic reviews (SRs) is a crucial step in evidence synthesis but is often time-consuming and labour-intensive. This study evaluates the effectiveness of various Machine Learning (ML) models and embedding techniques in automating this process. Five diverse datasets are utilized: Aceves-Martins (2021), comprising 1,258 excluded and 230 included abstracts on the utilization of animal models in depressive behaviour studies; Bannach-Brown (2016), with 896 excluded and 73 included abstracts focusing on the methodological rigour of environmental health systematic reviews; Meijboom (2021), containing 599 excluded and 32 included abstracts on the retransitioning of Etanercept in rheumatic disease patients; Menon (2022), with 896 excluded and 73 included abstracts on environmental health reviews; and a custom Clinical Review Paper Abstract (CRPA) dataset, featuring 500 excluded and 50 included abstracts. A significant research gap in abstract classification has been identified in previous literature, particularly in comparing Large Language Models (LLMs) with traditional ML and Natural Language Processing (NLP) techniques regarding scalability, adaptability, computational efficiency, and real-time application. Addressing this gap, this study employs GloVe for word embedding via matrix factorization, FastText for character n-gram representation, and Doc2Vec for capturing paragraph-level semantics. A novel Zero-BertXGB technique is introduced, integrating a transformer-based language model, zero-shot learning, and an ML classifier to enhance abstract screening and classification into “Include” or “Exclude” categories. This approach leverages contextual understanding and precision for efficient abstract processing. The Zero-BertXGB technique is compared against other prominent LLMs, including BERT, PaLM, LLaMA, GPT-3.5, and GPT-4, to validate its effectiveness. The Zero-BertXGB model achieved accuracy values of 99.3% for Aceves-Martins2021, 92.6% for Bannach-Brown2016, 85.7% for Meijboom2021, 94.1% for Menon2022, and 98.8% for CRPA. The findings indicate that the Zero-BertXGB model, alongside other LLMs, can deliver reliable results with minimal human intervention, enhancing abstract screening efficiency and potentially revolutionizing systematic review workflows.
format Article
id doaj-art-fdd7f32c339d412d85e52932f6a31647
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-fdd7f32c339d412d85e52932f6a316472025-01-31T00:00:56ZengIEEEIEEE Access2169-35362025-01-0113184181844010.1109/ACCESS.2025.353177810845770Zero-BertXGB: An Empirical Technique for Abstract Classification in Systematic ReviewsMohammad Shariful Islam0https://orcid.org/0009-0007-8442-1425Mohammad Abu Tareq Rony1https://orcid.org/0000-0002-0640-1425Md Rasel Hossain2Samah Alshathri3https://orcid.org/0000-0002-8805-7890Walid El-Shafai4https://orcid.org/0000-0001-7509-2120Department of Computer Science and Telecommunication Engineering, Noakhali Science and Technology University, Noakhali, BangladeshDepartment of Statistics, Noakhali Science and Technology University, Noakhali, BangladeshDepartment of Statistics, Noakhali Science and Technology University, Noakhali, BangladeshDepartment of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box, 84428, Riyadh, Saudi ArabiaComputer Science Department, Automated Systems and Soft Computing Laboratory (ASSCL), Prince Sultan University, Riyadh, Saudi ArabiaAbstract classification in systematic reviews (SRs) is a crucial step in evidence synthesis but is often time-consuming and labour-intensive. This study evaluates the effectiveness of various Machine Learning (ML) models and embedding techniques in automating this process. Five diverse datasets are utilized: Aceves-Martins (2021), comprising 1,258 excluded and 230 included abstracts on the utilization of animal models in depressive behaviour studies; Bannach-Brown (2016), with 896 excluded and 73 included abstracts focusing on the methodological rigour of environmental health systematic reviews; Meijboom (2021), containing 599 excluded and 32 included abstracts on the retransitioning of Etanercept in rheumatic disease patients; Menon (2022), with 896 excluded and 73 included abstracts on environmental health reviews; and a custom Clinical Review Paper Abstract (CRPA) dataset, featuring 500 excluded and 50 included abstracts. A significant research gap in abstract classification has been identified in previous literature, particularly in comparing Large Language Models (LLMs) with traditional ML and Natural Language Processing (NLP) techniques regarding scalability, adaptability, computational efficiency, and real-time application. Addressing this gap, this study employs GloVe for word embedding via matrix factorization, FastText for character n-gram representation, and Doc2Vec for capturing paragraph-level semantics. A novel Zero-BertXGB technique is introduced, integrating a transformer-based language model, zero-shot learning, and an ML classifier to enhance abstract screening and classification into “Include” or “Exclude” categories. This approach leverages contextual understanding and precision for efficient abstract processing. The Zero-BertXGB technique is compared against other prominent LLMs, including BERT, PaLM, LLaMA, GPT-3.5, and GPT-4, to validate its effectiveness. The Zero-BertXGB model achieved accuracy values of 99.3% for Aceves-Martins2021, 92.6% for Bannach-Brown2016, 85.7% for Meijboom2021, 94.1% for Menon2022, and 98.8% for CRPA. The findings indicate that the Zero-BertXGB model, alongside other LLMs, can deliver reliable results with minimal human intervention, enhancing abstract screening efficiency and potentially revolutionizing systematic review workflows.https://ieeexplore.ieee.org/document/10845770/Abstract classificationmachine learningnatural language processingzero-BertXGB methodslarge language models
spellingShingle Mohammad Shariful Islam
Mohammad Abu Tareq Rony
Md Rasel Hossain
Samah Alshathri
Walid El-Shafai
Zero-BertXGB: An Empirical Technique for Abstract Classification in Systematic Reviews
IEEE Access
Abstract classification
machine learning
natural language processing
zero-BertXGB methods
large language models
title Zero-BertXGB: An Empirical Technique for Abstract Classification in Systematic Reviews
title_full Zero-BertXGB: An Empirical Technique for Abstract Classification in Systematic Reviews
title_fullStr Zero-BertXGB: An Empirical Technique for Abstract Classification in Systematic Reviews
title_full_unstemmed Zero-BertXGB: An Empirical Technique for Abstract Classification in Systematic Reviews
title_short Zero-BertXGB: An Empirical Technique for Abstract Classification in Systematic Reviews
title_sort zero bertxgb an empirical technique for abstract classification in systematic reviews
topic Abstract classification
machine learning
natural language processing
zero-BertXGB methods
large language models
url https://ieeexplore.ieee.org/document/10845770/
work_keys_str_mv AT mohammadsharifulislam zerobertxgbanempiricaltechniqueforabstractclassificationinsystematicreviews
AT mohammadabutareqrony zerobertxgbanempiricaltechniqueforabstractclassificationinsystematicreviews
AT mdraselhossain zerobertxgbanempiricaltechniqueforabstractclassificationinsystematicreviews
AT samahalshathri zerobertxgbanempiricaltechniqueforabstractclassificationinsystematicreviews
AT walidelshafai zerobertxgbanempiricaltechniqueforabstractclassificationinsystematicreviews