Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery

Abstract Background The objective of this study was to evaluate the performance of ScholarGPT, ChatGPT-4o and Google Gemini in responding to queries pertaining to endodontic apical surgery, a subject that demands advanced specialist knowledge in endodontics. Methods A total of 30 questions, includin...

Full description

Saved in:
Bibliographic Details
Main Authors: Sevda Durust Baris, Kubilay Baris
Format: Article
Language:English
Published: BMC 2025-05-01
Series:BMC Oral Health
Subjects:
Online Access:https://doi.org/10.1186/s12903-025-06149-1
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1849309430258073600
author Sevda Durust Baris
Kubilay Baris
author_facet Sevda Durust Baris
Kubilay Baris
author_sort Sevda Durust Baris
collection DOAJ
description Abstract Background The objective of this study was to evaluate the performance of ScholarGPT, ChatGPT-4o and Google Gemini in responding to queries pertaining to endodontic apical surgery, a subject that demands advanced specialist knowledge in endodontics. Methods A total of 30 questions, including 12 binary and 18 open-ended queries, were formulated based on information on endodontic apical surgery taken from a well-known endodontic book called Cohen’s pathways of the pulp (12th edition). The questions were posed by two different researchers using different accounts on the ScholarGPT, ChatGPT-4o and Gemini platforms. The responses were then coded by the researchers and categorised as ‘correct’, ‘incorrect’, or ‘insufficient’. The Pearson chi-square test was used to assess the relationships between the platforms. Results A total of 5,400 responses were evaluated. Chi-square analysis revealed statistically significant differences between the accuracy of the responses provided applications (χ² = 22.61; p < 0.05). ScholarGPT demonstrated the highest rate of correct responses (97.7%), followed by ChatGPT-4o with 90.1%. Conversely, Gemini exhibited the lowest correct response rate (59.5%) among the applications examined. Conclusions ScholarGPT performed better overall on questions about endodontic apical surgery than ChatGPT-4o and Gemini. GPT models based on academic databases, such as ScholarGPT, may provide more accurate information about dentistry. However, additional research should be conducted to develop a GPT model that is specifically tailored to the field of endodontics.
format Article
id doaj-art-10fd5ec7b6cc4d6d904106971072596e
institution Kabale University
issn 1472-6831
language English
publishDate 2025-05-01
publisher BMC
record_format Article
series BMC Oral Health
spelling doaj-art-10fd5ec7b6cc4d6d904106971072596e2025-08-20T03:54:10ZengBMCBMC Oral Health1472-68312025-05-012511710.1186/s12903-025-06149-1Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgerySevda Durust Baris0Kubilay Baris1Kırıkkale UniversityKırıkkale UniversityAbstract Background The objective of this study was to evaluate the performance of ScholarGPT, ChatGPT-4o and Google Gemini in responding to queries pertaining to endodontic apical surgery, a subject that demands advanced specialist knowledge in endodontics. Methods A total of 30 questions, including 12 binary and 18 open-ended queries, were formulated based on information on endodontic apical surgery taken from a well-known endodontic book called Cohen’s pathways of the pulp (12th edition). The questions were posed by two different researchers using different accounts on the ScholarGPT, ChatGPT-4o and Gemini platforms. The responses were then coded by the researchers and categorised as ‘correct’, ‘incorrect’, or ‘insufficient’. The Pearson chi-square test was used to assess the relationships between the platforms. Results A total of 5,400 responses were evaluated. Chi-square analysis revealed statistically significant differences between the accuracy of the responses provided applications (χ² = 22.61; p < 0.05). ScholarGPT demonstrated the highest rate of correct responses (97.7%), followed by ChatGPT-4o with 90.1%. Conversely, Gemini exhibited the lowest correct response rate (59.5%) among the applications examined. Conclusions ScholarGPT performed better overall on questions about endodontic apical surgery than ChatGPT-4o and Gemini. GPT models based on academic databases, such as ScholarGPT, may provide more accurate information about dentistry. However, additional research should be conducted to develop a GPT model that is specifically tailored to the field of endodontics.https://doi.org/10.1186/s12903-025-06149-1Artificial intelligenceChatGPTEndodontic apical surgeryGeminiScholarGPT
spellingShingle Sevda Durust Baris
Kubilay Baris
Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery
BMC Oral Health
Artificial intelligence
ChatGPT
Endodontic apical surgery
Gemini
ScholarGPT
title Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery
title_full Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery
title_fullStr Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery
title_full_unstemmed Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery
title_short Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery
title_sort assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery
topic Artificial intelligence
ChatGPT
Endodontic apical surgery
Gemini
ScholarGPT
url https://doi.org/10.1186/s12903-025-06149-1
work_keys_str_mv AT sevdadurustbaris assessmentofvariousartificialintelligenceapplicationsinrespondingtotechnicalquestionsinendodonticsurgery
AT kubilaybaris assessmentofvariousartificialintelligenceapplicationsinrespondingtotechnicalquestionsinendodonticsurgery