Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery
Abstract Background The objective of this study was to evaluate the performance of ScholarGPT, ChatGPT-4o and Google Gemini in responding to queries pertaining to endodontic apical surgery, a subject that demands advanced specialist knowledge in endodontics. Methods A total of 30 questions, includin...
Saved in:
| Main Authors: | , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
BMC
2025-05-01
|
| Series: | BMC Oral Health |
| Subjects: | |
| Online Access: | https://doi.org/10.1186/s12903-025-06149-1 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| _version_ | 1849309430258073600 |
|---|---|
| author | Sevda Durust Baris Kubilay Baris |
| author_facet | Sevda Durust Baris Kubilay Baris |
| author_sort | Sevda Durust Baris |
| collection | DOAJ |
| description | Abstract Background The objective of this study was to evaluate the performance of ScholarGPT, ChatGPT-4o and Google Gemini in responding to queries pertaining to endodontic apical surgery, a subject that demands advanced specialist knowledge in endodontics. Methods A total of 30 questions, including 12 binary and 18 open-ended queries, were formulated based on information on endodontic apical surgery taken from a well-known endodontic book called Cohen’s pathways of the pulp (12th edition). The questions were posed by two different researchers using different accounts on the ScholarGPT, ChatGPT-4o and Gemini platforms. The responses were then coded by the researchers and categorised as ‘correct’, ‘incorrect’, or ‘insufficient’. The Pearson chi-square test was used to assess the relationships between the platforms. Results A total of 5,400 responses were evaluated. Chi-square analysis revealed statistically significant differences between the accuracy of the responses provided applications (χ² = 22.61; p < 0.05). ScholarGPT demonstrated the highest rate of correct responses (97.7%), followed by ChatGPT-4o with 90.1%. Conversely, Gemini exhibited the lowest correct response rate (59.5%) among the applications examined. Conclusions ScholarGPT performed better overall on questions about endodontic apical surgery than ChatGPT-4o and Gemini. GPT models based on academic databases, such as ScholarGPT, may provide more accurate information about dentistry. However, additional research should be conducted to develop a GPT model that is specifically tailored to the field of endodontics. |
| format | Article |
| id | doaj-art-10fd5ec7b6cc4d6d904106971072596e |
| institution | Kabale University |
| issn | 1472-6831 |
| language | English |
| publishDate | 2025-05-01 |
| publisher | BMC |
| record_format | Article |
| series | BMC Oral Health |
| spelling | doaj-art-10fd5ec7b6cc4d6d904106971072596e2025-08-20T03:54:10ZengBMCBMC Oral Health1472-68312025-05-012511710.1186/s12903-025-06149-1Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgerySevda Durust Baris0Kubilay Baris1Kırıkkale UniversityKırıkkale UniversityAbstract Background The objective of this study was to evaluate the performance of ScholarGPT, ChatGPT-4o and Google Gemini in responding to queries pertaining to endodontic apical surgery, a subject that demands advanced specialist knowledge in endodontics. Methods A total of 30 questions, including 12 binary and 18 open-ended queries, were formulated based on information on endodontic apical surgery taken from a well-known endodontic book called Cohen’s pathways of the pulp (12th edition). The questions were posed by two different researchers using different accounts on the ScholarGPT, ChatGPT-4o and Gemini platforms. The responses were then coded by the researchers and categorised as ‘correct’, ‘incorrect’, or ‘insufficient’. The Pearson chi-square test was used to assess the relationships between the platforms. Results A total of 5,400 responses were evaluated. Chi-square analysis revealed statistically significant differences between the accuracy of the responses provided applications (χ² = 22.61; p < 0.05). ScholarGPT demonstrated the highest rate of correct responses (97.7%), followed by ChatGPT-4o with 90.1%. Conversely, Gemini exhibited the lowest correct response rate (59.5%) among the applications examined. Conclusions ScholarGPT performed better overall on questions about endodontic apical surgery than ChatGPT-4o and Gemini. GPT models based on academic databases, such as ScholarGPT, may provide more accurate information about dentistry. However, additional research should be conducted to develop a GPT model that is specifically tailored to the field of endodontics.https://doi.org/10.1186/s12903-025-06149-1Artificial intelligenceChatGPTEndodontic apical surgeryGeminiScholarGPT |
| spellingShingle | Sevda Durust Baris Kubilay Baris Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery BMC Oral Health Artificial intelligence ChatGPT Endodontic apical surgery Gemini ScholarGPT |
| title | Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery |
| title_full | Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery |
| title_fullStr | Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery |
| title_full_unstemmed | Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery |
| title_short | Assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery |
| title_sort | assessment of various artificial intelligence applications in responding to technical questions in endodontic surgery |
| topic | Artificial intelligence ChatGPT Endodontic apical surgery Gemini ScholarGPT |
| url | https://doi.org/10.1186/s12903-025-06149-1 |
| work_keys_str_mv | AT sevdadurustbaris assessmentofvariousartificialintelligenceapplicationsinrespondingtotechnicalquestionsinendodonticsurgery AT kubilaybaris assessmentofvariousartificialintelligenceapplicationsinrespondingtotechnicalquestionsinendodonticsurgery |