ChatGPT-4 Omni’s superiority in answering multiple-choice oral radiology questions

Abstract Objectives This study evaluates and compares the performance of ChatGPT-3.5, ChatGPT-4 Omni (4o), Google Bard, and Microsoft Copilot in responding to text-based multiple-choice questions related to oral radiology, as featured in the Dental Specialty Admission Exam conducted in Türkiye. Mate...

Full description

Saved in:
Bibliographic Details
Main Author: Melek Tassoker
Format: Article
Language:English
Published: BMC 2025-02-01
Series:BMC Oral Health
Subjects:
Online Access:https://doi.org/10.1186/s12903-025-05554-w
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832571242926833664
author Melek Tassoker
author_facet Melek Tassoker
author_sort Melek Tassoker
collection DOAJ
description Abstract Objectives This study evaluates and compares the performance of ChatGPT-3.5, ChatGPT-4 Omni (4o), Google Bard, and Microsoft Copilot in responding to text-based multiple-choice questions related to oral radiology, as featured in the Dental Specialty Admission Exam conducted in Türkiye. Materials and methods A collection of text-based multiple-choice questions was sourced from the open-access question bank of the Turkish Dental Specialty Admission Exam, covering the years 2012 to 2021. The study included 123 questions, each with five options and one correct answer. The accuracy levels of ChatGPT-3.5, ChatGPT-4o, Google Bard, and Microsoft Copilot were compared using descriptive statistics, the Kruskal-Wallis test, Dunn’s post hoc test, and Cochran’s Q test. Results and discussion The accuracy of the responses generated by the four chatbots exhibited statistically significant differences (p = 0.000). ChatGPT-4o achieved the highest accuracy at 86.1%, followed by Google Bard at 61.8%. ChatGPT-3.5 demonstrated an accuracy rate of 43.9%, while Microsoft Copilot recorded a rate of 41.5%. Conclusion ChatGPT-4o showcases superior accuracy and advanced reasoning capabilities, positioning it as a promising educational tool. With regular updates, it has the potential to serve as a reliable source of information for both healthcare professionals and the general public. Clinical trial number Not applicable.
format Article
id doaj-art-73b35d97acf0451dae362d18737da2ac
institution Kabale University
issn 1472-6831
language English
publishDate 2025-02-01
publisher BMC
record_format Article
series BMC Oral Health
spelling doaj-art-73b35d97acf0451dae362d18737da2ac2025-02-02T12:45:04ZengBMCBMC Oral Health1472-68312025-02-012511810.1186/s12903-025-05554-wChatGPT-4 Omni’s superiority in answering multiple-choice oral radiology questionsMelek Tassoker0Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Necmettin Erbakan UniversityAbstract Objectives This study evaluates and compares the performance of ChatGPT-3.5, ChatGPT-4 Omni (4o), Google Bard, and Microsoft Copilot in responding to text-based multiple-choice questions related to oral radiology, as featured in the Dental Specialty Admission Exam conducted in Türkiye. Materials and methods A collection of text-based multiple-choice questions was sourced from the open-access question bank of the Turkish Dental Specialty Admission Exam, covering the years 2012 to 2021. The study included 123 questions, each with five options and one correct answer. The accuracy levels of ChatGPT-3.5, ChatGPT-4o, Google Bard, and Microsoft Copilot were compared using descriptive statistics, the Kruskal-Wallis test, Dunn’s post hoc test, and Cochran’s Q test. Results and discussion The accuracy of the responses generated by the four chatbots exhibited statistically significant differences (p = 0.000). ChatGPT-4o achieved the highest accuracy at 86.1%, followed by Google Bard at 61.8%. ChatGPT-3.5 demonstrated an accuracy rate of 43.9%, while Microsoft Copilot recorded a rate of 41.5%. Conclusion ChatGPT-4o showcases superior accuracy and advanced reasoning capabilities, positioning it as a promising educational tool. With regular updates, it has the potential to serve as a reliable source of information for both healthcare professionals and the general public. Clinical trial number Not applicable.https://doi.org/10.1186/s12903-025-05554-wChatGPTBardCopilotOral radiologyQuestion
spellingShingle Melek Tassoker
ChatGPT-4 Omni’s superiority in answering multiple-choice oral radiology questions
BMC Oral Health
ChatGPT
Bard
Copilot
Oral radiology
Question
title ChatGPT-4 Omni’s superiority in answering multiple-choice oral radiology questions
title_full ChatGPT-4 Omni’s superiority in answering multiple-choice oral radiology questions
title_fullStr ChatGPT-4 Omni’s superiority in answering multiple-choice oral radiology questions
title_full_unstemmed ChatGPT-4 Omni’s superiority in answering multiple-choice oral radiology questions
title_short ChatGPT-4 Omni’s superiority in answering multiple-choice oral radiology questions
title_sort chatgpt 4 omni s superiority in answering multiple choice oral radiology questions
topic ChatGPT
Bard
Copilot
Oral radiology
Question
url https://doi.org/10.1186/s12903-025-05554-w
work_keys_str_mv AT melektassoker chatgpt4omnissuperiorityinansweringmultiplechoiceoralradiologyquestions