Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional Study
Abstract BackgroundArtificial intelligence (AI) has become widely applied across many fields, including medical education. Content validation and its answers are based on training datasets and the optimization of each model. The accuracy of large language model (LLMs) in basic...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
JMIR Publications
2025-01-01
|
Series: | JMIR Medical Education |
Online Access: | https://mededu.jmir.org/2025/1/e58898 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832593392678207488 |
---|---|
author | Naritsaret Kaewboonlert Jiraphon Poontananggul Natthipong Pongsuwan Gun Bhakdisongkhram |
author_facet | Naritsaret Kaewboonlert Jiraphon Poontananggul Natthipong Pongsuwan Gun Bhakdisongkhram |
author_sort | Naritsaret Kaewboonlert |
collection | DOAJ |
description |
Abstract
BackgroundArtificial intelligence (AI) has become widely applied across many fields, including medical education. Content validation and its answers are based on training datasets and the optimization of each model. The accuracy of large language model (LLMs) in basic medical examinations and factors related to their accuracy have also been explored.
ObjectiveWe evaluated factors associated with the accuracy of LLMs (GPT-3.5, GPT-4, Google Bard, and Microsoft Bing) in answering multiple-choice questions from basic medical science examinations.
MethodsWe used questions that were closely aligned with the content and topic distribution of Thailand’s Step 1 National Medical Licensing Examination. Variables such as the difficulty index, discrimination index, and question characteristics were collected. These questions were then simultaneously input into ChatGPT (with GPT-3.5 and GPT-4), Microsoft Bing, and Google Bard, and their responses were recorded. The accuracy of these LLMs and the associated factors were analyzed using multivariable logistic regression. This analysis aimed to assess the effect of various factors on model accuracy, with results reported as odds ratios (ORs).
ResultsThe study revealed that GPT-4 was the top-performing model, with an overall accuracy of 89.07% (95% CI 84.76%‐92.41%), significantly outperforming the others (P
ConclusionsThe GPT-4 and Microsoft Bing models demonstrated equal and superior accuracy compared to GPT-3.5 and Google Bard in the domain of basic medical science. The accuracy of these models was significantly influenced by the item’s difficulty index, indicating that the LLMs are more accurate when answering easier questions. This suggests that the more accurate models, such as GPT-4 and Bing, can be valuable tools for understanding and learning basic medical science concepts. |
format | Article |
id | doaj-art-08a523418567447baae9f6e9bd6cfc30 |
institution | Kabale University |
issn | 2369-3762 |
language | English |
publishDate | 2025-01-01 |
publisher | JMIR Publications |
record_format | Article |
series | JMIR Medical Education |
spelling | doaj-art-08a523418567447baae9f6e9bd6cfc302025-01-20T16:15:54ZengJMIR PublicationsJMIR Medical Education2369-37622025-01-0111e58898e5889810.2196/58898Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional StudyNaritsaret Kaewboonlerthttp://orcid.org/0009-0004-2035-5631Jiraphon Poontananggulhttp://orcid.org/0009-0000-9566-7737Natthipong Pongsuwanhttp://orcid.org/0009-0002-0555-7767Gun Bhakdisongkhramhttp://orcid.org/0000-0001-7434-9262 Abstract BackgroundArtificial intelligence (AI) has become widely applied across many fields, including medical education. Content validation and its answers are based on training datasets and the optimization of each model. The accuracy of large language model (LLMs) in basic medical examinations and factors related to their accuracy have also been explored. ObjectiveWe evaluated factors associated with the accuracy of LLMs (GPT-3.5, GPT-4, Google Bard, and Microsoft Bing) in answering multiple-choice questions from basic medical science examinations. MethodsWe used questions that were closely aligned with the content and topic distribution of Thailand’s Step 1 National Medical Licensing Examination. Variables such as the difficulty index, discrimination index, and question characteristics were collected. These questions were then simultaneously input into ChatGPT (with GPT-3.5 and GPT-4), Microsoft Bing, and Google Bard, and their responses were recorded. The accuracy of these LLMs and the associated factors were analyzed using multivariable logistic regression. This analysis aimed to assess the effect of various factors on model accuracy, with results reported as odds ratios (ORs). ResultsThe study revealed that GPT-4 was the top-performing model, with an overall accuracy of 89.07% (95% CI 84.76%‐92.41%), significantly outperforming the others (P ConclusionsThe GPT-4 and Microsoft Bing models demonstrated equal and superior accuracy compared to GPT-3.5 and Google Bard in the domain of basic medical science. The accuracy of these models was significantly influenced by the item’s difficulty index, indicating that the LLMs are more accurate when answering easier questions. This suggests that the more accurate models, such as GPT-4 and Bing, can be valuable tools for understanding and learning basic medical science concepts.https://mededu.jmir.org/2025/1/e58898 |
spellingShingle | Naritsaret Kaewboonlert Jiraphon Poontananggul Natthipong Pongsuwan Gun Bhakdisongkhram Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional Study JMIR Medical Education |
title | Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional Study |
title_full | Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional Study |
title_fullStr | Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional Study |
title_full_unstemmed | Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional Study |
title_short | Factors Associated With the Accuracy of Large Language Models in Basic Medical Science Examinations: Cross-Sectional Study |
title_sort | factors associated with the accuracy of large language models in basic medical science examinations cross sectional study |
url | https://mededu.jmir.org/2025/1/e58898 |
work_keys_str_mv | AT naritsaretkaewboonlert factorsassociatedwiththeaccuracyoflargelanguagemodelsinbasicmedicalscienceexaminationscrosssectionalstudy AT jiraphonpoontananggul factorsassociatedwiththeaccuracyoflargelanguagemodelsinbasicmedicalscienceexaminationscrosssectionalstudy AT natthipongpongsuwan factorsassociatedwiththeaccuracyoflargelanguagemodelsinbasicmedicalscienceexaminationscrosssectionalstudy AT gunbhakdisongkhram factorsassociatedwiththeaccuracyoflargelanguagemodelsinbasicmedicalscienceexaminationscrosssectionalstudy |