Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study

Abstract BackgroundOpenAI released versions ChatGPT-3.5 and GPT-4 between 2022 and 2023. GPT-3.5 has demonstrated proficiency in various examinations, particularly the United States Medical Licensing Examination. However, GPT-4 has more advanced capabilities. Objec...

Full description

Saved in:
Bibliographic Details
Main Authors: Ying-Mei Wang, Hung-Wei Shen, Tzeng-Ji Chen, Shu-Chiung Chiang, Ting-Guan Lin
Format: Article
Language:English
Published: JMIR Publications 2025-01-01
Series:JMIR Medical Education
Online Access:https://mededu.jmir.org/2025/1/e56850
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832585111183294464
author Ying-Mei Wang
Hung-Wei Shen
Tzeng-Ji Chen
Shu-Chiung Chiang
Ting-Guan Lin
author_facet Ying-Mei Wang
Hung-Wei Shen
Tzeng-Ji Chen
Shu-Chiung Chiang
Ting-Guan Lin
author_sort Ying-Mei Wang
collection DOAJ
description Abstract BackgroundOpenAI released versions ChatGPT-3.5 and GPT-4 between 2022 and 2023. GPT-3.5 has demonstrated proficiency in various examinations, particularly the United States Medical Licensing Examination. However, GPT-4 has more advanced capabilities. ObjectiveThis study aims to examine the efficacy of GPT-3.5 and GPT-4 within the Taiwan National Pharmacist Licensing Examination and to ascertain their utility and potential application in clinical pharmacy and education. MethodsThe pharmacist examination in Taiwan consists of 2 stages: basic subjects and clinical subjects. In this study, exam questions were manually fed into the GPT-3.5 and GPT-4 models, and their responses were recorded; graphic-based questions were excluded. This study encompassed three steps: (1) determining the answering accuracy of GPT-3.5 and GPT-4, (2) categorizing question types and observing differences in model performance across these categories, and (3) comparing model performance on calculation and situational questions. Microsoft Excel and R software were used for statistical analyses. ResultsGPT-4 achieved an accuracy rate of 72.9%, overshadowing GPT-3.5, which achieved 59.1% (PP ConclusionsThis study demonstrates that GPT-4 outperforms GPT-3.5 in the Taiwan National Pharmacist Licensing Examination, particularly in basic subjects. While GPT-4 shows potential for use in clinical practice and pharmacy education, its limitations warrant caution. Future research should focus on refining prompts, improving model stability, integrating medical databases, and designing questions that better assess student competence and minimize guessing.
format Article
id doaj-art-03284571c2d84ba9ae90e09fbfc5de06
institution Kabale University
issn 2369-3762
language English
publishDate 2025-01-01
publisher JMIR Publications
record_format Article
series JMIR Medical Education
spelling doaj-art-03284571c2d84ba9ae90e09fbfc5de062025-01-27T02:52:29ZengJMIR PublicationsJMIR Medical Education2369-37622025-01-0111e56850e5685010.2196/56850Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation StudyYing-Mei Wanghttp://orcid.org/0009-0006-9823-5298Hung-Wei Shenhttp://orcid.org/0009-0002-6095-1185Tzeng-Ji Chenhttp://orcid.org/0000-0002-8350-0232Shu-Chiung Chianghttp://orcid.org/0000-0002-8795-2228Ting-Guan Linhttp://orcid.org/0009-0006-9138-9075 Abstract BackgroundOpenAI released versions ChatGPT-3.5 and GPT-4 between 2022 and 2023. GPT-3.5 has demonstrated proficiency in various examinations, particularly the United States Medical Licensing Examination. However, GPT-4 has more advanced capabilities. ObjectiveThis study aims to examine the efficacy of GPT-3.5 and GPT-4 within the Taiwan National Pharmacist Licensing Examination and to ascertain their utility and potential application in clinical pharmacy and education. MethodsThe pharmacist examination in Taiwan consists of 2 stages: basic subjects and clinical subjects. In this study, exam questions were manually fed into the GPT-3.5 and GPT-4 models, and their responses were recorded; graphic-based questions were excluded. This study encompassed three steps: (1) determining the answering accuracy of GPT-3.5 and GPT-4, (2) categorizing question types and observing differences in model performance across these categories, and (3) comparing model performance on calculation and situational questions. Microsoft Excel and R software were used for statistical analyses. ResultsGPT-4 achieved an accuracy rate of 72.9%, overshadowing GPT-3.5, which achieved 59.1% (PP ConclusionsThis study demonstrates that GPT-4 outperforms GPT-3.5 in the Taiwan National Pharmacist Licensing Examination, particularly in basic subjects. While GPT-4 shows potential for use in clinical practice and pharmacy education, its limitations warrant caution. Future research should focus on refining prompts, improving model stability, integrating medical databases, and designing questions that better assess student competence and minimize guessing.https://mededu.jmir.org/2025/1/e56850
spellingShingle Ying-Mei Wang
Hung-Wei Shen
Tzeng-Ji Chen
Shu-Chiung Chiang
Ting-Guan Lin
Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study
JMIR Medical Education
title Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study
title_full Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study
title_fullStr Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study
title_full_unstemmed Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study
title_short Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study
title_sort performance of chatgpt 3 5 and chatgpt 4 in the taiwan national pharmacist licensing examination comparative evaluation study
url https://mededu.jmir.org/2025/1/e56850
work_keys_str_mv AT yingmeiwang performanceofchatgpt35andchatgpt4inthetaiwannationalpharmacistlicensingexaminationcomparativeevaluationstudy
AT hungweishen performanceofchatgpt35andchatgpt4inthetaiwannationalpharmacistlicensingexaminationcomparativeevaluationstudy
AT tzengjichen performanceofchatgpt35andchatgpt4inthetaiwannationalpharmacistlicensingexaminationcomparativeevaluationstudy
AT shuchiungchiang performanceofchatgpt35andchatgpt4inthetaiwannationalpharmacistlicensingexaminationcomparativeevaluationstudy
AT tingguanlin performanceofchatgpt35andchatgpt4inthetaiwannationalpharmacistlicensingexaminationcomparativeevaluationstudy