Treffer: Comparing AI-Assisted Problem-Solving Ability With Internet Search Engine and e-Books in Medical Students With Variable Prior Subject Knowledge: Cross-Sectional Study.
AI Ethics. 2022;2(1):157-165. (PMID: 34790953)
Ann Acad Med Singap. 2024 Dec 12;53(12):754-757. (PMID: 39748174)
Stud Health Technol Inform. 2023 Jun 29;305:648-651. (PMID: 37387115)
J Adv Med Educ Prof. 2023 Jul;11(3):133-140. (PMID: 37469385)
BMC Med Educ. 2019 Apr 16;19(1):110. (PMID: 30991988)
Asian Bioeth Rev. 2024 May 25;16(3):513-526. (PMID: 39022373)
Cureus. 2023 Nov 28;15(11):e49604. (PMID: 38161821)
Res Social Adm Pharm. 2023 Aug;19(8):1236-1242. (PMID: 37321925)
Weitere Informationen
Background: Artificial intelligence (AI), particularly large language models (LLMs) such as ChatGPT (OpenAI), is rapidly influencing medical education. Its effectiveness for students with varying levels of prior knowledge remains underexplored.
Objective: This study aimed to evaluate the performance of medical students with and without formal pharmacology knowledge when using AI-LLM GPTs, internet search engines, e-books, or self-knowledge to solve multiple-choice questions (MCQs).
Methods: A cross-sectional study was conducted at a tertiary care teaching hospital with 100 medical students, divided into a "naive" group (n=50; no pharmacology training) and a "learned" group (n=50; completed pharmacology training). The study was started after approval from the Institutional Ethics Committee of Jawaharlal Nehru Medical College Hospital, Aligarh Muslim University (1018/IEC/23/8/23). Each participant answered 4 sets of 20 MCQs using self-knowledge, e-books, Google, or ChatGPT-4o. Scores were compared using analysis of covariance with self-knowledge scores as a covariate.
Results: Learned students significantly outperformed naive students across all methods (P<.001), with the largest effect size in the AI-LLM GPT set (partial η²=0.328). For both groups, the performance hierarchy was AI-LLM GPT > internet search engine > self-knowledge ≈ e-books. Notably, the naive students who used AI scored higher (mean 13.24, SD 3.31) than the learned students who used Google (mean 12.14, SD 2.01; P=.01) or e-books (mean 10.22, SD 3.12; P<.001).
Conclusions: AI-LLM GPTs can significantly enhance problem-solving performance in MCQ-based assessments, particularly for students with limited prior knowledge, even allowing them to outperform knowledgeable peers using traditional digital resources. This underscores the potential of AI to transform learning support in medical education, although its impact on deep learning and critical thinking requires further investigation.
(© Ajiith Xavier, Syed Shariq Naeem, Waseem Rizwi, Hiramani Rabha. Originally published in JMIR Medical Education (https://mededu.jmir.org).)