Anatomy exam model for the circulatory and respiratory systems using GPT-4: a medical school study
dc.contributor.author | Tekin, Ayla | |
dc.contributor.author | Karamus, Nizameddin Fatih | |
dc.contributor.author | Çolak, Tuncay | |
dc.date.accessioned | 2025-08-14T17:36:34Z | |
dc.date.available | 2025-08-14T17:36:34Z | |
dc.date.issued | 2025 | |
dc.department | Fakülteler, Tıp Fakültesi, Temel Tıp Bilimleri, Anatomi Ana Bilim Dalı | |
dc.description | Article Number : 158 | |
dc.description.abstract | Purpose: The study aimed to evaluate the effectiveness of anatomy multiple-choice questions (MCQs) generated by GPT-4, focused on their methodological appropriateness and alignment with the cognitive levels defined by Bloom's revised taxonomy to enhance assessment. Methods: The assessment questions developed for medical students were created utilizing GPT-4, comprising 240 MCQs organized into subcategories consistent with Bloom's revised taxonomy. When designing prompts to create MCQs, details about the lesson's purpose, learning objectives, and students' prior experiences were included to ensure the questions were contextually appropriate. A set of 30 MCQs was randomly selected from the generated questions for testing. A total of 280 students participated in the examination, which assessed the difficulty index of the MCQs, the item discrimination index, and the overall test difficulty level. Expert anatomists examined the taxonomy accuracy of GPT-4's questions. Results: Students achieved a median score of 50 (range, 36.67-60) points on the test. The test's internal consistency, assessed by KR-20, was 0.737. The average difficulty of the test was 0.5012. Results show difficulty and discrimination indices for each AI-generated question. Expert anatomists' taxonomy-based classifications matched GPT-4's 26.6%. Meanwhile, 80.9% of students found the questions were clear, and 85.8% showed interest in retaking the assessment exam. Conclusion: This study demonstrates GPT-4's significant potential for generating medical education exam questions. While it effectively assesses basic knowledge recall, it fails to sufficiently evaluate higher-order cognitive processes outlined in Bloom's revised taxonomy. Future research should consider alternative methods that combine AI with expert evaluation and specialized multimodal models. | |
dc.identifier.citation | Tekin, A., Karamus, N. F., & Çolak, T. (2025). Anatomy exam model for the circulatory and respiratory systems using GPT-4: a medical school study. Surgical and Radiologic Anatomy, 47(1), 158. 10.1007/s00276-025-03667-z | |
dc.identifier.doi | 10.1007/s00276-025-03667-z | |
dc.identifier.issn | 0930-1038 | |
dc.identifier.issn | 1279-8517 | |
dc.identifier.issue | 1 | |
dc.identifier.pmid | 40495075 | |
dc.identifier.scopus | 2-s2.0-105007648593 | |
dc.identifier.scopusquality | Q2 | |
dc.identifier.uri | https://hdl.handle.net/20.500.12939/5896 | |
dc.identifier.volume | 47 | |
dc.identifier.wos | WOS:001506184300001 | |
dc.identifier.wosquality | Q3 | |
dc.indekslendigikaynak | PubMed | |
dc.indekslendigikaynak | Web of Science | |
dc.indekslendigikaynak | Scopus | |
dc.institutionauthor | Karamus, Nizameddin Fatih | |
dc.language.iso | en | |
dc.publisher | Springer International | |
dc.relation.ispartof | Journal of Clinical Anatomy | |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | |
dc.rights | info:eu-repo/semantics/closedAccess | |
dc.subject | Anatomy assessment | |
dc.subject | Bloom's revised taxonomy | |
dc.subject | GPT-4 | |
dc.subject | Multiple-choice questions | |
dc.title | Anatomy exam model for the circulatory and respiratory systems using GPT-4: a medical school study | |
dc.type | Article |
Dosyalar
Lisans paketi
1 - 1 / 1
[ X ]
- İsim:
- license.txt
- Boyut:
- 1.17 KB
- Biçim:
- Item-specific license agreed upon to submission
- Açıklama: