Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Large Language Models for the National Radiological Technologist Licensure Examination in Japan: Cross-Sectional Comparative Benchmarking and Evaluation of Model-Generated Items Study
1
Zitationen
5
Autoren
2025
Jahr
Abstract
OpenAI o3 can generate radiological licensure items that align with national standards in terms of difficulty, factual correctness, and blueprint coverage. However, wording clarity and the pedagogical specificity of explanations were weaker and did not meet an adoptable threshold without further editorial refinement. These findings support a practical workflow in which LLMs draft syllabus-aligned items at scale, while faculty perform targeted edits to ensure clarity and formative feedback. Future studies should evaluate image-inclusive generation, use Application Programming Interface (API)-pinned model snapshots to increase reproducibility, and develop guidance to improve explanation quality for learner remediation.
Ähnliche Arbeiten
Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives
1999 · 103.486 Zit.
Common method biases in behavioral research: A critical review of the literature and recommended remedies.
2003 · 74.497 Zit.
Evaluating Structural Equation Models with Unobservable Variables and Measurement Error
1981 · 65.794 Zit.
Evaluating Structural Equation Models with Unobservable Variables and Measurement Error
1981 · 60.422 Zit.
Coefficient Alpha and the Internal Structure of Tests
1951 · 42.812 Zit.