Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
ChatGPT Performance on MCQ Exams in Higher Education. A Pragmatic Scoping Review
48
Zitationen
2
Autoren
2023
Jahr
Abstract
Background. Media coverage suggests that ChatGPT can pass exams based on Multiple Choice Questions (MCQs), including those used to qualify doctors, lawyers, scientists etc. Objectives. We aimed to review current research evidence regarding the performance of ChatGPT on MCQ-based exams, along with recommendations for how educators might address the challenges and benefits of ChatGPT.Eligibility Criteria. Studies testing the performance of ChatGPT on MCQ-based exams in Higher Education. All disciplines and languages were included although the research itself was published in English.Results. 53 studies were included, covering 114 question sets, totalling 49014 MCQs. Free versions of ChatGPT based upon GPT-3/3.5 performed better than random guessing but failed most exams, performing significantly worse than the average human student. GPT-4 passed most exams with a performance that was on a par with human subjects. Conclusions. All summative MCQ-based assessments should be conducted under secure conditions with restricted access to ChatGPT and similar tools, particularly those which assess foundational knowledge. Authentic assessments could incorporate student use of ChatGPT since, due to its excellent performance, graduates will almost certainly be using it in practice.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.292 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.143 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.539 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.452 Zit.