Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
WIP: Traditional Engineering Assessments Challenged by ChatGPT: An Evaluation of its Performance on a Fundamental Competencies Exam
0
Zitationen
3
Autoren
2024
Jahr
Abstract
Abstract ChatGPT, a chatbot which produces text with remarkable coherence, is leading higher education institutions to question the relevance of the current model of engineering education and, particularly, assessment. Among the many reasons behind this questioning is the fact that ChatGPT has been shown to be able to pass various engineering exams. In this research, the GPT-3.5 and GPT-4 models were used to solve different real-life versions of the Fundamental Competencies Exam (FCE), an exam used by a selective Latin American engineering school upon the completion of quintessential engineering courses like basic dynamics, ethics for engineers, and probability and statistics. The formulation of the questions seeks to demonstrate that the student has the fundamental knowledge of the discipline. We adopted a strategy in which the questions were extracted from the FCE modules, and translated to LaTeX. The statements of each question were presented in the absence of supplementary context to avoid influence between questions within the same exam. In addition, a comparative analysis of the effectiveness of the GPT-4 model is performed, evaluating its performance with and without image interpretation capability, due to the recent inclusion of the multimodal function of ChatGPT-4. The results obtained reveal that the difference in the pass rate of GPT-3.5 and GPT-4 is considerable, with 47.38% and 63.06% respectively. While the GPT-4 version without images achieved a passing rate sufficient to pass the exam in all modules, the results that include questions with images increased even more, reaching a 64.38% pass rate. We must continue to solve different versions of the exam. These data will allow us to perform multiple analyses related to the historical performance of the exam, providing a proxy to assess how difficulty has changed over the years. In light of these preliminary results, and given the tight constraints imposed on the model, it is imperative to question whether the FCE effectively assesses the fundamental skills required for an engineer and whether it is the best method for assessing foundational engineering competencies amidst the advent of innovative AI tools.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.436 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.311 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.753 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.523 Zit.