Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Is Artificial Intelligence (AI) the Missing Member of the Fellowship Interview Panel?
0
Zitationen
3
Autoren
2024
Jahr
Abstract
Background: Creating an objective summary of the interviewers' feedback to rank fellowship applicants is a challenging task that can be adversely affected by implicit bias. In this study, we examined the performance of ChatGPT (CGPT) in summarizing interviewers' evaluations of the applicants and compared the CGPT-generated ranking to our rank list. Methods: For each applicant, 4 faculty independently generated evaluation narratives, each about 100 words. PD use these narratives to create a list of strengths & weaknesses for the ranking meetings. In this study, we provided de-identified narratives to CGPT4.0. We compared the performance of two prompts: a simple prompt that requested CGPT to summarize the strengths and weaknesses and rank applicants without program-specific details and a complex prompt that described our program’s mission, aims, and desired applicant attributes. We used data from the 2022 & 2023 seasons. Spearman rank correlation coefficient was used to compare CGPT rank list to the final rank list by the program. The study was deemed exempt by IRB. Results: There were 29 applicants from 2022 and 26 from 2023. Narratives from 2022 were about 50% shorter. The performance of the complex CGPT prompt was superior to the simple prompt in detecting strengths and weaknesses and in the final ranking. The complex prompt-generated rank list had a correlation coefficient with our rank list of 0.95 (p<0.001) for 2022 and 0.99 (p<0.001) for 2023. The simple prompt correlation was 0.85 (p<0.001) and 0.83 (p<0.001). SeeFigure 1 Conclusion: A longer evaluation narrative and more complex prompt that included the program-specific information resulted in an excellent performance by CGPT in summarizing the applicants' strengths & weaknesses, which resulted in a final rank that correlated very closely with the program's rank list. AI limitations included missing subtle applicant attributes that a program may consider necessary but not defined in the prompt. Future research is needed to study the effect of AI-assisted ranking on implicit bias during ranking.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.245 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.100 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.466 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.429 Zit.