Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
88 Evaluating the impact of ChatGPT-4 on medical abstracts
2
Zitationen
5
Autoren
2024
Jahr
Abstract
Abstract Background Since its launch in November 2022, ChatGPT has received recognition for its use in medical writing. Some even suggested it could write scientific abstracts. However, its real-life effectiveness to help researchers improve conference abstracts remains unexplored. Objectives To evaluate whether ChatGPT 4.0 could improve the quality of abstracts submitted to a medical conference by clinical researchers with various experiences. Design/Methods In October 2023, we conducted an experimental study involving 24 international researchers who provided one “original” abstract intended for submission at the 2024 Pediatric Academic Society (PAS) conference. Our research team created a prompt asking ChatGPT-4 to improve the quality of the abstract while adhering PAS submission guidelines. Researchers received the revised version and were tasked with creating a “final” abstract. The quality of each version (original, ChatGPT and final) was evaluated by the researchers themselves using a numeric scale (0-100). Additionally, three co-investigators assessed abstracts blinded to the version. The primary analysis focused on the mean difference in scores between the final and original abstracts. We also compared scores of the three versions and determined the proportion of researchers reporting ChatGPT's impact on their abstract and its acceptance probability. For an effect size of 0.7, it was estimated that we needed at least 20 pairs of abstracts. Results Abstract quality varied between the three versions with mean scores of 82, 65 and 90 for the original, ChatGPT and final versions, respectively. Overall, the final version displayed significantly improved quality compared to the original (mean difference 8.0 points; 95% CI: 5.6-10.3). Of note, 3 (14%) abstracts remained unchanged between final and original versions. Independent ratings by the co-investigator confirmed statistical improvements (mean difference 0.94 points; 95% CI: 0.44-1.44). Researchers identified minor (n=10) and major (n=3) factual errors in ChatGPT’s abstracts. However, 18 (75%) participants reported that ChatGPT contributed to the enhancement of their final abstract. Ten (42%) participants believed it improved their abstract's chances of acceptance. Eighteen (75%) researchers reported being uncomfortable using ChatGPT version for submission. Conclusion While ChatGPT 4.0 does not produce abstracts of better quality then the one crafted by researchers, it serves as a valuable tool for researchers to enhance the quality of their own abstracts. The utilization of such tools is a potential strategy for researchers seeking to improve their abstracts.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.436 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.311 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.753 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.523 Zit.