Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
German medical students´ views regarding artificial intelligence in medicine: A cross-sectional survey
51
Zitationen
4
Autoren
2022
Jahr
Abstract
BACKGROUND: Medical students will likely be most impacted by the envisaged move to artificial intelligence (AI) driven digital medicine, and there is a need to better understand their views regarding the use of AI technology in medicine. This study aimed to explore German medical students´ views about AI in medicine. METHODS: A cross-sectional survey was conducted in October 2019 with all new medical students at the Ludwig Maximilian University of Munich and the Technical University Munich. This represented approximately 10% of all new medical students in Germany. RESULTS: A total of 844 medical students participated (91.9% response rate). Two thirds (64.4%) did not feel well informed about AI in medicine. Just over a half (57.4%) of students thought that AI has useful applications in medicine, particularly in drug research and development (82.5%), less so for clinical uses. Male students were more likely to agree with advantages of AI, and female participants were more likely to be concerned about disadvantages. The vast majority of students thought that when AI is used in medicine that it is important that there are legal rules regarding liability (97%) and oversight mechanisms (93.7%), that physicians should be consulted prior to implementation (96.8%), that developers should be able to explain to them the details of the algorithm (95.6%), that algorithms should use representative data (93.9%), and that patients should always be informed when AI is used (93.5%). CONCLUSIONS: Medical schools and continuing medical education organisers need to promptly develop programs to ensure that clinicians are able to fully realize the potential of AI technology. It is also important that legal rules and oversight are implemented to ensure that future clinicians are not faced with a workplace where important issues around responsibility are not clearly regulated.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.578 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.470 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.984 Zit.
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
2019 · 6.814 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.