OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 01.04.2026, 02:39

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Radiology Synthetic Confusion: How Generative Artificial Intelligence Amplifies Misunderstandings of Radiologists and Technologists in Patient-Facing Media

2025·1 Zitationen·Canadian Association of Radiologists JournalOpen Access
Volltext beim Verlag öffnen

1

Zitationen

5

Autoren

2025

Jahr

Abstract

<b>Rationale and Objectives:</b> Artificial intelligence (AI) tools, particularly generative models, are increasingly used to depict clinical roles in healthcare. This study evaluates whether generative AI systems accurately differentiate between radiologists and medical radiation technologists (MRTs), 2 roles often confused by patients and providers. <b>Materials and Methods:</b> We assessed 1380 images and videos generated by 8 text-to-image/video AI models. Five raters evaluated task-role accuracy, attire, equipment, lighting, isolation, and demographics. Statistical tests compared differences across models and roles. <b>Results:</b> MRTs were depicted accurately in 82.0% of outputs, while only 56.2% of radiologist images/videos were role-appropriate. Among inaccurate radiologist depictions, 79.1% misrepresented MRTs tasks. Radiologists were more often male (73.8%) and White (79.7%), while MRTs were more diverse. Stethoscope misuse, lack of disability/religious markers, and overuse of business attire for radiologists further reflected bias. <b>Conclusion:</b> Generative AI frequently misrepresents radiologist roles and demographics, reinforcing stereotypes and public confusion. Greater oversight and inclusion standards are needed to ensure equitable AI-generated healthcare content.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Artificial Intelligence in Healthcare and EducationRadiology practices and educationPatient-Provider Communication in Healthcare
Volltext beim Verlag öffnen