OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 24.03.2026, 18:29

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Who Is to Blame for the Bias in Visualizations, ChatGPT or DALL-E?

2025·7 Zitationen·AIOpen Access
Volltext beim Verlag öffnen

7

Zitationen

1

Autoren

2025

Jahr

Abstract

Due to range of factors in the development stage, generative artificial intelligence (AI) models cannot be completely free from bias. Some biases are introduced by the quality of training data, and developer influence during both design and training of the large language models (LLMs), while others are introduced in the text-to-image (T2I) visualization programs. The bias and initialization at the interface between LLMs and T2I applications has not been examined to date. This study analyzes 770 images of librarians and curators generated by DALL-E from ChatGPT-4o prompts to investigate the source of gender, ethnicity, and age biases in these visualizations. Comparing prompts generated by ChatGPT-4o with DALL-E’s visual interpretations, the research demonstrates that DALL-E primarily introduces biases when ChatGPT-4o provides non-specific prompts. This highlights the potential for generative AI to perpetuate and amplify harmful stereotypes related to gender, age, and ethnicity in professional roles.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Artificial Intelligence in Healthcare and EducationEthics and Social Impacts of AIComputational and Text Analysis Methods
Volltext beim Verlag öffnen