Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Teaching Parrots to See Red: Self-Audits of Generative Language Models Overlook Sociotechnical Harms
1
Zitationen
2
Autoren
2025
Jahr
Abstract
The release of ChatGPT as a “low-key research preview” and its viral growth spurred a gold rush among tech companies marketing generative AI (GenAI) as a universal tool. In 2023, the U.S. secured voluntary commitments from top AI developers, including OpenAI, Google, Meta, and Anthropic, to conduct self-audits ensuring model safety before release. However, these models exhibit widespread biases, including by race and gender, unjustly discriminating against users. To inspect this contradiction, we review ten corporate self-audits, finding a notable absence of real-world use cases in sectors like education, creative works, and public policy. Instead, audits focus on thwarting adversarial consumers in hypothetical scenarios and rely on GenAI models to approximate human impacts. This approach places consumers at risk by impairing the mitigation of representational, allocational, and quality-of-service harms. We conclude with recommendations to address audit gaps and protect GenAI consumers.
Ähnliche Arbeiten
The global landscape of AI ethics guidelines
2019 · 4.683 Zit.
The Limitations of Deep Learning in Adversarial Settings
2016 · 3.879 Zit.
Trust in Automation: Designing for Appropriate Reliance
2004 · 3.495 Zit.
Fairness through awareness
2012 · 3.298 Zit.
Mind over Machine: The Power of Human Intuition and Expertise in the Era of the Computer
1987 · 3.184 Zit.