Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Red Teaming for Generative AI, Report on a Copyright-Focused Exercise Completed in an Academic Medical Center
0
Zitationen
41
Autoren
2025
Jahr
Abstract
Background: Generative artificial intelligence (AI) deployment in academic medical settings raises copyright compliance concerns. Dana-Farber Cancer Institute implemented GPT4DFCI, an internal generative AI tool utilizing OpenAI models, that is approved for enterprise use in research and operations. Given (1) the exceptionally broad adoption of the tool in our organization, (2) our research mission, and (3) the shared responsibility model required to benefit from Customer Copyright Commitment in Azure OpenAI Service products, we deemed rigorous copyright compliance testing necessary. Case Description: We conducted a structured red teaming exercise in Nov. 2024, with 42 participants from academic, industry, and government institutions. Four teams attempted to extract copyrighted content from GPT4DFCI across four domains: literary works, news articles, scientific publications, and access-restricted clinical notes. Teams successfully extracted verbatim book dedications and near-exact passages through various strategies. News article extraction failed despite jailbreak attempts. Scientific article reproduction yielded only high-level summaries. Clinical note testing revealed appropriate privacy safeguards. Discussion: The successful extraction of literary content indicates potential copyrighted material presence in training data, necessitating inference-time filtering. Differential success rates across content types suggest varying protective mechanisms. The event led to implementation of a copyright-specific meta-prompt in GPT4DFCI; this mitigation has been in production since Jan. 2025. Conclusion: Systematic red teaming revealed specific vulnerabilities in generative AI copyright compliance, leading to concrete mitigation strategies. Academic medical institutions deploying generative AI should implement continuous testing protocols to ensure legal and ethical compliance.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.231 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.084 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.444 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.423 Zit.
Autoren
- James Wen
- Sahil Nalawade
- Zhigang Liang
- Catherine Bielick
- Marisa Ferrara Boston
- Alexander Chowdhury
- A. Collin
- Luigi De Angelis
- Jacob Ellen
- Heather Frase
- Rodrigo R. Gameiro
- Juan Manuel Parrilla Gutiérrez
- Pravin Bhaskar Kadam
- Murat Keceli
- Srikanth V. Krishnamurthy
- Anne Kwok
- Yanan Lance Lu
- Heather Mattie
- Liam G. McCoy
- Katherine Miller
- Ann W Morgan
- Marlene Louisa Moerig
- Trang T. Nguyen
- Alexander Owen-Post
- A. Ruiz-Jimeno
- Sreekar Reddy Puchala
- Soujanya Samineni
- Takeshi Tohyama
- Varun Ullanat
- Carmine Valenza
- Camilo Velez
- Pengcheng Wang
- Anna Wuest
- Yuxiang Zhou
- Yingde Zhu
- Jason M. Johnson
- Naomi Lenane
- Jennifer L. Willcox
- Francesco Vitiello
- Leo Anthony Celi
- Renato Umeton