Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Labeling messages as AI-generated does not reduce their persuasive effects
3
Zitationen
7
Autoren
2026
Jahr
Abstract
As generative AI enables the creation and dissemination of information at massive scale and speed, it is increasingly important to understand how people perceive AI-generated content. One prominent policy proposal requires explicitly labeling AI-generated content to increase transparency and encourage critical thinking about the information, but prior research has not yet tested the effects of such labels. To address this gap, we conducted a survey experiment ( <mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML"><mml:mi>N</mml:mi> <mml:mo>=</mml:mo> <mml:mn>1,601</mml:mn></mml:math> ) on a diverse sample of Americans, presenting participants with an AI-generated message about several public policies (e.g. allowing colleges to pay student-athletes), randomly assigning whether participants were told the message was generated by (i) an expert AI model, (ii) a human policy expert, or (iii) no label. We found that messages were generally persuasive, influencing participants' views of the policies by 9.74 percentage points on average. However, while 92.0% of participants assigned to the AI and human label conditions believed the authorship labels, labels had no significant effects on participants' attitude change toward the policies, judgments of message accuracy, nor intentions to share the message with others. These patterns were robust across a variety of participant characteristics, including prior knowledge of the policy, prior experience with AI, political party, education level, and age. Given current levels of trust in AI content, these results imply that, while authorship labels would likely enhance transparency, they are unlikely to substantially affect the persuasiveness of the labeled content, highlighting the need for alternative strategies to address challenges posed by AI-generated information.
Ähnliche Arbeiten
The global landscape of AI ethics guidelines
2019 · 4.611 Zit.
The Limitations of Deep Learning in Adversarial Settings
2016 · 3.877 Zit.
Trust in Automation: Designing for Appropriate Reliance
2004 · 3.431 Zit.
Fairness through awareness
2012 · 3.292 Zit.
Mind over Machine: The Power of Human Intuition and Expertise in the Era of the Computer
1987 · 3.184 Zit.