OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 14.03.2026, 16:23

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

A double-blind, crossover, non-inferiority randomized controlled trial where primary care providers and patients compare human- and AI-generated digital health messages: the AI-CARE study protocol

2025·0 ZitationenOpen Access
Volltext beim Verlag öffnen

0

Zitationen

9

Autoren

2025

Jahr

Abstract

ABSTRACT Introduction Primary care is facing multiple crises, including an increase in health misinformation. Digital health messaging by primary care providers has been shown to reach a diverse patient population. With the uptake of Generative Artificial Intelligence (GenAI) usage in healthcare, there is an important opportunity to rapidly create messages that are tailored to different populations and conditions. However, thoroughly assessing AI-generated content is essential, as GenAI raises concerns regarding its accuracy, understandability, actionability, and bias perpetuation. We aim to investigate whether digital health messages created by GenAI are evaluated as noninferior compared to those created by human experts. Methods and analysis The AI-CARE (AI to Create Accessible and Reliable patient Education materials) study is a double-blind, crossover, non-inferiority randomized controlled trial. Data collection began on May 30, 2025, and is expected to be completed at the end of April 2026. Over 12 months, 192 messages on 48 topics will be written: half by primary care and public health experts and half by a GenAI tool (OpenAI’s ChatGPT). Review Panels composed of 24 primary care providers and 24 patients will evaluate these messages using an Evaluation Grid developed to assess the messages’ quality of information, adaptation to the target audience, relevance and usefulness, and readiness to be shared with patients. Evaluations will be completed via online REDCap surveys and the order in which the 192 messages appear will be randomized and will vary between individuals. Participants and analysts will be blinded to the generation source. The primary outcome will be the Clarity and Understandability score. Ethics and dissemination The Research Ethics Boards of the Hôpital Montfort (24-25-11-038) and the University of Ottawa (S-12-24-11153) formally approved this study in December 2024. Reported data will be grouped and anonymized for dissemination in peer-reviewed scientific journals and conferences. Trial registration number NCT06997107 ARTICLE SUMMARY Strengths and limitations of this study The AI-CARE study allows for within-participant comparison between human- and AI-generated digital health messages, minimizing variability due to individual differences. The Review Panels are diverse and composed of primary care providers and patients currently practicing in or using the healthcare system in five Canadian provinces. The developed Evaluation Grid allows for the assessment of multiple aspects related to digital health messages: quality of information, adaptability to the target audience, relevance and usefulness, and readiness to be shared with patients. One limitation is that messages generated by AI are created using only one LLM (Open AI’s ChatGPT). Due to the nature and location of recruitment, we may introduce selection bias (participants already engaged in research and interested in digital communication and AI) and the racial diversity of our study population may be limited.

Ähnliche Arbeiten