Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Determinants of Trust in Artificial Intelligence (AI) for Health-Related Decision-Making Among Adults in Saudi Arabia: A Cross-Sectional Study
0
Zitationen
3
Autoren
2026
Jahr
Abstract
BACKGROUND/OBJECTIVES: Artificial intelligence (AI) is increasingly integrated into healthcare decision-making. Public trust in AI remains a critical determinant of its acceptance and effective use. Evidence on the factors shaping trust in AI within Middle Eastern contexts, particularly Saudi Arabia, remains limited. Therefore, we aimed to identify the determinants of trust in AI for health-related decision-making and to examine a theory-informed mediation pathway in which patient satisfaction mediates the association between patient-doctor relationships and trust in AI. METHODS: We conducted a cross-sectional, facility-based survey of adults in Saudi Arabia, using an electronic questionnaire distributed in four primary healthcare centers. We performed multiple linear regression to assess the association of trust in AI for health-related decision-making with patient satisfaction, patient-doctor relationships, sociodemographic characteristics, and healthcare-related factors. A mediation analysis was also employed to evaluate the indirect and direct association linking patient-doctor relationships, patient satisfaction, and trust in AI. RESULTS: Our findings showed that patient satisfaction was positively associated with trust in AI (β = 0.54, 95% CI: 0.18-0.90), while patient-doctor relationships showed an inverse association (β = -0.34, 95% CI: -0.48 to -0.20), possibly reflecting a greater reliance on physicians' clinical judgment and a reduced perceived need for AI-supported decision-making. Trust in AI varied across age groups, with a lower trust observed in older age categories compared with younger adults. No strong associations were observed for sex, education, body mass index, or healthcare-related factors. Patient-doctor relationship quality was indirectly associated with trust in AI via patient satisfaction (ACME = 0.138, 95% CI: 0.043-0.246), alongside a direct association with trust in AI (ADE = -0.313, 95% CI: -0.456 to -0.160). This means that patient-doctor relationships influenced trust in AI both directly and indirectly through patient satisfaction, suggesting that, while interpersonal care may reduce the reliance on AI (direct effect), enhancing patient satisfaction can partially offset this effect and promote trust in AI (indirect effect). CONCLUSIONS: These findings highlight that fostering patient-centered care and satisfaction may be crucial for promoting public trust in AI, which has important implications for AI governance, ethical deployment, and the design of AI-supported healthcare systems.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.635 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.543 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 8.051 Zit.
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
2019 · 6.844 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.