Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
The Potential of Generative AI to Support Medicare Decision Making
0
Zitationen
7
Autoren
2025
Jahr
Abstract
Abstract In 2022, ∼ 99% of American adults age 65+ yr were enrolled in Medicare, a complex, difficult to use insurance system. We describe a multi-pronged approach to assessing the potential value of AI to support Medicare decision-making processes that includes querying subject matter experts who provide Medicare advice, assessing Medicare users’ knowledge, preferences, and abilities through interviews and while interacting with knowledge sources, and evaluating existing AI tools for accuracy, reliability, and conciseness. In this presentation we present findings about the accuracy and reliability of digital assistants in answering Medicare questions, a new tool for assessing Medicare knowledge, and a study to assess preferences and performance with Medicare information sources. Generative AI (ChatGPT, Bard) were highly accurate (>90%) and reliable as well as superior to the average Medicare beneficiary, and much superior to digital home assistants (Alexa and Google Home Assistant). The Medicare Proficiency Questionnaire proved to be a short, reliable, and valid scale of older adults’ Medicare knowledge, differentiating between Medicare enrollees and non-enrollees. Prior access to the Medicare website and the Medicare and You Handbook mediated the relationship between knowledge and education level as well as knowledge and Medicare enrollment status, suggesting that understanding Medicare information sources may be critical for designing AI decision support tools and training. We also describe early results from an observational study with Medicare-enrolled and unenrolled older adults examining preferences for, and performance with, the Medicare.gov website, the Medicare and You Handbook, and Gemini (a Large Language Model).
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.292 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.143 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.539 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.452 Zit.