Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Machine Learning Models Disclosure from Trusted Research Environments\n (TRE), Challenges and Opportunities
4
Zitationen
5
Autoren
2021
Jahr
Abstract
Artificial intelligence (AI) applications in healthcare and medicine have\nincreased in recent years. To enable access to personal data, Trusted Research\nenvironments (TREs) provide safe and secure environments in which researchers\ncan access sensitive personal data and develop Artificial Intelligence (AI) and\nMachine Learning models. However currently few TREs support the use of\nautomated AI-based modelling using Machine Learning. Early attempts have been\nmade in the literature to present and introduce privacy preserving machine\nlearning from the design point of view [1]. However, there exists a gap in the\npractical decision-making guidance for TREs in handling models disclosure.\nSpecifically, the use of machine learning creates a need to disclose new types\nof outputs from TREs, such as trained machine learning models. Although TREs\nhave clear policies for the disclosure of statistical outputs, the extent to\nwhich trained models can leak personal training data once released is not well\nunderstood and guidelines do not exist within TREs for the safe disclosure of\nthese models.\n In this paper we introduce the challenge of disclosing trained machine\nlearning models from TREs. We first give an overview of machine learning models\nin general and describe some of their applications in healthcare and medicine.\nWe define the main vulnerabilities of trained machine learning models in\ngeneral. We also describe the main factors affecting the vulnerabilities of\ndisclosing machine learning models. This paper also provides insights and\nanalyses methods that could be introduced within TREs to mitigate the risk of\nprivacy breaches when disclosing trained models.\n
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.214 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.071 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.429 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.418 Zit.