Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Introducing a new Trusted Research Environment - the Safe Haven Artificial Platform (SHAIP).
4
Zitationen
5
Autoren
2022
Jahr
Abstract
ObjectivesScotland has an established Trusted Research Environment (TREs) through the Scottish Safe Haven Network. These Safe Havens traditionally service structured datasets, however, researchers increasingly require access to large multi-modal linked datasets that include medical imaging. We therefore introduced an equivalent to access anonymised NHS images and reports.
 ApproachA pan-Scotland collaboration of 15 partners from industry, NHS, and academia, collaborated to design and deploy a Safe Haven equivalent for SMEs and researchers interested in accessing anonymised NHS imaging and reports to allow them to develop, test, and validate AI algorithms for greater patient benefit. Two Safe Haven sites, worked with a leading medical software research and development team, to deliver a secure analytical platform for the research and development of AI for medical applications.
 ResultsThe Safe Haven Artificial Platform (SHAIP) is designed to support development of sophisticated AI components and has been used by several SME’s to undertake exemplar projects in stroke, breast screening, and x-ray interpretation for limb fracture and chest. These researchers were supported by the Safe Haven’s governance processes. SHAIP became the first research environment to be directly linked to the Scottish National PACS archive. Access was provided to project specific deidentified images using ‘Hidden in plain sight’; a privacy and data-structure preserving technique. The researchers were able to upload virtual-machines, or ‘dockers’, to bring their toolsets to their research workspaces. An annotation capability was provided to support ground-truth development for machine learning.
 ConclusionThe project has shown we can pull large scale medical images and reports into a TRE in such a way NHS Boards are reassured by our methods, creating a safe analytical platform for AI development that will benefit patient care by providing faster, cheaper and more accurate solutions.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.231 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.084 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.444 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.423 Zit.