Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Privacy preserving neural networks for electronic health records de-identification
12
Zitationen
4
Autoren
2021
Jahr
Abstract
Over the last decade, significant improvements and efforts in digitizing healthcare provided us with a sizeable collection of electronic medical records. These Electronic Health Records (EHRs), especially the clinical narratives, brimming with hidden knowledge to be discovered, contain sensitive information about the patient. For this reason, medical institutions are legally prohibited from publishing these data in the raw format to the public. Hence, the recent surge towards finding a solution for de-identification detects and removes sensitive information from clinical narratives. Since 2016, we have seen several deep learning-based approaches for de-identification, which achieved over 98% accuracy. However, these models are trained with sensitive information and can unwittingly memorize some of its training data, and a careful analysis of these models can reveal patients' data. In this work, we propose a differentially private ensemble framework for de-identification, allowing medical researchers to collaborate through publicly publishing the de-identification models. To the best of our knowledge, this is the first privacy-preserving machine learning approach for the de-identification of EHRs. Experiments in three different datasets showed competitive results compared to the state-of-the-art methods with guaranteed differential privacy.
Ähnliche Arbeiten
k-ANONYMITY: A MODEL FOR PROTECTING PRIVACY
2002 · 8.397 Zit.
Calibrating Noise to Sensitivity in Private Data Analysis
2006 · 6.878 Zit.
Deep Learning with Differential Privacy
2016 · 5.604 Zit.
Communication-Efficient Learning of Deep Networks from Decentralized\n Data
2016 · 5.592 Zit.
Large-Scale Machine Learning with Stochastic Gradient Descent
2010 · 5.569 Zit.