Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Can large language models be privacy preserving and fair medical coders?
0
Zitationen
4
Autoren
2024
Jahr
Abstract
Protecting patient data privacy is a critical concern when deploying machine learning algorithms in healthcare. Differential privacy (DP) is a common method for preserving privacy in such settings and, in this work, we examine two key trade-offs in applying DP to the NLP task of medical coding (ICD classification). Regarding the privacy-utility trade-off, we observe a significant performance drop in the privacy preserving models, with more than a 40% reduction in micro F1 scores on the top 50 labels in the MIMIC-III dataset. From the perspective of the privacy-fairness trade-off, we also observe an increase of over 3% in the recall gap between male and female patients in the DP models. Further understanding these trade-offs will help towards the challenges of real-world deployment.
Ähnliche Arbeiten
The meaning and use of the area under a receiver operating characteristic (ROC) curve.
1982 · 21.516 Zit.
Adapting a clinical comorbidity index for use with ICD-9-CM administrative databases
1992 · 10.482 Zit.
Coding Algorithms for Defining Comorbidities in ICD-9-CM and ICD-10 Administrative Data
2005 · 10.464 Zit.
Comorbidity Measures for Use with Administrative Data
1998 · 9.783 Zit.
Evaluating the added predictive ability of a new marker: From area under the ROC curve to reclassification and beyond
2007 · 6.232 Zit.