OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 13.03.2026, 11:02

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Inside the Black Box: Detecting Data Leakage in Pre-Trained Language Encoders

2024·1 Zitationen·Frontiers in artificial intelligence and applicationsOpen Access
Volltext beim Verlag öffnen

1

Zitationen

7

Autoren

2024

Jahr

Abstract

Despite being prevalent in the general field of Natural Language Processing (NLP), pre-trained language models inherently carry privacy and copyright concerns due to their nature of training on large-scale web-scraped data. In this paper, we pioneer a systematic exploration of such risks associated with pre-trained language encoders, specifically focusing on the membership leakage of pre-training data exposed through downstream models adapted from pre-trained language encoders–an aspect largely overlooked in existing literature. Our study encompasses comprehensive experiments across four types of pre-trained encoder architectures, three representative downstream tasks, and five benchmark datasets. Intriguingly, our evaluations reveal, for the first time, the existence of membership leakage even when only the black-box output of the downstream model is exposed, highlighting a privacy risk far greater than previously assumed. Alongside, we present in-depth analysis and insights toward guiding future researchers and practitioners in addressing the privacy considerations in developing pre-trained language models.

Ähnliche Arbeiten

Autoren

Institutionen

Themen

Online Learning and AnalyticsAdvanced Malware Detection TechniquesArtificial Intelligence in Healthcare and Education
Volltext beim Verlag öffnen