ADePT: Auto-encoder based Differentially Private Text Transformation

Satyapriya Krishna, Rahul Gupta, Christophe Dupuy

Machine Learning for NLP Short paper Paper

Gather-2D: Apr 22, Gather-2D: Apr 22 (13:00-15:00 UTC) [Join Gather Meeting]

You can open the pre-recorded video in separate windows.

Abstract: Privacy is an important concern when building statistical models on data containing personal information. Differential privacy offers a strong definition of privacy and can be used to solve several privacy concerns. Multiple solutions have been proposed for the differentially-private transformation of datasets containing sensitive information. However, such transformation algorithms offer poor utility in Natural Language Processing (NLP) tasks due to noise added in the process. This paper addresses this issue by providing a utility-preserving differentially private text transformation algorithm using auto-encoders. Our algorithm transforms text to offer robustness against attacks and produces transformations with high semantic quality that perform well on downstream NLP tasks. We prove our algorithm's theoretical privacy guarantee and assess its privacy leakage under Membership Inference Attacks (MIA) on models trained with transformed data. Our results show that the proposed model performs better against MIA attacks while offering lower to no degradation in the utility of the underlying transformation process compared to existing baselines.
NOTE: Video may display a random order of authors. Correct author list is at the top of this page.

Connected Papers in EACL2021

Similar Papers

On-Device Text Representations Robust To Misspellings via Projections
Chinnadhurai Sankar, Sujith Ravi, Zornitsa Kozareva,
Data Augmentation for Hypernymy Detection
Thomas Kober, Julie Weeds, Lorenzo Bertolini, David Weir,
Frequency-Guided Word Substitutions for Detecting Textual Adversarial Examples
Maximilian Mozes, Pontus Stenetorp, Bennett Kleinberg, Lewis Griffin,