DocumentCode
710131
Title
Conservative or liberal? Personalized differential privacy
Author
Jorgensen, Zach ; Ting Yu ; Cormode, Graham
Author_Institution
North Carolina State Univ., Raleigh, NC, USA
fYear
2015
fDate
13-17 April 2015
Firstpage
1023
Lastpage
1034
Abstract
Differential privacy is widely accepted as a powerful framework for providing strong, formal privacy guarantees for aggregate data analysis. A limitation of the model is that the same level of privacy protection is afforded for all individuals. However, it is common that the data subjects have quite different expectations regarding the acceptable level of privacy for their data. Consequently, differential privacy may lead to insufficient privacy protection for some users, while over-protecting others. We argue that by accepting that not all users require the same level of privacy, a higher level of utility can often be attained by not providing excess privacy to those who do not want it. We propose a new privacy definition called personalized differential privacy (PDP), a generalization of differential privacy in which users specify a personal privacy requirement for their data. We then introduce several novel mechanisms for achieving PDP. Our primary mechanism is a general one that automatically converts any existing differentially private algorithm into one that satisfies PDP. We also present a more direct approach for achieving PDP, inspired by the well-known exponential mechanism. We demonstrate our framework through extensive experiments on real and synthetic data.
Keywords
data protection; PDP; aggregate data analysis; exponential mechanism; formal privacy guarantees; personalized differential privacy; primary mechanism; privacy level; privacy protection; real data; synthetic data; utility level; Lead; Privacy;
fLanguage
English
Publisher
ieee
Conference_Titel
Data Engineering (ICDE), 2015 IEEE 31st International Conference on
Conference_Location
Seoul
Type
conf
DOI
10.1109/ICDE.2015.7113353
Filename
7113353
Link To Document