2015 IEEE 31st International Conference on Data Engineering (ICDE) (2015)
Seoul, South Korea
April 13, 2015 to April 17, 2015
Zach Jorgensen , North Carolina State University, USA
Ting Yu , Qatar Computing Research Institute, Qatar
Graham Cormode , University of Warwick, UK
Differential privacy is widely accepted as a powerful framework for providing strong, formal privacy guarantees for aggregate data analysis. A limitation of the model is that the same level of privacy protection is afforded for all individuals. However, it is common that the data subjects have quite different expectations regarding the acceptable level of privacy for their data. Consequently, differential privacy may lead to insufficient privacy protection for some users, while over-protecting others. We argue that by accepting that not all users require the same level of privacy, a higher level of utility can often be attained by not providing excess privacy to those who do not want it. We propose a new privacy definition called personalized differential privacy (PDP), a generalization of differential privacy in which users specify a personal privacy requirement for their data. We then introduce several novel mechanisms for achieving PDP. Our primary mechanism is a general one that automatically converts any existing differentially private algorithm into one that satisfies PDP. We also present a more direct approach for achieving PDP, inspired by the well-known exponential mechanism. We demonstrate our framework through extensive experiments on real and synthetic data.
Zach Jorgensen, Ting Yu, Graham Cormode, "Conservative or liberal? Personalized differential privacy", 2015 IEEE 31st International Conference on Data Engineering (ICDE), vol. 00, no. , pp. 1023-1034, 2015, doi:10.1109/ICDE.2015.7113353