Journal of Privacy and Confidentiality
Latest Publications


TOTAL DOCUMENTS

161
(FIVE YEARS 43)

H-INDEX

15
(FIVE YEARS 2)

Published By Journal Of Privacy And Confidentiality

2575-8527

2021 ◽  
Vol 11 (3) ◽  
Author(s):  
Ryan Rogers ◽  
Subbu Subramaniam ◽  
Sean Peng ◽  
David Durfee ◽  
Seunghyun Lee ◽  
...  

We present a privacy system that leverages differential privacy to protect LinkedIn members' data while also providing audience engagement insights to enable marketing analytics related applications. We detail the differentially private algorithms and other privacy safeguards used to provide results that can be used with existing real-time data analytics platforms, specifically with the open sourced Pinot system. Our privacy system provides user-level privacy guarantees. As part of our privacy system, we include a budget management service that enforces a strict differential privacy budget on the returned results to the analyst. This budget management service brings together the latest research in differential privacy into a product to maintain utility given a fixed differential privacy budget.


2021 ◽  
Vol 11 (3) ◽  
Author(s):  
Ryan McKenna ◽  
Gerome Miklau ◽  
Daniel Sheldon

We propose a general approach for differentially private synthetic data generation, that consists of three steps: (1) select a collection of low-dimensional marginals, (2) measure those marginals with a noise addition mechanism, and (3) generate synthetic data that preserves the measured marginals well. Central to this approach is Private-PGM, a post-processing method that is used to estimate a high-dimensional data distribution from noisy measurements of its marginals. We present two mechanisms, NIST-MST and MST, that are instances of this general approach. NIST-MST was the winning mechanism in the 2018 NIST differential privacy synthetic data competition, and MST is a new mechanism that can work in more general settings, while still performing comparably to NIST-MST. We believe our general approach should be of broad interest, and can be adopted in future mechanisms for synthetic data generation.


2021 ◽  
Vol 11 (3) ◽  
Author(s):  
Sivakanth Gopi ◽  
Pankaj Gulhane ◽  
Janardhan Kulkarni ◽  
Judy Hanwen Shen ◽  
Milad Shokouhi ◽  
...  

We study the basic operation of set union in the global model of differential privacy. In this problem, we are given a universe $U$ of items, possibly of infinite size, and a database $D$ of users. Each user $i$ contributes a subset $W_i \subseteq U$ of items. We want an ($\epsilon$,$\delta$)-differentially private algorithm which outputs a subset $S \subset \cup_i W_i$ such that the size of $S$ is as large as possible. The problem arises in countless real world applications; it is particularly ubiquitous in natural language processing (NLP) applications as vocabulary extraction. For example, discovering words, sentences, $n$-grams etc., from private text data belonging to users is an instance of the set union problem.Known algorithms for this problem proceed by collecting a subset of items from each user, taking the union of such subsets, and disclosing the items whose noisy counts fall above a certain threshold. Crucially, in the above process, the contribution of each individual user is always independent of the items held by other users, resulting in a wasteful aggregation process, where some item counts happen to be way above the threshold. We deviate from the above paradigm by allowing users to contribute their items in a {\em dependent fashion}, guided by a {\em policy}. In this new setting ensuring privacy is significantly delicate. We prove that any policy which has certain {\em contractive} properties would result in a differentially private algorithm. We design two new algorithms for differentially private set union, one using Laplace noise and other Gaussian noise, which use $\ell_1$-contractive and $\ell_2$-contractive policies respectively and provide concrete examples of such policies. Our experiments show that the new algorithms in combination with our policies significantly outperform previously known mechanisms for the problem.


2021 ◽  
Vol 11 (3) ◽  
Author(s):  
Ergute Bao ◽  
Xiaokui Xiao ◽  
Jun Zhao ◽  
Dongping Zhang ◽  
Bolin Ding

This paper describes PrivBayes, a differentially private method for generating synthetic datasets that was used in the 2018 Differential Privacy Synthetic Data Challenge organized by NIST.


2021 ◽  
Vol 11 (2) ◽  
Author(s):  
Lars Vilhuber

The present issue provides a diverse selection of articles. We introduced a new typeof article, “Perspectives,” in the previous issue, and continue with two such articles in thecurrent issue, both drawn again from presentations made at the October 2020 CanadianResearch Data Centre Network (CRDCN) conference. We also have a new article on the topic of “Privacy Challenges,” as well as  the first of several journal versions of contributions to TPDP 2020. We open with a regular article on the topic of  "Differentiallyprivate false discovery rate control."


2021 ◽  
Vol 11 (2) ◽  
Author(s):  
Yuval Dagan ◽  
Vitaly Feldman

Local differential privacy (LDP) is a model where users send privatized data to an untrusted central server whose goal it to solve some data analysis task. In the non-interactive version of this model the protocol consists of a single round in which a server sends requests to all users then receives their responses. This version is deployed in industry due to its practical advantages and has attracted significant research interest. Our main result is an exponential lower bound on the number of samples necessary to solve the standard task of learning a large-margin linear separator in the non-interactive LDP model. Via a standard reduction this lower bound implies an exponential lower bound for stochastic convex optimization and specifically, for learning linear models with a convex, Lipschitz and smooth loss. These results answer the questions posed by Smith, Thakurta, and Upadhyay (IEEE Symposium on Security and Privacy 2017) and Daniely and Feldman (NeurIPS 2019). Our lower bound relies on a new technique for constructing pairs of distributions with nearly matching moments but whose supports can be nearly separated by a large margin hyperplane. These lower bounds also hold in the model where communication from each user is limited and follow from a lower bound on learning using non-adaptive statistical queries.


2021 ◽  
Vol 11 (2) ◽  
Author(s):  
Cynthia Dwork ◽  
Weijie Su ◽  
Li Zhang

Differential privacy provides a rigorous framework for privacy-preserving data analysis. This paper proposes the first differentially private procedure for controlling the false discovery rate (FDR) in multiple hypothesis testing. Inspired by the Benjamini-Hochberg procedure (BHq), our approach is to first repeatedly add noise to the logarithms of the p-values to ensure differential privacy and to select an approximately smallest p-value serving as a promising candidate at each iteration; the selected p-values are further supplied to the BHq and our private procedure releases only the rejected ones. Moreover, we develop a new technique that is based on a backward submartingale for proving FDR control of a broad class of multiple testing procedures, including our private procedure, and both the BHq step- up and step-down procedures. As a novel aspect, the proof works for arbitrary dependence between the true null and false null test statistics, while FDR control is maintained up to a small multiplicative factor.


2021 ◽  
Vol 11 (2) ◽  
Author(s):  
Tianhao Wang ◽  
Ninghui Li ◽  
Zhikun Zhang

We summarize the experience of participating in two differential privacycompetitions organized by the National Institute of Standards and Technology (NIST). Inthis paper, we document our experiences in the competition, the approaches we have used,the lessons we have learned, and our call to the research community to further bridge thegap between theory and practice in DP research.


2021 ◽  
Vol 11 (2) ◽  
Author(s):  
Roxane SILBERMAN

Over the past twenty years, in various countries, secure access to data for the members of the research community was eased in a significant manner. Such data involve microdata and granular data from administrative records and detailed individual surveys. While some difficulties remain, the scene has been extensively redesigned, and new players emerged on both sides of the fence: data holders and users, both challenging what seemed to be well-established boundaries. In the French case, access to confidential data for research purposes has been carefully facilitated. The paper analyses the French developments and current achievements, providing insights into how obstacles can be overcome and newly emerging issues


Sign in / Sign up

Export Citation Format

Share Document