scholarly journals Concentrated Differential Privacy: Simplifications, Extensions, and Lower Bounds

Author(s):  
Mark Bun ◽  
Thomas Steinke
Author(s):  
Vitaly Feldman ◽  
Cristóbal Guzmán ◽  
Santosh Vempala

Stochastic convex optimization, by which the objective is the expectation of a random convex function, is an important and widely used method with numerous applications in machine learning, statistics, operations research, and other areas. We study the complexity of stochastic convex optimization given only statistical query (SQ) access to the objective function. We show that well-known and popular first-order iterative methods can be implemented using only statistical queries. For many cases of interest, we derive nearly matching upper and lower bounds on the estimation (sample) complexity, including linear optimization in the most general setting. We then present several consequences for machine learning, differential privacy, and proving concrete lower bounds on the power of convex optimization–based methods. The key ingredient of our work is SQ algorithms and lower bounds for estimating the mean vector of a distribution over vectors supported on a convex body in Rd. This natural problem has not been previously studied, and we show that our solutions can be used to get substantially improved SQ versions of Perceptron and other online algorithms for learning halfspaces.


2019 ◽  
Vol 9 (1) ◽  
Author(s):  
Mark Bun ◽  
Thomas Steinke ◽  
Jonathan Ullman

We consider the problem of answering queries about a sensitive dataset subject to differential privacy. The queries may be chosen adversarially from a larger set $Q$ of allowable queries in one of three ways, which we list in order from easiest to hardest to answer:Offline: The queries are chosen all at once and the differentially private mechanism answers the queries in a single batch. Online: The queries are chosen all at once, but the mechanism only receives the queries in a streaming fashion and must answer each query before seeing the next query. Adaptive: The queries are chosen one at a time and the mechanism must answer each query before the next query is chosen. In particular, each query may depend on the answers given to previous queries.Many differentially private mechanisms are just as efficient in the adaptive model as they are in the offline model. Meanwhile, most lower bounds for differential privacy hold in the offline setting. This suggests that the three models may be equivalent. We prove that these models are all, in fact, distinct. Specifically, we show that there is a family of statistical queries such that exponentially more queries from this family can be answered in the offline model than in the online model. We also exhibit a family of search queries such that exponentially more queries from this family can be answered in the online model than in the adaptive model. We also investigate whether such separations might hold for simple queries like threshold queries over the real line.


2007 ◽  
Author(s):  
T. Lee ◽  
A. Shraibman

Author(s):  
Parinya CHALERMSOOK ◽  
Hiroshi IMAI ◽  
Vorapong SUPPAKITPAISARN

2020 ◽  
Vol 148 (2) ◽  
pp. 321-327
Author(s):  
Rodolfo Gutiérrez-Romo ◽  
Carlos Matheus
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document