TY - JOUR
T1 - Statistical Query Algorithms for Mean Vector Estimation and Stochastic Convex Optimization
AU - Feldman, Vitaly
AU - Guzmán Paredes, Cristóbal Andrés
AU - Vempala, Santosh
N1 - Funding Information:
Funding: The work of C. Guzmán was partially supported by Grant FONDECYT 11160939, and by ANID – Millennium Science Initiative Program – NCN17_059.
Publisher Copyright:
© 2021 INFORMS.
PY - 2021/8
Y1 - 2021/8
N2 - Stochastic convex optimization, by which the objective is the expectation of a random convex function, is an important and widely used method with numerous applications in machine learning, statistics, operations research, and other areas. We study the complexity of stochastic convex optimization given only statistical query (SQ) access to the objective function. We show that well-known and popular first-order iterative methods can be implemented using only statistical queries. For many cases of interest, we derive nearly matching upper and lower bounds on the estimation (sample) complexity, including linear optimization in the most general setting. We then present several consequences for machine learning, differential privacy, and proving concrete lower bounds on the power of convex optimization–based methods. The key ingredient of our work is SQ algorithms and lower bounds for estimating the mean vector of a distribution over vectors supported on a convex body in Rd. This natural problem has not been previously studied, and we show that our solutions can be used to get substantially improved SQ versions of Perceptron and other online algorithms for learning halfspaces.
AB - Stochastic convex optimization, by which the objective is the expectation of a random convex function, is an important and widely used method with numerous applications in machine learning, statistics, operations research, and other areas. We study the complexity of stochastic convex optimization given only statistical query (SQ) access to the objective function. We show that well-known and popular first-order iterative methods can be implemented using only statistical queries. For many cases of interest, we derive nearly matching upper and lower bounds on the estimation (sample) complexity, including linear optimization in the most general setting. We then present several consequences for machine learning, differential privacy, and proving concrete lower bounds on the power of convex optimization–based methods. The key ingredient of our work is SQ algorithms and lower bounds for estimating the mean vector of a distribution over vectors supported on a convex body in Rd. This natural problem has not been previously studied, and we show that our solutions can be used to get substantially improved SQ versions of Perceptron and other online algorithms for learning halfspaces.
U2 - 10.1287/moor.2020.1111
DO - 10.1287/moor.2020.1111
M3 - Article
SN - 0364-765X
VL - 46
SP - 912
EP - 945
JO - Mathematics of operations research
JF - Mathematics of operations research
IS - 3
ER -