TY - JOUR
T1 - Convergence guarantees for forward gradient descent in the linear regression model
AU - Bos, Thijs
AU - Schmidt-Hieber, Johannes
N1 - Publisher Copyright:
© 2024 The Author(s)
PY - 2024/12
Y1 - 2024/12
N2 - Renewed interest in the relationship between artificial and biological neural networks motivates the study of gradient-free methods. Considering the linear regression model with random design, we theoretically analyze in this work the biologically motivated (weight-perturbed) forward gradient scheme that is based on random linear combination of the gradient. If d denotes the number of parameters and k the number of samples, we prove that the mean squared error of this method converges for k≳d2log(d) with rate d2log(d)/k. Compared to the dimension dependence d for stochastic gradient descent, an additional factor dlog(d) occurs.
AB - Renewed interest in the relationship between artificial and biological neural networks motivates the study of gradient-free methods. Considering the linear regression model with random design, we theoretically analyze in this work the biologically motivated (weight-perturbed) forward gradient scheme that is based on random linear combination of the gradient. If d denotes the number of parameters and k the number of samples, we prove that the mean squared error of this method converges for k≳d2log(d) with rate d2log(d)/k. Compared to the dimension dependence d for stochastic gradient descent, an additional factor dlog(d) occurs.
KW - Convergence rates
KW - Estimation
KW - Gradient descent
KW - Linear model
KW - Zeroth-order methods
UR - http://www.scopus.com/inward/record.url?scp=85189826291&partnerID=8YFLogxK
U2 - 10.1016/j.jspi.2024.106174
DO - 10.1016/j.jspi.2024.106174
M3 - Article
AN - SCOPUS:85189826291
SN - 0378-3758
VL - 233
JO - Journal of statistical planning and inference
JF - Journal of statistical planning and inference
M1 - 106174
ER -