Artículos de revistas
Extragradient method with variance reduction for stochastic variational inequalities
Fecha
2017Registro en:
SIAM J. Optim. Vol. 27, No. 2, pp. 686-724
10.1137/15M1031953
Autor
Iusem, A. N.
Jofré Cáceres, René
Oliveira, R. I.
Thompson, P.
Institución
Resumen
We propose an extragradient method with stepsizes bounded away from zero for stochastic variational inequalities requiring only pseudomonotonicity. We provide convergence and complexity analysis, allowing for an unbounded feasible set, unbounded operator, and nonuniform variance of the oracle, and, also, we do not require any regularization. Alongside the stochastic approximation procedure, we iteratively reduce the variance of the stochastic error. Our method attains the optimal oracle complexity O(1/is an element of(2)) (up to a logarithmic term) and a faster rate O(1/K) in terms of the mean (quadratic) natural residual and the D-gap function, where K is the number of iterations required for a given tolerance is an element of > 0. Such convergence rate represents an acceleration with respect to the stochastic error. The generated sequence also enjoys a new feature: the sequence is bounded in L-P if the stochastic error has finite p-moment. Explicit estimates for the convergence rate, the oracle complexity, and the p-moments are given depending on problem parameters and distance of the initial iterate to the solution set. Moreover, sharper constants are possible if the variance is uniform over the solution set or the feasible set. Our results provide new classes of stochastic variational inequalities for which a convergence rate of O(1/K) holds in terms of the mean-squared distance to the solution set. Our analysis includes the distributed solution of pseudomonotone Cartesian variational inequalities under partial coordination of parameters between users of a network.