Inexact proximal stochastic second-order methods for nonconvex composite optimization
Document Type
Article
Publication Date
7-3-2020
Abstract
In this paper, we propose a framework of Inexact Proximal Stochastic Second-order (IPSS) method for solving nonconvex optimization problems, whose objective function consists of an average of finitely many, possibly weakly, smooth functions and a convex but possibly nonsmooth function. At each iteration, IPSS inexactly solves a proximal subproblem constructed by using some positive definite matrix which could capture the second-order information of original problem. Proper tolerances are given for the subproblem solution in order to maintain global convergence and the desired overall complexity of the algorithm. Under mild conditions, we analyse the computational complexity related to the evaluations on the component gradient of the smooth function. We also investigate the number of evaluations of subgradient when using an iterative subgradient method to solve the subproblem. In addition, based on IPSS, we propose a linearly convergent algorithm under the proximal Polyak–Łojasiewicz condition. Finally, we extend the analysis to problems with weakly smooth function and obtain the computational complexity accordingly.
Publication Source (Journal or Book title)
Optimization Methods and Software
First Page
808
Last Page
835
Recommended Citation
Wang, X., & Zhang, H. (2020). Inexact proximal stochastic second-order methods for nonconvex composite optimization. Optimization Methods and Software, 35 (4), 808-835. https://doi.org/10.1080/10556788.2020.1713128