Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
Date
Availability
1-1 of 1
Xiyuan Wei
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
Obtaining Lower Query Complexities Through Lightweight Zeroth-Order Proximal Gradient Algorithms
UnavailablePublisher: Journals Gateway
Neural Computation (2024) 36 (5): 897–935.
Published: 23 April 2024
FIGURES
| View all 4
Abstract
View articletitled, Obtaining Lower Query Complexities Through Lightweight Zeroth-Order Proximal Gradient Algorithms
View
PDF
for article titled, Obtaining Lower Query Complexities Through Lightweight Zeroth-Order Proximal Gradient Algorithms
Zeroth-order (ZO) optimization is one key technique for machine learning problems where gradient calculation is expensive or impossible. Several variance, reduced ZO proximal algorithms have been proposed to speed up ZO optimization for nonsmooth problems, and all of them opted for the coordinated ZO estimator against the random ZO estimator when approximating the true gradient, since the former is more accurate. While the random ZO estimator introduces a larger error and makes convergence analysis more challenging compared to coordinated ZO estimator, it requires only O ( 1 ) computation, which is significantly less than O ( d ) computation of the coordinated ZO estimator, with d being dimension of the problem space. To take advantage of the computationally efficient nature of the random ZO estimator, we first propose a ZO objective decrease (ZOOD) property that can incorporate two different types of errors in the upper bound of convergence rate. Next, we propose two generic reduction frameworks for ZO optimization, which can automatically derive the convergence results for convex and nonconvex problems, respectively, as long as the convergence rate for the inner solver satisfies the ZOOD property. With the application of two reduction frameworks on our proposed ZOR-ProxSVRG and ZOR-ProxSAGA, two variance-reduced ZO proximal algorithms with fully random ZO estimators, we improve the state-of-the-art function query complexities from O min d n 1 / 2 ε 2 , d ε 3 to O ˜ n + d ε 2 under d > n 1 2 for nonconvex problems, and from O d ε 2 to O ˜ n log 1 ε + d ε for convex problems. Finally, we conduct experiments to verify the superiority of our proposed methods.