Loading…

Oracle Complexity Separation in Convex Optimization

Many convex optimization problems have structured objective function written as a sum of functions with different types of oracles (full gradient, coordinate derivative, stochastic gradient) and different evaluation complexity of these oracles. In the strongly convex case these functions also have d...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2022-03
Main Authors: Ivanova, Anastasiya, Vorontsova, Evgeniya, Pasechnyuk, Dmitry, Gasnikov, Alexander, Dvurechensky, Pavel, Dvinskikh, Darina, Tyurin, Alexander
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Many convex optimization problems have structured objective function written as a sum of functions with different types of oracles (full gradient, coordinate derivative, stochastic gradient) and different evaluation complexity of these oracles. In the strongly convex case these functions also have different condition numbers, which eventually define the iteration complexity of first-order methods and the number of oracle calls required to achieve given accuracy. Motivated by the desire to call more expensive oracle less number of times, in this paper we consider minimization of a sum of two functions and propose a generic algorithmic framework to separate oracle complexities for each component in the sum. As a specific example, for the \(\mu\)-strongly convex problem \(\min_{x\in \mathbb{R}^n} h(x) + g(x)\) with \(L_h\)-smooth function \(h\) and \(L_g\)-smooth function \(g\), a special case of our algorithm requires, up to a logarithmic factor, \(O(\sqrt{L_h/\mu})\) first-order oracle calls for \(h\) and \(O(\sqrt{L_g/\mu})\) first-order oracle calls for \(g\). Our general framework covers also the setting of strongly convex objectives, the setting when \(g\) is given by coordinate derivative oracle, and the setting when \(g\) has a finite-sum structure and is available through stochastic gradient oracle. In the latter two cases we obtain respectively accelerated random coordinate descent and accelerated variance reduction methods with oracle complexity separation.
ISSN:2331-8422