RSG: Beating Subgradient Method without Smoothness and Strong Convexity

نویسندگان

  • Tianbao Yang
  • Qihang Lin
چکیده

In this paper, we study the efficiency of a Restarted SubGradient (RSG) method that periodically restarts the standard subgradient method (SG). We show that, when applied to a broad class of convex optimization problems, RSG method can find an ǫ-optimal solution with a low complexity than SG method. In particular, we first show that RSG can reduce the dependence of SG’s iteration complexity on the distance between the initial solution and the optimal set to that between the ǫ-level set and the optimal set. In addition, we show the advantages of RSG over SG in solving three different families of convex optimization problems. (a) For the problems whose epigraph is a polyhedron, RSG is shown to converge linearly. (b) For the problems with local quadratic growth property, RSG has an O( ǫ log( ǫ )) iteration complexity. (c) For the problems that admit a local KurdykaLojasiewicz property with a power constant of β ∈ [0, 1), RSG has an O( 1 ǫ 2β log( 1 ǫ )) iteration complexity. On the contrary, with only the standard analysis, the iteration complexity of SG is known to be O( 1 ǫ 2 ) for these three classes of problems. The novelty of our analysis lies at exploiting the lower bound of the first-order optimality residual at the ǫ-level set. It is this novelty that allows us to explore the local properties of functions (e.g., local quadratic growth property, local KurdykaLojasiewicz property, more generally local error bounds) to develop the improved convergence of RSG. We demonstrate the effectiveness of the proposed algorithms on several machine learning tasks including regression and classification.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Convergence Rates for Deterministic and Stochastic Subgradient Methods Without Lipschitz Continuity

We extend the classic convergence rate theory for subgradient methods to apply to non-Lipschitz functions. For the deterministic projected subgradient method, we present a global O(1/ √ T ) convergence rate for any convex function which is locally Lipschitz around its minimizers. This approach is based on Shor’s classic subgradient analysis and implies generalizations of the standard convergenc...

متن کامل

Approximate Stochastic Subgradient Estimation Training for Support Vector Machines

Subgradient algorithms for training support vector machines have been quite successful for solving largescale and online learning problems. However, they have been restricted to linear kernels and strongly convex formulations. This paper describes efficient subgradient approaches without such limitations. Our approaches make use of randomized low-dimensional approximations to nonlinear kernels,...

متن کامل

ASSET: Approximate Stochastic Subgradient Estimation Training for Support Vector Machines

Subgradient methods for SVMs have been successful in solving the primal formulation with linear kernels. The approach is extended here to nonlinear kernels, and the assumption of strong convexity of the objective is dropped, allowing an intercept term to be used in the classifier.

متن کامل

OSGA: a fast subgradient algorithm with optimal complexity

This paper presents an algorithm for approximately minimizing a convex function in simple, not necessarily bounded convex, finite-dimensional domains, assuming only that function values and subgradients are available. No global information about the objective function is needed apart from a strong convexity parameter (which can be put to zero if only convexity is known). The worst case number o...

متن کامل

Identifying Active Manifolds

Determining the “active manifold” for a minimization problem is a large step towards solving the problem. Many researchers have studied under what conditions certain algorithms identify active manifolds in a finite number of iterations. In this work we outline a unifying framework encompassing many earlier results on identification via the Subgradient (Gradient) Projection Method, Newton-like M...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016