Convex programming under a linear optimization oracle The Complexity of Large-scale Convex Programming under a Linear Optimization Oracle

نویسنده

  • Guanghui Lan
چکیده

This paper considers a general class of iterative optimization algorithms, referred to as linearoptimization-based convex programming (LCP) methods, for solving large-scale convex programming (CP) problems. The LCP methods, covering the classic conditional gradient (CndG) method (a.k.a., Frank-Wolfe method) as a special case, can only solve a linear optimization subproblem at each iteration. In this paper, we first establish a series of lower complexity bounds for the LCP methods to solve different classes of CP problems, including smooth, nonsmooth and certain saddle-point problems. We then formally establish the theoretical optimality or nearly optimality, in the largescale case, for the CndG method and its variants to solve different classes of CP problems. We also introduce several new optimal LCP methods, obtained by properly modifying Nesterov’s accelerated gradient method, and demonstrate their possible advantages over the classic CndG for solving certain classes of large-scale CP problems.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

The Complexity of Large-scale Convex Programming under a Linear Optimization Oracle

This paper considers a general class of iterative optimization algorithms, referred to as linear-optimizationbased convex programming (LCP) methods, for solving large-scale convex programming (CP) problems. The LCP methods, covering the classic conditional gradient (CG) method (a.k.a., Frank-Wolfe method) as a special case, can only solve a linear optimization subproblem at each iteration. In t...

متن کامل

A Polynomial-Time Descent Method for Separable Convex Optimization Problems with Linear Constraints

We propose a polynomial algorithm for a separable convex optimization problem with linear constraints. We do not make any additional assumptions about the structure of the objective function except for polynomial computability. That is, the objective function can be non-differentiable. The running time of our algorithm is polynomial in the size of the input consisting of an instance of the prob...

متن کامل

On Convex Optimization, Fat Shattering and Learning

Oracle complexity of the problem under the oracle based optimization model introduced by Nemirovski & Yudin (1978) is considered. We show that the oracle complexity can be lower bounded by fat-shattering dimension introduced by Kearns & Schapire (1990), a key tool in learning theory. Using this result, we proceed to establish upper bounds on learning rates for agnostic PAC learning with linear ...

متن کامل

Integer convex minimization by mixed integer linear optimization

Minimizing a convex function over the integral points of a bounded convex set is polynomial in fixed dimension [6]. We provide an alternative, short, and geometrically motivated proof of this result. In particular, we present an oraclepolynomial algorithm based on a mixed integer linear optimization oracle.

متن کامل

A bi-level linear programming problem for computing the nadir point in MOLP

Computing the exact ideal and nadir criterion values is a very ‎important subject in ‎multi-‎objective linear programming (MOLP) ‎problems‎‎. In fact‎, ‎these values define the ideal and nadir points as lower and ‎upper bounds on the nondominated points‎. ‎Whereas determining the ‎ideal point is an easy work‎, ‎because it is equivalent to optimize a ‎convex function (linear function) over a con...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014