Global rates of convergence for nonconvex optimization on manifolds
نویسندگان
چکیده
منابع مشابه
Global Convergence of Splitting Methods for Nonconvex Composite Optimization
We consider the problem of minimizing the sum of a smooth function h with a bounded Hessian, and a nonsmooth function. We assume that the latter function is a composition of a proper closed function P and a surjective linear map M, with the proximal mappings of τP , τ > 0, simple to compute. This problem is nonconvex in general and encompasses many important applications in engineering and mach...
متن کاملGlobal Convergence of Langevin Dynamics Based Algorithms for Nonconvex Optimization
We present a unified framework to analyze the global convergence of Langevin dynamics based algorithms for nonconvex finite-sum optimization with n component functions. At the core of our analysis is a direct analysis of the ergodicity of the numerical approximations to Langevin dynamics, which leads to faster convergence rates. Specifically, we show that gradient Langevin dynamics (GLD) and st...
متن کاملGlobal Convergence of ADMM in Nonconvex Nonsmooth Optimization
In this paper, we analyze the convergence of the alternating direction method of multipliers (ADMM) for minimizing a nonconvex and possibly nonsmooth objective function, φ(x0, . . . , xp, y), subject to coupled linear equality constraints. Our ADMM updates each of the primal variables x0, . . . , xp, y, followed by updating the dual variable. We separate the variable y from xi’s as it has a spe...
متن کاملConvergence Rates of Efficient Global Optimization Algorithms
In the efficient global optimization problem, we minimize an unknown function f , using as few observations f (x) as possible. It can be considered a continuum-armed-bandit problem, with noiseless data, and simple regret. Expected-improvement algorithms are perhaps the most popular methods for solving the problem; in this paper, we provide theoretical results on their asymptotic behaviour. Impl...
متن کاملOn the Global Convergence of the PERRY-SHANNO Method for Nonconvex Unconstrained Optimization Problems
In this paper, we prove the global convergence of the Perry-Shanno’s memoryless quasi-Newton (PSMQN) method with a new inexact line search when applied to nonconvex unconstrained minimization problems. Preliminary numerical results show that the PSMQN with the particularly line search conditions are very promising.
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IMA Journal of Numerical Analysis
سال: 2018
ISSN: 0272-4979,1464-3642
DOI: 10.1093/imanum/drx080