Adaptive Online Gradient Descent
نویسندگان
چکیده
We study the rates of growth of the regret in online convex optimization. First, we show that a simple extension of the algorithm of Hazan et al eliminates the need for a priori knowledge of the lower bound on the second derivatives of the observed functions. We then provide an algorithm, Adaptive Online Gradient Descent, which interpolates between the results of Zinkevich for linear functions and of Hazan et al for strongly convex functions, achieving intermediate rates between √ T and log T . Furthermore, we show strong optimality of the algorithm. Finally, we provide an extension of our results to general norms.
منابع مشابه
Designing stable neural identifier based on Lyapunov method
The stability of learning rate in neural network identifiers and controllers is one of the challenging issues which attracts great interest from researchers of neural networks. This paper suggests adaptive gradient descent algorithm with stable learning laws for modified dynamic neural network (MDNN) and studies the stability of this algorithm. Also, stable learning algorithm for parameters of ...
متن کاملAdaptive Back-Propagation in On-Line Learning of Multilayer Networks
An adaptive back-propagation algorithm is studied and compared with gradient descent (standard back-propagation) for on-line learning in two-layer neural networks with an arbitrary number of hidden units. Within a statistical mechanics framework , both numerical studies and a rigorous analysis show that the adaptive back-propagation method results in faster training by breaking the symmetry bet...
متن کاملLess Regret via Online Conditioning
We analyze and evaluate an online gradient descent algorithm with adaptive per-coordinate adjustment of learning rates. Our algorithm can be thought of as an online version of batch gradient descent with a diagonal preconditioner. This approach leads to regret bounds that are stronger than those of standard online gradient descent for general online convex optimization problems. Experimentally,...
متن کاملOn-line Adaptive Learning Rate Bp Algorithm for Mlp and Application to an Identification Problem
An on-line algorithm that uses an adaptive learning rate is proposed. Its development is based on the analysis of the convergence of the conventional gradient descent method for threelayer BP neural networks. The effectiveness of the proposed algorithm applied to the identification and prediction of behavior of non-linear dynamic systems is demonstrated by simulation experiments.
متن کاملAdaptive fuzzy sliding mode and indirect radial-basis-function neural network controller for trajectory tracking control of a car-like robot
The ever-growing use of various vehicles for transportation, on the one hand, and the statistics ofsoaring road accidents resulting from human error, on the other hand, reminds us of the necessity toconduct more extensive research on the design, manufacturing and control of driver-less intelligentvehicles. For the automatic control of an autonomous vehicle, we need its dynamic...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2007