نتایج جستجو برای: stochastic gradient descent

تعداد نتایج: 258150  

Journal: :Journal of Machine Learning Research 2009
Antoine Bordes Léon Bottou Patrick Gallinari

The SGD-QN algorithm is a stochastic gradient descent algorithm that makes careful use of secondorder information and splits the parameter update into independently scheduled components. Thanks to this design, SGD-QN iterates nearly as fast as a first-order stochastic gradient descent but requires less iterations to achieve the same accuracy. This algorithm won the “Wild Track” of the first PAS...

2016
Hamed Karimi Julie Nutini Mark W. Schmidt

I Simple proof of linear convergence. I For convex functions, equivalent to several of the above conditions. I For non-convex functions, weakest assumption while still guaranteeing global minimizer. ? We generalize the PL condition to analyze proximal-gradient methods. ? We give simple new analyses in a variety of settings: I Least-squares and logistic regression. I Randomized coordinate descen...

Journal: :CoRR 2016
Mark Tygert

Poor (even random) starting points for learning/training/optimization are common in machine learning. In many settings, the method of Robbins and Monro (online stochastic gradient descent) is known to be optimal for good starting points, but may not be optimal for poor starting points — indeed, for poor starting points Nesterov acceleration can help during the initial iterations, even though Ne...

Journal: :CoRR 2017
Fan Zhou Guojing Cong

Despite their popularity, the practical performance of asynchronous stochastic gradient descent methods (ASGD) for solving large scale machine learning problems are not as good as theoretical results indicate. We adopt and analyze a synchronous K-step averaging stochastic gradient descent algorithm which we call K-AVG. We establish the convergence results of KAVG for nonconvex objectives, and s...

1997
C. RICHARD JOHNSON JAMES D. BEHM DONALD R. BROWN RAÚL A. CASAS

This paper provides a tutorial introduction to the constant modulus (CM) criterion for blind fractionally spaced equalizer (FSE) design via a (stochastic) gradient descent algorithm such as the constant modulus algorithm (CMA). The topical divisions utilized in this tutorial can be used to help catalog the emerging literature on the CM criterion and on the behavior of (stochastic) gradient desc...

2015
William Fedus Bobak Hashemi Matthew Burns

We train two convolutional neural networks on the POFA and NimStim datasets to identify individuals and identify emotions, respectively. In order to train these neural networks, we use two separate optimization procedures, the minFunc package and stochastic gradient descent. The minFunc optimization package achieved a 95.8% on the training set and achieved a 90.0% accuracy on the partitioned te...

Journal: :CoRR 2015
Qinxun Bai Henry Lam Stan Sclaroff

We propose a Bayesian approach for recursively estimating the classifier weights in online learning of a classifier ensemble. In contrast with past methods, such as stochastic gradient descent or online boosting, our approach estimates the weights by recursively updating its posterior distribution. For a specified class of loss functions, we show that it is possible to formulate a suitably defi...

Journal: :bulletin of the iranian mathematical society 2014
saman babaie-kafaki

‎based on an eigenvalue analysis‎, ‎a new proof for the sufficient‎ ‎descent property of the modified polak-ribière-polyak conjugate‎ ‎gradient method proposed by yu et al‎. ‎is presented‎.

Journal: :CoRR 2014
Huahua Wang Arindam Banerjee

Two types of low cost-per-iteration gradient descent methods have been extensively studied in parallel. One is online or stochastic gradient descent ( OGD/SGD), and the other is randomzied coordinate descent (RBCD). In this paper, we combine the two types of methods together and propose online randomized block coordinate descent (ORBCD). At each iteration, ORBCD only computes the partial gradie...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید