نتایج جستجو برای: stochastic gradient descent learning

تعداد نتایج: 840759  

2008
Léon Bottou Olivier Bousquet

This contribution develops a theoretical framework that takes into account the effect of approximate optimization on learning algorithms. The analysis shows distinct tradeoffs for the case of small-scale and large-scale learning problems. Small-scale learning problems are subject to the usual approximation– estimation tradeoff. Large-scale learning problems are subject to a qualitatively differ...

2016
Tim Roughgarden Gregory Valiant

Last lecture we covered the basics of gradient descent, with an emphasis on the intuition behind and geometry underlying the method, plus a concrete instantiation of it for the problem of linear regression (fitting the best hyperplane to a set of data points). This basic method is already interesting and useful in its own right (see Homework #3). This lecture we’ll cover two extensions that, wh...

2016
Joseph Sakaya Arto Klami

Automatic variational inference has recently become feasible as a scalable inference tool for probabilistic programming. The state-of-the-art algorithms are stochastic in two respects: they use stochastic gradient descent to optimize an expectation that is estimated with stochastic approximation. The core computation of such algorithms involves evaluating the loss and its automatically differen...

Journal: :Human movement science 2009
W I Schöllhorn G Mayer-Kress K M Newell M Michelbrink

In this paper, the major assumptions of influential approaches to the structure of variability in practice conditions are discussed from the perspective of a generalized evolving attractor landscape model of motor learning. The efficacy of the practice condition effects is considered in relation to the theoretical influence of stochastic perturbations in models of gradient descent learning of m...

2018
Hanlin Tang Xiangru Lian Ming Yan Ce Zhang Ji Liu

While training a machine learning model using multiple workers, each of which collects data from their own data sources, it would be most useful when the data collected from different workers can be unique and different. Ironically, recent analysis of decentralized parallel stochastic gradient descent (D-PSGD) relies on the assumption that the data hosted on different workers are not too differ...

2012
Shuang Wu Jun Sakuma

The traditional paradigm in machine learning has been that given a data set, the goal is to learn a target function or decision model (such as a classifier) from it. Many techniques in data mining and machine learning follow a gradient descent paradigm in the iterative process of discovering this target function or decision model. For instance, Linear regression can be resolved through a gradie...

Journal: :CoRR 2015
Guillaume Bouchard Théo Trouillon Julien Perez Adrien Gaidon

Stochastic Gradient Descent (SGD) is one of the most widely used techniques for online optimization in machine learning. In this work, we accelerate SGD by adaptively learning how to sample the most useful training examples at each time step. First, we show that SGD can be used to learn the best possible sampling distribution of an importance sampling estimator. Second, we show that the samplin...

2016
Chang Xu Tao Qin Gang Wang Tie-Yan Liu

Stochastic gradient descent (SGD), which updates the model parameters by adding a local gradient times a learning rate at each step, is widely used in model training of machine learning algorithms such as neural networks. It is observed that the models trained by SGD are sensitive to learning rates and good learning rates are problem specific. To avoid manually searching of learning rates, whic...

Journal: :Automatica 2009
Shalabh Bhatnagar Richard S. Sutton Mohammad Ghavamzadeh Mark Lee

We present four new reinforcement learning algorithms based on actor–critic, natural-gradient and function-approximation ideas, and we provide their convergence proofs. Actor–critic reinforcement learning methods are online approximations to policy iteration in which the value-function parameters are estimated using temporal difference learning and the policy parameters are updated by stochasti...

2007
Shalabh Bhatnagar Richard S. Sutton Mohammad Ghavamzadeh Mark Lee

We present four new reinforcement learning algorithms based on actor-critic and natural-gradient ideas, and provide their convergence proofs. Actor-critic reinforcement learning methods are online approximations to policy iteration in which the value-function parameters are estimated using temporal difference learning and the policy parameters are updated by stochastic gradient descent. Methods...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید