نتایج جستجو برای: sgd

تعداد نتایج: 1169  

Journal: :Limnology and oceanography letters 2022

Submarine groundwater discharge (SGD) can have a profound influence on marine environments and elemental biogeochemical cycles in coastal waters. We explored the feasibility of using Ba/Ca ratios benthic foraminiferal shells as proxy SGD. Dissolved barium (DBa) displayed enrichment behavior bottom waters indicating that SGD is major DBa source Changjiang (Yangtze) Estuary. Foraminifera lived wi...

Journal: :CoRR 2013
Tianbao Yang Lijun Zhang

We motivate this study from a recent work on a stochastic gradient descent (SGD) method with only one projection (Mahdavi et al., 2012), which aims at alleviating the computational bottleneck of the standard SGD method in performing the projection at each iteration, and enjoys an O(log T/T ) convergence rate for strongly convex optimization. In this paper, we make further contributions along th...

Journal: :CoRR 2018
Huishuai Zhang Wei Chen Tie-Yan Liu

Stochastic gradient descent (SGD) has achieved great success in training deep neural network, where the gradient is computed through backpropagation. However, the back-propagated values of different layers vary dramatically. This inconsistence of gradient magnitude across different layers renders optimization of deep neural network with a single learning rate problematic. We introduce the back-...

Journal: :CoRR 2017
Zijun Zhang Lin Ma Zongpeng Li Chuan Wu

Optimization algorithms for training deep models not only affects the convergence rate and stability of the training process, but are also highly related to the generalization performance of the models. While adaptive algorithms, such as Adam and RMSprop, have shown better optimization performance than stochastic gradient descent (SGD) in many scenarios, they often lead to worse generalization ...

2016
Chang Xu Tao Qin Gang Wang Tie-Yan Liu

Stochastic gradient descent (SGD), which updates the model parameters by adding a local gradient times a learning rate at each step, is widely used in model training of machine learning algorithms such as neural networks. It is observed that the models trained by SGD are sensitive to learning rates and good learning rates are problem specific. To avoid manually searching of learning rates, whic...

Journal: :Augmentative and alternative communication 2013
Manon Robillard Chantal Mayer-Crittenden Annie Roy-Charland Michèle Minor-Corriveau Roxanne Bélanger

This study examined the impact of cognition on young children's ability to navigate a speech-generating device (SGD) with dynamic paging. Knowledge of which cognitive factors impact navigational skills could help clinicians select the most appropriate SGD for children who have complex communication needs. A total of 65 typically developing children aged 48-77 months were assessed using the Leit...

2018
Sanghamitra Dutta Gauri Joshi Soumyadip Ghosh Parijat Dube Priya Nagpurkar

Distributed Stochastic Gradient Descent (SGD) when run in a synchronous manner, suffers from delays in waiting for the slowest learners (stragglers). Asynchronous methods can alleviate stragglers, but cause gradient staleness that can adversely affect convergence. In this work we present the first theoretical characterization of the speed-up offered by asynchronous methods by analyzing the trad...

Journal: :CoRR 2017
Chang Xu Tao Qin Gang Wang Tie-Yan Liu

Stochastic gradient descent (SGD), which updates the model parameters by adding a local gradient times a learning rate at each step, is widely used in model training of machine learning algorithms such as neural networks. It is observed that the models trained by SGD are sensitive to learning rates and good learning rates are problem specific. We propose an algorithm to automatically learn lear...

Journal: :Advances in neural information processing systems 2015
Christopher De Sa Ce Zhang Kunle Olukotun Christopher Ré

Stochastic gradient descent (SGD) is a ubiquitous algorithm for a variety of machine learning problems. Researchers and industry have developed several techniques to optimize SGD's runtime performance, including asynchronous execution and reduced precision. Our main result is a martingale-based analysis that enables us to capture the rich noise models that may arise from such techniques. Specif...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید