نتایج جستجو برای: sgd

تعداد نتایج: 1169  

2017
Yuanzhi Li Yang Yuan

In recent years, stochastic gradient descent (SGD) based techniques has become the standard tools for training neural networks. However, formal theoretical understanding of why SGD can train neural networks in practice is largely missing. In this paper, we make progress on understanding this mystery by providing a convergence analysis for SGD on a rich subset of two-layer feedforward networks w...

Journal: :Proceedings of the National Academy of Sciences of the United States of America 2015
Valentí Rodellas Jordi Garcia-Orellana Pere Masqué Mor Feldman Yishai Weinstein

The Mediterranean Sea (MS) is a semienclosed basin that is considered one of the most oligotrophic seas in the world. In such an environment, inputs of allochthonous nutrients and micronutrients play an important role in sustaining primary productivity. Atmospheric deposition and riverine runoff have been traditionally considered the main external sources of nutrients to the MS, whereas the rol...

2016
Tanmay Gupta Aditya Deshpande

In this paper, we mainly study the convergence properties of stochastic gradient descent (SGD) as described in Needell et al. [2]. The function to be minimized with SGD is assumed to be strongly convex. Also, its gradients are assumed to be Lipschitz continuous. First, we discuss the superior bound on convergence (of standard SGD) obtained by Needell et al. [2] as opposed to the previous work o...

2018
Pavel Izmailov Dmitrii Podoprikhin Timur Garipov Dmitry Vetrov Andrew Gordon Wilson

Deep neural networks are typically trained by optimizing a loss function with an SGD variant, in conjunction with a decaying learning rate, until convergence. We show that simple averaging of multiple points along the trajectory of SGD, with a cyclical or constant learning rate, leads to better generalization than conventional training. We also show that this Stochastic Weight Averaging (SWA) p...

Journal: :Nucleic acids research 2002
Selina S. Dwight Midori A. Harris Kara Dolinski Catherine A. Ball Gail Binkley Karen R. Christie Dianna G. Fisk Laurie Issel-Tarver Mark Schroeder Gavin Sherlock Anand Sethuraman Shuai Weng David Botstein J. Michael Cherry

The Saccharomyces Genome Database (SGD) resources, ranging from genetic and physical maps to genome-wide analysis tools, reflect the scientific progress in identifying genes and their functions over the last decade. As emphasis shifts from identification of the genes to identification of the role of their gene products in the cell, SGD seeks to provide its users with annotations that will allow...

2016
Shen-Yi Zhao Wu-Jun Li

Stochastic gradient descent (SGD) and its variants have become more and more popular in machine learning due to their efficiency and effectiveness. To handle large-scale problems, researchers have recently proposed several parallel SGD methods for multicore systems. However, existing parallel SGD methods cannot achieve satisfactory performance in real applications. In this paper, we propose a f...

Journal: :Palliative medicine 2014
W-S Kelvin Teo Anusha Govinda Raj Woan Shin Tan Charis Wei Ling Ng Bee Hoon Heng Ian Yi-Onn Leong

BACKGROUND Due to limited end-of-life discussions and the absence of palliative care, hospitalisations are frequent at the end of life among nursing home residents in Singapore, resulting in high health-care costs. AIM Our objective was to evaluate the economic impact of Project Care at the End-of-Life for Residents in homes for the Elderly (CARE) programme on nursing home residents compared ...

Journal: : 2021

? ?????? ??????????????? ?????? ??????????? ??????????????? ???????? ???????? ???????. ?????? ?????? ???? ??????????? ? ???????? ????????, ? ????? ????? ????????? ???? ?????? ??????????. ?? ???????? ??? ?? ??????? ?????? ???????????? ????????? ???? ??????????????? ???????????? ?????? (SGD). ????? ?????? ???????????? ?????? ????? ?????? ? ?????????????? ?????? ??????????? ????-?????????. ???????...

Journal: :CoRR 2017
Xingwen Zhang Jeff Clune Kenneth O. Stanley

Because stochastic gradient descent (SGD) has shown promise optimizing neural networks with millions of parameters and few if any alternatives are known to exist, it has moved to the heart of leading approaches to reinforcement learning (RL). For that reason, the recent result from OpenAI showing that a particular kind of evolution strategy (ES) can rival the performance of SGD-based deep RL me...

Journal: :CoRR 2016
Prateek Jain Sham M. Kakade Rahul Kidambi Praneeth Netrapalli Aaron Sidford

This work characterizes the benefits of averaging techniques widely used in conjunction with stochastic gradient descent (SGD). In particular, this work sharply analyzes: (1) mini-batching, a method of averaging many samples of the gradient to both reduce the variance of a stochastic gradient estimate and for parallelizing SGD and (2) tail-averaging, a method involving averaging the final few i...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید