نتایج جستجو برای: sgd

تعداد نتایج: 1169  

2012
J. Michael Cherry Eurie L. Hong Craig Amundsen Rama Balakrishnan Gail Binkley Esther T. Chan Karen R. Christie Maria C. Costanzo Selina S. Dwight Stacia R. Engel Dianna G. Fisk Jodi E. Hirschman Benjamin C. Hitz Kalpana Karra Cynthia J. Krieger Stuart R. Miyasato Robert S. Nash Julie Park Marek S. Skrzypek Matt Simison Shuai Weng Edith D. Wong

The Saccharomyces Genome Database (SGD, http://www.yeastgenome.org) is the community resource for the budding yeast Saccharomyces cerevisiae. The SGD project provides the highest-quality manually curated information from peer-reviewed literature. The experimental results reported in the literature are extracted and integrated within a well-developed database. These data are combined with qualit...

2008
Chun-Nan Hsu Han-Shen Huang Yu-Ming Chang

Previously, Bottou and LeCun [1] established that the second-order stochastic gradient descent (SGD) method can potentially achieve generalization performance as well as empirical optimum in a single pass through the training examples. However, second-order SGD requires computing the inverse of the Hessian matrix of the loss function, which is usually prohibitively expensive. Recently, we inven...

2017
Ashia C. Wilson Rebecca Roelofs Mitchell Stern Nathan Srebro Benjamin Recht

Adaptive optimization methods, which perform local optimization with a metric constructed from the history of iterates, are becoming increasingly popular for training deep neural networks. Examples include AdaGrad, RMSProp, and Adam. We show that for simple overparameterized problems, adaptive methods often find drastically different solutions than gradient descent (GD) or stochastic gradient d...

2017
Amit Daniely

We show that the standard stochastic gradient decent (SGD) algorithm is guaranteed to learn, in polynomial time, a function that is competitive with the best function in the conjugate kernel space of the network, as defined in Daniely et al. [13]. The result holds for log-depth networks from a rich family of architectures. To the best of our knowledge, it is the first polynomial-time guarantee ...

2015
Stephan Mandt Matthew D. Hoffman David M. Blei

Stochastic Gradient Descent (SGD) is an important algorithm in machine learning. With constant learning rates, it is a stochastic process that reaches a stationary distribution. We revisit an analysis of SGD in terms of stochastic differential equations in the limit of small constant gradient steps. This limit, which we feel is not appreciated in the machine learning community, allows us to app...

Journal: :CoRR 2016
Valentin Dalibard Michael Schaarschmidt Eiko Yoneki

We present an optimizer which uses Bayesian optimization to tune the system parameters of distributed stochastic gradient descent (SGD). Given a specific context, our goal is to quickly find efficient configurations which appropriately balance the load between the available machines to minimize the average SGD iteration time. Our experiments consider setups with over thirty parameters. Traditio...

2014
Fanglin Li Bin Wu Liutong Xu Chuan Shi Jing Shi

The accuracy and effectiveness of matrix factorization technique were well demonstrated in the Netflix movie recommendation contest. Among the numerous solutions for matrix factorization, Stochastic Gradient Descent (SGD) is one of the most widely used algorithms. However, as a sequential approach, SGD algorithm cannot directly be used in the Distributed Cluster Environment (DCE). In this paper...

2012
Caitlin Young

Submarine groundwater discharge (SGD) nutrient fluxes from un-sewered suburban landscapes pose a significant eutrophication risk to coastal waters. In unconsolidated coastal aquifers, nutrients recharged to groundwater undergo diffuse discharge to the coastline during SGD. Defining pathways of nutrient attenuation is required for land management professionals to take effective nutrient remediat...

2011
Benjamin Recht Christopher Ré Stephen J. Wright Feng Niu

Stochastic Gradient Descent (SGD) is a popular algorithm that can achieve stateof-the-art performance on a variety of machine learning tasks. Several researchers have recently proposed schemes to parallelize SGD, but all require performancedestroying memory locking and synchronization. This work aims to show using novel theoretical analysis, algorithms, and implementation that SGD can be implem...

2016
Andrew Cotter Maya R. Gupta Jan Pfeifer

Projected stochastic gradient descent (SGD) is often the default choice for large-scale optimization in machine learning, but requires a projection after each update. For heavily-constrained objectives, we propose an efficient extension of SGD that stays close to the feasible region while only applying constraints probabilistically at each iteration. Theoretical analysis shows a good trade-off ...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید