نتایج جستجو برای: deep seq2seq network

تعداد نتایج: 847003  

2012
Ebru Arisoy Tara N. Sainath Brian Kingsbury Bhuvana Ramabhadran

In recent years, neural network language models (NNLMs) have shown success in both peplexity and word error rate (WER) compared to conventional n-gram language models. Most NNLMs are trained with one hidden layer. Deep neural networks (DNNs) with more hidden layers have been shown to capture higher-level discriminative information about input features, and thus produce better networks. Motivate...

2016
Somshubra Majumdar Ishaan Jain Anand Madhavan Li Wan Matthew Zeiler Sixin Zhang Yann LeCun Jie Huang

Recent developments in the field of deep learning have shown that convolutional networks with several layers can approach human level accuracy in tasks such as handwritten digit classification and object recognition. It is observed that the state-of-the-art performance is obtained from model ensembles, where several models are trained on the same data and their predictions probabilities are ave...

Journal: :CoRR 2016
Zhen Hu Zhuyin Xue Tong Cui Shiqiang Zong ChengLong He

Pretraining is widely used in deep neutral network and one of the most famous pretraining models is Deep Belief Network (DBN). The optimization formulas are different during the pretraining process for different pretraining models. In this paper, we pretrained deep neutral network by different pretraining models and hence investigated the difference between DBN and Stacked Denoising Autoencoder...

2015
Tejas D. Kulkarni William F. Whitney Pushmeet Kohli Joshua B. Tenenbaum

This paper presents the Deep Convolution Inverse Graphics Network (DC-IGN) that aims to learn an interpretable representation of images that is disentangled with respect to various transformations such as object out-of-plane rotations, lighting variations, and texture. The DC-IGN model is composed of multiple layers of convolution and de-convolution operators and is trained using the Stochastic...

Journal: :CoRR 2015
Guoqiang Zhong Pan Yang Sijiang Wang Junyu Dong

Hashing-based methods seek compact and efficient binary codes that preserve the neighborhood structure in the original data space. For most existing hashing methods, an image is first encoded as a vector of hand-crafted visual feature, followed by a hash projection and quantization step to get the compact binary vector. Most of the hand-crafted features just encode the low-level information of ...

Journal: :CoRR 2017
Jaya Thomas Sonia Thomas Lee Sael

Background: Should we input known genome sequence features or input sequence itself in deep learning framework? As deep learning more popular in various applications, researchers often come to question whether to generate features or use raw sequences for deep learning. To answer this question, we study the prediction accuracy of precursor miRNA prediction of feature-based deep belief network a...

2016
Tiehang Duan Sargur N. Srihari

A computationally efficient method to improve classification performance of a Deep Belief Network(DBN) is introduced. In the Pseudo Boost Deep Belief Network(PB-DBN), top layers are boosted while lower layers of the base classifiers share weights for feature extraction. PB-DBN maintains the same time complexity as a DBN with fast convergence to optimality by introducing the mechanism of pseudo ...

2017
Yue Wu Wael AbdAlmageed Prem Natarajan

Image splicing is a very common image manipulation technique that is sometimes used for malicious purposes. A splicing detection and localization algorithm usually takes an input image and produces a binary decision indicating whether the input image has been manipulated, and also a segmentation mask that corresponds to the spliced region. Most existing splicing detection and localization pipel...

2016
Yuan Gao Dorota Glowacka

This paper explores the possibility of using multiplicative gate to build two recurrent neural network structures. These two structures are called Deep Simple Gated Unit (DSGU) and Simple Gated Unit (SGU), which are structures for learning long-term dependencies. Compared to traditional Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU), both structures require fewer parameters and le...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید