نتایج جستجو برای: lstm

تعداد نتایج: 6907  

2017
Limin Wang Qian Yan Shoushan Li Guodong Zhou

In the last decades, named entity recognition has been extensively studied with various supervised learning approaches depend on massive labeled data. In this paper, we focus on person name recognition in judgment documents. Owing to the lack of human-annotated data, we propose a joint learning approach, namely Aux-LSTM, to use a large scale of auto-annotated data to help human-annotated data (...

2013
Jürgen T. Geiger Florian Eyben Björn W. Schuller Gerhard Rigoll

Detecting segments of overlapping speech (when two or more speakers are active at the same time) is a challenging problem. Previously, mostly HMM-based systems have been used for overlap detection, employing various different audio features. In this work, we propose a novel overlap detection system using Long Short-Term Memory (LSTM) recurrent neural networks. LSTMs are used to generate framewi...

Journal: :CoRR 2018
Chao Zhang Philip C. Woodland

Vanishing long-term gradients are a major issue in training standard recurrent neural networks (RNNs), which can be alleviated by long short-term memory (LSTM) models with memory cells. However, the extra parameters associated with the memory cells mean an LSTM layer has four times as many parameters as an RNN with the same hidden vector size. This paper addresses the vanishing gradient problem...

2016
Chuanhai Dong Jiajun Zhang Chengqing Zong Masanori Hattori Hui Di

State-of-the-art systems of Chinese Named Entity Recognition (CNER) require large amounts of hand-crafted features and domainspecific knowledge to achieve high performance. In this paper, we apply a bidirectional LSTM-CRF neural network that utilizes both characterlevel and radical-level representations. We are the first to use characterbased BLSTM-CRF neural architecture for CNER. By contrasti...

2015
Yan Xu Lili Mou Ge Li Yunchuan Chen Hao Peng Zhi Jin

Relation classification is an important research arena in the field of natural language processing (NLP). In this paper, we present SDP-LSTM, a novel neural network to classify the relation of two entities in a sentence. Our neural architecture leverages the shortest dependency path (SDP) between two entities; multichannel recurrent neural networks, with long short term memory (LSTM) units, pic...

Journal: :CoRR 2018
Zixiang Ding Rui Xia Jianfei Yu Xiang Li Jian Yang

Deep neural networks have recently been shown to achieve highly competitive performance in many computer vision tasks due to their abilities of exploring in a much larger hypothesis space. However, since most deep architectures like stacked RNNs tend to suffer from the vanishing-gradient and overfitting problems, their effects are still understudied in many NLP tasks. Inspired by this, we propo...

2016
Connor Schenck Dieter Fox

Recent advances in AI and robotics have claimed many incredible results with deep learning, yet no work to date has applied deep learning to the problem of liquid perception and reasoning. In this paper, we apply fully-convolutional deep neural networks to the tasks of detecting and tracking liquids. We evaluate three models: a single-frame network, multi-frame network, and a LSTM recurrent net...

2016
Ruben Zazo Alicia Lozano-Diez Joaquin Gonzalez-Rodriguez

Long Short-Term Memory recurrent neural networks (LSTM RNNs) provide an outstanding performance in language identification (LID) due to its ability to model speech sequences. So far, previously published LSTM RNNs solutions for LID deal with highly controlled scenarios, balanced datasets and limited channel variability. In this paper we evaluate an endto-end LSTM LID system, comparing it agains...

2016
Huiwei Zhou Junli Xu Yunlong Yang Huijie Deng Long Chen Degen Huang

Hedge detection aims to distinguish factual and uncertain information, which is important in information extraction. The task of hedge detection contains two subtasks: identifying hedge cues and detecting their linguistic scopes. Hedge scope detection is dependent on syntactic and semantic information. Previous researches usually use lexical and syntactic information and ignore deep semantic in...

Journal: :CoRR 2017
Abdelhadi Azzouni Guy Pujolle

This paper presents NeuTM, a framework for network Traffic Matrix (TM) prediction based on Long Short-Term Memory Recurrent Neural Networks (LSTM RNNs). TM prediction is defined as the problem of estimating future network traffic matrix from the previous and achieved network traffic data. It is widely used in network planning, resource management and network security. Long Short-Term Memory (LS...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید