نتایج جستجو برای: deep seq2seq network

تعداد نتایج: 847003  

Journal: :CoRR 2017
Michael Zhu Suyog Gupta

Model pruning seeks to induce sparsity in a deep neural network’s various connection matrices, thereby reducing the number of nonzero-valued parameters in the model. Recent reports (Han et al., 2015a; Narang et al., 2017) prune deep networks at the cost of only a marginal loss in accuracy and achieve a sizable reduction in model size. This hints at the possibility that the baseline models in th...

Journal: :CoRR 2017
Marjan Ghazvininejad Chris Brockett Ming-Wei Chang William B. Dolan Jianfeng Gao Wen-tau Yih Michel Galley

Neural network models are capable of generating extremely natural sounding conversational interactions. Nevertheless, these models have yet to demonstrate that they can incorporate content in the form of factual information or entity-grounded opinion that would enable them to serve in more task-oriented conversational applications. This paper presents a novel, fully data-driven, and knowledge-g...

Journal: :CoRR 2018
Myeongjun Jang Seungwan Seo Pilsung Kang

Sequence-to-sequence (Seq2seq) models have played an import role in the recent success of various natural language processing methods, such as machine translation, text summarization, and speech recognition. However, current Seq2seq models have trouble preserving global latent information from a long sequence of words. Variational autoencoder (VAE) alleviates this problem by learning a continuo...

2017
Minghui Qiu Feng-Lin Li Siyu Wang Xing Gao Yan Chen Weipeng Zhao Haiqing Chen Jun Huang Wei Chu

We propose AliMe Chat, an open-domain chatbot engine that integrates the joint results of Information Retrieval (IR) and Sequence to Sequence (Seq2Seq) based generation models. AliMe Chat uses an attentive Seq2Seq based rerank model to optimize the joint results. Extensive experiments show our engine outperforms both IR and generation based models. We launch AliMe Chat for a real-world industri...

2017
Prajit Ramachandran Peter J. Liu Quoc V. Le

This work presents a general unsupervised learning method to improve the accuracy of sequence to sequence (seq2seq) models. In our method, the weights of the encoder and decoder of a seq2seq model are initialized with the pretrained weights of two language models and then fine-tuned with labeled data. We apply this method to challenging benchmarks in machine translation and abstractive summariz...

Journal: :CoRR 2018
Wei Li Zheng Yang Xu Sun

Traditional Chinese Medicine (TCM) is an influential form of medical treatment in China and surrounding areas. In this paper, we propose a TCM prescription generation task that aims to automatically generate a herbal medicine prescription based on textual symptom descriptions. Sequence-to-sequence (seq2seq) model has been successful in dealing with conditional sequence generation tasks like dia...

Journal: :Soft Computing 2022

Recent advanced deep learning architectures, such as neural seq2seq and transformer, have demonstrated remarkable improvements in multi-typed sentiment classification tasks. Even though recent transformer-based seq2seq-based models successfully enabled to capture rich contextual information of texts, they still lacked attention on incorporating global semantic which enables sufficiently leverag...

Journal: :CoRR 2016
Kun Xiong Anqi Cui Zefeng Zhang Ming Li

Neural conversational models tend to produce generic or safe responses in different contexts, e.g., reply “Of course” to narrative statements or “I don’t know” to questions. In this paper, we propose an end-to-end approach to avoid such problem in neural generative models. Additional memory mechanisms have been introduced to standard sequence-to-sequence (seq2seq) models, so that context can be...

Journal: :CoRR 2017
Hareesh Bahuleyan Lili Mou Olga Vechtomova Pascal Poupart

The variational encoder-decoder (VED) encodes source information as a set of random variables using a neural network, which in turn is decoded into target data using another neural network. In natural language processing, sequenceto-sequence (Seq2Seq) models typically serve as encoder-decoder networks. When combined with a traditional (deterministic) attention mechanism, the variational latent ...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید