Matrix product operators for sequence-to-sequence learning

نویسندگان
چکیده

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A new sequence space and norm of certain matrix operators on this space

In the present paper, we introduce the sequence space [{l_p}(E,Delta) = left{ x = (x_n)_{n = 1}^infty : sum_{n = 1}^infty left|  sum_{j in {E_n}} x_j - sum_{j in E_{n + 1}} x_jright| ^p < infty right},] where $E=(E_n)$ is a partition of finite subsets of the positive integers and $pge 1$. We investigate its topological properties and inclusion relations. Moreover, we consider the problem of fin...

متن کامل

Some inequalities involving lower bounds of operators on weighted sequence spaces by a matrix norm

Let A = (an;k)n;k1 and B = (bn;k)n;k1 be two non-negative ma-trices. Denote by Lv;p;q;B(A), the supremum of those L, satisfying the followinginequality:k Ax kv;B(q) L k x kv;B(p);where x 0 and x 2 lp(v;B) and also v = (vn)1n=1 is an increasing, non-negativesequence of real numbers. In this paper, we obtain a Hardy-type formula forLv;p;q;B(H), where H is the Hausdor matrix and 0 < q p 1. Also...

متن کامل

Sequence to Sequence Learning for Event Prediction

This paper presents an approach to the task of predicting an event description from a preceding sentence in a text. Our approach explores sequence-to-sequence learning using a bidirectional multi-layer recurrent neural network. Our approach substantially outperforms previous work in terms of the BLEU score on two datasets derived from WIKIHOW and DESCRIPT respectively. Since the BLEU score is n...

متن کامل

Unsupervised Pretraining for Sequence to Sequence Learning

This work presents a general unsupervised learning method to improve the accuracy of sequence to sequence (seq2seq) models. In our method, the weights of the encoder and decoder of a seq2seq model are initialized with the pretrained weights of two language models and then fine-tuned with labeled data. We apply this method to challenging benchmarks in machine translation and abstractive summariz...

متن کامل

Convolutional Sequence to Sequence Learning

A. Weight Initialization We derive a weight initialization scheme tailored to the GLU activation function similar to Glorot & Bengio (2010); He et al. (2015b) by focusing on the variance of activations within the network for both forward and backward passes. We also detail how we modify the weight initialization for dropout. A.1. Forward Pass Assuming that the inputs x l of a convolutional laye...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Physical Review E

سال: 2018

ISSN: 2470-0045,2470-0053

DOI: 10.1103/physreve.98.042114