Multi-Task Learning for Improved Discriminative Training in SMT
نویسندگان
چکیده
Multi-task learning has been shown to be effective in various applications, including discriminative SMT. We present an experimental evaluation of the question whether multi-task learning depends on a “natural” division of data into tasks that balance shared and individual knowledge, or whether its inherent regularization makes multi-task learning a broadly applicable remedy against overfitting. To investigate this question, we compare “natural” tasks defined as sections of the International Patent Classification versus “random” tasks defined as random shards in the context of patent SMT. We find that both versions of multi-task learning improve equally well over independent and pooled baselines, and gain nearly 2 BLEU points over standard MERT tuning.
منابع مشابه
Multi-Task Minimum Error Rate Training for SMT
We present experiments on multi-task learning for discriminative training in statistical machine translation (SMT), extending standardminimum-error-rate training (MERT) by techniques that take advantage of the similarity of related tasks. We apply our techniques to German-toEnglish translation of patents from 8 tasks according to the International Patent Classification (IPC) system. Our experim...
متن کاملThe HDU Discriminative SMT System for Constrained Data PatentMT at NTCIR10
We describe the statistical machine translation (SMT) systems developed at Heidelberg University for the Chinese-toEnglish and Japanese-to-English PatentMT subtasks at the NTCIR10 workshop. The core system used in both subtasks is a combination of hierarchical phrase-based translation and discriminative training using either large feature sets and `1/`2 regularization (for Japanese-to-English) ...
متن کاملLatent Structure Discriminative Learning for Natural Language Processing
Natural language is rich with layers of implicit structure, and previous research has shown that we can take advantage of this structure to make more accurate models. Most attempts to utilize forms of implicit natural language structure for natural language processing tasks have assumed a pre-defined structural analysis before training the task-specific model. However, rather than fixing the la...
متن کاملExploiting Linguistic Resources for Neural Machine Translation Using Multi-task Learning
Linguistic resources such as part-ofspeech (POS) tags have been extensively used in statistical machine translation (SMT) frameworks and have yielded better performances. However, usage of such linguistic annotations in neural machine translation (NMT) systems has been left under-explored. In this work, we show that multi-task learning is a successful and a easy approach to introduce an additio...
متن کاملJoint Feature Selection in Distributed Stochastic Learning for Large-Scale Discriminative Training in SMT
With a few exceptions, discriminative training in statistical machine translation (SMT) has been content with tuning weights for large feature sets on small development data. Evidence from machine learning indicates that increasing the training sample size results in better prediction. The goal of this paper is to show that this common wisdom can also be brought to bear upon SMT. We deploy loca...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2013