DKPLM: Decomposable Knowledge-Enhanced Pre-trained Language Model for Natural Language Understanding

نویسندگان

چکیده

Knowledge-Enhanced Pre-trained Language Models (KEPLMs) are pre-trained models with relation triples injecting from knowledge graphs to improve language understanding abilities.Experiments show that our model outperforms other KEPLMs significantly over zero-shot probing tasks and multiple knowledge-aware tasks. To guarantee effective injection, previous studies integrate encoders for representing retrieved graphs. The operations retrieval encoding bring significant computational burdens, restricting the usage of such in real-world applications require high inference speed. In this paper, we propose a novel KEPLM named DKPLM decomposes injection process pre-training, fine-tuning stages, which facilitates scenarios. Specifically, first detect long-tail entities as target enhancing KEPLMs' semantic abilities avoiding redundant information. embeddings replaced by ``pseudo token representations'' formed relevant triples. We further design relational decoding task pre-training force truly understand injected triple reconstruction. Experiments has higher speed than competing due decomposing mechanism.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Natural Language Understanding with Knowledge

This paper examines the problem of extracting structured knowledge from unstructured free text. The extraction process is modeled after construction grammars, essentially providing a means of putting together form and meaning. The knowledge base is not simply treated as a destination, but also an important partner in the extraction process. In particular, the ideas are implemented as a module c...

متن کامل

Knowledge-Aware Natural Language Understanding

Natural Language Understanding (NLU) systems need to encode human generated text (or speech) and reason over it at a deep semantic level. Any NLU system typically involves two main components: The first is an encoder, which composes words (or other basic linguistic units) within the input utterances to compute encoded representations, that are then used as features in the second component, a pr...

متن کامل

Enriching Knowledge Sources for Natural Language Understanding

This paper presents the complete and consistent ontological annotation of the nominal part of WordNet. The annotation has been carried out using the semantic features defined in the EuroWordNet Top Concept Ontology and made available to the NLP community. Up to now only an initial core set of 1,024 synsets, the so-called Base Concepts, was ontologized in such a way. The work has been achieved b...

متن کامل

A Lexical Knowledge Representation Model for Natural Language Understanding

Knowledge representation is essential for semantics modeling and intelligent information processing. For decades researchers have proposed many knowledge representation techniques. However, it is a daunting problem how to capture deep semantic information effectively and support the construction of a large-scale knowledge base efficiently. This paper describes a new knowledge representation mod...

متن کامل

Configuring Domain Knowledge for Natural Language Understanding

Knowledge-based configuration has been used for numerous applications including natural language processing (NLP). By formalising property grammars as a configuration problem, it has been shown that configuration can provide a flexible, nondeterministic, method of parsing natural language. However, it focuses only on syntactic parsing. In contrast, configuration is usually performed using knowl...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2022

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v36i10.21425