DKPLM: Decomposable Knowledge-Enhanced Pre-trained Language Model for Natural Language Understanding
نویسندگان
چکیده
Knowledge-Enhanced Pre-trained Language Models (KEPLMs) are pre-trained models with relation triples injecting from knowledge graphs to improve language understanding abilities.Experiments show that our model outperforms other KEPLMs significantly over zero-shot probing tasks and multiple knowledge-aware tasks. To guarantee effective injection, previous studies integrate encoders for representing retrieved graphs. The operations retrieval encoding bring significant computational burdens, restricting the usage of such in real-world applications require high inference speed. In this paper, we propose a novel KEPLM named DKPLM decomposes injection process pre-training, fine-tuning stages, which facilitates scenarios. Specifically, first detect long-tail entities as target enhancing KEPLMs' semantic abilities avoiding redundant information. embeddings replaced by ``pseudo token representations'' formed relevant triples. We further design relational decoding task pre-training force truly understand injected triple reconstruction. Experiments has higher speed than competing due decomposing mechanism.
منابع مشابه
Natural Language Understanding with Knowledge
This paper examines the problem of extracting structured knowledge from unstructured free text. The extraction process is modeled after construction grammars, essentially providing a means of putting together form and meaning. The knowledge base is not simply treated as a destination, but also an important partner in the extraction process. In particular, the ideas are implemented as a module c...
متن کاملKnowledge-Aware Natural Language Understanding
Natural Language Understanding (NLU) systems need to encode human generated text (or speech) and reason over it at a deep semantic level. Any NLU system typically involves two main components: The first is an encoder, which composes words (or other basic linguistic units) within the input utterances to compute encoded representations, that are then used as features in the second component, a pr...
متن کاملEnriching Knowledge Sources for Natural Language Understanding
This paper presents the complete and consistent ontological annotation of the nominal part of WordNet. The annotation has been carried out using the semantic features defined in the EuroWordNet Top Concept Ontology and made available to the NLP community. Up to now only an initial core set of 1,024 synsets, the so-called Base Concepts, was ontologized in such a way. The work has been achieved b...
متن کاملA Lexical Knowledge Representation Model for Natural Language Understanding
Knowledge representation is essential for semantics modeling and intelligent information processing. For decades researchers have proposed many knowledge representation techniques. However, it is a daunting problem how to capture deep semantic information effectively and support the construction of a large-scale knowledge base efficiently. This paper describes a new knowledge representation mod...
متن کاملConfiguring Domain Knowledge for Natural Language Understanding
Knowledge-based configuration has been used for numerous applications including natural language processing (NLP). By formalising property grammars as a configuration problem, it has been shown that configuration can provide a flexible, nondeterministic, method of parsing natural language. However, it focuses only on syntactic parsing. In contrast, configuration is usually performed using knowl...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2022
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v36i10.21425