Inferring Selectional Preferences from Part-Of-Speech N-grams
نویسندگان
چکیده
We present the PONG method to compute selectional preferences using part-of-speech (POS) N-grams. From a corpus labeled with grammatical dependencies, PONG learns the distribution of word relations for each POS N-gram. From the much larger but unlabeled Google N-grams corpus, PONG learns the distribution of POS N-grams for a given pair of words. We derive the probability that one word has a given grammatical relation to the other. PONG estimates this probability by combining both distributions, whether or not either word occurs in the labeled corpus. PONG achieves higher average precision on 16 relations than a state-of-the-art baseline in a pseudo-disambiguation task, but lower coverage and recall.
منابع مشابه
Exploiting Web-Derived Selectional Preference to Improve Statistical Dependency Parsing
In this paper, we present a novel approach which incorporates the web-derived selectional preferences to improve statistical dependency parsing. Conventional selectional preference learning methods have usually focused on word-to-class relations, e.g., a verb selects as its subject a given nominal class. This paper extends previous work to wordto-word selectional preferences by using webscale d...
متن کاملWord similarity using constructions as contextual features
1 We propose and implement an alternative source of contextual features for word similarity detection based on the notion of lexicogrammatical construction. On the assumption that selectional restrictions provide indicators of the semantic similarity of words attested in selected positions, we extend the notion of selection beyond that of single selecting heads to multiword constructions exerti...
متن کاملA Latent Dirichlet Allocation Method for Selectional Preferences
Computation of selectional preferences, the admissible argument values for a relation, is a well studied NLP task with wide applicability. We present LDA-SP, the first LDA-based approach to computing selectional preferences. By simultaneously inferring latent topics and topic distributions over relations, LDA-SP combines the benefits of previous approaches: it is competitive with the non-class-...
متن کاملIf You Have It, Flaunt It: Using Full Ontological Knowledge for Word Sense Disambiguation
Word sense disambiguation continues to be a difficult problem in natural language processing. Current methods, such as marker passing and spreading activation, for applying world knowledge in the form of selectional preferences to solve this problem do not make effective use of available knowledge. Moreover, their effectiveness decreases as the knowledge is made richer by acquiring more and mor...
متن کاملImproving Finite-State Spell-Checker Suggestions with Part of Speech N-Grams
We demonstrate a finite-state implementation of context-aware spell checking utilizing an N-gram based part of speech (POS) tagger to rerank the suggestions from a simple edit-distance based spell-checker. We demonstrate the benefits of context-aware spellchecking for English and Finnish and introduce modifications that are necessary to make traditional N-gram models work for morphologically mo...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2012