Probabilistic Grammars as Models of Gradience in Language Processing
نویسندگان
چکیده
This article deals with gradience in human sentence processing. We review the experimental evidence for the role of experience in guiding the decisions of the sentence processor. Based on this evidence, we argue that the gradient behavior observed in the processing of certain syntactic constructions can be traced back to the amount of past experience that the processor has had with these constructions. In modeling terms, linguistic experience can be approximated using large, balanced corpora. We give an overview of corpus-based and probabilistic models in the literature that have exploited this fact, and hence are well placed to make gradient predictions about processing behavior. Finally, we discuss a number of questions regarding the relationship between gradience in sentence processing and gradient grammaticality, and come to the conclusion that these two phenomena should be treated separately in conceptual and modeling
منابع مشابه
Studying impressive parameters on the performance of Persian probabilistic context free grammar parser
In linguistics, a tree bank is a parsed text corpus that annotates syntactic or semantic sentence structure. The exploitation of tree bank data has been important ever since the first large-scale tree bank, The Penn Treebank, was published. However, although originating in computational linguistics, the value of tree bank is becoming more widely appreciated in linguistics research as a whole. F...
متن کاملComputational Complexity of Probabilistic Disambiguation
Recent models of natural language processing employ statistical reasoning for dealing with the ambiguity of formal grammars. In this approach, statistics, concerning the various linguistic phenomena of interest, are gathered from actual linguistic data and used to estimate the probabilities of the various entities that are generated by a given grammar, e.g., derivations, parse-trees and sentenc...
متن کاملComputational Learning of Probabilistic Grammars in the Unsupervised Setting
With the rising amount of available multilingual text data, computational linguistics faces an opportunity and a challenge. This text can enrich the domains of NLP applications and improve their performance. Traditional supervised learning for this kind of data would require annotation of part of this text for induction of natural language structure. For these large amounts of rich text, such a...
متن کاملAn Empirical Evaluation of Probabilistic Lexicalized Tree Insertion Grammars
We present an empirical study of the applicability of Probabilistic Lexicalized Tree Insertion Grammars (PLTIG), a lexicalized counterpart to Probabilistic Context-Free Grammars (PCFG), to problems in stochastic naturallanguage processing. Comparing the performance of PLTIGs with non-hierarchicalN -gram models and PCFGs, we show that PLTIG combines the best aspects of both, with language modeli...
متن کاملGrammaticality, Acceptability, and Probability: A Probabilistic View of Linguistic Knowledge
The question of whether humans represent grammatical knowledge as a binary condition on membership in a set of well-formed sentences, or as a probabilistic property has been the subject of debate among linguists, psychologists, and cognitive scientists for many decades. Acceptability judgments present a serious problem for both classical binary and probabilistic theories of grammaticality. Thes...
متن کامل