Probabilistic Word Alignment under the $L_0$-norm
نویسنده
چکیده
This paper makes two contributions to the area of single-word based word alignment for bilingual sentence pairs. Firstly, it integrates the – seemingly rather different – works of (Bodrumlu et al., 2009) and the standard probabilistic ones into a single framework. Secondly, we present two algorithms to optimize the arising task. The first is an iterative scheme similar to Viterbi training, able to handle large tasks. The second is based on the inexact solution of an integer program. While it can handle only small corpora, it allows more insight into the quality of the model and the performance of the iterative scheme. Finally, we present an alternative way to handle prior dictionary knowledge and discuss connections to computing IBM-3 Viterbi alignments.
منابع مشابه
Correspondence between probabilistic norms and fuzzy norms
In this paper, the connection between Menger probabilistic norms and H"{o}hle probabilistic norms is discussed. In addition, the correspondence between probabilistic norms and Wu-Fang fuzzy (semi-) norms is established. It is shown that a probabilistic norm (with triangular norm $min$) can generate a Wu-Fang fuzzy semi-norm and conversely, a Wu-Fang fuzzy norm can generate a probabilistic norm.
متن کاملSampling Alignment Structure under a Bayesian Translation Model
We describe the first tractable Gibbs sampling procedure for estimating phrase pair frequencies under a probabilistic model of phrase alignment. We propose and evaluate two nonparametric priors that successfully avoid the degenerate behavior noted in previous work, where overly large phrases memorize the training data. Phrase table weights learned under our model yield an increase in BLEU score...
متن کاملPerformance Analysis of $l_0$ Norm Constrained Recursive Least Squares Algorithm
Performance analysis of l0 norm constrained Recursive least Squares (RLS) algorithm is attempted in this paper. Though the performance pretty attractive compared to its various alternatives, no thorough study of theoretical analysis has been performed. Like the popular l0 Least Mean Squares (LMS) algorithm, in l0 RLS, a l0 norm penalty is added to provide zero tap attractions on the instantaneo...
متن کاملWord Alignment with Synonym Regularization
We present a novel framework for word alignment that incorporates synonym knowledge collected from monolingual linguistic resources in a bilingual probabilistic model. Synonym information is helpful for word alignment because we can expect a synonym to correspond to the same word in a different language. We design a generative model for word alignment that uses synonym information as a regulari...
متن کاملProbabilistic Normed Groups
In this paper, we introduce the probabilistic normed groups. Among other results, we investigate the continuityof inner automorphisms of a group and the continuity of left and right shifts in probabilistic group-norm. We also study midconvex functions defined on probabilistic normed groups and give some results about locally boundedness of such functions.
متن کامل