Probabilistic Word Alignment under the $L_0$-norm

نویسنده

  • Thomas Schoenemann
چکیده

This paper makes two contributions to the area of single-word based word alignment for bilingual sentence pairs. Firstly, it integrates the – seemingly rather different – works of (Bodrumlu et al., 2009) and the standard probabilistic ones into a single framework. Secondly, we present two algorithms to optimize the arising task. The first is an iterative scheme similar to Viterbi training, able to handle large tasks. The second is based on the inexact solution of an integer program. While it can handle only small corpora, it allows more insight into the quality of the model and the performance of the iterative scheme. Finally, we present an alternative way to handle prior dictionary knowledge and discuss connections to computing IBM-3 Viterbi alignments.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Correspondence between probabilistic norms and fuzzy norms

In this paper, the connection between Menger probabilistic norms and H"{o}hle probabilistic norms is discussed. In addition, the correspondence between probabilistic norms and Wu-Fang fuzzy (semi-) norms is established. It is shown that a probabilistic norm (with triangular norm $min$) can generate a Wu-Fang fuzzy semi-norm and conversely, a Wu-Fang fuzzy norm can generate a probabilistic norm.

متن کامل

Sampling Alignment Structure under a Bayesian Translation Model

We describe the first tractable Gibbs sampling procedure for estimating phrase pair frequencies under a probabilistic model of phrase alignment. We propose and evaluate two nonparametric priors that successfully avoid the degenerate behavior noted in previous work, where overly large phrases memorize the training data. Phrase table weights learned under our model yield an increase in BLEU score...

متن کامل

Performance Analysis of $l_0$ Norm Constrained Recursive Least Squares Algorithm

Performance analysis of l0 norm constrained Recursive least Squares (RLS) algorithm is attempted in this paper. Though the performance pretty attractive compared to its various alternatives, no thorough study of theoretical analysis has been performed. Like the popular l0 Least Mean Squares (LMS) algorithm, in l0 RLS, a l0 norm penalty is added to provide zero tap attractions on the instantaneo...

متن کامل

Word Alignment with Synonym Regularization

We present a novel framework for word alignment that incorporates synonym knowledge collected from monolingual linguistic resources in a bilingual probabilistic model. Synonym information is helpful for word alignment because we can expect a synonym to correspond to the same word in a different language. We design a generative model for word alignment that uses synonym information as a regulari...

متن کامل

Probabilistic Normed Groups

In this paper, we introduce the  probabilistic normed groups. Among other results, we investigate the continuityof inner automorphisms of a group and the continuity of left and right shifts in probabilistic group-norm. We also study midconvex functions defined  on probabilistic normed groups and  give  some results about locally boundedness of such  functions.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2011