Incremental Methods to Select Test Sentences for Evaluating Translation Ability
نویسندگان
چکیده
This paper addresses the problem of selecting test sentences for automatically evaluating language learners’ translation ability within a smaller error. In this paper, the ability to translate is measured as a TOEIC score. The existing selection methods only check whether an individual test sentence contributes to the estimation of the ability to translate or that of more general academic abilities, although combinations of test sentences may be used to contribute the estimation. This paper proposes two methods that solve the selection problem. The first method selects test sentences to minimize the estimation errors of learners’ TOEIC scores. The second method selects test sentences to maximize the correlation coefficient between the number of correct translations and learners’ estimated TOEIC scores. The optimization technique used in both of the proposed methods is the gradient technique in mathematical programming. The proposed methods proved to be more accurate than any of the existing methods we tested, and they estimated each TOEIC score within a permissible error of 69 points.
منابع مشابه
The Effects of Implementing Summative Assessment, Formative Assessment and Dynamic Assessment on Iranian EFL Learners’ Listening Ability and Listening Strategy Use
Every society has its own culture, values, and ideology. Translations convey the meaning as well as cultural and ideological values, beliefs, ideas and norms from source culture to target culture. One type of translation which nowadays is popular among people, and can attract so many audiences in different ages is Audiovisual Translation (AVT), particularly dubbing. If there is difference betwe...
متن کاملEvaluating Discourse Phenomena in Neural Machine Translation
For machine translation to tackle discourse phenomena, models must have access to extrasentential linguistic context. There has been recent interest in modelling context in neural machine translation (NMT), but models have been principally evaluated with standard automatic metrics, poorly adapted to evaluating discourse phenomena. In this article, we present hand-crafted, discourse test sets, d...
متن کاملIncremental Prediction of Sentence-final Verbs: Humans versus Machines
Verb prediction is important in human sentence processing and, practically, in simultaneous machine translation. In verb-final languages, speakers select the final verb before it is uttered, and listeners predict it before it is uttered. Simultaneous interpreters must do the same to translate in real-time. Motivated by the problem of SOV-SVO simultaneous machine translation, we provide a study ...
متن کاملCS562/CS662 (Natural Language Processing): Evaluating machine translation quality with BLEU
The gold standard for measuring machine translation quality is the rating of candidate sentences by by experienced translators. However, automated measures are necessary for rapid iterative development. BLEU (Papineni et al. 2002) is the best-known automatic measure of translation quality. BLEU and related measures are used to automatically evaluate machine translation (MT) systems, as well as ...
متن کاملThe Correlation of Machine Translation Evaluation Metrics with Human Judgement on Persian Language
Machine Translation Evaluation Metrics (MTEMs) are the central core of Machine Translation (MT) engines as they are developed based on frequent evaluation. Although MTEMs are widespread today, their validity and quality for many languages is still under question. The aim of this research study was to examine the validity and assess the quality of MTEMs from Lexical Similarity set on machine tra...
متن کامل