Title

Training a Sentence-Level Machine Translation Confidence Measure

Author(s)

Christopher B. Quirk

Microsoft Research

Session

O21-EW

Abstract

We present a supervised method for training a sentence level confidence measure on translation output using a human-annotated corpus. We evaluate a variety of machine learning methods. The resultant measure, while trained on a very small dataset, correlates well with human judgments, and proves to be effective on one task based evaluation. Although the experiments have only been run on one MT system, we believe the nature of the features gathered are general enough that the approach will also work well on other systems.

Keyword(s)

confidence estimation; machine translation; mt evaluation

Language(s)

English, Spanish

Full Paper

426.pdf