Summary of the paper

Title Evaluation of Machine Translation Performance Across Multiple Genres and Languages
Authors Marlies Van der Wees, Arianna Bisazza and Christof Monz
Abstract In this paper, we present evaluation corpora covering four genres for four language pairs that we harvested from the web in an automated fashion. We use these multi-genre benchmarks to evaluate the impact of genre differences on machine translation (MT). We observe that BLEU score differences between genres can be large and that, for all genres and all language pairs, translation quality improves when using four genre-optimized systems rather than a single genre-agnostic system. Finally, we train and use genre classifiers to route test documents to the most appropriate genre systems. The results of these experiments show that our multi-genre benchmarks can serve to advance research on text genre adaptation for MT.
Topics Corpus (Creation, Annotation, Etc.), Other
Full paper Evaluation of Machine Translation Performance Across Multiple Genres and Languages
Bibtex @InProceedings{VAN DER WEES18.853,
  author = {Marlies Van der Wees and Arianna Bisazza and Christof Monz},
  title = "{Evaluation of Machine Translation Performance Across Multiple Genres and Languages}",
  booktitle = {Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)},
  year = {2018},
  month = {May 7-12, 2018},
  address = {Miyazaki, Japan},
  editor = {Nicoletta Calzolari (Conference chair) and Khalid Choukri and Christopher Cieri and Thierry Declerck and Sara Goggi and Koiti Hasida and Hitoshi Isahara and Bente Maegaard and Joseph Mariani and Hélène Mazo and Asuncion Moreno and Jan Odijk and Stelios Piperidis and Takenobu Tokunaga},
  publisher = {European Language Resources Association (ELRA)},
  isbn = {979-10-95546-00-9},
  language = {english}
  }
Powered by ELDA © 2018 ELDA/ELRA