The tara corpus of human-annotated machine translations

Eleftherios Avramidis, Aljoscha Burchardt, Sabine Hunsicker, Maja Popović, Cindy Tscherwinka, David Vilar, Hans Uszkoreit


Abstract
Human translators are the key to evaluating machine translation (MT) quality and also to addressing the so far unanswered question when and how to use MT in professional translation workflows. This paper describes the corpus developed as a result of a detailed large scale human evaluation consisting of three tightly connected tasks: ranking, error classification and post-editing.
Anthology ID:
L14-1347
Volume:
Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)
Month:
May
Year:
2014
Address:
Reykjavik, Iceland
Editors:
Nicoletta Calzolari, Khalid Choukri, Thierry Declerck, Hrafn Loftsson, Bente Maegaard, Joseph Mariani, Asuncion Moreno, Jan Odijk, Stelios Piperidis
Venue:
LREC
SIG:
Publisher:
European Language Resources Association (ELRA)
Note:
Pages:
2679–2682
Language:
URL:
http://www.lrec-conf.org/proceedings/lrec2014/pdf/401_Paper.pdf
DOI:
Bibkey:
Cite (ACL):
Eleftherios Avramidis, Aljoscha Burchardt, Sabine Hunsicker, Maja Popović, Cindy Tscherwinka, David Vilar, and Hans Uszkoreit. 2014. The taraXÜ corpus of human-annotated machine translations. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14), pages 2679–2682, Reykjavik, Iceland. European Language Resources Association (ELRA).
Cite (Informal):
The taraXÜ corpus of human-annotated machine translations (Avramidis et al., LREC 2014)
Copy Citation:
PDF:
http://www.lrec-conf.org/proceedings/lrec2014/pdf/401_Paper.pdf