Measuring comprehension and perception of neural machine translated texts : a pilot study
- Author
- Lieve Macken (UGent) and Iris Ghyselen
- Organization
- Abstract
- In this paper we compare the results of reading comprehension tests on both human translated and raw (unedited) machine translated texts. We selected three texts of the English Machine Translation Evaluation version (CREG-MT-eval) of the Corpus of Reading Comprehension Exercises (CREG), for which we produced three different translations: a manual translation and two automatic translations generated by two state-of-the-art neural machine translation engines, viz. DeepL and Google Translate. The experiment was conducted via a SurveyMonkey questionnaire, which 99 participants filled in. Participants were asked to read the translation very carefully after which they had to answer the comprehension questions without having access to the translated text. Apart from assessing comprehension, we posed additional questions to get information on the participants’ perception of the machine translations. The results show that 74% of the participants can tell whether a translation was produced by a human or a machine. Human translations received the best overall clarity scores, but the reading comprehension tests provided much less unequivocal results. The errors that bother readers most relate to grammar, sentence length, level of idiomaticity and incoherence.
- Keywords
- machine translation, comprehensibility, LT3
Downloads
-
TC4013.pdf
- full text
- |
- open access
- |
- |
- 175.26 KB
Citation
Please use this url to cite or link to this publication: http://hdl.handle.net/1854/LU-8580951
- MLA
- Macken, Lieve, and Iris Ghyselen. “Measuring Comprehension and Perception of Neural Machine Translated Texts : A Pilot Study.” Translating and the Computer 40 (TC40) : Proceedings, Editions Tradulex, 2018, pp. 120–26.
- APA
- Macken, L., & Ghyselen, I. (2018). Measuring comprehension and perception of neural machine translated texts : a pilot study. Translating and the Computer 40 (TC40) : Proceedings, 120–126. Geneva: Editions Tradulex.
- Chicago author-date
- Macken, Lieve, and Iris Ghyselen. 2018. “Measuring Comprehension and Perception of Neural Machine Translated Texts : A Pilot Study.” In Translating and the Computer 40 (TC40) : Proceedings, 120–26. Geneva: Editions Tradulex.
- Chicago author-date (all authors)
- Macken, Lieve, and Iris Ghyselen. 2018. “Measuring Comprehension and Perception of Neural Machine Translated Texts : A Pilot Study.” In Translating and the Computer 40 (TC40) : Proceedings, 120–126. Geneva: Editions Tradulex.
- Vancouver
- 1.Macken L, Ghyselen I. Measuring comprehension and perception of neural machine translated texts : a pilot study. In: Translating and the computer 40 (TC40) : proceedings. Geneva: Editions Tradulex; 2018. p. 120–6.
- IEEE
- [1]L. Macken and I. Ghyselen, “Measuring comprehension and perception of neural machine translated texts : a pilot study,” in Translating and the computer 40 (TC40) : proceedings, London, 2018, pp. 120–126.
@inproceedings{8580951, abstract = {{In this paper we compare the results of reading comprehension tests on both human translated and raw (unedited) machine translated texts. We selected three texts of the English Machine Translation Evaluation version (CREG-MT-eval) of the Corpus of Reading Comprehension Exercises (CREG), for which we produced three different translations: a manual translation and two automatic translations generated by two state-of-the-art neural machine translation engines, viz. DeepL and Google Translate. The experiment was conducted via a SurveyMonkey questionnaire, which 99 participants filled in. Participants were asked to read the translation very carefully after which they had to answer the comprehension questions without having access to the translated text. Apart from assessing comprehension, we posed additional questions to get information on the participants’ perception of the machine translations. The results show that 74% of the participants can tell whether a translation was produced by a human or a machine. Human translations received the best overall clarity scores, but the reading comprehension tests provided much less unequivocal results. The errors that bother readers most relate to grammar, sentence length, level of idiomaticity and incoherence.}}, author = {{Macken, Lieve and Ghyselen, Iris}}, booktitle = {{Translating and the computer 40 (TC40) : proceedings}}, isbn = {{9782970109556}}, keywords = {{machine translation,comprehensibility,LT3}}, language = {{eng}}, location = {{London}}, pages = {{120--126}}, publisher = {{Editions Tradulex}}, title = {{Measuring comprehension and perception of neural machine translated texts : a pilot study}}, url = {{https://www.asling.org/tc40/wp-content/uploads/TC40-Proceedings.pdf}}, year = {{2018}}, }