Please use this identifier to cite or link to this item: https://evnuir.vnu.edu.ua/handle/123456789/30553
Title: Representing emotive discourse in Ukrainian-English literary translation: A multi-method performance evaluation of Large Language Models, Neural Machine Translation and Computer-Assisted Translation tools
Authors: Karpina, Olena
Zasiekin, Serhii
Affiliation: Lesya Ukrainka Volyn National University, Ukraine
University College London, UK
Bibliographic description (Ukraine): Karpina, O., & Zasiekin, S. (2025). Representing emotive discourse in Ukrainian-English literary translation: A multi-method performance evaluation of Large Language Models, Neural Machine Translation and Computer-Assisted Translation tools. East European Journal of Psycholinguistics , 12(2), 178-203. https://doi.org/10.29038/kar
Journal/Collection: East European Journal of Psycholinguistics
Issue Date: Dec-2025
Date of entry: 2-Mar-2026
Publisher: Lesya Ukrainka Volyn National University
Country (code): UA
Place of the edition/event: Lesya Ukrainka Volyn National University
DOI: https://doi.org/10.29038/kar
Keywords: literary translation
emotive discourse
Large Language Model
quality evaluation
neural machine translation
CAT tools
BLEU
Page range: 178-203
Abstract: The study examines the capacity of modern translation technologies to render Ukrainian literary texts into English. Lesya Ukrainka’s Letter to Serhii Merzhynskyi was chosen as the original text for translation analysis. It is a piece of emotive discourse marked by vivid imagery, nuanced stylistic features, expressive syntactic patterns and archaic vocabulary. Six translation services were tested. They included general-purpose Neural Machine Translation services, Computer-Assisted Translation tools and Large Language Models. Their output was evaluated using a three-step methodological framework. First, automatic evaluation was conducted using a Bilingual Evaluation Understudy (BLEU) metric to provide initial quantitative comparability across the systems’ output. Second, a qualitative analysis was undertaken through the concept of literariness, focusing on literature-specific features, aesthetic and stylistic peculiarities that distinguish literary texts from non-literary ones. In the final stage, human evaluation was employed, with five human annotators – native speakers with advanced linguistic proficiency, professional translators and scholars – ranking sentences to assess MT performance. The results of human evaluation and qualitative analysis revealed that the top-performing translation technologies were LLMs ChatGPT-5 and DeepSeek, which not only met a baseline level of translation adequacy but also consistently surpassed human translation in contextual and emotional sensitivity and overall naturalness and fluency. By contrast, automatic evaluation using the BLEU metric assigned the highest score to Google Translate output, highlighting the metric's limitations for literary text. Despite the notable efficiency of modern translation technologies, certain errors persist to varying degrees across all tested tools. These errors are connected with rendering imagery, handling syntactic constructions with long-range dependencies, translating pronouns, handling register mismatches, disrupting tone and other similar issues.
URI: https://evnuir.vnu.edu.ua/handle/123456789/30553
Copyright owner: © East European Journal of Psycholinguistics, 2025
Content type: Article
Appears in Collections:East European Journal of Psycholinguistics, 2025, Volume 12, Number 2

Files in This Item:
File Description SizeFormat 
eejpl_12_2_2025_Karpina_Zasiekin.pdf358,23 kBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.