Artigo Acesso aberto Revisado por pares

Translation Performance from the User’s Perspective of Large Language Models and Neural Machine Translation Systems

2023; Multidisciplinary Digital Publishing Institute; Volume: 14; Issue: 10 Linguagem: Inglês

10.3390/info14100574

ISSN

2078-2489

Autores

Jungha Son, Boyoung Kim,

Tópico(s)

Artificial Intelligence in Healthcare and Education

Resumo

The rapid global expansion of ChatGPT, which plays a crucial role in interactive knowledge sharing and translation, underscores the importance of comparative performance assessments in artificial intelligence (AI) technology. This study concentrated on this crucial issue by exploring and contrasting the translation performances of large language models (LLMs) and neural machine translation (NMT) systems. For this aim, the APIs of Google Translate, Microsoft Translator, and OpenAI’s ChatGPT were utilized, leveraging parallel corpora from the Workshop on Machine Translation (WMT) 2018 and 2020 benchmarks. By applying recognized evaluation metrics such as BLEU, chrF, and TER, a comprehensive performance analysis across a variety of language pairs, translation directions, and reference token sizes was conducted. The findings reveal that while Google Translate and Microsoft Translator generally surpass ChatGPT in terms of their BLEU, chrF, and TER scores, ChatGPT exhibits superior performance in specific language pairs. Translations from non-English to English consistently yielded better results across all three systems compared with translations from English to non-English. Significantly, an improvement in translation system performance was observed as the token size increased, hinting at the potential benefits of training models on larger token sizes.

Referência(s)
Altmetric
PlumX