Abstract
In this paper, we describe a tool for debugging the output and attention weights of neural machine translation (NMT) systems and for improved estimations of confidence about the output based on the attention. The purpose of the tool is to help researchers and developers find weak and faulty example translations that their NMT systems produce without the need for reference translations. Our tool also includes an option to directly compare translation outputs from two different NMT engines or experiments. In addition, we present a demo website of our tool with examples of good and bad translations: <a href="http://attention.lielakeda.lv">this http URL</a>
Abstract (translated)
在本文中,我们描述了一种用于调试神经机器翻译(NMT)系统的输出和注意权重的工具,以及基于注意力改进的输出置信度估计。该工具的目的是帮助研究人员和开发人员找到他们的NMT系统产生的弱且错误的示例翻译,而无需参考翻译。我们的工具还包括直接比较两个不同NMT引擎或实验的翻译输出的选项。此外,我们还提供了我们工具的演示网站,其中包含好的和坏的翻译示例:<a href="http://attention.lielakeda.lv">此http网址</a>
URL
https://arxiv.org/abs/1808.02733