Paper ID | HLT-8.4 | ||
Paper Title | IMAGE-ASSISTED TRANSFORMER IN ZERO-RESOURCE MULTI-MODAL TRANSLATION | ||
Authors | Ping Huang, Shiliang Sun, East China Normal University, China; Hao Yang, Huawei Technologies CO., LTD, China | ||
Session | HLT-8: Speech Translation 2: Aspects | ||
Location | Gather.Town | ||
Session Time: | Wednesday, 09 June, 14:00 - 14:45 | ||
Presentation Time: | Wednesday, 09 June, 14:00 - 14:45 | ||
Presentation | Poster | ||
Topic | Human Language Technology: [HLT-MTSW] Machine Translation for Spoken and Written Language | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | Humans learn language speaking and translation with the help of common knowledge of the external world, while standard machine translation depends only on parallel corpora. Therefore, zero-resource translation has been proposed to explore a way to make models learn to translate without any parallel corpora but with external knowledge. Current models in the field usually combine an image encoder with a textual decoder, which leads to extra efforts to make use of the original model in machine translation. On the other hand, Transformers have achieved great progress in neural language processing while they are rarely utilized in the field of zero-resource translation with image pivot. In this paper, we investigate how to use visual information as an auxiliary hint for a Transformer-based system in a zero-resource translation scenario. Our model achieves state-of-the-art BLEU scores in the field of zero-resource machine translation with image pivot. |