OAK

Impact of Sentence Representation Matching in Neural Machine Translation

Metadata Downloads
Abstract
Most neural machine translation models are implemented as a conditional language model framework composed of encoder and decoder models. This framework learns complex and long-distant dependencies, but its deep structure causes inefficiency in training. Matching vector representations of source and target sentences improves the inefficiency by shortening the depth from parameters to costs and generalizes NMTs with a different perspective to cross-entropy loss. In this paper, we propose matching methods to derive the cost based on constant word-embedding vectors of source and target sentences. To find the best method, we analyze the impact of the methods with varying structures, distance metrics, and model capacity in a French to English translation task. An optimally configured method is applied to English translation tasks from and to French, Spanish, and German. In the tasks, the method showed performance improvement by 3.23 BLEU at maximum, with an improvement of 0.71 on average. We evaluated the robustness of this method to various embedding distributions and models, such as conventional gated structures and transformer networks, and empirical results showed that it has a higher chance to improve performance in those models.
Author(s)
Jung, HeeseungKim, KangilShin, Jong-HunNa, Seung-HoonJung, SangkeunWoo, Sangmin
Issued Date
2022-02
Type
Article
DOI
10.3390/app12031313
URI
https://scholar.gist.ac.kr/handle/local/11021
Publisher
MDPI
Citation
APPLIED SCIENCES-BASEL, v.12, no.3
ISSN
2076-3417
Appears in Collections:
Department of AI Convergence > 1. Journal Articles
공개 및 라이선스
  • 공개 구분공개
파일 목록

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.