ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.08092
50
1

Languages Transferred Within the Encoder: On Representation Transfer in Zero-Shot Multilingual Translation

12 June 2024
Zhi Qu
Chenchen Ding
Taro Watanabe
ArXivPDFHTML
Abstract

Understanding representation transfer in multilingual neural machine translation (MNMT) can reveal the reason for the zero-shot translation deficiency. In this work, we systematically analyze the representational issue of MNMT models. We first introduce the identity pair, translating a sentence to itself, to address the lack of the base measure in multilingual investigations, as the identity pair can reflect the representation of a language within the model. Then, we demonstrate that the encoder transfers the source language to the representational subspace of the target language instead of the language-agnostic state. Thus, the zero-shot translation deficiency arises because the representation of a translation is entangled with other languages and not transferred to the target language effectively. Based on our findings, we propose two methods: 1) low-rank language-specific embedding at the encoder, and 2) language-specific contrastive learning of the representation at the decoder. The experimental results on Europarl-15, TED-19, and OPUS-100 datasets show that our methods substantially enhance the performance of zero-shot translations without sacrifices in supervised directions by improving language transfer capacity, thereby providing practical evidence to support our conclusions. Codes are available atthis https URL.

View on arXiv
@article{qu2025_2406.08092,
  title={ Languages Transferred Within the Encoder: On Representation Transfer in Zero-Shot Multilingual Translation },
  author={ Zhi Qu and Chenchen Ding and Taro Watanabe },
  journal={arXiv preprint arXiv:2406.08092},
  year={ 2025 }
}
Comments on this paper