Loading…

Bridging the Domain Gaps in Context Representations for k-Nearest Neighbor Neural Machine Translation

\(k\)-Nearest neighbor machine translation (\(k\)NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains. By using an upstream NMT model to traverse the downstream training corpus, it is equipped with a datastore containing vectorized key-v...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2023-05
Main Authors: Cao, Zhiwei, Yang, Baosong, Lin, Huan, Wu, Suhang, Wei, Xiangpeng, Liu, Dayiheng, Xie, Jun, Zhang, Min, Su, Jinsong
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Cao, Zhiwei
Yang, Baosong
Lin, Huan
Wu, Suhang
Wei, Xiangpeng
Liu, Dayiheng
Xie, Jun
Zhang, Min
Su, Jinsong
description \(k\)-Nearest neighbor machine translation (\(k\)NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains. By using an upstream NMT model to traverse the downstream training corpus, it is equipped with a datastore containing vectorized key-value pairs, which are retrieved during inference to benefit translation. However, there often exists a significant gap between upstream and downstream domains, which hurts the retrieval accuracy and the final translation quality. To deal with this issue, we propose a novel approach to boost the datastore retrieval of \(k\)NN-MT by reconstructing the original datastore. Concretely, we design a reviser to revise the key representations, making them better fit for the downstream domain. The reviser is trained using the collected semantically-related key-queries pairs, and optimized by two proposed losses: one is the key-queries semantic distance ensuring each revised key representation is semantically related to its corresponding queries, and the other is an L2-norm loss encouraging revised key representations to effectively retain the knowledge learned by the upstream NMT model. Extensive experiments on domain adaptation tasks demonstrate that our method can effectively boost the datastore retrieval and translation quality of \(k\)NN-MT.\footnote{Our code is available at \url{https://github.com/DeepLearnXMU/RevisedKey-knn-mt}.}
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2820199191</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2820199191</sourcerecordid><originalsourceid>FETCH-proquest_journals_28201991913</originalsourceid><addsrcrecordid>eNqNjN0KgjAYhkcQJOU9fNCxsM0sPc3-TvIgPJdVnzqzzbYJXX4juoCOHnjenwkJeByzKF1xPiOhtR2llK83PEnigODWyHsjVQOuRdjpp5AKjmKw4Jlr5fDt4IKDQYvKCSe1slBrA4-oQOGtgwJl0169KnA0ooezuLVSIZRGKNt_JwsyrUVvMfxxTpaHfZmfosHo1-hPqk6PRvmo4imnLMtYxuL_Wh8_lUaU</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2820199191</pqid></control><display><type>article</type><title>Bridging the Domain Gaps in Context Representations for k-Nearest Neighbor Neural Machine Translation</title><source>Publicly Available Content Database</source><creator>Cao, Zhiwei ; Yang, Baosong ; Lin, Huan ; Wu, Suhang ; Wei, Xiangpeng ; Liu, Dayiheng ; Xie, Jun ; Zhang, Min ; Su, Jinsong</creator><creatorcontrib>Cao, Zhiwei ; Yang, Baosong ; Lin, Huan ; Wu, Suhang ; Wei, Xiangpeng ; Liu, Dayiheng ; Xie, Jun ; Zhang, Min ; Su, Jinsong</creatorcontrib><description>\(k\)-Nearest neighbor machine translation (\(k\)NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains. By using an upstream NMT model to traverse the downstream training corpus, it is equipped with a datastore containing vectorized key-value pairs, which are retrieved during inference to benefit translation. However, there often exists a significant gap between upstream and downstream domains, which hurts the retrieval accuracy and the final translation quality. To deal with this issue, we propose a novel approach to boost the datastore retrieval of \(k\)NN-MT by reconstructing the original datastore. Concretely, we design a reviser to revise the key representations, making them better fit for the downstream domain. The reviser is trained using the collected semantically-related key-queries pairs, and optimized by two proposed losses: one is the key-queries semantic distance ensuring each revised key representation is semantically related to its corresponding queries, and the other is an L2-norm loss encouraging revised key representations to effectively retain the knowledge learned by the upstream NMT model. Extensive experiments on domain adaptation tasks demonstrate that our method can effectively boost the datastore retrieval and translation quality of \(k\)NN-MT.\footnote{Our code is available at \url{https://github.com/DeepLearnXMU/RevisedKey-knn-mt}.}</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Domains ; Machine translation ; Queries ; Representations ; Retrieval ; Translations ; Upstream</subject><ispartof>arXiv.org, 2023-05</ispartof><rights>2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2820199191?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>776,780,25731,36989,44566</link.rule.ids></links><search><creatorcontrib>Cao, Zhiwei</creatorcontrib><creatorcontrib>Yang, Baosong</creatorcontrib><creatorcontrib>Lin, Huan</creatorcontrib><creatorcontrib>Wu, Suhang</creatorcontrib><creatorcontrib>Wei, Xiangpeng</creatorcontrib><creatorcontrib>Liu, Dayiheng</creatorcontrib><creatorcontrib>Xie, Jun</creatorcontrib><creatorcontrib>Zhang, Min</creatorcontrib><creatorcontrib>Su, Jinsong</creatorcontrib><title>Bridging the Domain Gaps in Context Representations for k-Nearest Neighbor Neural Machine Translation</title><title>arXiv.org</title><description>\(k\)-Nearest neighbor machine translation (\(k\)NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains. By using an upstream NMT model to traverse the downstream training corpus, it is equipped with a datastore containing vectorized key-value pairs, which are retrieved during inference to benefit translation. However, there often exists a significant gap between upstream and downstream domains, which hurts the retrieval accuracy and the final translation quality. To deal with this issue, we propose a novel approach to boost the datastore retrieval of \(k\)NN-MT by reconstructing the original datastore. Concretely, we design a reviser to revise the key representations, making them better fit for the downstream domain. The reviser is trained using the collected semantically-related key-queries pairs, and optimized by two proposed losses: one is the key-queries semantic distance ensuring each revised key representation is semantically related to its corresponding queries, and the other is an L2-norm loss encouraging revised key representations to effectively retain the knowledge learned by the upstream NMT model. Extensive experiments on domain adaptation tasks demonstrate that our method can effectively boost the datastore retrieval and translation quality of \(k\)NN-MT.\footnote{Our code is available at \url{https://github.com/DeepLearnXMU/RevisedKey-knn-mt}.}</description><subject>Domains</subject><subject>Machine translation</subject><subject>Queries</subject><subject>Representations</subject><subject>Retrieval</subject><subject>Translations</subject><subject>Upstream</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNjN0KgjAYhkcQJOU9fNCxsM0sPc3-TvIgPJdVnzqzzbYJXX4juoCOHnjenwkJeByzKF1xPiOhtR2llK83PEnigODWyHsjVQOuRdjpp5AKjmKw4Jlr5fDt4IKDQYvKCSe1slBrA4-oQOGtgwJl0169KnA0ooezuLVSIZRGKNt_JwsyrUVvMfxxTpaHfZmfosHo1-hPqk6PRvmo4imnLMtYxuL_Wh8_lUaU</recordid><startdate>20230526</startdate><enddate>20230526</enddate><creator>Cao, Zhiwei</creator><creator>Yang, Baosong</creator><creator>Lin, Huan</creator><creator>Wu, Suhang</creator><creator>Wei, Xiangpeng</creator><creator>Liu, Dayiheng</creator><creator>Xie, Jun</creator><creator>Zhang, Min</creator><creator>Su, Jinsong</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20230526</creationdate><title>Bridging the Domain Gaps in Context Representations for k-Nearest Neighbor Neural Machine Translation</title><author>Cao, Zhiwei ; Yang, Baosong ; Lin, Huan ; Wu, Suhang ; Wei, Xiangpeng ; Liu, Dayiheng ; Xie, Jun ; Zhang, Min ; Su, Jinsong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_28201991913</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Domains</topic><topic>Machine translation</topic><topic>Queries</topic><topic>Representations</topic><topic>Retrieval</topic><topic>Translations</topic><topic>Upstream</topic><toplevel>online_resources</toplevel><creatorcontrib>Cao, Zhiwei</creatorcontrib><creatorcontrib>Yang, Baosong</creatorcontrib><creatorcontrib>Lin, Huan</creatorcontrib><creatorcontrib>Wu, Suhang</creatorcontrib><creatorcontrib>Wei, Xiangpeng</creatorcontrib><creatorcontrib>Liu, Dayiheng</creatorcontrib><creatorcontrib>Xie, Jun</creatorcontrib><creatorcontrib>Zhang, Min</creatorcontrib><creatorcontrib>Su, Jinsong</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Cao, Zhiwei</au><au>Yang, Baosong</au><au>Lin, Huan</au><au>Wu, Suhang</au><au>Wei, Xiangpeng</au><au>Liu, Dayiheng</au><au>Xie, Jun</au><au>Zhang, Min</au><au>Su, Jinsong</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Bridging the Domain Gaps in Context Representations for k-Nearest Neighbor Neural Machine Translation</atitle><jtitle>arXiv.org</jtitle><date>2023-05-26</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>\(k\)-Nearest neighbor machine translation (\(k\)NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains. By using an upstream NMT model to traverse the downstream training corpus, it is equipped with a datastore containing vectorized key-value pairs, which are retrieved during inference to benefit translation. However, there often exists a significant gap between upstream and downstream domains, which hurts the retrieval accuracy and the final translation quality. To deal with this issue, we propose a novel approach to boost the datastore retrieval of \(k\)NN-MT by reconstructing the original datastore. Concretely, we design a reviser to revise the key representations, making them better fit for the downstream domain. The reviser is trained using the collected semantically-related key-queries pairs, and optimized by two proposed losses: one is the key-queries semantic distance ensuring each revised key representation is semantically related to its corresponding queries, and the other is an L2-norm loss encouraging revised key representations to effectively retain the knowledge learned by the upstream NMT model. Extensive experiments on domain adaptation tasks demonstrate that our method can effectively boost the datastore retrieval and translation quality of \(k\)NN-MT.\footnote{Our code is available at \url{https://github.com/DeepLearnXMU/RevisedKey-knn-mt}.}</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2023-05
issn 2331-8422
language eng
recordid cdi_proquest_journals_2820199191
source Publicly Available Content Database
subjects Domains
Machine translation
Queries
Representations
Retrieval
Translations
Upstream
title Bridging the Domain Gaps in Context Representations for k-Nearest Neighbor Neural Machine Translation
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T14%3A55%3A53IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Bridging%20the%20Domain%20Gaps%20in%20Context%20Representations%20for%20k-Nearest%20Neighbor%20Neural%20Machine%20Translation&rft.jtitle=arXiv.org&rft.au=Cao,%20Zhiwei&rft.date=2023-05-26&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2820199191%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_28201991913%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2820199191&rft_id=info:pmid/&rfr_iscdi=true