Loading…
A Quantitative Approach to Understand Self-Supervised Models as Cross-lingual Feature Extractors
In this work, we study the features extracted by English self-supervised learning (SSL) models in cross-lingual contexts and propose a new metric to predict the quality of feature representations. Using automatic speech recognition (ASR) as a downstream task, we analyze the effect of model size, tra...
Saved in:
Published in: | arXiv.org 2023-11 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Li, Shuyue Stella Xu, Beining Zhang, Xiangyu Liu, Hexin Chao, Wenhan Leibny Paola Garcia |
description | In this work, we study the features extracted by English self-supervised learning (SSL) models in cross-lingual contexts and propose a new metric to predict the quality of feature representations. Using automatic speech recognition (ASR) as a downstream task, we analyze the effect of model size, training objectives, and model architecture on the models' performance as a feature extractor for a set of topologically diverse corpora. We develop a novel metric, the Phonetic-Syntax Ratio (PSR), to measure the phonetic and synthetic information in the extracted representations using deep generalized canonical correlation analysis. Results show the contrastive loss in the wav2vec2.0 objective facilitates more effective cross-lingual feature extraction. There is a positive correlation between PSR scores and ASR performance, suggesting that phonetic information extracted by monolingual SSL models can be used for downstream tasks in cross-lingual settings. The proposed metric is an effective indicator of the quality of the representations and can be useful for model selection. |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2894588358</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2894588358</sourcerecordid><originalsourceid>FETCH-proquest_journals_28945883583</originalsourceid><addsrcrecordid>eNqNjbEKwjAQQIMgKOo_HDgXamI1jkUUFwdR53rYU1tCUnMX8fN18AOc3vAevJ4aamNmmZ1rPVAT5jbPc71Y6qIwQ3Up4ZDQSyMozYug7LoY8PoACXD2NUUW9DUcyd2yY-oovhqmGvahJseADOsYmDPX-HtCB1tCSZFg85aIVwmRx6p_Q8c0-XGkptvNab3Lvp9nIpaqDSn6r6q0Xc0La01hzX_VB_2PRT0</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2894588358</pqid></control><display><type>article</type><title>A Quantitative Approach to Understand Self-Supervised Models as Cross-lingual Feature Extractors</title><source>Publicly Available Content Database</source><creator>Li, Shuyue Stella ; Xu, Beining ; Zhang, Xiangyu ; Liu, Hexin ; Chao, Wenhan ; Leibny Paola Garcia</creator><creatorcontrib>Li, Shuyue Stella ; Xu, Beining ; Zhang, Xiangyu ; Liu, Hexin ; Chao, Wenhan ; Leibny Paola Garcia</creatorcontrib><description>In this work, we study the features extracted by English self-supervised learning (SSL) models in cross-lingual contexts and propose a new metric to predict the quality of feature representations. Using automatic speech recognition (ASR) as a downstream task, we analyze the effect of model size, training objectives, and model architecture on the models' performance as a feature extractor for a set of topologically diverse corpora. We develop a novel metric, the Phonetic-Syntax Ratio (PSR), to measure the phonetic and synthetic information in the extracted representations using deep generalized canonical correlation analysis. Results show the contrastive loss in the wav2vec2.0 objective facilitates more effective cross-lingual feature extraction. There is a positive correlation between PSR scores and ASR performance, suggesting that phonetic information extracted by monolingual SSL models can be used for downstream tasks in cross-lingual settings. The proposed metric is an effective indicator of the quality of the representations and can be useful for model selection.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Automatic speech recognition ; Correlation analysis ; Feature extraction ; Machine learning ; Phonetics ; Representations ; Self-supervised learning</subject><ispartof>arXiv.org, 2023-11</ispartof><rights>2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2894588358?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>776,780,25732,36991,44569</link.rule.ids></links><search><creatorcontrib>Li, Shuyue Stella</creatorcontrib><creatorcontrib>Xu, Beining</creatorcontrib><creatorcontrib>Zhang, Xiangyu</creatorcontrib><creatorcontrib>Liu, Hexin</creatorcontrib><creatorcontrib>Chao, Wenhan</creatorcontrib><creatorcontrib>Leibny Paola Garcia</creatorcontrib><title>A Quantitative Approach to Understand Self-Supervised Models as Cross-lingual Feature Extractors</title><title>arXiv.org</title><description>In this work, we study the features extracted by English self-supervised learning (SSL) models in cross-lingual contexts and propose a new metric to predict the quality of feature representations. Using automatic speech recognition (ASR) as a downstream task, we analyze the effect of model size, training objectives, and model architecture on the models' performance as a feature extractor for a set of topologically diverse corpora. We develop a novel metric, the Phonetic-Syntax Ratio (PSR), to measure the phonetic and synthetic information in the extracted representations using deep generalized canonical correlation analysis. Results show the contrastive loss in the wav2vec2.0 objective facilitates more effective cross-lingual feature extraction. There is a positive correlation between PSR scores and ASR performance, suggesting that phonetic information extracted by monolingual SSL models can be used for downstream tasks in cross-lingual settings. The proposed metric is an effective indicator of the quality of the representations and can be useful for model selection.</description><subject>Automatic speech recognition</subject><subject>Correlation analysis</subject><subject>Feature extraction</subject><subject>Machine learning</subject><subject>Phonetics</subject><subject>Representations</subject><subject>Self-supervised learning</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNjbEKwjAQQIMgKOo_HDgXamI1jkUUFwdR53rYU1tCUnMX8fN18AOc3vAevJ4aamNmmZ1rPVAT5jbPc71Y6qIwQ3Up4ZDQSyMozYug7LoY8PoACXD2NUUW9DUcyd2yY-oovhqmGvahJseADOsYmDPX-HtCB1tCSZFg85aIVwmRx6p_Q8c0-XGkptvNab3Lvp9nIpaqDSn6r6q0Xc0La01hzX_VB_2PRT0</recordid><startdate>20231127</startdate><enddate>20231127</enddate><creator>Li, Shuyue Stella</creator><creator>Xu, Beining</creator><creator>Zhang, Xiangyu</creator><creator>Liu, Hexin</creator><creator>Chao, Wenhan</creator><creator>Leibny Paola Garcia</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20231127</creationdate><title>A Quantitative Approach to Understand Self-Supervised Models as Cross-lingual Feature Extractors</title><author>Li, Shuyue Stella ; Xu, Beining ; Zhang, Xiangyu ; Liu, Hexin ; Chao, Wenhan ; Leibny Paola Garcia</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_28945883583</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Automatic speech recognition</topic><topic>Correlation analysis</topic><topic>Feature extraction</topic><topic>Machine learning</topic><topic>Phonetics</topic><topic>Representations</topic><topic>Self-supervised learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Li, Shuyue Stella</creatorcontrib><creatorcontrib>Xu, Beining</creatorcontrib><creatorcontrib>Zhang, Xiangyu</creatorcontrib><creatorcontrib>Liu, Hexin</creatorcontrib><creatorcontrib>Chao, Wenhan</creatorcontrib><creatorcontrib>Leibny Paola Garcia</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Li, Shuyue Stella</au><au>Xu, Beining</au><au>Zhang, Xiangyu</au><au>Liu, Hexin</au><au>Chao, Wenhan</au><au>Leibny Paola Garcia</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>A Quantitative Approach to Understand Self-Supervised Models as Cross-lingual Feature Extractors</atitle><jtitle>arXiv.org</jtitle><date>2023-11-27</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>In this work, we study the features extracted by English self-supervised learning (SSL) models in cross-lingual contexts and propose a new metric to predict the quality of feature representations. Using automatic speech recognition (ASR) as a downstream task, we analyze the effect of model size, training objectives, and model architecture on the models' performance as a feature extractor for a set of topologically diverse corpora. We develop a novel metric, the Phonetic-Syntax Ratio (PSR), to measure the phonetic and synthetic information in the extracted representations using deep generalized canonical correlation analysis. Results show the contrastive loss in the wav2vec2.0 objective facilitates more effective cross-lingual feature extraction. There is a positive correlation between PSR scores and ASR performance, suggesting that phonetic information extracted by monolingual SSL models can be used for downstream tasks in cross-lingual settings. The proposed metric is an effective indicator of the quality of the representations and can be useful for model selection.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2023-11 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2894588358 |
source | Publicly Available Content Database |
subjects | Automatic speech recognition Correlation analysis Feature extraction Machine learning Phonetics Representations Self-supervised learning |
title | A Quantitative Approach to Understand Self-Supervised Models as Cross-lingual Feature Extractors |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T17%3A04%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=A%20Quantitative%20Approach%20to%20Understand%20Self-Supervised%20Models%20as%20Cross-lingual%20Feature%20Extractors&rft.jtitle=arXiv.org&rft.au=Li,%20Shuyue%20Stella&rft.date=2023-11-27&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2894588358%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_28945883583%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2894588358&rft_id=info:pmid/&rfr_iscdi=true |