Loading…

The use of Voice Source Features for Sung Speech Recognition

In this paper, we ask whether vocal source features (pitch, shimmer, jitter, etc) can improve the performance of automatic sung speech recognition, arguing that conclusions previously drawn from spoken speech studies may not be valid in the sung speech domain. We first use a parallel singing/speakin...

Full description

Saved in:
Bibliographic Details
Main Authors: Dabike, Gerardo Roa, Barker, Jon
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 6517
container_issue
container_start_page 6513
container_title
container_volume
creator Dabike, Gerardo Roa
Barker, Jon
description In this paper, we ask whether vocal source features (pitch, shimmer, jitter, etc) can improve the performance of automatic sung speech recognition, arguing that conclusions previously drawn from spoken speech studies may not be valid in the sung speech domain. We first use a parallel singing/speaking corpus (NUS-48E) to illustrate differences in sung vs spoken voicing characteristics including pitch range, syllables duration, vibrato, jitter and shimmer. We then use this analysis to inform speech recognition experiments on the sung speech DSing corpus, using a state of the art acoustic model and augmenting conventional features with various voice source parameters. Experiments are run with three standard (increasingly large) training sets, DSing1 (15.1 hours), DSing3 (44.7 hours) and DS-ing30 (149.1 hours). Pitch combined with degree of voicing produces a significant decrease in WER from 38.1% to 36.7% when training with DSing1 however smaller decreases in WER observed when training with the larger more varied DSing3 and DSing30 sets were not seen to be statistically significant. Voicing quality characteristics did not improve recognition performance although analysis suggests that they do contribute to an improved discrimination between voiced/unvoiced phoneme pairs.
doi_str_mv 10.1109/ICASSP39728.2021.9414950
format conference_proceeding
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_9414950</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9414950</ieee_id><sourcerecordid>9414950</sourcerecordid><originalsourceid>FETCH-LOGICAL-i203t-89b081eac147f71614a7ff1b48ae8de1c9ebd681e65ac52895c00e176edc4cbb3</originalsourceid><addsrcrecordid>eNotj81KAzEUhaNQsLZ9Ajd5gRnvzSSTBNxIsSoUFKeKu5LJ3LQRnZT5Wfj2DtjVB4fD4TuMcYQcEezt8_q-ql4Lq4XJBQjMrURpFVywldUGpxh1CUpdsrkotM3QwucVu-77LwAwWpo5u9sdiY898RT4R4qeeJXGbsKG3DB21POQOl6N7YFXJyJ_5G_k06GNQ0ztks2C--5pdeaCvW8eduunbPvyOLltsyigGDJjazBIzqPUQWOJ0ukQsJbGkWkIvaW6KadGqZxXwljlAWhSp8ZLX9fFgt3870Yi2p-6-OO63_35bPEH_Q5Jzw</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>The use of Voice Source Features for Sung Speech Recognition</title><source>IEEE Xplore All Conference Series</source><creator>Dabike, Gerardo Roa ; Barker, Jon</creator><creatorcontrib>Dabike, Gerardo Roa ; Barker, Jon</creatorcontrib><description>In this paper, we ask whether vocal source features (pitch, shimmer, jitter, etc) can improve the performance of automatic sung speech recognition, arguing that conclusions previously drawn from spoken speech studies may not be valid in the sung speech domain. We first use a parallel singing/speaking corpus (NUS-48E) to illustrate differences in sung vs spoken voicing characteristics including pitch range, syllables duration, vibrato, jitter and shimmer. We then use this analysis to inform speech recognition experiments on the sung speech DSing corpus, using a state of the art acoustic model and augmenting conventional features with various voice source parameters. Experiments are run with three standard (increasingly large) training sets, DSing1 (15.1 hours), DSing3 (44.7 hours) and DS-ing30 (149.1 hours). Pitch combined with degree of voicing produces a significant decrease in WER from 38.1% to 36.7% when training with DSing1 however smaller decreases in WER observed when training with the larger more varied DSing3 and DSing30 sets were not seen to be statistically significant. Voicing quality characteristics did not improve recognition performance although analysis suggests that they do contribute to an improved discrimination between voiced/unvoiced phoneme pairs.</description><identifier>EISSN: 2379-190X</identifier><identifier>EISBN: 9781728176055</identifier><identifier>EISBN: 1728176050</identifier><identifier>DOI: 10.1109/ICASSP39728.2021.9414950</identifier><language>eng</language><publisher>IEEE</publisher><subject>Acoustics ; Jitter ; Phonetics ; Signal processing ; Speech recognition ; Sung speech ; Training ; Training data ; voice source</subject><ispartof>ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021, p.6513-6517</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9414950$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,27925,54555,54932</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9414950$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Dabike, Gerardo Roa</creatorcontrib><creatorcontrib>Barker, Jon</creatorcontrib><title>The use of Voice Source Features for Sung Speech Recognition</title><title>ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</title><addtitle>ICASSP</addtitle><description>In this paper, we ask whether vocal source features (pitch, shimmer, jitter, etc) can improve the performance of automatic sung speech recognition, arguing that conclusions previously drawn from spoken speech studies may not be valid in the sung speech domain. We first use a parallel singing/speaking corpus (NUS-48E) to illustrate differences in sung vs spoken voicing characteristics including pitch range, syllables duration, vibrato, jitter and shimmer. We then use this analysis to inform speech recognition experiments on the sung speech DSing corpus, using a state of the art acoustic model and augmenting conventional features with various voice source parameters. Experiments are run with three standard (increasingly large) training sets, DSing1 (15.1 hours), DSing3 (44.7 hours) and DS-ing30 (149.1 hours). Pitch combined with degree of voicing produces a significant decrease in WER from 38.1% to 36.7% when training with DSing1 however smaller decreases in WER observed when training with the larger more varied DSing3 and DSing30 sets were not seen to be statistically significant. Voicing quality characteristics did not improve recognition performance although analysis suggests that they do contribute to an improved discrimination between voiced/unvoiced phoneme pairs.</description><subject>Acoustics</subject><subject>Jitter</subject><subject>Phonetics</subject><subject>Signal processing</subject><subject>Speech recognition</subject><subject>Sung speech</subject><subject>Training</subject><subject>Training data</subject><subject>voice source</subject><issn>2379-190X</issn><isbn>9781728176055</isbn><isbn>1728176050</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2021</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNotj81KAzEUhaNQsLZ9Ajd5gRnvzSSTBNxIsSoUFKeKu5LJ3LQRnZT5Wfj2DtjVB4fD4TuMcYQcEezt8_q-ql4Lq4XJBQjMrURpFVywldUGpxh1CUpdsrkotM3QwucVu-77LwAwWpo5u9sdiY898RT4R4qeeJXGbsKG3DB21POQOl6N7YFXJyJ_5G_k06GNQ0ztks2C--5pdeaCvW8eduunbPvyOLltsyigGDJjazBIzqPUQWOJ0ukQsJbGkWkIvaW6KadGqZxXwljlAWhSp8ZLX9fFgt3870Yi2p-6-OO63_35bPEH_Q5Jzw</recordid><startdate>20210606</startdate><enddate>20210606</enddate><creator>Dabike, Gerardo Roa</creator><creator>Barker, Jon</creator><general>IEEE</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>20210606</creationdate><title>The use of Voice Source Features for Sung Speech Recognition</title><author>Dabike, Gerardo Roa ; Barker, Jon</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i203t-89b081eac147f71614a7ff1b48ae8de1c9ebd681e65ac52895c00e176edc4cbb3</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Acoustics</topic><topic>Jitter</topic><topic>Phonetics</topic><topic>Signal processing</topic><topic>Speech recognition</topic><topic>Sung speech</topic><topic>Training</topic><topic>Training data</topic><topic>voice source</topic><toplevel>online_resources</toplevel><creatorcontrib>Dabike, Gerardo Roa</creatorcontrib><creatorcontrib>Barker, Jon</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Xplore</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Dabike, Gerardo Roa</au><au>Barker, Jon</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>The use of Voice Source Features for Sung Speech Recognition</atitle><btitle>ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</btitle><stitle>ICASSP</stitle><date>2021-06-06</date><risdate>2021</risdate><spage>6513</spage><epage>6517</epage><pages>6513-6517</pages><eissn>2379-190X</eissn><eisbn>9781728176055</eisbn><eisbn>1728176050</eisbn><abstract>In this paper, we ask whether vocal source features (pitch, shimmer, jitter, etc) can improve the performance of automatic sung speech recognition, arguing that conclusions previously drawn from spoken speech studies may not be valid in the sung speech domain. We first use a parallel singing/speaking corpus (NUS-48E) to illustrate differences in sung vs spoken voicing characteristics including pitch range, syllables duration, vibrato, jitter and shimmer. We then use this analysis to inform speech recognition experiments on the sung speech DSing corpus, using a state of the art acoustic model and augmenting conventional features with various voice source parameters. Experiments are run with three standard (increasingly large) training sets, DSing1 (15.1 hours), DSing3 (44.7 hours) and DS-ing30 (149.1 hours). Pitch combined with degree of voicing produces a significant decrease in WER from 38.1% to 36.7% when training with DSing1 however smaller decreases in WER observed when training with the larger more varied DSing3 and DSing30 sets were not seen to be statistically significant. Voicing quality characteristics did not improve recognition performance although analysis suggests that they do contribute to an improved discrimination between voiced/unvoiced phoneme pairs.</abstract><pub>IEEE</pub><doi>10.1109/ICASSP39728.2021.9414950</doi><tpages>5</tpages></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2379-190X
ispartof ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021, p.6513-6517
issn 2379-190X
language eng
recordid cdi_ieee_primary_9414950
source IEEE Xplore All Conference Series
subjects Acoustics
Jitter
Phonetics
Signal processing
Speech recognition
Sung speech
Training
Training data
voice source
title The use of Voice Source Features for Sung Speech Recognition
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T12%3A44%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=The%20use%20of%20Voice%20Source%20Features%20for%20Sung%20Speech%20Recognition&rft.btitle=ICASSP%202021%20-%202021%20IEEE%20International%20Conference%20on%20Acoustics,%20Speech%20and%20Signal%20Processing%20(ICASSP)&rft.au=Dabike,%20Gerardo%20Roa&rft.date=2021-06-06&rft.spage=6513&rft.epage=6517&rft.pages=6513-6517&rft.eissn=2379-190X&rft_id=info:doi/10.1109/ICASSP39728.2021.9414950&rft.eisbn=9781728176055&rft.eisbn_list=1728176050&rft_dat=%3Cieee_CHZPO%3E9414950%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i203t-89b081eac147f71614a7ff1b48ae8de1c9ebd681e65ac52895c00e176edc4cbb3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=9414950&rfr_iscdi=true