Loading…

Latent neural dynamics encode temporal context in speech

•We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cy...

Full description

Saved in:
Bibliographic Details
Published in:Hearing research 2023-09, Vol.437, p.108838-108838, Article 108838
Main Authors: Stephen, Emily P, Li, Yuanning, Metzger, Sean, Oganian, Yulia, Chang, Edward F
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63
cites cdi_FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63
container_end_page 108838
container_issue
container_start_page 108838
container_title Hearing research
container_volume 437
creator Stephen, Emily P
Li, Yuanning
Metzger, Sean
Oganian, Yulia
Chang, Edward F
description •We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cyclical dynamics.•The structure of these latent dynamics could bind phonetic features across time. Direct neural recordings from human auditory cortex have demonstrated encoding for acoustic-phonetic features of consonants and vowels. Neural responses also encode distinct acoustic amplitude cues related to timing, such as those that occur at the onset of a sentence after a silent period or the onset of the vowel in each syllable. Here, we used a group reduced rank regression model to show that distributed cortical responses support a low-dimensional latent state representation of temporal context in speech. The timing cues each capture more unique variance than all other phonetic features and exhibit rotational or cyclical dynamics in latent space from activity that is widespread over the superior temporal gyrus. We propose that these spatially distributed timing signals could serve to provide temporal context for, and possibly bind across time, the concurrent processing of individual phonetic features, to compose higher-order phonological (e.g. word-level) representations.
doi_str_mv 10.1016/j.heares.2023.108838
format article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_11182421</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0378595523001508</els_id><sourcerecordid>2838242848</sourcerecordid><originalsourceid>FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63</originalsourceid><addsrcrecordid>eNp9kEtPwzAQhC0EoqXwDxDKkUuKX0mcCwhVvKRKXOBsOc6GukriYicV_fc4Sqngwskr7-zM6EPokuA5wSS9Wc9XoBz4OcWUhS8hmDhCUyIyESciJ8doitkw50kyQWferzEmCeP0FE1YxjkRAk-RWKoO2i5qoXeqjspdqxqjfQSttiVEHTQbOyy0bTv46iLTRn4DoFfn6KRStYeL_TtD748Pb4vnePn69LK4X8aap7yLdUgqU55RSAWwilGcV4LTjJCMcMaY4JymIEie4KpKmSogqSArdKGyAjJI2Qzdjb6bvmig1KFs6CM3zjTK7aRVRv7dtGYlP-xWEkIE5ZQEh-u9g7OfPfhONsZrqGvVgu29pAFcEAougpSPUu2s9w6qQw7BcqAu13KkLgfqcqQezq5-dzwc_WAOgttRAIHU1oCTXpuAGErjQHeytOb_hG9IypVH</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2838242848</pqid></control><display><type>article</type><title>Latent neural dynamics encode temporal context in speech</title><source>ScienceDirect Journals</source><creator>Stephen, Emily P ; Li, Yuanning ; Metzger, Sean ; Oganian, Yulia ; Chang, Edward F</creator><creatorcontrib>Stephen, Emily P ; Li, Yuanning ; Metzger, Sean ; Oganian, Yulia ; Chang, Edward F</creatorcontrib><description>•We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cyclical dynamics.•The structure of these latent dynamics could bind phonetic features across time. Direct neural recordings from human auditory cortex have demonstrated encoding for acoustic-phonetic features of consonants and vowels. Neural responses also encode distinct acoustic amplitude cues related to timing, such as those that occur at the onset of a sentence after a silent period or the onset of the vowel in each syllable. Here, we used a group reduced rank regression model to show that distributed cortical responses support a low-dimensional latent state representation of temporal context in speech. The timing cues each capture more unique variance than all other phonetic features and exhibit rotational or cyclical dynamics in latent space from activity that is widespread over the superior temporal gyrus. We propose that these spatially distributed timing signals could serve to provide temporal context for, and possibly bind across time, the concurrent processing of individual phonetic features, to compose higher-order phonological (e.g. word-level) representations.</description><identifier>ISSN: 0378-5955</identifier><identifier>EISSN: 1878-5891</identifier><identifier>DOI: 10.1016/j.heares.2023.108838</identifier><identifier>PMID: 37441880</identifier><language>eng</language><publisher>Netherlands: Elsevier B.V</publisher><subject>Acoustic Stimulation ; Auditory ; Auditory Cortex - physiology ; Electrocorticography ; Humans ; Latent state ; Phonetics ; Reduced-rank regression ; Speech - physiology ; Speech Perception - physiology ; Superior temporal gyrus ; Temporal Lobe - physiology</subject><ispartof>Hearing research, 2023-09, Vol.437, p.108838-108838, Article 108838</ispartof><rights>2023</rights><rights>Copyright © 2023. Published by Elsevier B.V.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63</citedby><cites>FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63</cites><orcidid>0000-0003-1978-9622 ; 0000-0003-3277-0600 ; 0000-0002-3451-8938 ; 0000-0003-2379-6358</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,780,784,885,27924,27925</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37441880$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Stephen, Emily P</creatorcontrib><creatorcontrib>Li, Yuanning</creatorcontrib><creatorcontrib>Metzger, Sean</creatorcontrib><creatorcontrib>Oganian, Yulia</creatorcontrib><creatorcontrib>Chang, Edward F</creatorcontrib><title>Latent neural dynamics encode temporal context in speech</title><title>Hearing research</title><addtitle>Hear Res</addtitle><description>•We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cyclical dynamics.•The structure of these latent dynamics could bind phonetic features across time. Direct neural recordings from human auditory cortex have demonstrated encoding for acoustic-phonetic features of consonants and vowels. Neural responses also encode distinct acoustic amplitude cues related to timing, such as those that occur at the onset of a sentence after a silent period or the onset of the vowel in each syllable. Here, we used a group reduced rank regression model to show that distributed cortical responses support a low-dimensional latent state representation of temporal context in speech. The timing cues each capture more unique variance than all other phonetic features and exhibit rotational or cyclical dynamics in latent space from activity that is widespread over the superior temporal gyrus. We propose that these spatially distributed timing signals could serve to provide temporal context for, and possibly bind across time, the concurrent processing of individual phonetic features, to compose higher-order phonological (e.g. word-level) representations.</description><subject>Acoustic Stimulation</subject><subject>Auditory</subject><subject>Auditory Cortex - physiology</subject><subject>Electrocorticography</subject><subject>Humans</subject><subject>Latent state</subject><subject>Phonetics</subject><subject>Reduced-rank regression</subject><subject>Speech - physiology</subject><subject>Speech Perception - physiology</subject><subject>Superior temporal gyrus</subject><subject>Temporal Lobe - physiology</subject><issn>0378-5955</issn><issn>1878-5891</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kEtPwzAQhC0EoqXwDxDKkUuKX0mcCwhVvKRKXOBsOc6GukriYicV_fc4Sqngwskr7-zM6EPokuA5wSS9Wc9XoBz4OcWUhS8hmDhCUyIyESciJ8doitkw50kyQWferzEmCeP0FE1YxjkRAk-RWKoO2i5qoXeqjspdqxqjfQSttiVEHTQbOyy0bTv46iLTRn4DoFfn6KRStYeL_TtD748Pb4vnePn69LK4X8aap7yLdUgqU55RSAWwilGcV4LTjJCMcMaY4JymIEie4KpKmSogqSArdKGyAjJI2Qzdjb6bvmig1KFs6CM3zjTK7aRVRv7dtGYlP-xWEkIE5ZQEh-u9g7OfPfhONsZrqGvVgu29pAFcEAougpSPUu2s9w6qQw7BcqAu13KkLgfqcqQezq5-dzwc_WAOgttRAIHU1oCTXpuAGErjQHeytOb_hG9IypVH</recordid><startdate>20230915</startdate><enddate>20230915</enddate><creator>Stephen, Emily P</creator><creator>Li, Yuanning</creator><creator>Metzger, Sean</creator><creator>Oganian, Yulia</creator><creator>Chang, Edward F</creator><general>Elsevier B.V</general><scope>6I.</scope><scope>AAFTH</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0000-0003-1978-9622</orcidid><orcidid>https://orcid.org/0000-0003-3277-0600</orcidid><orcidid>https://orcid.org/0000-0002-3451-8938</orcidid><orcidid>https://orcid.org/0000-0003-2379-6358</orcidid></search><sort><creationdate>20230915</creationdate><title>Latent neural dynamics encode temporal context in speech</title><author>Stephen, Emily P ; Li, Yuanning ; Metzger, Sean ; Oganian, Yulia ; Chang, Edward F</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Acoustic Stimulation</topic><topic>Auditory</topic><topic>Auditory Cortex - physiology</topic><topic>Electrocorticography</topic><topic>Humans</topic><topic>Latent state</topic><topic>Phonetics</topic><topic>Reduced-rank regression</topic><topic>Speech - physiology</topic><topic>Speech Perception - physiology</topic><topic>Superior temporal gyrus</topic><topic>Temporal Lobe - physiology</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Stephen, Emily P</creatorcontrib><creatorcontrib>Li, Yuanning</creatorcontrib><creatorcontrib>Metzger, Sean</creatorcontrib><creatorcontrib>Oganian, Yulia</creatorcontrib><creatorcontrib>Chang, Edward F</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Hearing research</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Stephen, Emily P</au><au>Li, Yuanning</au><au>Metzger, Sean</au><au>Oganian, Yulia</au><au>Chang, Edward F</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Latent neural dynamics encode temporal context in speech</atitle><jtitle>Hearing research</jtitle><addtitle>Hear Res</addtitle><date>2023-09-15</date><risdate>2023</risdate><volume>437</volume><spage>108838</spage><epage>108838</epage><pages>108838-108838</pages><artnum>108838</artnum><issn>0378-5955</issn><eissn>1878-5891</eissn><abstract>•We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cyclical dynamics.•The structure of these latent dynamics could bind phonetic features across time. Direct neural recordings from human auditory cortex have demonstrated encoding for acoustic-phonetic features of consonants and vowels. Neural responses also encode distinct acoustic amplitude cues related to timing, such as those that occur at the onset of a sentence after a silent period or the onset of the vowel in each syllable. Here, we used a group reduced rank regression model to show that distributed cortical responses support a low-dimensional latent state representation of temporal context in speech. The timing cues each capture more unique variance than all other phonetic features and exhibit rotational or cyclical dynamics in latent space from activity that is widespread over the superior temporal gyrus. We propose that these spatially distributed timing signals could serve to provide temporal context for, and possibly bind across time, the concurrent processing of individual phonetic features, to compose higher-order phonological (e.g. word-level) representations.</abstract><cop>Netherlands</cop><pub>Elsevier B.V</pub><pmid>37441880</pmid><doi>10.1016/j.heares.2023.108838</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0003-1978-9622</orcidid><orcidid>https://orcid.org/0000-0003-3277-0600</orcidid><orcidid>https://orcid.org/0000-0002-3451-8938</orcidid><orcidid>https://orcid.org/0000-0003-2379-6358</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0378-5955
ispartof Hearing research, 2023-09, Vol.437, p.108838-108838, Article 108838
issn 0378-5955
1878-5891
language eng
recordid cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_11182421
source ScienceDirect Journals
subjects Acoustic Stimulation
Auditory
Auditory Cortex - physiology
Electrocorticography
Humans
Latent state
Phonetics
Reduced-rank regression
Speech - physiology
Speech Perception - physiology
Superior temporal gyrus
Temporal Lobe - physiology
title Latent neural dynamics encode temporal context in speech
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T23%3A08%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Latent%20neural%20dynamics%20encode%20temporal%20context%20in%20speech&rft.jtitle=Hearing%20research&rft.au=Stephen,%20Emily%20P&rft.date=2023-09-15&rft.volume=437&rft.spage=108838&rft.epage=108838&rft.pages=108838-108838&rft.artnum=108838&rft.issn=0378-5955&rft.eissn=1878-5891&rft_id=info:doi/10.1016/j.heares.2023.108838&rft_dat=%3Cproquest_pubme%3E2838242848%3C/proquest_pubme%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2838242848&rft_id=info:pmid/37441880&rfr_iscdi=true