Loading…
Latent neural dynamics encode temporal context in speech
•We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cy...
Saved in:
Published in: | Hearing research 2023-09, Vol.437, p.108838-108838, Article 108838 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63 |
---|---|
cites | cdi_FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63 |
container_end_page | 108838 |
container_issue | |
container_start_page | 108838 |
container_title | Hearing research |
container_volume | 437 |
creator | Stephen, Emily P Li, Yuanning Metzger, Sean Oganian, Yulia Chang, Edward F |
description | •We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cyclical dynamics.•The structure of these latent dynamics could bind phonetic features across time.
Direct neural recordings from human auditory cortex have demonstrated encoding for acoustic-phonetic features of consonants and vowels. Neural responses also encode distinct acoustic amplitude cues related to timing, such as those that occur at the onset of a sentence after a silent period or the onset of the vowel in each syllable. Here, we used a group reduced rank regression model to show that distributed cortical responses support a low-dimensional latent state representation of temporal context in speech. The timing cues each capture more unique variance than all other phonetic features and exhibit rotational or cyclical dynamics in latent space from activity that is widespread over the superior temporal gyrus. We propose that these spatially distributed timing signals could serve to provide temporal context for, and possibly bind across time, the concurrent processing of individual phonetic features, to compose higher-order phonological (e.g. word-level) representations. |
doi_str_mv | 10.1016/j.heares.2023.108838 |
format | article |
fullrecord | <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_11182421</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0378595523001508</els_id><sourcerecordid>2838242848</sourcerecordid><originalsourceid>FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63</originalsourceid><addsrcrecordid>eNp9kEtPwzAQhC0EoqXwDxDKkUuKX0mcCwhVvKRKXOBsOc6GukriYicV_fc4Sqngwskr7-zM6EPokuA5wSS9Wc9XoBz4OcWUhS8hmDhCUyIyESciJ8doitkw50kyQWferzEmCeP0FE1YxjkRAk-RWKoO2i5qoXeqjspdqxqjfQSttiVEHTQbOyy0bTv46iLTRn4DoFfn6KRStYeL_TtD748Pb4vnePn69LK4X8aap7yLdUgqU55RSAWwilGcV4LTjJCMcMaY4JymIEie4KpKmSogqSArdKGyAjJI2Qzdjb6bvmig1KFs6CM3zjTK7aRVRv7dtGYlP-xWEkIE5ZQEh-u9g7OfPfhONsZrqGvVgu29pAFcEAougpSPUu2s9w6qQw7BcqAu13KkLgfqcqQezq5-dzwc_WAOgttRAIHU1oCTXpuAGErjQHeytOb_hG9IypVH</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2838242848</pqid></control><display><type>article</type><title>Latent neural dynamics encode temporal context in speech</title><source>ScienceDirect Journals</source><creator>Stephen, Emily P ; Li, Yuanning ; Metzger, Sean ; Oganian, Yulia ; Chang, Edward F</creator><creatorcontrib>Stephen, Emily P ; Li, Yuanning ; Metzger, Sean ; Oganian, Yulia ; Chang, Edward F</creatorcontrib><description>•We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cyclical dynamics.•The structure of these latent dynamics could bind phonetic features across time.
Direct neural recordings from human auditory cortex have demonstrated encoding for acoustic-phonetic features of consonants and vowels. Neural responses also encode distinct acoustic amplitude cues related to timing, such as those that occur at the onset of a sentence after a silent period or the onset of the vowel in each syllable. Here, we used a group reduced rank regression model to show that distributed cortical responses support a low-dimensional latent state representation of temporal context in speech. The timing cues each capture more unique variance than all other phonetic features and exhibit rotational or cyclical dynamics in latent space from activity that is widespread over the superior temporal gyrus. We propose that these spatially distributed timing signals could serve to provide temporal context for, and possibly bind across time, the concurrent processing of individual phonetic features, to compose higher-order phonological (e.g. word-level) representations.</description><identifier>ISSN: 0378-5955</identifier><identifier>EISSN: 1878-5891</identifier><identifier>DOI: 10.1016/j.heares.2023.108838</identifier><identifier>PMID: 37441880</identifier><language>eng</language><publisher>Netherlands: Elsevier B.V</publisher><subject>Acoustic Stimulation ; Auditory ; Auditory Cortex - physiology ; Electrocorticography ; Humans ; Latent state ; Phonetics ; Reduced-rank regression ; Speech - physiology ; Speech Perception - physiology ; Superior temporal gyrus ; Temporal Lobe - physiology</subject><ispartof>Hearing research, 2023-09, Vol.437, p.108838-108838, Article 108838</ispartof><rights>2023</rights><rights>Copyright © 2023. Published by Elsevier B.V.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63</citedby><cites>FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63</cites><orcidid>0000-0003-1978-9622 ; 0000-0003-3277-0600 ; 0000-0002-3451-8938 ; 0000-0003-2379-6358</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,780,784,885,27924,27925</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37441880$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Stephen, Emily P</creatorcontrib><creatorcontrib>Li, Yuanning</creatorcontrib><creatorcontrib>Metzger, Sean</creatorcontrib><creatorcontrib>Oganian, Yulia</creatorcontrib><creatorcontrib>Chang, Edward F</creatorcontrib><title>Latent neural dynamics encode temporal context in speech</title><title>Hearing research</title><addtitle>Hear Res</addtitle><description>•We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cyclical dynamics.•The structure of these latent dynamics could bind phonetic features across time.
Direct neural recordings from human auditory cortex have demonstrated encoding for acoustic-phonetic features of consonants and vowels. Neural responses also encode distinct acoustic amplitude cues related to timing, such as those that occur at the onset of a sentence after a silent period or the onset of the vowel in each syllable. Here, we used a group reduced rank regression model to show that distributed cortical responses support a low-dimensional latent state representation of temporal context in speech. The timing cues each capture more unique variance than all other phonetic features and exhibit rotational or cyclical dynamics in latent space from activity that is widespread over the superior temporal gyrus. We propose that these spatially distributed timing signals could serve to provide temporal context for, and possibly bind across time, the concurrent processing of individual phonetic features, to compose higher-order phonological (e.g. word-level) representations.</description><subject>Acoustic Stimulation</subject><subject>Auditory</subject><subject>Auditory Cortex - physiology</subject><subject>Electrocorticography</subject><subject>Humans</subject><subject>Latent state</subject><subject>Phonetics</subject><subject>Reduced-rank regression</subject><subject>Speech - physiology</subject><subject>Speech Perception - physiology</subject><subject>Superior temporal gyrus</subject><subject>Temporal Lobe - physiology</subject><issn>0378-5955</issn><issn>1878-5891</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNp9kEtPwzAQhC0EoqXwDxDKkUuKX0mcCwhVvKRKXOBsOc6GukriYicV_fc4Sqngwskr7-zM6EPokuA5wSS9Wc9XoBz4OcWUhS8hmDhCUyIyESciJ8doitkw50kyQWferzEmCeP0FE1YxjkRAk-RWKoO2i5qoXeqjspdqxqjfQSttiVEHTQbOyy0bTv46iLTRn4DoFfn6KRStYeL_TtD748Pb4vnePn69LK4X8aap7yLdUgqU55RSAWwilGcV4LTjJCMcMaY4JymIEie4KpKmSogqSArdKGyAjJI2Qzdjb6bvmig1KFs6CM3zjTK7aRVRv7dtGYlP-xWEkIE5ZQEh-u9g7OfPfhONsZrqGvVgu29pAFcEAougpSPUu2s9w6qQw7BcqAu13KkLgfqcqQezq5-dzwc_WAOgttRAIHU1oCTXpuAGErjQHeytOb_hG9IypVH</recordid><startdate>20230915</startdate><enddate>20230915</enddate><creator>Stephen, Emily P</creator><creator>Li, Yuanning</creator><creator>Metzger, Sean</creator><creator>Oganian, Yulia</creator><creator>Chang, Edward F</creator><general>Elsevier B.V</general><scope>6I.</scope><scope>AAFTH</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0000-0003-1978-9622</orcidid><orcidid>https://orcid.org/0000-0003-3277-0600</orcidid><orcidid>https://orcid.org/0000-0002-3451-8938</orcidid><orcidid>https://orcid.org/0000-0003-2379-6358</orcidid></search><sort><creationdate>20230915</creationdate><title>Latent neural dynamics encode temporal context in speech</title><author>Stephen, Emily P ; Li, Yuanning ; Metzger, Sean ; Oganian, Yulia ; Chang, Edward F</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Acoustic Stimulation</topic><topic>Auditory</topic><topic>Auditory Cortex - physiology</topic><topic>Electrocorticography</topic><topic>Humans</topic><topic>Latent state</topic><topic>Phonetics</topic><topic>Reduced-rank regression</topic><topic>Speech - physiology</topic><topic>Speech Perception - physiology</topic><topic>Superior temporal gyrus</topic><topic>Temporal Lobe - physiology</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Stephen, Emily P</creatorcontrib><creatorcontrib>Li, Yuanning</creatorcontrib><creatorcontrib>Metzger, Sean</creatorcontrib><creatorcontrib>Oganian, Yulia</creatorcontrib><creatorcontrib>Chang, Edward F</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Hearing research</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Stephen, Emily P</au><au>Li, Yuanning</au><au>Metzger, Sean</au><au>Oganian, Yulia</au><au>Chang, Edward F</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Latent neural dynamics encode temporal context in speech</atitle><jtitle>Hearing research</jtitle><addtitle>Hear Res</addtitle><date>2023-09-15</date><risdate>2023</risdate><volume>437</volume><spage>108838</spage><epage>108838</epage><pages>108838-108838</pages><artnum>108838</artnum><issn>0378-5955</issn><eissn>1878-5891</eissn><abstract>•We recorded auditory neural responses to speech using electrocorticography.•Reduced-rank regression captures responses with low-dimensional latent states.•Responses to timing cues are more widespread than phonetic feature responses.•Responses to sentence-level and syllable-level timing cues have cyclical dynamics.•The structure of these latent dynamics could bind phonetic features across time.
Direct neural recordings from human auditory cortex have demonstrated encoding for acoustic-phonetic features of consonants and vowels. Neural responses also encode distinct acoustic amplitude cues related to timing, such as those that occur at the onset of a sentence after a silent period or the onset of the vowel in each syllable. Here, we used a group reduced rank regression model to show that distributed cortical responses support a low-dimensional latent state representation of temporal context in speech. The timing cues each capture more unique variance than all other phonetic features and exhibit rotational or cyclical dynamics in latent space from activity that is widespread over the superior temporal gyrus. We propose that these spatially distributed timing signals could serve to provide temporal context for, and possibly bind across time, the concurrent processing of individual phonetic features, to compose higher-order phonological (e.g. word-level) representations.</abstract><cop>Netherlands</cop><pub>Elsevier B.V</pub><pmid>37441880</pmid><doi>10.1016/j.heares.2023.108838</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0003-1978-9622</orcidid><orcidid>https://orcid.org/0000-0003-3277-0600</orcidid><orcidid>https://orcid.org/0000-0002-3451-8938</orcidid><orcidid>https://orcid.org/0000-0003-2379-6358</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0378-5955 |
ispartof | Hearing research, 2023-09, Vol.437, p.108838-108838, Article 108838 |
issn | 0378-5955 1878-5891 |
language | eng |
recordid | cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_11182421 |
source | ScienceDirect Journals |
subjects | Acoustic Stimulation Auditory Auditory Cortex - physiology Electrocorticography Humans Latent state Phonetics Reduced-rank regression Speech - physiology Speech Perception - physiology Superior temporal gyrus Temporal Lobe - physiology |
title | Latent neural dynamics encode temporal context in speech |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T23%3A08%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Latent%20neural%20dynamics%20encode%20temporal%20context%20in%20speech&rft.jtitle=Hearing%20research&rft.au=Stephen,%20Emily%20P&rft.date=2023-09-15&rft.volume=437&rft.spage=108838&rft.epage=108838&rft.pages=108838-108838&rft.artnum=108838&rft.issn=0378-5955&rft.eissn=1878-5891&rft_id=info:doi/10.1016/j.heares.2023.108838&rft_dat=%3Cproquest_pubme%3E2838242848%3C/proquest_pubme%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c464t-c744d6472e68e3f3209f84271171433384426e81950ff63abe5fe7bcba7be7e63%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2838242848&rft_id=info:pmid/37441880&rfr_iscdi=true |