Loading…

Symbolic gestures and spoken language are processed by a common neural system

Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and func...

Full description

Saved in:
Bibliographic Details
Published in:Proceedings of the National Academy of Sciences - PNAS 2009-12, Vol.106 (49), p.20664-20669
Main Authors: Xu, Jiang, Gannon, Patrick J, Emmorey, Karen, Smith, Jason F, Braun, Allen R
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563
cites cdi_FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563
container_end_page 20669
container_issue 49
container_start_page 20664
container_title Proceedings of the National Academy of Sciences - PNAS
container_volume 106
creator Xu, Jiang
Gannon, Patrick J
Emmorey, Karen
Smith, Jason F
Braun, Allen R
description Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and function as complete utterances in their own right. The relationship between these gestures and spoken language remains unclear. We used functional MRI to investigate whether these two forms of communication are processed by the same system in the human brain. Responses to symbolic gestures, to their spoken glosses (expressing the gestures' meaning in English), and to visually and acoustically matched control stimuli were compared in a randomized block design. General Linear Models (GLM) contrasts identified shared and unique activations and functional connectivity analyses delineated regional interactions associated with each condition. Results support a model in which bilateral modality-specific areas in superior and inferior temporal cortices extract salient features from vocal-auditory and gestural-visual stimuli respectively. However, both classes of stimuli activate a common, left-lateralized network of inferior frontal and posterior temporal regions in which symbolic gestures and spoken words may be mapped onto common, corresponding conceptual representations. We suggest that these anterior and posterior perisylvian areas, identified since the mid-19th century as the core of the brain's language system, are not in fact committed to language processing, but may function as a modality-independent semiotic system that plays a broader role in human communication, linking meaning with symbols whether these are words, gestures, images, sounds, or objects.
doi_str_mv 10.1073/pnas.0909197106
format article
fullrecord <record><control><sourceid>jstor_proqu</sourceid><recordid>TN_cdi_proquest_miscellaneous_748961467</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><jstor_id>40536054</jstor_id><sourcerecordid>40536054</sourcerecordid><originalsourceid>FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563</originalsourceid><addsrcrecordid>eNqFkc1v1DAQxS0EokvhzAmwuCAOacfxV3xBQhVfUhGH0rPlZCchS2IHO0Hsf4-jXXWBCydbmt-8eU-PkKcMLhhofjl5ly7AgGFGM1D3yIblf6GEgftkA1DqohKlOCOPUtoBgJEVPCRnzJiSC6425PPNfqzD0De0wzQvERN1fkvTFL6jp4Pz3eI6pC4inWJoMCXc0npPHW3COAZPPS7RDTTt04zjY_KgdUPCJ8f3nNy-f_f16mNx_eXDp6u310UjSzEXQoETUjoltJCVUiVIx0RdZk91zeoWtg22rK252GKVV5AxZ4CjbrlopFT8nLw56E5LPWKm_ZxN2Cn2o4t7G1xv_574_pvtwk9bam1K4Fng1VEghh9LTm7HPjU45MAYlmS1qIxiQulMvvyH3IUl-pzOlsAESKNFhi4PUBNDShHbOysM7FqUXYuyp6LyxvM_E5z4YzMZoEdg3TzJKStMvqzUevX1fxDbLsMw4685s88O7C7NId7B2T9XIFetF4d564J1XeyTvb3JATkwzTIi-G_kJbpq</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>201405974</pqid></control><display><type>article</type><title>Symbolic gestures and spoken language are processed by a common neural system</title><source>JSTOR Archival Journals and Primary Sources Collection</source><source>PubMed Central</source><creator>Xu, Jiang ; Gannon, Patrick J ; Emmorey, Karen ; Smith, Jason F ; Braun, Allen R</creator><creatorcontrib>Xu, Jiang ; Gannon, Patrick J ; Emmorey, Karen ; Smith, Jason F ; Braun, Allen R</creatorcontrib><description>Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and function as complete utterances in their own right. The relationship between these gestures and spoken language remains unclear. We used functional MRI to investigate whether these two forms of communication are processed by the same system in the human brain. Responses to symbolic gestures, to their spoken glosses (expressing the gestures' meaning in English), and to visually and acoustically matched control stimuli were compared in a randomized block design. General Linear Models (GLM) contrasts identified shared and unique activations and functional connectivity analyses delineated regional interactions associated with each condition. Results support a model in which bilateral modality-specific areas in superior and inferior temporal cortices extract salient features from vocal-auditory and gestural-visual stimuli respectively. However, both classes of stimuli activate a common, left-lateralized network of inferior frontal and posterior temporal regions in which symbolic gestures and spoken words may be mapped onto common, corresponding conceptual representations. We suggest that these anterior and posterior perisylvian areas, identified since the mid-19th century as the core of the brain's language system, are not in fact committed to language processing, but may function as a modality-independent semiotic system that plays a broader role in human communication, linking meaning with symbols whether these are words, gestures, images, sounds, or objects.</description><identifier>ISSN: 0027-8424</identifier><identifier>EISSN: 1091-6490</identifier><identifier>DOI: 10.1073/pnas.0909197106</identifier><identifier>PMID: 19923436</identifier><language>eng</language><publisher>United States: National Academy of Sciences</publisher><subject>American sign language ; Brain ; Brain Mapping ; Cerebral hemispheres ; Connectivity ; Female ; Gestures ; Human subjects ; Humans ; Language ; Language comprehension ; Magnetic resonance imaging ; Mental Processes ; Nerve Net - physiology ; Nonverbal communication ; Pantomime ; Semantics ; Social Sciences ; Symbolic communication ; Symbolism ; Verbal communication</subject><ispartof>Proceedings of the National Academy of Sciences - PNAS, 2009-12, Vol.106 (49), p.20664-20669</ispartof><rights>Copyright National Academy of Sciences Dec 8, 2009</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563</citedby><cites>FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Uhttp://www.pnas.org/content/106/49.cover.gif</thumbnail><linktopdf>$$Uhttps://www.jstor.org/stable/pdf/40536054$$EPDF$$P50$$Gjstor$$H</linktopdf><linktohtml>$$Uhttps://www.jstor.org/stable/40536054$$EHTML$$P50$$Gjstor$$H</linktohtml><link.rule.ids>230,314,723,776,780,881,27903,27904,53770,53772,58217,58450</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/19923436$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Xu, Jiang</creatorcontrib><creatorcontrib>Gannon, Patrick J</creatorcontrib><creatorcontrib>Emmorey, Karen</creatorcontrib><creatorcontrib>Smith, Jason F</creatorcontrib><creatorcontrib>Braun, Allen R</creatorcontrib><title>Symbolic gestures and spoken language are processed by a common neural system</title><title>Proceedings of the National Academy of Sciences - PNAS</title><addtitle>Proc Natl Acad Sci U S A</addtitle><description>Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and function as complete utterances in their own right. The relationship between these gestures and spoken language remains unclear. We used functional MRI to investigate whether these two forms of communication are processed by the same system in the human brain. Responses to symbolic gestures, to their spoken glosses (expressing the gestures' meaning in English), and to visually and acoustically matched control stimuli were compared in a randomized block design. General Linear Models (GLM) contrasts identified shared and unique activations and functional connectivity analyses delineated regional interactions associated with each condition. Results support a model in which bilateral modality-specific areas in superior and inferior temporal cortices extract salient features from vocal-auditory and gestural-visual stimuli respectively. However, both classes of stimuli activate a common, left-lateralized network of inferior frontal and posterior temporal regions in which symbolic gestures and spoken words may be mapped onto common, corresponding conceptual representations. We suggest that these anterior and posterior perisylvian areas, identified since the mid-19th century as the core of the brain's language system, are not in fact committed to language processing, but may function as a modality-independent semiotic system that plays a broader role in human communication, linking meaning with symbols whether these are words, gestures, images, sounds, or objects.</description><subject>American sign language</subject><subject>Brain</subject><subject>Brain Mapping</subject><subject>Cerebral hemispheres</subject><subject>Connectivity</subject><subject>Female</subject><subject>Gestures</subject><subject>Human subjects</subject><subject>Humans</subject><subject>Language</subject><subject>Language comprehension</subject><subject>Magnetic resonance imaging</subject><subject>Mental Processes</subject><subject>Nerve Net - physiology</subject><subject>Nonverbal communication</subject><subject>Pantomime</subject><subject>Semantics</subject><subject>Social Sciences</subject><subject>Symbolic communication</subject><subject>Symbolism</subject><subject>Verbal communication</subject><issn>0027-8424</issn><issn>1091-6490</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2009</creationdate><recordtype>article</recordtype><recordid>eNqFkc1v1DAQxS0EokvhzAmwuCAOacfxV3xBQhVfUhGH0rPlZCchS2IHO0Hsf4-jXXWBCydbmt-8eU-PkKcMLhhofjl5ly7AgGFGM1D3yIblf6GEgftkA1DqohKlOCOPUtoBgJEVPCRnzJiSC6425PPNfqzD0De0wzQvERN1fkvTFL6jp4Pz3eI6pC4inWJoMCXc0npPHW3COAZPPS7RDTTt04zjY_KgdUPCJ8f3nNy-f_f16mNx_eXDp6u310UjSzEXQoETUjoltJCVUiVIx0RdZk91zeoWtg22rK252GKVV5AxZ4CjbrlopFT8nLw56E5LPWKm_ZxN2Cn2o4t7G1xv_574_pvtwk9bam1K4Fng1VEghh9LTm7HPjU45MAYlmS1qIxiQulMvvyH3IUl-pzOlsAESKNFhi4PUBNDShHbOysM7FqUXYuyp6LyxvM_E5z4YzMZoEdg3TzJKStMvqzUevX1fxDbLsMw4685s88O7C7NId7B2T9XIFetF4d564J1XeyTvb3JATkwzTIi-G_kJbpq</recordid><startdate>20091208</startdate><enddate>20091208</enddate><creator>Xu, Jiang</creator><creator>Gannon, Patrick J</creator><creator>Emmorey, Karen</creator><creator>Smith, Jason F</creator><creator>Braun, Allen R</creator><general>National Academy of Sciences</general><general>National Acad Sciences</general><scope>FBQ</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QG</scope><scope>7QL</scope><scope>7QP</scope><scope>7QR</scope><scope>7SN</scope><scope>7SS</scope><scope>7T5</scope><scope>7TK</scope><scope>7TM</scope><scope>7TO</scope><scope>7U9</scope><scope>8FD</scope><scope>C1K</scope><scope>FR3</scope><scope>H94</scope><scope>M7N</scope><scope>P64</scope><scope>RC3</scope><scope>7X8</scope><scope>5PM</scope></search><sort><creationdate>20091208</creationdate><title>Symbolic gestures and spoken language are processed by a common neural system</title><author>Xu, Jiang ; Gannon, Patrick J ; Emmorey, Karen ; Smith, Jason F ; Braun, Allen R</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2009</creationdate><topic>American sign language</topic><topic>Brain</topic><topic>Brain Mapping</topic><topic>Cerebral hemispheres</topic><topic>Connectivity</topic><topic>Female</topic><topic>Gestures</topic><topic>Human subjects</topic><topic>Humans</topic><topic>Language</topic><topic>Language comprehension</topic><topic>Magnetic resonance imaging</topic><topic>Mental Processes</topic><topic>Nerve Net - physiology</topic><topic>Nonverbal communication</topic><topic>Pantomime</topic><topic>Semantics</topic><topic>Social Sciences</topic><topic>Symbolic communication</topic><topic>Symbolism</topic><topic>Verbal communication</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Xu, Jiang</creatorcontrib><creatorcontrib>Gannon, Patrick J</creatorcontrib><creatorcontrib>Emmorey, Karen</creatorcontrib><creatorcontrib>Smith, Jason F</creatorcontrib><creatorcontrib>Braun, Allen R</creatorcontrib><collection>AGRIS</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Animal Behavior Abstracts</collection><collection>Bacteriology Abstracts (Microbiology B)</collection><collection>Calcium &amp; Calcified Tissue Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Ecology Abstracts</collection><collection>Entomology Abstracts (Full archive)</collection><collection>Immunology Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Nucleic Acids Abstracts</collection><collection>Oncogenes and Growth Factors Abstracts</collection><collection>Virology and AIDS Abstracts</collection><collection>Technology Research Database</collection><collection>Environmental Sciences and Pollution Management</collection><collection>Engineering Research Database</collection><collection>AIDS and Cancer Research Abstracts</collection><collection>Algology Mycology and Protozoology Abstracts (Microbiology C)</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>Genetics Abstracts</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Proceedings of the National Academy of Sciences - PNAS</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Xu, Jiang</au><au>Gannon, Patrick J</au><au>Emmorey, Karen</au><au>Smith, Jason F</au><au>Braun, Allen R</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Symbolic gestures and spoken language are processed by a common neural system</atitle><jtitle>Proceedings of the National Academy of Sciences - PNAS</jtitle><addtitle>Proc Natl Acad Sci U S A</addtitle><date>2009-12-08</date><risdate>2009</risdate><volume>106</volume><issue>49</issue><spage>20664</spage><epage>20669</epage><pages>20664-20669</pages><issn>0027-8424</issn><eissn>1091-6490</eissn><abstract>Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and function as complete utterances in their own right. The relationship between these gestures and spoken language remains unclear. We used functional MRI to investigate whether these two forms of communication are processed by the same system in the human brain. Responses to symbolic gestures, to their spoken glosses (expressing the gestures' meaning in English), and to visually and acoustically matched control stimuli were compared in a randomized block design. General Linear Models (GLM) contrasts identified shared and unique activations and functional connectivity analyses delineated regional interactions associated with each condition. Results support a model in which bilateral modality-specific areas in superior and inferior temporal cortices extract salient features from vocal-auditory and gestural-visual stimuli respectively. However, both classes of stimuli activate a common, left-lateralized network of inferior frontal and posterior temporal regions in which symbolic gestures and spoken words may be mapped onto common, corresponding conceptual representations. We suggest that these anterior and posterior perisylvian areas, identified since the mid-19th century as the core of the brain's language system, are not in fact committed to language processing, but may function as a modality-independent semiotic system that plays a broader role in human communication, linking meaning with symbols whether these are words, gestures, images, sounds, or objects.</abstract><cop>United States</cop><pub>National Academy of Sciences</pub><pmid>19923436</pmid><doi>10.1073/pnas.0909197106</doi><tpages>6</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0027-8424
ispartof Proceedings of the National Academy of Sciences - PNAS, 2009-12, Vol.106 (49), p.20664-20669
issn 0027-8424
1091-6490
language eng
recordid cdi_proquest_miscellaneous_748961467
source JSTOR Archival Journals and Primary Sources Collection; PubMed Central
subjects American sign language
Brain
Brain Mapping
Cerebral hemispheres
Connectivity
Female
Gestures
Human subjects
Humans
Language
Language comprehension
Magnetic resonance imaging
Mental Processes
Nerve Net - physiology
Nonverbal communication
Pantomime
Semantics
Social Sciences
Symbolic communication
Symbolism
Verbal communication
title Symbolic gestures and spoken language are processed by a common neural system
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T17%3A13%3A05IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-jstor_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Symbolic%20gestures%20and%20spoken%20language%20are%20processed%20by%20a%20common%20neural%20system&rft.jtitle=Proceedings%20of%20the%20National%20Academy%20of%20Sciences%20-%20PNAS&rft.au=Xu,%20Jiang&rft.date=2009-12-08&rft.volume=106&rft.issue=49&rft.spage=20664&rft.epage=20669&rft.pages=20664-20669&rft.issn=0027-8424&rft.eissn=1091-6490&rft_id=info:doi/10.1073/pnas.0909197106&rft_dat=%3Cjstor_proqu%3E40536054%3C/jstor_proqu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=201405974&rft_id=info:pmid/19923436&rft_jstor_id=40536054&rfr_iscdi=true