Loading…
Symbolic gestures and spoken language are processed by a common neural system
Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and func...
Saved in:
Published in: | Proceedings of the National Academy of Sciences - PNAS 2009-12, Vol.106 (49), p.20664-20669 |
---|---|
Main Authors: | , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563 |
---|---|
cites | cdi_FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563 |
container_end_page | 20669 |
container_issue | 49 |
container_start_page | 20664 |
container_title | Proceedings of the National Academy of Sciences - PNAS |
container_volume | 106 |
creator | Xu, Jiang Gannon, Patrick J Emmorey, Karen Smith, Jason F Braun, Allen R |
description | Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and function as complete utterances in their own right. The relationship between these gestures and spoken language remains unclear. We used functional MRI to investigate whether these two forms of communication are processed by the same system in the human brain. Responses to symbolic gestures, to their spoken glosses (expressing the gestures' meaning in English), and to visually and acoustically matched control stimuli were compared in a randomized block design. General Linear Models (GLM) contrasts identified shared and unique activations and functional connectivity analyses delineated regional interactions associated with each condition. Results support a model in which bilateral modality-specific areas in superior and inferior temporal cortices extract salient features from vocal-auditory and gestural-visual stimuli respectively. However, both classes of stimuli activate a common, left-lateralized network of inferior frontal and posterior temporal regions in which symbolic gestures and spoken words may be mapped onto common, corresponding conceptual representations. We suggest that these anterior and posterior perisylvian areas, identified since the mid-19th century as the core of the brain's language system, are not in fact committed to language processing, but may function as a modality-independent semiotic system that plays a broader role in human communication, linking meaning with symbols whether these are words, gestures, images, sounds, or objects. |
doi_str_mv | 10.1073/pnas.0909197106 |
format | article |
fullrecord | <record><control><sourceid>jstor_proqu</sourceid><recordid>TN_cdi_proquest_miscellaneous_748961467</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><jstor_id>40536054</jstor_id><sourcerecordid>40536054</sourcerecordid><originalsourceid>FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563</originalsourceid><addsrcrecordid>eNqFkc1v1DAQxS0EokvhzAmwuCAOacfxV3xBQhVfUhGH0rPlZCchS2IHO0Hsf4-jXXWBCydbmt-8eU-PkKcMLhhofjl5ly7AgGFGM1D3yIblf6GEgftkA1DqohKlOCOPUtoBgJEVPCRnzJiSC6425PPNfqzD0De0wzQvERN1fkvTFL6jp4Pz3eI6pC4inWJoMCXc0npPHW3COAZPPS7RDTTt04zjY_KgdUPCJ8f3nNy-f_f16mNx_eXDp6u310UjSzEXQoETUjoltJCVUiVIx0RdZk91zeoWtg22rK252GKVV5AxZ4CjbrlopFT8nLw56E5LPWKm_ZxN2Cn2o4t7G1xv_574_pvtwk9bam1K4Fng1VEghh9LTm7HPjU45MAYlmS1qIxiQulMvvyH3IUl-pzOlsAESKNFhi4PUBNDShHbOysM7FqUXYuyp6LyxvM_E5z4YzMZoEdg3TzJKStMvqzUevX1fxDbLsMw4685s88O7C7NId7B2T9XIFetF4d564J1XeyTvb3JATkwzTIi-G_kJbpq</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>201405974</pqid></control><display><type>article</type><title>Symbolic gestures and spoken language are processed by a common neural system</title><source>JSTOR Archival Journals and Primary Sources Collection</source><source>PubMed Central</source><creator>Xu, Jiang ; Gannon, Patrick J ; Emmorey, Karen ; Smith, Jason F ; Braun, Allen R</creator><creatorcontrib>Xu, Jiang ; Gannon, Patrick J ; Emmorey, Karen ; Smith, Jason F ; Braun, Allen R</creatorcontrib><description>Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and function as complete utterances in their own right. The relationship between these gestures and spoken language remains unclear. We used functional MRI to investigate whether these two forms of communication are processed by the same system in the human brain. Responses to symbolic gestures, to their spoken glosses (expressing the gestures' meaning in English), and to visually and acoustically matched control stimuli were compared in a randomized block design. General Linear Models (GLM) contrasts identified shared and unique activations and functional connectivity analyses delineated regional interactions associated with each condition. Results support a model in which bilateral modality-specific areas in superior and inferior temporal cortices extract salient features from vocal-auditory and gestural-visual stimuli respectively. However, both classes of stimuli activate a common, left-lateralized network of inferior frontal and posterior temporal regions in which symbolic gestures and spoken words may be mapped onto common, corresponding conceptual representations. We suggest that these anterior and posterior perisylvian areas, identified since the mid-19th century as the core of the brain's language system, are not in fact committed to language processing, but may function as a modality-independent semiotic system that plays a broader role in human communication, linking meaning with symbols whether these are words, gestures, images, sounds, or objects.</description><identifier>ISSN: 0027-8424</identifier><identifier>EISSN: 1091-6490</identifier><identifier>DOI: 10.1073/pnas.0909197106</identifier><identifier>PMID: 19923436</identifier><language>eng</language><publisher>United States: National Academy of Sciences</publisher><subject>American sign language ; Brain ; Brain Mapping ; Cerebral hemispheres ; Connectivity ; Female ; Gestures ; Human subjects ; Humans ; Language ; Language comprehension ; Magnetic resonance imaging ; Mental Processes ; Nerve Net - physiology ; Nonverbal communication ; Pantomime ; Semantics ; Social Sciences ; Symbolic communication ; Symbolism ; Verbal communication</subject><ispartof>Proceedings of the National Academy of Sciences - PNAS, 2009-12, Vol.106 (49), p.20664-20669</ispartof><rights>Copyright National Academy of Sciences Dec 8, 2009</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563</citedby><cites>FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Uhttp://www.pnas.org/content/106/49.cover.gif</thumbnail><linktopdf>$$Uhttps://www.jstor.org/stable/pdf/40536054$$EPDF$$P50$$Gjstor$$H</linktopdf><linktohtml>$$Uhttps://www.jstor.org/stable/40536054$$EHTML$$P50$$Gjstor$$H</linktohtml><link.rule.ids>230,314,723,776,780,881,27903,27904,53770,53772,58217,58450</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/19923436$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Xu, Jiang</creatorcontrib><creatorcontrib>Gannon, Patrick J</creatorcontrib><creatorcontrib>Emmorey, Karen</creatorcontrib><creatorcontrib>Smith, Jason F</creatorcontrib><creatorcontrib>Braun, Allen R</creatorcontrib><title>Symbolic gestures and spoken language are processed by a common neural system</title><title>Proceedings of the National Academy of Sciences - PNAS</title><addtitle>Proc Natl Acad Sci U S A</addtitle><description>Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and function as complete utterances in their own right. The relationship between these gestures and spoken language remains unclear. We used functional MRI to investigate whether these two forms of communication are processed by the same system in the human brain. Responses to symbolic gestures, to their spoken glosses (expressing the gestures' meaning in English), and to visually and acoustically matched control stimuli were compared in a randomized block design. General Linear Models (GLM) contrasts identified shared and unique activations and functional connectivity analyses delineated regional interactions associated with each condition. Results support a model in which bilateral modality-specific areas in superior and inferior temporal cortices extract salient features from vocal-auditory and gestural-visual stimuli respectively. However, both classes of stimuli activate a common, left-lateralized network of inferior frontal and posterior temporal regions in which symbolic gestures and spoken words may be mapped onto common, corresponding conceptual representations. We suggest that these anterior and posterior perisylvian areas, identified since the mid-19th century as the core of the brain's language system, are not in fact committed to language processing, but may function as a modality-independent semiotic system that plays a broader role in human communication, linking meaning with symbols whether these are words, gestures, images, sounds, or objects.</description><subject>American sign language</subject><subject>Brain</subject><subject>Brain Mapping</subject><subject>Cerebral hemispheres</subject><subject>Connectivity</subject><subject>Female</subject><subject>Gestures</subject><subject>Human subjects</subject><subject>Humans</subject><subject>Language</subject><subject>Language comprehension</subject><subject>Magnetic resonance imaging</subject><subject>Mental Processes</subject><subject>Nerve Net - physiology</subject><subject>Nonverbal communication</subject><subject>Pantomime</subject><subject>Semantics</subject><subject>Social Sciences</subject><subject>Symbolic communication</subject><subject>Symbolism</subject><subject>Verbal communication</subject><issn>0027-8424</issn><issn>1091-6490</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2009</creationdate><recordtype>article</recordtype><recordid>eNqFkc1v1DAQxS0EokvhzAmwuCAOacfxV3xBQhVfUhGH0rPlZCchS2IHO0Hsf4-jXXWBCydbmt-8eU-PkKcMLhhofjl5ly7AgGFGM1D3yIblf6GEgftkA1DqohKlOCOPUtoBgJEVPCRnzJiSC6425PPNfqzD0De0wzQvERN1fkvTFL6jp4Pz3eI6pC4inWJoMCXc0npPHW3COAZPPS7RDTTt04zjY_KgdUPCJ8f3nNy-f_f16mNx_eXDp6u310UjSzEXQoETUjoltJCVUiVIx0RdZk91zeoWtg22rK252GKVV5AxZ4CjbrlopFT8nLw56E5LPWKm_ZxN2Cn2o4t7G1xv_574_pvtwk9bam1K4Fng1VEghh9LTm7HPjU45MAYlmS1qIxiQulMvvyH3IUl-pzOlsAESKNFhi4PUBNDShHbOysM7FqUXYuyp6LyxvM_E5z4YzMZoEdg3TzJKStMvqzUevX1fxDbLsMw4685s88O7C7NId7B2T9XIFetF4d564J1XeyTvb3JATkwzTIi-G_kJbpq</recordid><startdate>20091208</startdate><enddate>20091208</enddate><creator>Xu, Jiang</creator><creator>Gannon, Patrick J</creator><creator>Emmorey, Karen</creator><creator>Smith, Jason F</creator><creator>Braun, Allen R</creator><general>National Academy of Sciences</general><general>National Acad Sciences</general><scope>FBQ</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QG</scope><scope>7QL</scope><scope>7QP</scope><scope>7QR</scope><scope>7SN</scope><scope>7SS</scope><scope>7T5</scope><scope>7TK</scope><scope>7TM</scope><scope>7TO</scope><scope>7U9</scope><scope>8FD</scope><scope>C1K</scope><scope>FR3</scope><scope>H94</scope><scope>M7N</scope><scope>P64</scope><scope>RC3</scope><scope>7X8</scope><scope>5PM</scope></search><sort><creationdate>20091208</creationdate><title>Symbolic gestures and spoken language are processed by a common neural system</title><author>Xu, Jiang ; Gannon, Patrick J ; Emmorey, Karen ; Smith, Jason F ; Braun, Allen R</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2009</creationdate><topic>American sign language</topic><topic>Brain</topic><topic>Brain Mapping</topic><topic>Cerebral hemispheres</topic><topic>Connectivity</topic><topic>Female</topic><topic>Gestures</topic><topic>Human subjects</topic><topic>Humans</topic><topic>Language</topic><topic>Language comprehension</topic><topic>Magnetic resonance imaging</topic><topic>Mental Processes</topic><topic>Nerve Net - physiology</topic><topic>Nonverbal communication</topic><topic>Pantomime</topic><topic>Semantics</topic><topic>Social Sciences</topic><topic>Symbolic communication</topic><topic>Symbolism</topic><topic>Verbal communication</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Xu, Jiang</creatorcontrib><creatorcontrib>Gannon, Patrick J</creatorcontrib><creatorcontrib>Emmorey, Karen</creatorcontrib><creatorcontrib>Smith, Jason F</creatorcontrib><creatorcontrib>Braun, Allen R</creatorcontrib><collection>AGRIS</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Animal Behavior Abstracts</collection><collection>Bacteriology Abstracts (Microbiology B)</collection><collection>Calcium & Calcified Tissue Abstracts</collection><collection>Chemoreception Abstracts</collection><collection>Ecology Abstracts</collection><collection>Entomology Abstracts (Full archive)</collection><collection>Immunology Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Nucleic Acids Abstracts</collection><collection>Oncogenes and Growth Factors Abstracts</collection><collection>Virology and AIDS Abstracts</collection><collection>Technology Research Database</collection><collection>Environmental Sciences and Pollution Management</collection><collection>Engineering Research Database</collection><collection>AIDS and Cancer Research Abstracts</collection><collection>Algology Mycology and Protozoology Abstracts (Microbiology C)</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>Genetics Abstracts</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Proceedings of the National Academy of Sciences - PNAS</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Xu, Jiang</au><au>Gannon, Patrick J</au><au>Emmorey, Karen</au><au>Smith, Jason F</au><au>Braun, Allen R</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Symbolic gestures and spoken language are processed by a common neural system</atitle><jtitle>Proceedings of the National Academy of Sciences - PNAS</jtitle><addtitle>Proc Natl Acad Sci U S A</addtitle><date>2009-12-08</date><risdate>2009</risdate><volume>106</volume><issue>49</issue><spage>20664</spage><epage>20669</epage><pages>20664-20669</pages><issn>0027-8424</issn><eissn>1091-6490</eissn><abstract>Symbolic gestures, such as pantomimes that signify actions (e.g., threading a needle) or emblems that facilitate social transactions (e.g., finger to lips indicating "be quiet"), play an important role in human communication. They are autonomous, can fully take the place of words, and function as complete utterances in their own right. The relationship between these gestures and spoken language remains unclear. We used functional MRI to investigate whether these two forms of communication are processed by the same system in the human brain. Responses to symbolic gestures, to their spoken glosses (expressing the gestures' meaning in English), and to visually and acoustically matched control stimuli were compared in a randomized block design. General Linear Models (GLM) contrasts identified shared and unique activations and functional connectivity analyses delineated regional interactions associated with each condition. Results support a model in which bilateral modality-specific areas in superior and inferior temporal cortices extract salient features from vocal-auditory and gestural-visual stimuli respectively. However, both classes of stimuli activate a common, left-lateralized network of inferior frontal and posterior temporal regions in which symbolic gestures and spoken words may be mapped onto common, corresponding conceptual representations. We suggest that these anterior and posterior perisylvian areas, identified since the mid-19th century as the core of the brain's language system, are not in fact committed to language processing, but may function as a modality-independent semiotic system that plays a broader role in human communication, linking meaning with symbols whether these are words, gestures, images, sounds, or objects.</abstract><cop>United States</cop><pub>National Academy of Sciences</pub><pmid>19923436</pmid><doi>10.1073/pnas.0909197106</doi><tpages>6</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0027-8424 |
ispartof | Proceedings of the National Academy of Sciences - PNAS, 2009-12, Vol.106 (49), p.20664-20669 |
issn | 0027-8424 1091-6490 |
language | eng |
recordid | cdi_proquest_miscellaneous_748961467 |
source | JSTOR Archival Journals and Primary Sources Collection; PubMed Central |
subjects | American sign language Brain Brain Mapping Cerebral hemispheres Connectivity Female Gestures Human subjects Humans Language Language comprehension Magnetic resonance imaging Mental Processes Nerve Net - physiology Nonverbal communication Pantomime Semantics Social Sciences Symbolic communication Symbolism Verbal communication |
title | Symbolic gestures and spoken language are processed by a common neural system |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T17%3A13%3A05IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-jstor_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Symbolic%20gestures%20and%20spoken%20language%20are%20processed%20by%20a%20common%20neural%20system&rft.jtitle=Proceedings%20of%20the%20National%20Academy%20of%20Sciences%20-%20PNAS&rft.au=Xu,%20Jiang&rft.date=2009-12-08&rft.volume=106&rft.issue=49&rft.spage=20664&rft.epage=20669&rft.pages=20664-20669&rft.issn=0027-8424&rft.eissn=1091-6490&rft_id=info:doi/10.1073/pnas.0909197106&rft_dat=%3Cjstor_proqu%3E40536054%3C/jstor_proqu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c524t-460a455a64745866205a14b2234bb1bf0dcef1fb34de8c52e11a903e7f34c5563%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=201405974&rft_id=info:pmid/19923436&rft_jstor_id=40536054&rfr_iscdi=true |