Loading…

Putting the Face to the Voice': Matching Identity across Modality

Speech perception provides compelling examples of a strong link between auditory and visual modalities [1, 2]. This link originates in the mechanics of speech production, which, in shaping the vocal tract, determine the movement of the face as well as the sound of the voice [3, 4]. In this paper, we...

Full description

Saved in:
Bibliographic Details
Published in:Current biology 2003-09, Vol.13 (19), p.1709-1714
Main Authors: Kamachi, Miyuki, Hill, Harold, Lander, Karen, Vatikiotis-Bateson, Eric
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 1714
container_issue 19
container_start_page 1709
container_title Current biology
container_volume 13
creator Kamachi, Miyuki
Hill, Harold
Lander, Karen
Vatikiotis-Bateson, Eric
description Speech perception provides compelling examples of a strong link between auditory and visual modalities [1, 2]. This link originates in the mechanics of speech production, which, in shaping the vocal tract, determine the movement of the face as well as the sound of the voice [3, 4]. In this paper, we present evidence that equivalent information about identity is available cross-modally from both the face and voice. Using a delayed matching to sample task, XAB, we show that people can match the video of an unfamiliar face, X, to an unfamiliar voice, A or B, and vice versa, but only when stimuli are moving and are played forward. The critical role of time-varying information is underlined by the ability to match faces to voices containing only the coarse spatial and temporal information provided by sine wave speech [5]. The effect of varying sentence content across modalities was small, showing that identity-specific information is not closely tied to particular utterances. We conclude that the physical constraints linking faces to voices result in bimodally available dynamic information, not only about what is being said, but also about who is saying it.
doi_str_mv 10.1016/j.cub.2003.09.005
format article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_proquest_miscellaneous_75742604</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0960982203006638</els_id><sourcerecordid>19228473</sourcerecordid><originalsourceid>FETCH-LOGICAL-e385t-7ed59160d1bbf3cf4dc16cf123f6dd5c94115478c974caf638cc2d3729504da83</originalsourceid><addsrcrecordid>eNqFkctOwzAQRS0EglL4ADYoK1gljF-xDStU8agEggWwtdyxA67apsQOEn9PymPNamako5HuPYQcUago0PpsXmE_qxgAr8BUAHKLjKhWpgQh5DYZgamhNJqxPbKf0hyAMm3qXbJHhWRUczUil499znH1WuS3UFw7DEVuv_eXNmI4PS_uXca3DTD1YZVj_iwcdm1KxX3r3WK4D8hO4xYpHP7OMXm-vnqa3JZ3DzfTyeVdGbiWuVTBS0Nr8HQ2azg2wiOtsaGMN7X3Eo2gVAql0SiBrqm5RmSeK2YkCO80H5OTn7_rrn3vQ8p2GROGxcKtQtsnq6QSrAbxL0gNY1ooPoDHv2A_WwZv111cuu7T_rUzABc_QBhyfcTQ2YQxrDD42AXM1rfRUrAbF3ZuBxd248KCsYML_gX_oXoF</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>19228473</pqid></control><display><type>article</type><title>Putting the Face to the Voice': Matching Identity across Modality</title><source>BACON - Elsevier - GLOBAL_SCIENCEDIRECT-OPENACCESS</source><creator>Kamachi, Miyuki ; Hill, Harold ; Lander, Karen ; Vatikiotis-Bateson, Eric</creator><creatorcontrib>Kamachi, Miyuki ; Hill, Harold ; Lander, Karen ; Vatikiotis-Bateson, Eric</creatorcontrib><description>Speech perception provides compelling examples of a strong link between auditory and visual modalities [1, 2]. This link originates in the mechanics of speech production, which, in shaping the vocal tract, determine the movement of the face as well as the sound of the voice [3, 4]. In this paper, we present evidence that equivalent information about identity is available cross-modally from both the face and voice. Using a delayed matching to sample task, XAB, we show that people can match the video of an unfamiliar face, X, to an unfamiliar voice, A or B, and vice versa, but only when stimuli are moving and are played forward. The critical role of time-varying information is underlined by the ability to match faces to voices containing only the coarse spatial and temporal information provided by sine wave speech [5]. The effect of varying sentence content across modalities was small, showing that identity-specific information is not closely tied to particular utterances. We conclude that the physical constraints linking faces to voices result in bimodally available dynamic information, not only about what is being said, but also about who is saying it.</description><identifier>ISSN: 0960-9822</identifier><identifier>EISSN: 1879-0445</identifier><identifier>DOI: 10.1016/j.cub.2003.09.005</identifier><identifier>PMID: 14521837</identifier><language>eng</language><publisher>England: Elsevier Inc</publisher><subject>Adult ; Facial Expression ; Female ; Humans ; Individuality ; Japan ; Male ; Speech Perception - physiology ; Videotape Recording ; Visual Perception - physiology ; Voice - physiology</subject><ispartof>Current biology, 2003-09, Vol.13 (19), p.1709-1714</ispartof><rights>2003 Cell Press</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/14521837$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Kamachi, Miyuki</creatorcontrib><creatorcontrib>Hill, Harold</creatorcontrib><creatorcontrib>Lander, Karen</creatorcontrib><creatorcontrib>Vatikiotis-Bateson, Eric</creatorcontrib><title>Putting the Face to the Voice': Matching Identity across Modality</title><title>Current biology</title><addtitle>Curr Biol</addtitle><description>Speech perception provides compelling examples of a strong link between auditory and visual modalities [1, 2]. This link originates in the mechanics of speech production, which, in shaping the vocal tract, determine the movement of the face as well as the sound of the voice [3, 4]. In this paper, we present evidence that equivalent information about identity is available cross-modally from both the face and voice. Using a delayed matching to sample task, XAB, we show that people can match the video of an unfamiliar face, X, to an unfamiliar voice, A or B, and vice versa, but only when stimuli are moving and are played forward. The critical role of time-varying information is underlined by the ability to match faces to voices containing only the coarse spatial and temporal information provided by sine wave speech [5]. The effect of varying sentence content across modalities was small, showing that identity-specific information is not closely tied to particular utterances. We conclude that the physical constraints linking faces to voices result in bimodally available dynamic information, not only about what is being said, but also about who is saying it.</description><subject>Adult</subject><subject>Facial Expression</subject><subject>Female</subject><subject>Humans</subject><subject>Individuality</subject><subject>Japan</subject><subject>Male</subject><subject>Speech Perception - physiology</subject><subject>Videotape Recording</subject><subject>Visual Perception - physiology</subject><subject>Voice - physiology</subject><issn>0960-9822</issn><issn>1879-0445</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2003</creationdate><recordtype>article</recordtype><recordid>eNqFkctOwzAQRS0EglL4ADYoK1gljF-xDStU8agEggWwtdyxA67apsQOEn9PymPNamako5HuPYQcUago0PpsXmE_qxgAr8BUAHKLjKhWpgQh5DYZgamhNJqxPbKf0hyAMm3qXbJHhWRUczUil499znH1WuS3UFw7DEVuv_eXNmI4PS_uXca3DTD1YZVj_iwcdm1KxX3r3WK4D8hO4xYpHP7OMXm-vnqa3JZ3DzfTyeVdGbiWuVTBS0Nr8HQ2azg2wiOtsaGMN7X3Eo2gVAql0SiBrqm5RmSeK2YkCO80H5OTn7_rrn3vQ8p2GROGxcKtQtsnq6QSrAbxL0gNY1ooPoDHv2A_WwZv111cuu7T_rUzABc_QBhyfcTQ2YQxrDD42AXM1rfRUrAbF3ZuBxd248KCsYML_gX_oXoF</recordid><startdate>20030930</startdate><enddate>20030930</enddate><creator>Kamachi, Miyuki</creator><creator>Hill, Harold</creator><creator>Lander, Karen</creator><creator>Vatikiotis-Bateson, Eric</creator><general>Elsevier Inc</general><scope>6I.</scope><scope>AAFTH</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>7TK</scope><scope>7X8</scope></search><sort><creationdate>20030930</creationdate><title>Putting the Face to the Voice': Matching Identity across Modality</title><author>Kamachi, Miyuki ; Hill, Harold ; Lander, Karen ; Vatikiotis-Bateson, Eric</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-e385t-7ed59160d1bbf3cf4dc16cf123f6dd5c94115478c974caf638cc2d3729504da83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2003</creationdate><topic>Adult</topic><topic>Facial Expression</topic><topic>Female</topic><topic>Humans</topic><topic>Individuality</topic><topic>Japan</topic><topic>Male</topic><topic>Speech Perception - physiology</topic><topic>Videotape Recording</topic><topic>Visual Perception - physiology</topic><topic>Voice - physiology</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Kamachi, Miyuki</creatorcontrib><creatorcontrib>Hill, Harold</creatorcontrib><creatorcontrib>Lander, Karen</creatorcontrib><creatorcontrib>Vatikiotis-Bateson, Eric</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>Neurosciences Abstracts</collection><collection>MEDLINE - Academic</collection><jtitle>Current biology</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Kamachi, Miyuki</au><au>Hill, Harold</au><au>Lander, Karen</au><au>Vatikiotis-Bateson, Eric</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Putting the Face to the Voice': Matching Identity across Modality</atitle><jtitle>Current biology</jtitle><addtitle>Curr Biol</addtitle><date>2003-09-30</date><risdate>2003</risdate><volume>13</volume><issue>19</issue><spage>1709</spage><epage>1714</epage><pages>1709-1714</pages><issn>0960-9822</issn><eissn>1879-0445</eissn><abstract>Speech perception provides compelling examples of a strong link between auditory and visual modalities [1, 2]. This link originates in the mechanics of speech production, which, in shaping the vocal tract, determine the movement of the face as well as the sound of the voice [3, 4]. In this paper, we present evidence that equivalent information about identity is available cross-modally from both the face and voice. Using a delayed matching to sample task, XAB, we show that people can match the video of an unfamiliar face, X, to an unfamiliar voice, A or B, and vice versa, but only when stimuli are moving and are played forward. The critical role of time-varying information is underlined by the ability to match faces to voices containing only the coarse spatial and temporal information provided by sine wave speech [5]. The effect of varying sentence content across modalities was small, showing that identity-specific information is not closely tied to particular utterances. We conclude that the physical constraints linking faces to voices result in bimodally available dynamic information, not only about what is being said, but also about who is saying it.</abstract><cop>England</cop><pub>Elsevier Inc</pub><pmid>14521837</pmid><doi>10.1016/j.cub.2003.09.005</doi><tpages>6</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0960-9822
ispartof Current biology, 2003-09, Vol.13 (19), p.1709-1714
issn 0960-9822
1879-0445
language eng
recordid cdi_proquest_miscellaneous_75742604
source BACON - Elsevier - GLOBAL_SCIENCEDIRECT-OPENACCESS
subjects Adult
Facial Expression
Female
Humans
Individuality
Japan
Male
Speech Perception - physiology
Videotape Recording
Visual Perception - physiology
Voice - physiology
title Putting the Face to the Voice': Matching Identity across Modality
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T22%3A53%3A00IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Putting%20the%20Face%20to%20the%20Voice':%20Matching%20Identity%20across%20Modality&rft.jtitle=Current%20biology&rft.au=Kamachi,%20Miyuki&rft.date=2003-09-30&rft.volume=13&rft.issue=19&rft.spage=1709&rft.epage=1714&rft.pages=1709-1714&rft.issn=0960-9822&rft.eissn=1879-0445&rft_id=info:doi/10.1016/j.cub.2003.09.005&rft_dat=%3Cproquest_pubme%3E19228473%3C/proquest_pubme%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-e385t-7ed59160d1bbf3cf4dc16cf123f6dd5c94115478c974caf638cc2d3729504da83%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=19228473&rft_id=info:pmid/14521837&rfr_iscdi=true