Loading…

Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images

Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and impr...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on multimedia 2022, Vol.24, p.3327-3339
Main Authors: Huang, Wei, Zhang, Siyuan, Zhang, Peng, Zha, Yufei, Fang, Yuming, Zhang, Yanning
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793
cites cdi_FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793
container_end_page 3339
container_issue
container_start_page 3327
container_title IEEE transactions on multimedia
container_volume 24
creator Huang, Wei
Zhang, Siyuan
Zhang, Peng
Zha, Yufei
Fang, Yuming
Zhang, Yanning
description Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and improper measures of facial expressions, can further contribute to the above dilemma. To solve these problems, a novel identity-aware method is proposed in this study. Furthermore, this study also represents the first attempt to fulfill the challenging person-dependent facial expression recognition task based on deep metric learning and facial image synthesis techniques. Technically, a StarGAN is incorporated to synthesize facial images depicting different but complete basic emotions for each identity to augment the training data. Then, a deep-convolutional-neural-network-based network is employed to automatically extract latent features from both real facial images and all synthesized facial images. Next, a Mahalanobis metric network trained based on extracted latent features outputs a learned metric that measures facial expression differences between images, and the recognition task can thus be realized. Extensive experiments based on several well-known publicly available datasets are carried out in this study for performance evaluations. Person-dependent datasets, including CK+, Oulu (all 6 subdatasets), MMI, ISAFE, ISED, etc., are all incorporated. After comparing the new method with several popular or state-of-the-art facial expression recognition methods, its superiority in person-dependent facial expression recognition can be proposed from a statistical point of view.
doi_str_mv 10.1109/TMM.2021.3096068
format article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2688683897</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9479695</ieee_id><sourcerecordid>2688683897</sourcerecordid><originalsourceid>FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793</originalsourceid><addsrcrecordid>eNo9kE1LAzEQhoMoWKt3wUvA89ZJdjcfx6qtFloErV5DNjtbU9rdmmzR-uvd0uJp3oHnnYGHkGsGA8ZA381nswEHzgYpaAFCnZAe0xlLAKQ87XLOIdGcwTm5iHEJwLIcZI_YSYl169tdMvy2AenYOm9XdPSzCRijb2r6iq5Z1L7d5w9v6SPihs6wDd7RKdpQ-3pB723EknbE265uPzH6326drO0C4yU5q-wq4tVx9sn7eDR_eE6mL0-Th-E0cVyzNuFoUWAhoFIqq5xkiLzI8jQrlHAlc5VzsqyKAlzuZIGuzLB0KtWKW4aZ1Gmf3B7ubkLztcXYmmWzDXX30nChlFCp0rKj4EC50MQYsDKb4Nc27AwDsxdpOpFmL9IcRXaVm0PFI-I_rrufQufpH0mycHw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2688683897</pqid></control><display><type>article</type><title>Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images</title><source>IEEE Xplore (Online service)</source><creator>Huang, Wei ; Zhang, Siyuan ; Zhang, Peng ; Zha, Yufei ; Fang, Yuming ; Zhang, Yanning</creator><creatorcontrib>Huang, Wei ; Zhang, Siyuan ; Zhang, Peng ; Zha, Yufei ; Fang, Yuming ; Zhang, Yanning</creatorcontrib><description>Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and improper measures of facial expressions, can further contribute to the above dilemma. To solve these problems, a novel identity-aware method is proposed in this study. Furthermore, this study also represents the first attempt to fulfill the challenging person-dependent facial expression recognition task based on deep metric learning and facial image synthesis techniques. Technically, a StarGAN is incorporated to synthesize facial images depicting different but complete basic emotions for each identity to augment the training data. Then, a deep-convolutional-neural-network-based network is employed to automatically extract latent features from both real facial images and all synthesized facial images. Next, a Mahalanobis metric network trained based on extracted latent features outputs a learned metric that measures facial expression differences between images, and the recognition task can thus be realized. Extensive experiments based on several well-known publicly available datasets are carried out in this study for performance evaluations. Person-dependent datasets, including CK+, Oulu (all 6 subdatasets), MMI, ISAFE, ISED, etc., are all incorporated. After comparing the new method with several popular or state-of-the-art facial expression recognition methods, its superiority in person-dependent facial expression recognition can be proposed from a statistical point of view.</description><identifier>ISSN: 1520-9210</identifier><identifier>EISSN: 1941-0077</identifier><identifier>DOI: 10.1109/TMM.2021.3096068</identifier><identifier>CODEN: ITMUF8</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Datasets ; Deep learning ; Face recognition ; facial expression recognition ; Feature extraction ; Generative adversarial networks ; Image recognition ; Image synthesis ; Learning ; Measurement ; metric learning ; Neural networks ; Object recognition ; Performance evaluation ; person-dependent ; Synthesis ; Task analysis ; Training</subject><ispartof>IEEE transactions on multimedia, 2022, Vol.24, p.3327-3339</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793</citedby><cites>FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793</cites><orcidid>0000-0002-2977-8057 ; 0000-0002-0541-8612 ; 0000-0001-9690-7026</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9479695$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,777,781,4010,27904,27905,27906,54777</link.rule.ids></links><search><creatorcontrib>Huang, Wei</creatorcontrib><creatorcontrib>Zhang, Siyuan</creatorcontrib><creatorcontrib>Zhang, Peng</creatorcontrib><creatorcontrib>Zha, Yufei</creatorcontrib><creatorcontrib>Fang, Yuming</creatorcontrib><creatorcontrib>Zhang, Yanning</creatorcontrib><title>Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images</title><title>IEEE transactions on multimedia</title><addtitle>TMM</addtitle><description>Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and improper measures of facial expressions, can further contribute to the above dilemma. To solve these problems, a novel identity-aware method is proposed in this study. Furthermore, this study also represents the first attempt to fulfill the challenging person-dependent facial expression recognition task based on deep metric learning and facial image synthesis techniques. Technically, a StarGAN is incorporated to synthesize facial images depicting different but complete basic emotions for each identity to augment the training data. Then, a deep-convolutional-neural-network-based network is employed to automatically extract latent features from both real facial images and all synthesized facial images. Next, a Mahalanobis metric network trained based on extracted latent features outputs a learned metric that measures facial expression differences between images, and the recognition task can thus be realized. Extensive experiments based on several well-known publicly available datasets are carried out in this study for performance evaluations. Person-dependent datasets, including CK+, Oulu (all 6 subdatasets), MMI, ISAFE, ISED, etc., are all incorporated. After comparing the new method with several popular or state-of-the-art facial expression recognition methods, its superiority in person-dependent facial expression recognition can be proposed from a statistical point of view.</description><subject>Datasets</subject><subject>Deep learning</subject><subject>Face recognition</subject><subject>facial expression recognition</subject><subject>Feature extraction</subject><subject>Generative adversarial networks</subject><subject>Image recognition</subject><subject>Image synthesis</subject><subject>Learning</subject><subject>Measurement</subject><subject>metric learning</subject><subject>Neural networks</subject><subject>Object recognition</subject><subject>Performance evaluation</subject><subject>person-dependent</subject><subject>Synthesis</subject><subject>Task analysis</subject><subject>Training</subject><issn>1520-9210</issn><issn>1941-0077</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNo9kE1LAzEQhoMoWKt3wUvA89ZJdjcfx6qtFloErV5DNjtbU9rdmmzR-uvd0uJp3oHnnYGHkGsGA8ZA381nswEHzgYpaAFCnZAe0xlLAKQ87XLOIdGcwTm5iHEJwLIcZI_YSYl169tdMvy2AenYOm9XdPSzCRijb2r6iq5Z1L7d5w9v6SPihs6wDd7RKdpQ-3pB723EknbE265uPzH6326drO0C4yU5q-wq4tVx9sn7eDR_eE6mL0-Th-E0cVyzNuFoUWAhoFIqq5xkiLzI8jQrlHAlc5VzsqyKAlzuZIGuzLB0KtWKW4aZ1Gmf3B7ubkLztcXYmmWzDXX30nChlFCp0rKj4EC50MQYsDKb4Nc27AwDsxdpOpFmL9IcRXaVm0PFI-I_rrufQufpH0mycHw</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Huang, Wei</creator><creator>Zhang, Siyuan</creator><creator>Zhang, Peng</creator><creator>Zha, Yufei</creator><creator>Fang, Yuming</creator><creator>Zhang, Yanning</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0002-2977-8057</orcidid><orcidid>https://orcid.org/0000-0002-0541-8612</orcidid><orcidid>https://orcid.org/0000-0001-9690-7026</orcidid></search><sort><creationdate>2022</creationdate><title>Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images</title><author>Huang, Wei ; Zhang, Siyuan ; Zhang, Peng ; Zha, Yufei ; Fang, Yuming ; Zhang, Yanning</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Datasets</topic><topic>Deep learning</topic><topic>Face recognition</topic><topic>facial expression recognition</topic><topic>Feature extraction</topic><topic>Generative adversarial networks</topic><topic>Image recognition</topic><topic>Image synthesis</topic><topic>Learning</topic><topic>Measurement</topic><topic>metric learning</topic><topic>Neural networks</topic><topic>Object recognition</topic><topic>Performance evaluation</topic><topic>person-dependent</topic><topic>Synthesis</topic><topic>Task analysis</topic><topic>Training</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Huang, Wei</creatorcontrib><creatorcontrib>Zhang, Siyuan</creatorcontrib><creatorcontrib>Zhang, Peng</creatorcontrib><creatorcontrib>Zha, Yufei</creatorcontrib><creatorcontrib>Fang, Yuming</creatorcontrib><creatorcontrib>Zhang, Yanning</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998–Present</collection><collection>IEEE Xplore</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on multimedia</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Huang, Wei</au><au>Zhang, Siyuan</au><au>Zhang, Peng</au><au>Zha, Yufei</au><au>Fang, Yuming</au><au>Zhang, Yanning</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images</atitle><jtitle>IEEE transactions on multimedia</jtitle><stitle>TMM</stitle><date>2022</date><risdate>2022</risdate><volume>24</volume><spage>3327</spage><epage>3339</epage><pages>3327-3339</pages><issn>1520-9210</issn><eissn>1941-0077</eissn><coden>ITMUF8</coden><abstract>Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and improper measures of facial expressions, can further contribute to the above dilemma. To solve these problems, a novel identity-aware method is proposed in this study. Furthermore, this study also represents the first attempt to fulfill the challenging person-dependent facial expression recognition task based on deep metric learning and facial image synthesis techniques. Technically, a StarGAN is incorporated to synthesize facial images depicting different but complete basic emotions for each identity to augment the training data. Then, a deep-convolutional-neural-network-based network is employed to automatically extract latent features from both real facial images and all synthesized facial images. Next, a Mahalanobis metric network trained based on extracted latent features outputs a learned metric that measures facial expression differences between images, and the recognition task can thus be realized. Extensive experiments based on several well-known publicly available datasets are carried out in this study for performance evaluations. Person-dependent datasets, including CK+, Oulu (all 6 subdatasets), MMI, ISAFE, ISED, etc., are all incorporated. After comparing the new method with several popular or state-of-the-art facial expression recognition methods, its superiority in person-dependent facial expression recognition can be proposed from a statistical point of view.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/TMM.2021.3096068</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0002-2977-8057</orcidid><orcidid>https://orcid.org/0000-0002-0541-8612</orcidid><orcidid>https://orcid.org/0000-0001-9690-7026</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1520-9210
ispartof IEEE transactions on multimedia, 2022, Vol.24, p.3327-3339
issn 1520-9210
1941-0077
language eng
recordid cdi_proquest_journals_2688683897
source IEEE Xplore (Online service)
subjects Datasets
Deep learning
Face recognition
facial expression recognition
Feature extraction
Generative adversarial networks
Image recognition
Image synthesis
Learning
Measurement
metric learning
Neural networks
Object recognition
Performance evaluation
person-dependent
Synthesis
Task analysis
Training
title Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-19T21%3A54%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Identity-Aware%20Facial%20Expression%20Recognition%20Via%20Deep%20Metric%20Learning%20Based%20on%20Synthesized%20Images&rft.jtitle=IEEE%20transactions%20on%20multimedia&rft.au=Huang,%20Wei&rft.date=2022&rft.volume=24&rft.spage=3327&rft.epage=3339&rft.pages=3327-3339&rft.issn=1520-9210&rft.eissn=1941-0077&rft.coden=ITMUF8&rft_id=info:doi/10.1109/TMM.2021.3096068&rft_dat=%3Cproquest_ieee_%3E2688683897%3C/proquest_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2688683897&rft_id=info:pmid/&rft_ieee_id=9479695&rfr_iscdi=true