Loading…
Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images
Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and impr...
Saved in:
Published in: | IEEE transactions on multimedia 2022, Vol.24, p.3327-3339 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793 |
---|---|
cites | cdi_FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793 |
container_end_page | 3339 |
container_issue | |
container_start_page | 3327 |
container_title | IEEE transactions on multimedia |
container_volume | 24 |
creator | Huang, Wei Zhang, Siyuan Zhang, Peng Zha, Yufei Fang, Yuming Zhang, Yanning |
description | Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and improper measures of facial expressions, can further contribute to the above dilemma. To solve these problems, a novel identity-aware method is proposed in this study. Furthermore, this study also represents the first attempt to fulfill the challenging person-dependent facial expression recognition task based on deep metric learning and facial image synthesis techniques. Technically, a StarGAN is incorporated to synthesize facial images depicting different but complete basic emotions for each identity to augment the training data. Then, a deep-convolutional-neural-network-based network is employed to automatically extract latent features from both real facial images and all synthesized facial images. Next, a Mahalanobis metric network trained based on extracted latent features outputs a learned metric that measures facial expression differences between images, and the recognition task can thus be realized. Extensive experiments based on several well-known publicly available datasets are carried out in this study for performance evaluations. Person-dependent datasets, including CK+, Oulu (all 6 subdatasets), MMI, ISAFE, ISED, etc., are all incorporated. After comparing the new method with several popular or state-of-the-art facial expression recognition methods, its superiority in person-dependent facial expression recognition can be proposed from a statistical point of view. |
doi_str_mv | 10.1109/TMM.2021.3096068 |
format | article |
fullrecord | <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2688683897</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9479695</ieee_id><sourcerecordid>2688683897</sourcerecordid><originalsourceid>FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793</originalsourceid><addsrcrecordid>eNo9kE1LAzEQhoMoWKt3wUvA89ZJdjcfx6qtFloErV5DNjtbU9rdmmzR-uvd0uJp3oHnnYGHkGsGA8ZA381nswEHzgYpaAFCnZAe0xlLAKQ87XLOIdGcwTm5iHEJwLIcZI_YSYl169tdMvy2AenYOm9XdPSzCRijb2r6iq5Z1L7d5w9v6SPihs6wDd7RKdpQ-3pB723EknbE265uPzH6326drO0C4yU5q-wq4tVx9sn7eDR_eE6mL0-Th-E0cVyzNuFoUWAhoFIqq5xkiLzI8jQrlHAlc5VzsqyKAlzuZIGuzLB0KtWKW4aZ1Gmf3B7ubkLztcXYmmWzDXX30nChlFCp0rKj4EC50MQYsDKb4Nc27AwDsxdpOpFmL9IcRXaVm0PFI-I_rrufQufpH0mycHw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2688683897</pqid></control><display><type>article</type><title>Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images</title><source>IEEE Xplore (Online service)</source><creator>Huang, Wei ; Zhang, Siyuan ; Zhang, Peng ; Zha, Yufei ; Fang, Yuming ; Zhang, Yanning</creator><creatorcontrib>Huang, Wei ; Zhang, Siyuan ; Zhang, Peng ; Zha, Yufei ; Fang, Yuming ; Zhang, Yanning</creatorcontrib><description>Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and improper measures of facial expressions, can further contribute to the above dilemma. To solve these problems, a novel identity-aware method is proposed in this study. Furthermore, this study also represents the first attempt to fulfill the challenging person-dependent facial expression recognition task based on deep metric learning and facial image synthesis techniques. Technically, a StarGAN is incorporated to synthesize facial images depicting different but complete basic emotions for each identity to augment the training data. Then, a deep-convolutional-neural-network-based network is employed to automatically extract latent features from both real facial images and all synthesized facial images. Next, a Mahalanobis metric network trained based on extracted latent features outputs a learned metric that measures facial expression differences between images, and the recognition task can thus be realized. Extensive experiments based on several well-known publicly available datasets are carried out in this study for performance evaluations. Person-dependent datasets, including CK+, Oulu (all 6 subdatasets), MMI, ISAFE, ISED, etc., are all incorporated. After comparing the new method with several popular or state-of-the-art facial expression recognition methods, its superiority in person-dependent facial expression recognition can be proposed from a statistical point of view.</description><identifier>ISSN: 1520-9210</identifier><identifier>EISSN: 1941-0077</identifier><identifier>DOI: 10.1109/TMM.2021.3096068</identifier><identifier>CODEN: ITMUF8</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Datasets ; Deep learning ; Face recognition ; facial expression recognition ; Feature extraction ; Generative adversarial networks ; Image recognition ; Image synthesis ; Learning ; Measurement ; metric learning ; Neural networks ; Object recognition ; Performance evaluation ; person-dependent ; Synthesis ; Task analysis ; Training</subject><ispartof>IEEE transactions on multimedia, 2022, Vol.24, p.3327-3339</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793</citedby><cites>FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793</cites><orcidid>0000-0002-2977-8057 ; 0000-0002-0541-8612 ; 0000-0001-9690-7026</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9479695$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,777,781,4010,27904,27905,27906,54777</link.rule.ids></links><search><creatorcontrib>Huang, Wei</creatorcontrib><creatorcontrib>Zhang, Siyuan</creatorcontrib><creatorcontrib>Zhang, Peng</creatorcontrib><creatorcontrib>Zha, Yufei</creatorcontrib><creatorcontrib>Fang, Yuming</creatorcontrib><creatorcontrib>Zhang, Yanning</creatorcontrib><title>Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images</title><title>IEEE transactions on multimedia</title><addtitle>TMM</addtitle><description>Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and improper measures of facial expressions, can further contribute to the above dilemma. To solve these problems, a novel identity-aware method is proposed in this study. Furthermore, this study also represents the first attempt to fulfill the challenging person-dependent facial expression recognition task based on deep metric learning and facial image synthesis techniques. Technically, a StarGAN is incorporated to synthesize facial images depicting different but complete basic emotions for each identity to augment the training data. Then, a deep-convolutional-neural-network-based network is employed to automatically extract latent features from both real facial images and all synthesized facial images. Next, a Mahalanobis metric network trained based on extracted latent features outputs a learned metric that measures facial expression differences between images, and the recognition task can thus be realized. Extensive experiments based on several well-known publicly available datasets are carried out in this study for performance evaluations. Person-dependent datasets, including CK+, Oulu (all 6 subdatasets), MMI, ISAFE, ISED, etc., are all incorporated. After comparing the new method with several popular or state-of-the-art facial expression recognition methods, its superiority in person-dependent facial expression recognition can be proposed from a statistical point of view.</description><subject>Datasets</subject><subject>Deep learning</subject><subject>Face recognition</subject><subject>facial expression recognition</subject><subject>Feature extraction</subject><subject>Generative adversarial networks</subject><subject>Image recognition</subject><subject>Image synthesis</subject><subject>Learning</subject><subject>Measurement</subject><subject>metric learning</subject><subject>Neural networks</subject><subject>Object recognition</subject><subject>Performance evaluation</subject><subject>person-dependent</subject><subject>Synthesis</subject><subject>Task analysis</subject><subject>Training</subject><issn>1520-9210</issn><issn>1941-0077</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNo9kE1LAzEQhoMoWKt3wUvA89ZJdjcfx6qtFloErV5DNjtbU9rdmmzR-uvd0uJp3oHnnYGHkGsGA8ZA381nswEHzgYpaAFCnZAe0xlLAKQ87XLOIdGcwTm5iHEJwLIcZI_YSYl169tdMvy2AenYOm9XdPSzCRijb2r6iq5Z1L7d5w9v6SPihs6wDd7RKdpQ-3pB723EknbE265uPzH6326drO0C4yU5q-wq4tVx9sn7eDR_eE6mL0-Th-E0cVyzNuFoUWAhoFIqq5xkiLzI8jQrlHAlc5VzsqyKAlzuZIGuzLB0KtWKW4aZ1Gmf3B7ubkLztcXYmmWzDXX30nChlFCp0rKj4EC50MQYsDKb4Nc27AwDsxdpOpFmL9IcRXaVm0PFI-I_rrufQufpH0mycHw</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Huang, Wei</creator><creator>Zhang, Siyuan</creator><creator>Zhang, Peng</creator><creator>Zha, Yufei</creator><creator>Fang, Yuming</creator><creator>Zhang, Yanning</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0002-2977-8057</orcidid><orcidid>https://orcid.org/0000-0002-0541-8612</orcidid><orcidid>https://orcid.org/0000-0001-9690-7026</orcidid></search><sort><creationdate>2022</creationdate><title>Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images</title><author>Huang, Wei ; Zhang, Siyuan ; Zhang, Peng ; Zha, Yufei ; Fang, Yuming ; Zhang, Yanning</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Datasets</topic><topic>Deep learning</topic><topic>Face recognition</topic><topic>facial expression recognition</topic><topic>Feature extraction</topic><topic>Generative adversarial networks</topic><topic>Image recognition</topic><topic>Image synthesis</topic><topic>Learning</topic><topic>Measurement</topic><topic>metric learning</topic><topic>Neural networks</topic><topic>Object recognition</topic><topic>Performance evaluation</topic><topic>person-dependent</topic><topic>Synthesis</topic><topic>Task analysis</topic><topic>Training</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Huang, Wei</creatorcontrib><creatorcontrib>Zhang, Siyuan</creatorcontrib><creatorcontrib>Zhang, Peng</creatorcontrib><creatorcontrib>Zha, Yufei</creatorcontrib><creatorcontrib>Fang, Yuming</creatorcontrib><creatorcontrib>Zhang, Yanning</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998–Present</collection><collection>IEEE Xplore</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on multimedia</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Huang, Wei</au><au>Zhang, Siyuan</au><au>Zhang, Peng</au><au>Zha, Yufei</au><au>Fang, Yuming</au><au>Zhang, Yanning</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images</atitle><jtitle>IEEE transactions on multimedia</jtitle><stitle>TMM</stitle><date>2022</date><risdate>2022</risdate><volume>24</volume><spage>3327</spage><epage>3339</epage><pages>3327-3339</pages><issn>1520-9210</issn><eissn>1941-0077</eissn><coden>ITMUF8</coden><abstract>Person-dependent facial expression recognition has received considerable research attention in recent years. Unfortunately, different identities can adversely influence recognition accuracy, and the recognition task becomes challenging. Other adverse factors, including limited training data and improper measures of facial expressions, can further contribute to the above dilemma. To solve these problems, a novel identity-aware method is proposed in this study. Furthermore, this study also represents the first attempt to fulfill the challenging person-dependent facial expression recognition task based on deep metric learning and facial image synthesis techniques. Technically, a StarGAN is incorporated to synthesize facial images depicting different but complete basic emotions for each identity to augment the training data. Then, a deep-convolutional-neural-network-based network is employed to automatically extract latent features from both real facial images and all synthesized facial images. Next, a Mahalanobis metric network trained based on extracted latent features outputs a learned metric that measures facial expression differences between images, and the recognition task can thus be realized. Extensive experiments based on several well-known publicly available datasets are carried out in this study for performance evaluations. Person-dependent datasets, including CK+, Oulu (all 6 subdatasets), MMI, ISAFE, ISED, etc., are all incorporated. After comparing the new method with several popular or state-of-the-art facial expression recognition methods, its superiority in person-dependent facial expression recognition can be proposed from a statistical point of view.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/TMM.2021.3096068</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0002-2977-8057</orcidid><orcidid>https://orcid.org/0000-0002-0541-8612</orcidid><orcidid>https://orcid.org/0000-0001-9690-7026</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1520-9210 |
ispartof | IEEE transactions on multimedia, 2022, Vol.24, p.3327-3339 |
issn | 1520-9210 1941-0077 |
language | eng |
recordid | cdi_proquest_journals_2688683897 |
source | IEEE Xplore (Online service) |
subjects | Datasets Deep learning Face recognition facial expression recognition Feature extraction Generative adversarial networks Image recognition Image synthesis Learning Measurement metric learning Neural networks Object recognition Performance evaluation person-dependent Synthesis Task analysis Training |
title | Identity-Aware Facial Expression Recognition Via Deep Metric Learning Based on Synthesized Images |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-19T21%3A54%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Identity-Aware%20Facial%20Expression%20Recognition%20Via%20Deep%20Metric%20Learning%20Based%20on%20Synthesized%20Images&rft.jtitle=IEEE%20transactions%20on%20multimedia&rft.au=Huang,%20Wei&rft.date=2022&rft.volume=24&rft.spage=3327&rft.epage=3339&rft.pages=3327-3339&rft.issn=1520-9210&rft.eissn=1941-0077&rft.coden=ITMUF8&rft_id=info:doi/10.1109/TMM.2021.3096068&rft_dat=%3Cproquest_ieee_%3E2688683897%3C/proquest_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c291t-2eae6eb60f884fc71ee2b4534b86cd1cfcc7dfbb0c5c7becd4edc83982a1e4793%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2688683897&rft_id=info:pmid/&rft_ieee_id=9479695&rfr_iscdi=true |