Loading…

Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation

We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to...

Full description

Saved in:
Bibliographic Details
Published in:Electronics (Basel) 2024-12, Vol.13 (23), p.4761
Main Authors: Lo, Shih-Lun, Cheng, Hsu-Yung, Yu, Chih-Chang
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703
container_end_page
container_issue 23
container_start_page 4761
container_title Electronics (Basel)
container_volume 13
creator Lo, Shih-Lun
Cheng, Hsu-Yung
Yu, Chih-Chang
description We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to calculate the error between the generated image and the real human face image. To avoid obvious color deviation in the generated images, we introduced perceptual color loss into the loss function. In addition, due to the lack of reasonable metrics to evaluate the quality of the animated images, we propose the use of Fréchet anime inception distance to calculate the distance between the distribution of the generated animated images and the real animated images in high-dimensional space, so as to understand the quality of the generated animated images. In the user survey, up to 74.46% of users think that the image produced by the proposed method is the best compared with other models. Also, the proposed method reaches a score of 126.05 for Fréchet anime inception distance. Our model performs the best in both user studies and FAID, showing that we have achieved better performance in human visual perception and model distribution. According to the experimental results and user feedback, our proposed method can generate results with better quality compared to existing methods.
doi_str_mv 10.3390/electronics13234761
format article
fullrecord <record><control><sourceid>gale_proqu</sourceid><recordid>TN_cdi_proquest_journals_3144067734</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A819847728</galeid><sourcerecordid>A819847728</sourcerecordid><originalsourceid>FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703</originalsourceid><addsrcrecordid>eNptUU1LAzEQXURB0f4CLwHP1XxsN5tjqbYKRUUUj8tsdraNbpOapC39Lf5ZU6vgwZnDzDzem8cwWXbO6KUQil5hhzp6Z40OTHCRy4IdZCecStVXXPHDP_1x1gvhjaZQTJSCnmSfY4S48khe0czmERsy2uoOyQQteohmjWTYrNEH8AY6co9x4_w72Zg4J2PQO2wKtllAAp9Qu5k10ThLEkYe0WtcxlXijFznPLk2IYLVSNo03Nj5rm92a5KJNQv4Vv46O3uWHbXQBez91NPsZXzzPLrtTx8md6PhtK-Z4qwv24LRXDFV0IIyUcMAaqXlgCM2qLmkraol54q1muYwkHVRgCilFiB5TSUVp9nFfu_Su48Vhli9uZW3ybISLM9pIaXIE-tyz5pBh5WxrYsedMoGF0Y7i61J-LBkqsyl5GUSiL1AexeCx7Za-nSk31aMVrvPVf98TnwBc6-Qdg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3144067734</pqid></control><display><type>article</type><title>Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation</title><source>Publicly Available Content Database</source><creator>Lo, Shih-Lun ; Cheng, Hsu-Yung ; Yu, Chih-Chang</creator><creatorcontrib>Lo, Shih-Lun ; Cheng, Hsu-Yung ; Yu, Chih-Chang</creatorcontrib><description>We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to calculate the error between the generated image and the real human face image. To avoid obvious color deviation in the generated images, we introduced perceptual color loss into the loss function. In addition, due to the lack of reasonable metrics to evaluate the quality of the animated images, we propose the use of Fréchet anime inception distance to calculate the distance between the distribution of the generated animated images and the real animated images in high-dimensional space, so as to understand the quality of the generated animated images. In the user survey, up to 74.46% of users think that the image produced by the proposed method is the best compared with other models. Also, the proposed method reaches a score of 126.05 for Fréchet anime inception distance. Our model performs the best in both user studies and FAID, showing that we have achieved better performance in human visual perception and model distribution. According to the experimental results and user feedback, our proposed method can generate results with better quality compared to existing methods.</description><identifier>ISSN: 2079-9292</identifier><identifier>EISSN: 2079-9292</identifier><identifier>DOI: 10.3390/electronics13234761</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Analysis ; Animation ; Anime ; Anime (Animation) ; Color ; Datasets ; Deep learning ; Generative adversarial networks ; Human performance ; Image quality ; Social networks ; Visual perception</subject><ispartof>Electronics (Basel), 2024-12, Vol.13 (23), p.4761</ispartof><rights>COPYRIGHT 2024 MDPI AG</rights><rights>2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703</cites><orcidid>0000-0002-8342-7450 ; 0000-0003-1611-0223</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/3144067734/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/3144067734?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,25752,27923,27924,37011,44589,74997</link.rule.ids></links><search><creatorcontrib>Lo, Shih-Lun</creatorcontrib><creatorcontrib>Cheng, Hsu-Yung</creatorcontrib><creatorcontrib>Yu, Chih-Chang</creatorcontrib><title>Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation</title><title>Electronics (Basel)</title><description>We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to calculate the error between the generated image and the real human face image. To avoid obvious color deviation in the generated images, we introduced perceptual color loss into the loss function. In addition, due to the lack of reasonable metrics to evaluate the quality of the animated images, we propose the use of Fréchet anime inception distance to calculate the distance between the distribution of the generated animated images and the real animated images in high-dimensional space, so as to understand the quality of the generated animated images. In the user survey, up to 74.46% of users think that the image produced by the proposed method is the best compared with other models. Also, the proposed method reaches a score of 126.05 for Fréchet anime inception distance. Our model performs the best in both user studies and FAID, showing that we have achieved better performance in human visual perception and model distribution. According to the experimental results and user feedback, our proposed method can generate results with better quality compared to existing methods.</description><subject>Analysis</subject><subject>Animation</subject><subject>Anime</subject><subject>Anime (Animation)</subject><subject>Color</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Generative adversarial networks</subject><subject>Human performance</subject><subject>Image quality</subject><subject>Social networks</subject><subject>Visual perception</subject><issn>2079-9292</issn><issn>2079-9292</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNptUU1LAzEQXURB0f4CLwHP1XxsN5tjqbYKRUUUj8tsdraNbpOapC39Lf5ZU6vgwZnDzDzem8cwWXbO6KUQil5hhzp6Z40OTHCRy4IdZCecStVXXPHDP_1x1gvhjaZQTJSCnmSfY4S48khe0czmERsy2uoOyQQteohmjWTYrNEH8AY6co9x4_w72Zg4J2PQO2wKtllAAp9Qu5k10ThLEkYe0WtcxlXijFznPLk2IYLVSNo03Nj5rm92a5KJNQv4Vv46O3uWHbXQBez91NPsZXzzPLrtTx8md6PhtK-Z4qwv24LRXDFV0IIyUcMAaqXlgCM2qLmkraol54q1muYwkHVRgCilFiB5TSUVp9nFfu_Su48Vhli9uZW3ybISLM9pIaXIE-tyz5pBh5WxrYsedMoGF0Y7i61J-LBkqsyl5GUSiL1AexeCx7Za-nSk31aMVrvPVf98TnwBc6-Qdg</recordid><startdate>20241201</startdate><enddate>20241201</enddate><creator>Lo, Shih-Lun</creator><creator>Cheng, Hsu-Yung</creator><creator>Yu, Chih-Chang</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SP</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L7M</scope><scope>P5Z</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><orcidid>https://orcid.org/0000-0002-8342-7450</orcidid><orcidid>https://orcid.org/0000-0003-1611-0223</orcidid></search><sort><creationdate>20241201</creationdate><title>Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation</title><author>Lo, Shih-Lun ; Cheng, Hsu-Yung ; Yu, Chih-Chang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Analysis</topic><topic>Animation</topic><topic>Anime</topic><topic>Anime (Animation)</topic><topic>Color</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Generative adversarial networks</topic><topic>Human performance</topic><topic>Image quality</topic><topic>Social networks</topic><topic>Visual perception</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Lo, Shih-Lun</creatorcontrib><creatorcontrib>Cheng, Hsu-Yung</creatorcontrib><creatorcontrib>Yu, Chih-Chang</creatorcontrib><collection>CrossRef</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><jtitle>Electronics (Basel)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Lo, Shih-Lun</au><au>Cheng, Hsu-Yung</au><au>Yu, Chih-Chang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation</atitle><jtitle>Electronics (Basel)</jtitle><date>2024-12-01</date><risdate>2024</risdate><volume>13</volume><issue>23</issue><spage>4761</spage><pages>4761-</pages><issn>2079-9292</issn><eissn>2079-9292</eissn><abstract>We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to calculate the error between the generated image and the real human face image. To avoid obvious color deviation in the generated images, we introduced perceptual color loss into the loss function. In addition, due to the lack of reasonable metrics to evaluate the quality of the animated images, we propose the use of Fréchet anime inception distance to calculate the distance between the distribution of the generated animated images and the real animated images in high-dimensional space, so as to understand the quality of the generated animated images. In the user survey, up to 74.46% of users think that the image produced by the proposed method is the best compared with other models. Also, the proposed method reaches a score of 126.05 for Fréchet anime inception distance. Our model performs the best in both user studies and FAID, showing that we have achieved better performance in human visual perception and model distribution. According to the experimental results and user feedback, our proposed method can generate results with better quality compared to existing methods.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/electronics13234761</doi><orcidid>https://orcid.org/0000-0002-8342-7450</orcidid><orcidid>https://orcid.org/0000-0003-1611-0223</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2079-9292
ispartof Electronics (Basel), 2024-12, Vol.13 (23), p.4761
issn 2079-9292
2079-9292
language eng
recordid cdi_proquest_journals_3144067734
source Publicly Available Content Database
subjects Analysis
Animation
Anime
Anime (Animation)
Color
Datasets
Deep learning
Generative adversarial networks
Human performance
Image quality
Social networks
Visual perception
title Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T01%3A54%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Feature%20Weighted%20Cycle%20Generative%20Adversarial%20Network%20with%20Facial%20Landmark%20Recognition%20and%20Perceptual%20Color%20Distance%20for%20Enhanced%20Face%20Animation%20Generation&rft.jtitle=Electronics%20(Basel)&rft.au=Lo,%20Shih-Lun&rft.date=2024-12-01&rft.volume=13&rft.issue=23&rft.spage=4761&rft.pages=4761-&rft.issn=2079-9292&rft.eissn=2079-9292&rft_id=info:doi/10.3390/electronics13234761&rft_dat=%3Cgale_proqu%3EA819847728%3C/gale_proqu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3144067734&rft_id=info:pmid/&rft_galeid=A819847728&rfr_iscdi=true