Loading…
Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation
We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to...
Saved in:
Published in: | Electronics (Basel) 2024-12, Vol.13 (23), p.4761 |
---|---|
Main Authors: | , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703 |
container_end_page | |
container_issue | 23 |
container_start_page | 4761 |
container_title | Electronics (Basel) |
container_volume | 13 |
creator | Lo, Shih-Lun Cheng, Hsu-Yung Yu, Chih-Chang |
description | We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to calculate the error between the generated image and the real human face image. To avoid obvious color deviation in the generated images, we introduced perceptual color loss into the loss function. In addition, due to the lack of reasonable metrics to evaluate the quality of the animated images, we propose the use of Fréchet anime inception distance to calculate the distance between the distribution of the generated animated images and the real animated images in high-dimensional space, so as to understand the quality of the generated animated images. In the user survey, up to 74.46% of users think that the image produced by the proposed method is the best compared with other models. Also, the proposed method reaches a score of 126.05 for Fréchet anime inception distance. Our model performs the best in both user studies and FAID, showing that we have achieved better performance in human visual perception and model distribution. According to the experimental results and user feedback, our proposed method can generate results with better quality compared to existing methods. |
doi_str_mv | 10.3390/electronics13234761 |
format | article |
fullrecord | <record><control><sourceid>gale_proqu</sourceid><recordid>TN_cdi_proquest_journals_3144067734</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A819847728</galeid><sourcerecordid>A819847728</sourcerecordid><originalsourceid>FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703</originalsourceid><addsrcrecordid>eNptUU1LAzEQXURB0f4CLwHP1XxsN5tjqbYKRUUUj8tsdraNbpOapC39Lf5ZU6vgwZnDzDzem8cwWXbO6KUQil5hhzp6Z40OTHCRy4IdZCecStVXXPHDP_1x1gvhjaZQTJSCnmSfY4S48khe0czmERsy2uoOyQQteohmjWTYrNEH8AY6co9x4_w72Zg4J2PQO2wKtllAAp9Qu5k10ThLEkYe0WtcxlXijFznPLk2IYLVSNo03Nj5rm92a5KJNQv4Vv46O3uWHbXQBez91NPsZXzzPLrtTx8md6PhtK-Z4qwv24LRXDFV0IIyUcMAaqXlgCM2qLmkraol54q1muYwkHVRgCilFiB5TSUVp9nFfu_Su48Vhli9uZW3ybISLM9pIaXIE-tyz5pBh5WxrYsedMoGF0Y7i61J-LBkqsyl5GUSiL1AexeCx7Za-nSk31aMVrvPVf98TnwBc6-Qdg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3144067734</pqid></control><display><type>article</type><title>Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation</title><source>Publicly Available Content Database</source><creator>Lo, Shih-Lun ; Cheng, Hsu-Yung ; Yu, Chih-Chang</creator><creatorcontrib>Lo, Shih-Lun ; Cheng, Hsu-Yung ; Yu, Chih-Chang</creatorcontrib><description>We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to calculate the error between the generated image and the real human face image. To avoid obvious color deviation in the generated images, we introduced perceptual color loss into the loss function. In addition, due to the lack of reasonable metrics to evaluate the quality of the animated images, we propose the use of Fréchet anime inception distance to calculate the distance between the distribution of the generated animated images and the real animated images in high-dimensional space, so as to understand the quality of the generated animated images. In the user survey, up to 74.46% of users think that the image produced by the proposed method is the best compared with other models. Also, the proposed method reaches a score of 126.05 for Fréchet anime inception distance. Our model performs the best in both user studies and FAID, showing that we have achieved better performance in human visual perception and model distribution. According to the experimental results and user feedback, our proposed method can generate results with better quality compared to existing methods.</description><identifier>ISSN: 2079-9292</identifier><identifier>EISSN: 2079-9292</identifier><identifier>DOI: 10.3390/electronics13234761</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Analysis ; Animation ; Anime ; Anime (Animation) ; Color ; Datasets ; Deep learning ; Generative adversarial networks ; Human performance ; Image quality ; Social networks ; Visual perception</subject><ispartof>Electronics (Basel), 2024-12, Vol.13 (23), p.4761</ispartof><rights>COPYRIGHT 2024 MDPI AG</rights><rights>2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703</cites><orcidid>0000-0002-8342-7450 ; 0000-0003-1611-0223</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.proquest.com/docview/3144067734/fulltextPDF?pq-origsite=primo$$EPDF$$P50$$Gproquest$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.proquest.com/docview/3144067734?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,25752,27923,27924,37011,44589,74997</link.rule.ids></links><search><creatorcontrib>Lo, Shih-Lun</creatorcontrib><creatorcontrib>Cheng, Hsu-Yung</creatorcontrib><creatorcontrib>Yu, Chih-Chang</creatorcontrib><title>Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation</title><title>Electronics (Basel)</title><description>We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to calculate the error between the generated image and the real human face image. To avoid obvious color deviation in the generated images, we introduced perceptual color loss into the loss function. In addition, due to the lack of reasonable metrics to evaluate the quality of the animated images, we propose the use of Fréchet anime inception distance to calculate the distance between the distribution of the generated animated images and the real animated images in high-dimensional space, so as to understand the quality of the generated animated images. In the user survey, up to 74.46% of users think that the image produced by the proposed method is the best compared with other models. Also, the proposed method reaches a score of 126.05 for Fréchet anime inception distance. Our model performs the best in both user studies and FAID, showing that we have achieved better performance in human visual perception and model distribution. According to the experimental results and user feedback, our proposed method can generate results with better quality compared to existing methods.</description><subject>Analysis</subject><subject>Animation</subject><subject>Anime</subject><subject>Anime (Animation)</subject><subject>Color</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Generative adversarial networks</subject><subject>Human performance</subject><subject>Image quality</subject><subject>Social networks</subject><subject>Visual perception</subject><issn>2079-9292</issn><issn>2079-9292</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNptUU1LAzEQXURB0f4CLwHP1XxsN5tjqbYKRUUUj8tsdraNbpOapC39Lf5ZU6vgwZnDzDzem8cwWXbO6KUQil5hhzp6Z40OTHCRy4IdZCecStVXXPHDP_1x1gvhjaZQTJSCnmSfY4S48khe0czmERsy2uoOyQQteohmjWTYrNEH8AY6co9x4_w72Zg4J2PQO2wKtllAAp9Qu5k10ThLEkYe0WtcxlXijFznPLk2IYLVSNo03Nj5rm92a5KJNQv4Vv46O3uWHbXQBez91NPsZXzzPLrtTx8md6PhtK-Z4qwv24LRXDFV0IIyUcMAaqXlgCM2qLmkraol54q1muYwkHVRgCilFiB5TSUVp9nFfu_Su48Vhli9uZW3ybISLM9pIaXIE-tyz5pBh5WxrYsedMoGF0Y7i61J-LBkqsyl5GUSiL1AexeCx7Za-nSk31aMVrvPVf98TnwBc6-Qdg</recordid><startdate>20241201</startdate><enddate>20241201</enddate><creator>Lo, Shih-Lun</creator><creator>Cheng, Hsu-Yung</creator><creator>Yu, Chih-Chang</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SP</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L7M</scope><scope>P5Z</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><orcidid>https://orcid.org/0000-0002-8342-7450</orcidid><orcidid>https://orcid.org/0000-0003-1611-0223</orcidid></search><sort><creationdate>20241201</creationdate><title>Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation</title><author>Lo, Shih-Lun ; Cheng, Hsu-Yung ; Yu, Chih-Chang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Analysis</topic><topic>Animation</topic><topic>Anime</topic><topic>Anime (Animation)</topic><topic>Color</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Generative adversarial networks</topic><topic>Human performance</topic><topic>Image quality</topic><topic>Social networks</topic><topic>Visual perception</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Lo, Shih-Lun</creatorcontrib><creatorcontrib>Cheng, Hsu-Yung</creatorcontrib><creatorcontrib>Yu, Chih-Chang</creatorcontrib><collection>CrossRef</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><jtitle>Electronics (Basel)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Lo, Shih-Lun</au><au>Cheng, Hsu-Yung</au><au>Yu, Chih-Chang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation</atitle><jtitle>Electronics (Basel)</jtitle><date>2024-12-01</date><risdate>2024</risdate><volume>13</volume><issue>23</issue><spage>4761</spage><pages>4761-</pages><issn>2079-9292</issn><eissn>2079-9292</eissn><abstract>We propose an anime style transfer model to generate anime faces from human face images. We improve the model by modifying the normalization function to obtain more feature information. To make the face feature position of the anime face similar to the human face, we propose facial landmark loss to calculate the error between the generated image and the real human face image. To avoid obvious color deviation in the generated images, we introduced perceptual color loss into the loss function. In addition, due to the lack of reasonable metrics to evaluate the quality of the animated images, we propose the use of Fréchet anime inception distance to calculate the distance between the distribution of the generated animated images and the real animated images in high-dimensional space, so as to understand the quality of the generated animated images. In the user survey, up to 74.46% of users think that the image produced by the proposed method is the best compared with other models. Also, the proposed method reaches a score of 126.05 for Fréchet anime inception distance. Our model performs the best in both user studies and FAID, showing that we have achieved better performance in human visual perception and model distribution. According to the experimental results and user feedback, our proposed method can generate results with better quality compared to existing methods.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/electronics13234761</doi><orcidid>https://orcid.org/0000-0002-8342-7450</orcidid><orcidid>https://orcid.org/0000-0003-1611-0223</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2079-9292 |
ispartof | Electronics (Basel), 2024-12, Vol.13 (23), p.4761 |
issn | 2079-9292 2079-9292 |
language | eng |
recordid | cdi_proquest_journals_3144067734 |
source | Publicly Available Content Database |
subjects | Analysis Animation Anime Anime (Animation) Color Datasets Deep learning Generative adversarial networks Human performance Image quality Social networks Visual perception |
title | Feature Weighted Cycle Generative Adversarial Network with Facial Landmark Recognition and Perceptual Color Distance for Enhanced Face Animation Generation |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T01%3A54%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Feature%20Weighted%20Cycle%20Generative%20Adversarial%20Network%20with%20Facial%20Landmark%20Recognition%20and%20Perceptual%20Color%20Distance%20for%20Enhanced%20Face%20Animation%20Generation&rft.jtitle=Electronics%20(Basel)&rft.au=Lo,%20Shih-Lun&rft.date=2024-12-01&rft.volume=13&rft.issue=23&rft.spage=4761&rft.pages=4761-&rft.issn=2079-9292&rft.eissn=2079-9292&rft_id=info:doi/10.3390/electronics13234761&rft_dat=%3Cgale_proqu%3EA819847728%3C/gale_proqu%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c1921-7f6104919606013ba5ab9c752eedec270f9b72291fc04a57b66a387c3a72b0703%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3144067734&rft_id=info:pmid/&rft_galeid=A819847728&rfr_iscdi=true |