Loading…

3D Face Tracking from 2D Video through Iterative Dense UV to Image Flow

When working with 3D facial data, improving fidelity and avoiding the uncanny valley effect is critically dependent on accurate 3D facial performance capture. Because such methods are expensive and due to the widespread availability of 2D videos, recent methods have focused on how to perform monocul...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2024-04
Main Authors: Taubner, Felix, Raina, Prashant, Tuli, Mathieu, Teh, Eu Wern, Lee, Chul, Huang, Jinmiao
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Taubner, Felix
Raina, Prashant
Tuli, Mathieu
Teh, Eu Wern
Lee, Chul
Huang, Jinmiao
description When working with 3D facial data, improving fidelity and avoiding the uncanny valley effect is critically dependent on accurate 3D facial performance capture. Because such methods are expensive and due to the widespread availability of 2D videos, recent methods have focused on how to perform monocular 3D face tracking. However, these methods often fall short in capturing precise facial movements due to limitations in their network architecture, training, and evaluation processes. Addressing these challenges, we propose a novel face tracker, FlowFace, that introduces an innovative 2D alignment network for dense per-vertex alignment. Unlike prior work, FlowFace is trained on high-quality 3D scan annotations rather than weak supervision or synthetic data. Our 3D model fitting module jointly fits a 3D face model from one or many observations, integrating existing neutral shape priors for enhanced identity and expression disentanglement and per-vertex deformations for detailed facial feature reconstruction. Additionally, we propose a novel metric and benchmark for assessing tracking accuracy. Our method exhibits superior performance on both custom and publicly available benchmarks. We further validate the effectiveness of our tracker by generating high-quality 3D data from 2D videos, which leads to performance gains on downstream tasks.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3039627919</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3039627919</sourcerecordid><originalsourceid>FETCH-proquest_journals_30396279193</originalsourceid><addsrcrecordid>eNqNyrsOgjAUgOHGxESivMNJnElKy0VmEWVHVtLg4SZwtC36-jr4AE7_8P0r5ggpfe8QCLFhrjED51xEsQhD6bCzTCFTNUKhVX3v5xYaTROIFMr-hgS207S0HeQWtbL9CyHF2SBcS7AE-aRahGyk946tGzUadH_dsn12Ko4X76HpuaCx1UCLnr9USS6TSMSJn8j_rg_gOzmO</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3039627919</pqid></control><display><type>article</type><title>3D Face Tracking from 2D Video through Iterative Dense UV to Image Flow</title><source>Publicly Available Content Database</source><creator>Taubner, Felix ; Raina, Prashant ; Tuli, Mathieu ; Teh, Eu Wern ; Lee, Chul ; Huang, Jinmiao</creator><creatorcontrib>Taubner, Felix ; Raina, Prashant ; Tuli, Mathieu ; Teh, Eu Wern ; Lee, Chul ; Huang, Jinmiao</creatorcontrib><description>When working with 3D facial data, improving fidelity and avoiding the uncanny valley effect is critically dependent on accurate 3D facial performance capture. Because such methods are expensive and due to the widespread availability of 2D videos, recent methods have focused on how to perform monocular 3D face tracking. However, these methods often fall short in capturing precise facial movements due to limitations in their network architecture, training, and evaluation processes. Addressing these challenges, we propose a novel face tracker, FlowFace, that introduces an innovative 2D alignment network for dense per-vertex alignment. Unlike prior work, FlowFace is trained on high-quality 3D scan annotations rather than weak supervision or synthetic data. Our 3D model fitting module jointly fits a 3D face model from one or many observations, integrating existing neutral shape priors for enhanced identity and expression disentanglement and per-vertex deformations for detailed facial feature reconstruction. Additionally, we propose a novel metric and benchmark for assessing tracking accuracy. Our method exhibits superior performance on both custom and publicly available benchmarks. We further validate the effectiveness of our tracker by generating high-quality 3D data from 2D videos, which leads to performance gains on downstream tasks.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Alignment ; Annotations ; Availability ; Benchmarks ; Image reconstruction ; Synthetic data ; Three dimensional flow ; Three dimensional models ; Tracking ; Video</subject><ispartof>arXiv.org, 2024-04</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/3039627919?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25744,37003,44581</link.rule.ids></links><search><creatorcontrib>Taubner, Felix</creatorcontrib><creatorcontrib>Raina, Prashant</creatorcontrib><creatorcontrib>Tuli, Mathieu</creatorcontrib><creatorcontrib>Teh, Eu Wern</creatorcontrib><creatorcontrib>Lee, Chul</creatorcontrib><creatorcontrib>Huang, Jinmiao</creatorcontrib><title>3D Face Tracking from 2D Video through Iterative Dense UV to Image Flow</title><title>arXiv.org</title><description>When working with 3D facial data, improving fidelity and avoiding the uncanny valley effect is critically dependent on accurate 3D facial performance capture. Because such methods are expensive and due to the widespread availability of 2D videos, recent methods have focused on how to perform monocular 3D face tracking. However, these methods often fall short in capturing precise facial movements due to limitations in their network architecture, training, and evaluation processes. Addressing these challenges, we propose a novel face tracker, FlowFace, that introduces an innovative 2D alignment network for dense per-vertex alignment. Unlike prior work, FlowFace is trained on high-quality 3D scan annotations rather than weak supervision or synthetic data. Our 3D model fitting module jointly fits a 3D face model from one or many observations, integrating existing neutral shape priors for enhanced identity and expression disentanglement and per-vertex deformations for detailed facial feature reconstruction. Additionally, we propose a novel metric and benchmark for assessing tracking accuracy. Our method exhibits superior performance on both custom and publicly available benchmarks. We further validate the effectiveness of our tracker by generating high-quality 3D data from 2D videos, which leads to performance gains on downstream tasks.</description><subject>Alignment</subject><subject>Annotations</subject><subject>Availability</subject><subject>Benchmarks</subject><subject>Image reconstruction</subject><subject>Synthetic data</subject><subject>Three dimensional flow</subject><subject>Three dimensional models</subject><subject>Tracking</subject><subject>Video</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNyrsOgjAUgOHGxESivMNJnElKy0VmEWVHVtLg4SZwtC36-jr4AE7_8P0r5ggpfe8QCLFhrjED51xEsQhD6bCzTCFTNUKhVX3v5xYaTROIFMr-hgS207S0HeQWtbL9CyHF2SBcS7AE-aRahGyk946tGzUadH_dsn12Ko4X76HpuaCx1UCLnr9USS6TSMSJn8j_rg_gOzmO</recordid><startdate>20240415</startdate><enddate>20240415</enddate><creator>Taubner, Felix</creator><creator>Raina, Prashant</creator><creator>Tuli, Mathieu</creator><creator>Teh, Eu Wern</creator><creator>Lee, Chul</creator><creator>Huang, Jinmiao</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20240415</creationdate><title>3D Face Tracking from 2D Video through Iterative Dense UV to Image Flow</title><author>Taubner, Felix ; Raina, Prashant ; Tuli, Mathieu ; Teh, Eu Wern ; Lee, Chul ; Huang, Jinmiao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_30396279193</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Alignment</topic><topic>Annotations</topic><topic>Availability</topic><topic>Benchmarks</topic><topic>Image reconstruction</topic><topic>Synthetic data</topic><topic>Three dimensional flow</topic><topic>Three dimensional models</topic><topic>Tracking</topic><topic>Video</topic><toplevel>online_resources</toplevel><creatorcontrib>Taubner, Felix</creatorcontrib><creatorcontrib>Raina, Prashant</creatorcontrib><creatorcontrib>Tuli, Mathieu</creatorcontrib><creatorcontrib>Teh, Eu Wern</creatorcontrib><creatorcontrib>Lee, Chul</creatorcontrib><creatorcontrib>Huang, Jinmiao</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Taubner, Felix</au><au>Raina, Prashant</au><au>Tuli, Mathieu</au><au>Teh, Eu Wern</au><au>Lee, Chul</au><au>Huang, Jinmiao</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>3D Face Tracking from 2D Video through Iterative Dense UV to Image Flow</atitle><jtitle>arXiv.org</jtitle><date>2024-04-15</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>When working with 3D facial data, improving fidelity and avoiding the uncanny valley effect is critically dependent on accurate 3D facial performance capture. Because such methods are expensive and due to the widespread availability of 2D videos, recent methods have focused on how to perform monocular 3D face tracking. However, these methods often fall short in capturing precise facial movements due to limitations in their network architecture, training, and evaluation processes. Addressing these challenges, we propose a novel face tracker, FlowFace, that introduces an innovative 2D alignment network for dense per-vertex alignment. Unlike prior work, FlowFace is trained on high-quality 3D scan annotations rather than weak supervision or synthetic data. Our 3D model fitting module jointly fits a 3D face model from one or many observations, integrating existing neutral shape priors for enhanced identity and expression disentanglement and per-vertex deformations for detailed facial feature reconstruction. Additionally, we propose a novel metric and benchmark for assessing tracking accuracy. Our method exhibits superior performance on both custom and publicly available benchmarks. We further validate the effectiveness of our tracker by generating high-quality 3D data from 2D videos, which leads to performance gains on downstream tasks.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-04
issn 2331-8422
language eng
recordid cdi_proquest_journals_3039627919
source Publicly Available Content Database
subjects Alignment
Annotations
Availability
Benchmarks
Image reconstruction
Synthetic data
Three dimensional flow
Three dimensional models
Tracking
Video
title 3D Face Tracking from 2D Video through Iterative Dense UV to Image Flow
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-15T05%3A56%3A49IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=3D%20Face%20Tracking%20from%202D%20Video%20through%20Iterative%20Dense%20UV%20to%20Image%20Flow&rft.jtitle=arXiv.org&rft.au=Taubner,%20Felix&rft.date=2024-04-15&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3039627919%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_30396279193%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=3039627919&rft_id=info:pmid/&rfr_iscdi=true