Loading…

Implicit Functions in Feature Space for 3D Shape Reconstruction and Completion

While many works focus on 3D reconstruction from images, in this paper, we focus on 3D shape reconstruction and completion from a variety of 3D inputs, which are deficient in some respect: low and high resolution voxels, sparse and dense point clouds, complete or incomplete. Processing of such 3D in...

Full description

Saved in:
Bibliographic Details
Main Authors: Chibane, Julian, Alldieck, Thiemo, Pons-Moll, Gerard
Format: Conference Proceeding
Language:English
Subjects:
Online Access:Request full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 6979
container_issue
container_start_page 6968
container_title
container_volume
creator Chibane, Julian
Alldieck, Thiemo
Pons-Moll, Gerard
description While many works focus on 3D reconstruction from images, in this paper, we focus on 3D shape reconstruction and completion from a variety of 3D inputs, which are deficient in some respect: low and high resolution voxels, sparse and dense point clouds, complete or incomplete. Processing of such 3D inputs is an increasingly important problem as they are the output of 3D scanners, which are becoming more accessible, and are the intermediate output of 3D computer vision algorithms. Recently, learned implicit functions have shown great promise as they produce continuous reconstructions. However, we identified two limitations in reconstruction from 3D inputs: 1) details present in the input data are not retained, and 2) poor reconstruction of articulated humans. To solve this, we propose Implicit Feature Networks (IF-Nets), which deliver continuous outputs, can handle multiple topologies, and complete shapes for missing or sparse input data retaining the nice properties of recent learned implicit functions, but critically they can also retain detail when it is present in the input data, and can reconstruct articulated humans. Our work differs from prior work in two crucial aspects. First, instead of using a single vector to encode a 3D shape, we extract a learnable 3-dimensional multi-scale tensor of deep features, which is aligned with the original Euclidean space embedding the shape. Second, instead of classifying x-y-z point coordinates directly, we classify deep features extracted from the tensor at a continuous query point. We show that this forces our model to make decisions based on global and local shape structure, as opposed to point coordinates, which are arbitrary under Euclidean transformations. Experiments demonstrate that IF-Nets outperform prior work in 3D object reconstruction in ShapeNet, and obtain significantly more accurate 3D human reconstructions. Code and project website is available at https://virtualhumans.mpi-inf.mpg.de/ifnets/.
doi_str_mv 10.1109/CVPR42600.2020.00700
format conference_proceeding
fullrecord <record><control><sourceid>ieee_CHZPO</sourceid><recordid>TN_cdi_ieee_primary_9156292</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9156292</ieee_id><sourcerecordid>9156292</sourcerecordid><originalsourceid>FETCH-LOGICAL-i249t-701f91eb7bcd5c659aef8e54042af97601e2ff0072c087f31eb8c1f0cabf09c3</originalsourceid><addsrcrecordid>eNotj91KxDAUhKMguKx9Ar3IC7SekzZJcynVuguLyu7i7ZJmTzCy25b-XPj2xp-LYRj4GGYYu0PIEMHcV-9v20IogEyAgAxAA1ywxOgStYhCVcpLthBSy1SDltcsGcdPAMgFojLlgr2sz_0puDDxem7dFLp25KHlNdlpHojveuuI-27g-SPffdie-JZchKZh_qW5bY-86mIJ_cQbduXtaaTk35dsXz_tq1W6eX1eVw-bNIjCTHELeoPU6MYdpVPSWPIlyQIKYb3RCpCE9_GOcFBqn0e0dOjB2caDcfmS3f7VBiI69EM42-HrYFAqYUT-DY_ET7o</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Implicit Functions in Feature Space for 3D Shape Reconstruction and Completion</title><source>IEEE Xplore All Conference Series</source><creator>Chibane, Julian ; Alldieck, Thiemo ; Pons-Moll, Gerard</creator><creatorcontrib>Chibane, Julian ; Alldieck, Thiemo ; Pons-Moll, Gerard</creatorcontrib><description>While many works focus on 3D reconstruction from images, in this paper, we focus on 3D shape reconstruction and completion from a variety of 3D inputs, which are deficient in some respect: low and high resolution voxels, sparse and dense point clouds, complete or incomplete. Processing of such 3D inputs is an increasingly important problem as they are the output of 3D scanners, which are becoming more accessible, and are the intermediate output of 3D computer vision algorithms. Recently, learned implicit functions have shown great promise as they produce continuous reconstructions. However, we identified two limitations in reconstruction from 3D inputs: 1) details present in the input data are not retained, and 2) poor reconstruction of articulated humans. To solve this, we propose Implicit Feature Networks (IF-Nets), which deliver continuous outputs, can handle multiple topologies, and complete shapes for missing or sparse input data retaining the nice properties of recent learned implicit functions, but critically they can also retain detail when it is present in the input data, and can reconstruct articulated humans. Our work differs from prior work in two crucial aspects. First, instead of using a single vector to encode a 3D shape, we extract a learnable 3-dimensional multi-scale tensor of deep features, which is aligned with the original Euclidean space embedding the shape. Second, instead of classifying x-y-z point coordinates directly, we classify deep features extracted from the tensor at a continuous query point. We show that this forces our model to make decisions based on global and local shape structure, as opposed to point coordinates, which are arbitrary under Euclidean transformations. Experiments demonstrate that IF-Nets outperform prior work in 3D object reconstruction in ShapeNet, and obtain significantly more accurate 3D human reconstructions. Code and project website is available at https://virtualhumans.mpi-inf.mpg.de/ifnets/.</description><identifier>EISSN: 2575-7075</identifier><identifier>EISBN: 9781728171685</identifier><identifier>EISBN: 1728171687</identifier><identifier>DOI: 10.1109/CVPR42600.2020.00700</identifier><identifier>CODEN: IEEPAD</identifier><language>eng</language><publisher>IEEE</publisher><subject>Computer vision ; Feature extraction ; Image reconstruction ; Shape ; Tensile stress ; Three-dimensional displays ; Topology</subject><ispartof>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, p.6968-6979</ispartof><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9156292$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,776,780,785,786,27904,54534,54911</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9156292$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Chibane, Julian</creatorcontrib><creatorcontrib>Alldieck, Thiemo</creatorcontrib><creatorcontrib>Pons-Moll, Gerard</creatorcontrib><title>Implicit Functions in Feature Space for 3D Shape Reconstruction and Completion</title><title>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</title><addtitle>CVPR</addtitle><description>While many works focus on 3D reconstruction from images, in this paper, we focus on 3D shape reconstruction and completion from a variety of 3D inputs, which are deficient in some respect: low and high resolution voxels, sparse and dense point clouds, complete or incomplete. Processing of such 3D inputs is an increasingly important problem as they are the output of 3D scanners, which are becoming more accessible, and are the intermediate output of 3D computer vision algorithms. Recently, learned implicit functions have shown great promise as they produce continuous reconstructions. However, we identified two limitations in reconstruction from 3D inputs: 1) details present in the input data are not retained, and 2) poor reconstruction of articulated humans. To solve this, we propose Implicit Feature Networks (IF-Nets), which deliver continuous outputs, can handle multiple topologies, and complete shapes for missing or sparse input data retaining the nice properties of recent learned implicit functions, but critically they can also retain detail when it is present in the input data, and can reconstruct articulated humans. Our work differs from prior work in two crucial aspects. First, instead of using a single vector to encode a 3D shape, we extract a learnable 3-dimensional multi-scale tensor of deep features, which is aligned with the original Euclidean space embedding the shape. Second, instead of classifying x-y-z point coordinates directly, we classify deep features extracted from the tensor at a continuous query point. We show that this forces our model to make decisions based on global and local shape structure, as opposed to point coordinates, which are arbitrary under Euclidean transformations. Experiments demonstrate that IF-Nets outperform prior work in 3D object reconstruction in ShapeNet, and obtain significantly more accurate 3D human reconstructions. Code and project website is available at https://virtualhumans.mpi-inf.mpg.de/ifnets/.</description><subject>Computer vision</subject><subject>Feature extraction</subject><subject>Image reconstruction</subject><subject>Shape</subject><subject>Tensile stress</subject><subject>Three-dimensional displays</subject><subject>Topology</subject><issn>2575-7075</issn><isbn>9781728171685</isbn><isbn>1728171687</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2020</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><recordid>eNotj91KxDAUhKMguKx9Ar3IC7SekzZJcynVuguLyu7i7ZJmTzCy25b-XPj2xp-LYRj4GGYYu0PIEMHcV-9v20IogEyAgAxAA1ywxOgStYhCVcpLthBSy1SDltcsGcdPAMgFojLlgr2sz_0puDDxem7dFLp25KHlNdlpHojveuuI-27g-SPffdie-JZchKZh_qW5bY-86mIJ_cQbduXtaaTk35dsXz_tq1W6eX1eVw-bNIjCTHELeoPU6MYdpVPSWPIlyQIKYb3RCpCE9_GOcFBqn0e0dOjB2caDcfmS3f7VBiI69EM42-HrYFAqYUT-DY_ET7o</recordid><startdate>20200101</startdate><enddate>20200101</enddate><creator>Chibane, Julian</creator><creator>Alldieck, Thiemo</creator><creator>Pons-Moll, Gerard</creator><general>IEEE</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>20200101</creationdate><title>Implicit Functions in Feature Space for 3D Shape Reconstruction and Completion</title><author>Chibane, Julian ; Alldieck, Thiemo ; Pons-Moll, Gerard</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i249t-701f91eb7bcd5c659aef8e54042af97601e2ff0072c087f31eb8c1f0cabf09c3</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer vision</topic><topic>Feature extraction</topic><topic>Image reconstruction</topic><topic>Shape</topic><topic>Tensile stress</topic><topic>Three-dimensional displays</topic><topic>Topology</topic><toplevel>online_resources</toplevel><creatorcontrib>Chibane, Julian</creatorcontrib><creatorcontrib>Alldieck, Thiemo</creatorcontrib><creatorcontrib>Pons-Moll, Gerard</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE/IET Electronic Library (IEL)</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chibane, Julian</au><au>Alldieck, Thiemo</au><au>Pons-Moll, Gerard</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Implicit Functions in Feature Space for 3D Shape Reconstruction and Completion</atitle><btitle>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</btitle><stitle>CVPR</stitle><date>2020-01-01</date><risdate>2020</risdate><spage>6968</spage><epage>6979</epage><pages>6968-6979</pages><eissn>2575-7075</eissn><eisbn>9781728171685</eisbn><eisbn>1728171687</eisbn><coden>IEEPAD</coden><abstract>While many works focus on 3D reconstruction from images, in this paper, we focus on 3D shape reconstruction and completion from a variety of 3D inputs, which are deficient in some respect: low and high resolution voxels, sparse and dense point clouds, complete or incomplete. Processing of such 3D inputs is an increasingly important problem as they are the output of 3D scanners, which are becoming more accessible, and are the intermediate output of 3D computer vision algorithms. Recently, learned implicit functions have shown great promise as they produce continuous reconstructions. However, we identified two limitations in reconstruction from 3D inputs: 1) details present in the input data are not retained, and 2) poor reconstruction of articulated humans. To solve this, we propose Implicit Feature Networks (IF-Nets), which deliver continuous outputs, can handle multiple topologies, and complete shapes for missing or sparse input data retaining the nice properties of recent learned implicit functions, but critically they can also retain detail when it is present in the input data, and can reconstruct articulated humans. Our work differs from prior work in two crucial aspects. First, instead of using a single vector to encode a 3D shape, we extract a learnable 3-dimensional multi-scale tensor of deep features, which is aligned with the original Euclidean space embedding the shape. Second, instead of classifying x-y-z point coordinates directly, we classify deep features extracted from the tensor at a continuous query point. We show that this forces our model to make decisions based on global and local shape structure, as opposed to point coordinates, which are arbitrary under Euclidean transformations. Experiments demonstrate that IF-Nets outperform prior work in 3D object reconstruction in ShapeNet, and obtain significantly more accurate 3D human reconstructions. Code and project website is available at https://virtualhumans.mpi-inf.mpg.de/ifnets/.</abstract><pub>IEEE</pub><doi>10.1109/CVPR42600.2020.00700</doi><tpages>12</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 2575-7075
ispartof 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, p.6968-6979
issn 2575-7075
language eng
recordid cdi_ieee_primary_9156292
source IEEE Xplore All Conference Series
subjects Computer vision
Feature extraction
Image reconstruction
Shape
Tensile stress
Three-dimensional displays
Topology
title Implicit Functions in Feature Space for 3D Shape Reconstruction and Completion
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-23T05%3A33%3A23IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_CHZPO&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Implicit%20Functions%20in%20Feature%20Space%20for%203D%20Shape%20Reconstruction%20and%20Completion&rft.btitle=2020%20IEEE/CVF%20Conference%20on%20Computer%20Vision%20and%20Pattern%20Recognition%20(CVPR)&rft.au=Chibane,%20Julian&rft.date=2020-01-01&rft.spage=6968&rft.epage=6979&rft.pages=6968-6979&rft.eissn=2575-7075&rft.coden=IEEPAD&rft_id=info:doi/10.1109/CVPR42600.2020.00700&rft.eisbn=9781728171685&rft.eisbn_list=1728171687&rft_dat=%3Cieee_CHZPO%3E9156292%3C/ieee_CHZPO%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-i249t-701f91eb7bcd5c659aef8e54042af97601e2ff0072c087f31eb8c1f0cabf09c3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=9156292&rfr_iscdi=true