Loading…

T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition

The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements a...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on cybernetics 2024-11, p.1-11
Main Authors: Yang, Zhiyuan, Zhou, Yunjiao, Xie, Lihua, Yang, Jianfei
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page 11
container_issue
container_start_page 1
container_title IEEE transactions on cybernetics
container_volume
creator Yang, Zhiyuan
Zhou, Yunjiao
Xie, Lihua
Yang, Jianfei
description The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements and nonreal-time latency. There has been a lack of research on how to compress 3-D point cloud models into lightweight models. In this article, we propose a method called T3DNet (tiny 3-D network with augmentation and distillation) to address this issue. We find that the tiny model after network augmentation is much easier for a teacher to distill. Instead of gradually reducing the parameters through techniques, such as pruning or quantization, we predefine a tiny model and improve its performance through auxiliary supervision from augmented networks and the original model. We evaluate our method on several public datasets, including ModelNet40, ShapeNet, and ScanObjectNN. Our method can achieve high compression rates without significant accuracy sacrifice, achieving state-of-the-art performances on three datasets against existing methods. Amazingly, our T3DNet is 58 \times smaller and 54 \times faster than the original model yet with only 1.4 \% accuracy descent on the ModelNet40 dataset. Our code is available at https://github.com/Zhiyuan002/T3DNet.
doi_str_mv 10.1109/TCYB.2024.3487220
format article
fullrecord <record><control><sourceid>crossref_ieee_</sourceid><recordid>TN_cdi_ieee_primary_10760265</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10760265</ieee_id><sourcerecordid>10_1109_TCYB_2024_3487220</sourcerecordid><originalsourceid>FETCH-LOGICAL-c191t-4fb1aab7710a5f1a61d3d46e811b7600e0fe6ff2c913e5e10b27e7780c338f273</originalsourceid><addsrcrecordid>eNpNkN1KxDAQhYMouKz7AIIXeYHWTNImqXfadVWoP0i98Kr0Z1Ij3WZpKuLb27KLOBdnhmHOGfgIOQcWArDkMk_fb0LOeBSKSCvO2RFZcJA64FzFx3-zVKdk5f0nm0pPq0QvyCYX6yccr2jqtrsBvbd9S1-c7Ueadu6roY-uwc5T4waa2fZj_MZZqQjW9BVr1_Z2tK4_Iyem7DyuDn1J3ja3eXofZM93D-l1FtSQwBhEpoKyrJQCVsYGSgmNaCKJGqBSkjFkBqUxvE5AYIzAKq5QKc1qIbThSiwJ7HPrwXk_oCl2g92Ww08BrJhZFDOLYmZRHFhMnou9xyLiv_vpIZex-AX6FFmY</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Yang, Zhiyuan ; Zhou, Yunjiao ; Xie, Lihua ; Yang, Jianfei</creator><creatorcontrib>Yang, Zhiyuan ; Zhou, Yunjiao ; Xie, Lihua ; Yang, Jianfei</creatorcontrib><description><![CDATA[The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements and nonreal-time latency. There has been a lack of research on how to compress 3-D point cloud models into lightweight models. In this article, we propose a method called T3DNet (tiny 3-D network with augmentation and distillation) to address this issue. We find that the tiny model after network augmentation is much easier for a teacher to distill. Instead of gradually reducing the parameters through techniques, such as pruning or quantization, we predefine a tiny model and improve its performance through auxiliary supervision from augmented networks and the original model. We evaluate our method on several public datasets, including ModelNet40, ShapeNet, and ScanObjectNN. Our method can achieve high compression rates without significant accuracy sacrifice, achieving state-of-the-art performances on three datasets against existing methods. Amazingly, our T3DNet is 58<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> smaller and 54<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> faster than the original model yet with only 1.4<inline-formula> <tex-math notation="LaTeX">\%</tex-math> </inline-formula> accuracy descent on the ModelNet40 dataset. Our code is available at https://github.com/Zhiyuan002/T3DNet.]]></description><identifier>ISSN: 2168-2267</identifier><identifier>EISSN: 2168-2275</identifier><identifier>DOI: 10.1109/TCYB.2024.3487220</identifier><identifier>CODEN: ITCEB8</identifier><language>eng</language><publisher>IEEE</publisher><subject>3-D model compression ; Accuracy ; Computational modeling ; Feature extraction ; Image coding ; knowledge distillation (KD) ; Knowledge engineering ; network augmentation ; Point cloud compression ; point cloud models ; Quantization (signal) ; Sensors ; Solid modeling ; Three-dimensional displays</subject><ispartof>IEEE transactions on cybernetics, 2024-11, p.1-11</ispartof><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>yang0674@ntu.edu.sg ; elhxie@ntu.edu.sg ; yunjiao001@e.ntu.edu.sg ; YANG0478@e.ntu.edu.sg</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10760265$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Yang, Zhiyuan</creatorcontrib><creatorcontrib>Zhou, Yunjiao</creatorcontrib><creatorcontrib>Xie, Lihua</creatorcontrib><creatorcontrib>Yang, Jianfei</creatorcontrib><title>T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition</title><title>IEEE transactions on cybernetics</title><addtitle>TCYB</addtitle><description><![CDATA[The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements and nonreal-time latency. There has been a lack of research on how to compress 3-D point cloud models into lightweight models. In this article, we propose a method called T3DNet (tiny 3-D network with augmentation and distillation) to address this issue. We find that the tiny model after network augmentation is much easier for a teacher to distill. Instead of gradually reducing the parameters through techniques, such as pruning or quantization, we predefine a tiny model and improve its performance through auxiliary supervision from augmented networks and the original model. We evaluate our method on several public datasets, including ModelNet40, ShapeNet, and ScanObjectNN. Our method can achieve high compression rates without significant accuracy sacrifice, achieving state-of-the-art performances on three datasets against existing methods. Amazingly, our T3DNet is 58<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> smaller and 54<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> faster than the original model yet with only 1.4<inline-formula> <tex-math notation="LaTeX">\%</tex-math> </inline-formula> accuracy descent on the ModelNet40 dataset. Our code is available at https://github.com/Zhiyuan002/T3DNet.]]></description><subject>3-D model compression</subject><subject>Accuracy</subject><subject>Computational modeling</subject><subject>Feature extraction</subject><subject>Image coding</subject><subject>knowledge distillation (KD)</subject><subject>Knowledge engineering</subject><subject>network augmentation</subject><subject>Point cloud compression</subject><subject>point cloud models</subject><subject>Quantization (signal)</subject><subject>Sensors</subject><subject>Solid modeling</subject><subject>Three-dimensional displays</subject><issn>2168-2267</issn><issn>2168-2275</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNpNkN1KxDAQhYMouKz7AIIXeYHWTNImqXfadVWoP0i98Kr0Z1Ij3WZpKuLb27KLOBdnhmHOGfgIOQcWArDkMk_fb0LOeBSKSCvO2RFZcJA64FzFx3-zVKdk5f0nm0pPq0QvyCYX6yccr2jqtrsBvbd9S1-c7Ueadu6roY-uwc5T4waa2fZj_MZZqQjW9BVr1_Z2tK4_Iyem7DyuDn1J3ja3eXofZM93D-l1FtSQwBhEpoKyrJQCVsYGSgmNaCKJGqBSkjFkBqUxvE5AYIzAKq5QKc1qIbThSiwJ7HPrwXk_oCl2g92Ww08BrJhZFDOLYmZRHFhMnou9xyLiv_vpIZex-AX6FFmY</recordid><startdate>20241121</startdate><enddate>20241121</enddate><creator>Yang, Zhiyuan</creator><creator>Zhou, Yunjiao</creator><creator>Xie, Lihua</creator><creator>Yang, Jianfei</creator><general>IEEE</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/yang0674@ntu.edu.sg</orcidid><orcidid>https://orcid.org/elhxie@ntu.edu.sg</orcidid><orcidid>https://orcid.org/yunjiao001@e.ntu.edu.sg</orcidid><orcidid>https://orcid.org/YANG0478@e.ntu.edu.sg</orcidid></search><sort><creationdate>20241121</creationdate><title>T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition</title><author>Yang, Zhiyuan ; Zhou, Yunjiao ; Xie, Lihua ; Yang, Jianfei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c191t-4fb1aab7710a5f1a61d3d46e811b7600e0fe6ff2c913e5e10b27e7780c338f273</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>3-D model compression</topic><topic>Accuracy</topic><topic>Computational modeling</topic><topic>Feature extraction</topic><topic>Image coding</topic><topic>knowledge distillation (KD)</topic><topic>Knowledge engineering</topic><topic>network augmentation</topic><topic>Point cloud compression</topic><topic>point cloud models</topic><topic>Quantization (signal)</topic><topic>Sensors</topic><topic>Solid modeling</topic><topic>Three-dimensional displays</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yang, Zhiyuan</creatorcontrib><creatorcontrib>Zhou, Yunjiao</creatorcontrib><creatorcontrib>Xie, Lihua</creatorcontrib><creatorcontrib>Yang, Jianfei</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) Online</collection><collection>IEEE</collection><collection>CrossRef</collection><jtitle>IEEE transactions on cybernetics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yang, Zhiyuan</au><au>Zhou, Yunjiao</au><au>Xie, Lihua</au><au>Yang, Jianfei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition</atitle><jtitle>IEEE transactions on cybernetics</jtitle><stitle>TCYB</stitle><date>2024-11-21</date><risdate>2024</risdate><spage>1</spage><epage>11</epage><pages>1-11</pages><issn>2168-2267</issn><eissn>2168-2275</eissn><coden>ITCEB8</coden><abstract><![CDATA[The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements and nonreal-time latency. There has been a lack of research on how to compress 3-D point cloud models into lightweight models. In this article, we propose a method called T3DNet (tiny 3-D network with augmentation and distillation) to address this issue. We find that the tiny model after network augmentation is much easier for a teacher to distill. Instead of gradually reducing the parameters through techniques, such as pruning or quantization, we predefine a tiny model and improve its performance through auxiliary supervision from augmented networks and the original model. We evaluate our method on several public datasets, including ModelNet40, ShapeNet, and ScanObjectNN. Our method can achieve high compression rates without significant accuracy sacrifice, achieving state-of-the-art performances on three datasets against existing methods. Amazingly, our T3DNet is 58<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> smaller and 54<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> faster than the original model yet with only 1.4<inline-formula> <tex-math notation="LaTeX">\%</tex-math> </inline-formula> accuracy descent on the ModelNet40 dataset. Our code is available at https://github.com/Zhiyuan002/T3DNet.]]></abstract><pub>IEEE</pub><doi>10.1109/TCYB.2024.3487220</doi><tpages>11</tpages><orcidid>https://orcid.org/yang0674@ntu.edu.sg</orcidid><orcidid>https://orcid.org/elhxie@ntu.edu.sg</orcidid><orcidid>https://orcid.org/yunjiao001@e.ntu.edu.sg</orcidid><orcidid>https://orcid.org/YANG0478@e.ntu.edu.sg</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2168-2267
ispartof IEEE transactions on cybernetics, 2024-11, p.1-11
issn 2168-2267
2168-2275
language eng
recordid cdi_ieee_primary_10760265
source IEEE Electronic Library (IEL) Journals
subjects 3-D model compression
Accuracy
Computational modeling
Feature extraction
Image coding
knowledge distillation (KD)
Knowledge engineering
network augmentation
Point cloud compression
point cloud models
Quantization (signal)
Sensors
Solid modeling
Three-dimensional displays
title T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T19%3A06%3A34IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-crossref_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=T3DNet:%20Compressing%20Point%20Cloud%20Models%20for%20Lightweight%203-D%20Recognition&rft.jtitle=IEEE%20transactions%20on%20cybernetics&rft.au=Yang,%20Zhiyuan&rft.date=2024-11-21&rft.spage=1&rft.epage=11&rft.pages=1-11&rft.issn=2168-2267&rft.eissn=2168-2275&rft.coden=ITCEB8&rft_id=info:doi/10.1109/TCYB.2024.3487220&rft_dat=%3Ccrossref_ieee_%3E10_1109_TCYB_2024_3487220%3C/crossref_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c191t-4fb1aab7710a5f1a61d3d46e811b7600e0fe6ff2c913e5e10b27e7780c338f273%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10760265&rfr_iscdi=true