Loading…
T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition
The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements a...
Saved in:
Published in: | IEEE transactions on cybernetics 2024-11, p.1-11 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | 11 |
container_issue | |
container_start_page | 1 |
container_title | IEEE transactions on cybernetics |
container_volume | |
creator | Yang, Zhiyuan Zhou, Yunjiao Xie, Lihua Yang, Jianfei |
description | The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements and nonreal-time latency. There has been a lack of research on how to compress 3-D point cloud models into lightweight models. In this article, we propose a method called T3DNet (tiny 3-D network with augmentation and distillation) to address this issue. We find that the tiny model after network augmentation is much easier for a teacher to distill. Instead of gradually reducing the parameters through techniques, such as pruning or quantization, we predefine a tiny model and improve its performance through auxiliary supervision from augmented networks and the original model. We evaluate our method on several public datasets, including ModelNet40, ShapeNet, and ScanObjectNN. Our method can achieve high compression rates without significant accuracy sacrifice, achieving state-of-the-art performances on three datasets against existing methods. Amazingly, our T3DNet is 58 \times smaller and 54 \times faster than the original model yet with only 1.4 \% accuracy descent on the ModelNet40 dataset. Our code is available at https://github.com/Zhiyuan002/T3DNet. |
doi_str_mv | 10.1109/TCYB.2024.3487220 |
format | article |
fullrecord | <record><control><sourceid>crossref_ieee_</sourceid><recordid>TN_cdi_ieee_primary_10760265</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10760265</ieee_id><sourcerecordid>10_1109_TCYB_2024_3487220</sourcerecordid><originalsourceid>FETCH-LOGICAL-c191t-4fb1aab7710a5f1a61d3d46e811b7600e0fe6ff2c913e5e10b27e7780c338f273</originalsourceid><addsrcrecordid>eNpNkN1KxDAQhYMouKz7AIIXeYHWTNImqXfadVWoP0i98Kr0Z1Ij3WZpKuLb27KLOBdnhmHOGfgIOQcWArDkMk_fb0LOeBSKSCvO2RFZcJA64FzFx3-zVKdk5f0nm0pPq0QvyCYX6yccr2jqtrsBvbd9S1-c7Ueadu6roY-uwc5T4waa2fZj_MZZqQjW9BVr1_Z2tK4_Iyem7DyuDn1J3ja3eXofZM93D-l1FtSQwBhEpoKyrJQCVsYGSgmNaCKJGqBSkjFkBqUxvE5AYIzAKq5QKc1qIbThSiwJ7HPrwXk_oCl2g92Ww08BrJhZFDOLYmZRHFhMnou9xyLiv_vpIZex-AX6FFmY</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition</title><source>IEEE Electronic Library (IEL) Journals</source><creator>Yang, Zhiyuan ; Zhou, Yunjiao ; Xie, Lihua ; Yang, Jianfei</creator><creatorcontrib>Yang, Zhiyuan ; Zhou, Yunjiao ; Xie, Lihua ; Yang, Jianfei</creatorcontrib><description><![CDATA[The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements and nonreal-time latency. There has been a lack of research on how to compress 3-D point cloud models into lightweight models. In this article, we propose a method called T3DNet (tiny 3-D network with augmentation and distillation) to address this issue. We find that the tiny model after network augmentation is much easier for a teacher to distill. Instead of gradually reducing the parameters through techniques, such as pruning or quantization, we predefine a tiny model and improve its performance through auxiliary supervision from augmented networks and the original model. We evaluate our method on several public datasets, including ModelNet40, ShapeNet, and ScanObjectNN. Our method can achieve high compression rates without significant accuracy sacrifice, achieving state-of-the-art performances on three datasets against existing methods. Amazingly, our T3DNet is 58<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> smaller and 54<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> faster than the original model yet with only 1.4<inline-formula> <tex-math notation="LaTeX">\%</tex-math> </inline-formula> accuracy descent on the ModelNet40 dataset. Our code is available at https://github.com/Zhiyuan002/T3DNet.]]></description><identifier>ISSN: 2168-2267</identifier><identifier>EISSN: 2168-2275</identifier><identifier>DOI: 10.1109/TCYB.2024.3487220</identifier><identifier>CODEN: ITCEB8</identifier><language>eng</language><publisher>IEEE</publisher><subject>3-D model compression ; Accuracy ; Computational modeling ; Feature extraction ; Image coding ; knowledge distillation (KD) ; Knowledge engineering ; network augmentation ; Point cloud compression ; point cloud models ; Quantization (signal) ; Sensors ; Solid modeling ; Three-dimensional displays</subject><ispartof>IEEE transactions on cybernetics, 2024-11, p.1-11</ispartof><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>yang0674@ntu.edu.sg ; elhxie@ntu.edu.sg ; yunjiao001@e.ntu.edu.sg ; YANG0478@e.ntu.edu.sg</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10760265$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,54796</link.rule.ids></links><search><creatorcontrib>Yang, Zhiyuan</creatorcontrib><creatorcontrib>Zhou, Yunjiao</creatorcontrib><creatorcontrib>Xie, Lihua</creatorcontrib><creatorcontrib>Yang, Jianfei</creatorcontrib><title>T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition</title><title>IEEE transactions on cybernetics</title><addtitle>TCYB</addtitle><description><![CDATA[The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements and nonreal-time latency. There has been a lack of research on how to compress 3-D point cloud models into lightweight models. In this article, we propose a method called T3DNet (tiny 3-D network with augmentation and distillation) to address this issue. We find that the tiny model after network augmentation is much easier for a teacher to distill. Instead of gradually reducing the parameters through techniques, such as pruning or quantization, we predefine a tiny model and improve its performance through auxiliary supervision from augmented networks and the original model. We evaluate our method on several public datasets, including ModelNet40, ShapeNet, and ScanObjectNN. Our method can achieve high compression rates without significant accuracy sacrifice, achieving state-of-the-art performances on three datasets against existing methods. Amazingly, our T3DNet is 58<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> smaller and 54<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> faster than the original model yet with only 1.4<inline-formula> <tex-math notation="LaTeX">\%</tex-math> </inline-formula> accuracy descent on the ModelNet40 dataset. Our code is available at https://github.com/Zhiyuan002/T3DNet.]]></description><subject>3-D model compression</subject><subject>Accuracy</subject><subject>Computational modeling</subject><subject>Feature extraction</subject><subject>Image coding</subject><subject>knowledge distillation (KD)</subject><subject>Knowledge engineering</subject><subject>network augmentation</subject><subject>Point cloud compression</subject><subject>point cloud models</subject><subject>Quantization (signal)</subject><subject>Sensors</subject><subject>Solid modeling</subject><subject>Three-dimensional displays</subject><issn>2168-2267</issn><issn>2168-2275</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNpNkN1KxDAQhYMouKz7AIIXeYHWTNImqXfadVWoP0i98Kr0Z1Ij3WZpKuLb27KLOBdnhmHOGfgIOQcWArDkMk_fb0LOeBSKSCvO2RFZcJA64FzFx3-zVKdk5f0nm0pPq0QvyCYX6yccr2jqtrsBvbd9S1-c7Ueadu6roY-uwc5T4waa2fZj_MZZqQjW9BVr1_Z2tK4_Iyem7DyuDn1J3ja3eXofZM93D-l1FtSQwBhEpoKyrJQCVsYGSgmNaCKJGqBSkjFkBqUxvE5AYIzAKq5QKc1qIbThSiwJ7HPrwXk_oCl2g92Ww08BrJhZFDOLYmZRHFhMnou9xyLiv_vpIZex-AX6FFmY</recordid><startdate>20241121</startdate><enddate>20241121</enddate><creator>Yang, Zhiyuan</creator><creator>Zhou, Yunjiao</creator><creator>Xie, Lihua</creator><creator>Yang, Jianfei</creator><general>IEEE</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/yang0674@ntu.edu.sg</orcidid><orcidid>https://orcid.org/elhxie@ntu.edu.sg</orcidid><orcidid>https://orcid.org/yunjiao001@e.ntu.edu.sg</orcidid><orcidid>https://orcid.org/YANG0478@e.ntu.edu.sg</orcidid></search><sort><creationdate>20241121</creationdate><title>T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition</title><author>Yang, Zhiyuan ; Zhou, Yunjiao ; Xie, Lihua ; Yang, Jianfei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c191t-4fb1aab7710a5f1a61d3d46e811b7600e0fe6ff2c913e5e10b27e7780c338f273</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>3-D model compression</topic><topic>Accuracy</topic><topic>Computational modeling</topic><topic>Feature extraction</topic><topic>Image coding</topic><topic>knowledge distillation (KD)</topic><topic>Knowledge engineering</topic><topic>network augmentation</topic><topic>Point cloud compression</topic><topic>point cloud models</topic><topic>Quantization (signal)</topic><topic>Sensors</topic><topic>Solid modeling</topic><topic>Three-dimensional displays</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yang, Zhiyuan</creatorcontrib><creatorcontrib>Zhou, Yunjiao</creatorcontrib><creatorcontrib>Xie, Lihua</creatorcontrib><creatorcontrib>Yang, Jianfei</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) Online</collection><collection>IEEE</collection><collection>CrossRef</collection><jtitle>IEEE transactions on cybernetics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yang, Zhiyuan</au><au>Zhou, Yunjiao</au><au>Xie, Lihua</au><au>Yang, Jianfei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition</atitle><jtitle>IEEE transactions on cybernetics</jtitle><stitle>TCYB</stitle><date>2024-11-21</date><risdate>2024</risdate><spage>1</spage><epage>11</epage><pages>1-11</pages><issn>2168-2267</issn><eissn>2168-2275</eissn><coden>ITCEB8</coden><abstract><![CDATA[The 3-D point cloud has been widely used in many mobile application scenarios, including autonomous driving and 3-D sensing on mobile devices. However, existing 3-D point cloud models tend to be large and cumbersome, making them hard to deploy on edged devices due to their high memory requirements and nonreal-time latency. There has been a lack of research on how to compress 3-D point cloud models into lightweight models. In this article, we propose a method called T3DNet (tiny 3-D network with augmentation and distillation) to address this issue. We find that the tiny model after network augmentation is much easier for a teacher to distill. Instead of gradually reducing the parameters through techniques, such as pruning or quantization, we predefine a tiny model and improve its performance through auxiliary supervision from augmented networks and the original model. We evaluate our method on several public datasets, including ModelNet40, ShapeNet, and ScanObjectNN. Our method can achieve high compression rates without significant accuracy sacrifice, achieving state-of-the-art performances on three datasets against existing methods. Amazingly, our T3DNet is 58<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> smaller and 54<inline-formula> <tex-math notation="LaTeX">\times</tex-math> </inline-formula> faster than the original model yet with only 1.4<inline-formula> <tex-math notation="LaTeX">\%</tex-math> </inline-formula> accuracy descent on the ModelNet40 dataset. Our code is available at https://github.com/Zhiyuan002/T3DNet.]]></abstract><pub>IEEE</pub><doi>10.1109/TCYB.2024.3487220</doi><tpages>11</tpages><orcidid>https://orcid.org/yang0674@ntu.edu.sg</orcidid><orcidid>https://orcid.org/elhxie@ntu.edu.sg</orcidid><orcidid>https://orcid.org/yunjiao001@e.ntu.edu.sg</orcidid><orcidid>https://orcid.org/YANG0478@e.ntu.edu.sg</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2168-2267 |
ispartof | IEEE transactions on cybernetics, 2024-11, p.1-11 |
issn | 2168-2267 2168-2275 |
language | eng |
recordid | cdi_ieee_primary_10760265 |
source | IEEE Electronic Library (IEL) Journals |
subjects | 3-D model compression Accuracy Computational modeling Feature extraction Image coding knowledge distillation (KD) Knowledge engineering network augmentation Point cloud compression point cloud models Quantization (signal) Sensors Solid modeling Three-dimensional displays |
title | T3DNet: Compressing Point Cloud Models for Lightweight 3-D Recognition |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T19%3A06%3A34IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-crossref_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=T3DNet:%20Compressing%20Point%20Cloud%20Models%20for%20Lightweight%203-D%20Recognition&rft.jtitle=IEEE%20transactions%20on%20cybernetics&rft.au=Yang,%20Zhiyuan&rft.date=2024-11-21&rft.spage=1&rft.epage=11&rft.pages=1-11&rft.issn=2168-2267&rft.eissn=2168-2275&rft.coden=ITCEB8&rft_id=info:doi/10.1109/TCYB.2024.3487220&rft_dat=%3Ccrossref_ieee_%3E10_1109_TCYB_2024_3487220%3C/crossref_ieee_%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c191t-4fb1aab7710a5f1a61d3d46e811b7600e0fe6ff2c913e5e10b27e7780c338f273%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10760265&rfr_iscdi=true |