Loading…

Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data

Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning,...

Full description

Saved in:
Bibliographic Details
Published in:Applied computing review : a publication of the Special Interest Group on Applied Computing 2024-05, Vol.24 (1), p.14-23
Main Authors: Hadi, Abir Mohammad, Won, Kwanghee
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3
container_end_page 23
container_issue 1
container_start_page 14
container_title Applied computing review : a publication of the Special Interest Group on Applied Computing
container_volume 24
creator Hadi, Abir Mohammad
Won, Kwanghee
description Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning, and fine-tuning steps. In this study, we introduce an approach utilizing a deep reinforcement learning-based agent to dynamically compress a deep convolutional neural network throughout its training process. We integrate the concept of the intrinsic dimension of the training data to provide the agent with insights into the task's complexity. The agent employs two distinct ranking criteria, L1-norm-based and attention-based measures, to selectively prune filters from each layer as it determines necessary. In the experiments, we used the CIFAR-10 dataset and its subsets (2-class and 5-class subsets) to model the task complexity and showed that the agent learns different policies depending on the intrinsic dimension. The agent, on average, pruned off 78.48%, 77.9%, and 83.12% filters from all the layers of VGG-16 network for CIFAR-10 full, 5-class, and 2-class subsets respectively.
doi_str_mv 10.1145/3663652.3663654
format article
fullrecord <record><control><sourceid>acm_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1145_3663652_3663654</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3663654</sourcerecordid><originalsourceid>FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3</originalsourceid><addsrcrecordid>eNo9kM1OwzAQhC0EEqVwRuLkF0jr_8RH1EKpVMGl92jrrMHQOJWdgnh7Qls4fSPNzGo1hNxyNuFc6ak0RhotJkeqMzLiVvKCccPPB621LYzl-pJc5fzOGK-4ESPi5og7OuviZ7fd96GLsKXPuE8H9F9d-hjMdpcw58GkG8jY0EH0b0iXsU8h5uDoPLQYD4HOH6x1ghBDfKVz6OGaXHjYZrw5cUzWjw_r2VOxelksZ_erArRVRSlE5X0prHUcASxIabw2w_vSKCus2RjXlKLEBsuKyQatAgdMNa6y0juQYzI9nnWpyzmhr3cptJC-a87q34nq00QnqqFxd2yAa__Df-YP3uViQQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data</title><source>Association for Computing Machinery:Jisc Collections:ACM OPEN Journals 2023-2025 (reading list)</source><creator>Hadi, Abir Mohammad ; Won, Kwanghee</creator><creatorcontrib>Hadi, Abir Mohammad ; Won, Kwanghee</creatorcontrib><description>Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning, and fine-tuning steps. In this study, we introduce an approach utilizing a deep reinforcement learning-based agent to dynamically compress a deep convolutional neural network throughout its training process. We integrate the concept of the intrinsic dimension of the training data to provide the agent with insights into the task's complexity. The agent employs two distinct ranking criteria, L1-norm-based and attention-based measures, to selectively prune filters from each layer as it determines necessary. In the experiments, we used the CIFAR-10 dataset and its subsets (2-class and 5-class subsets) to model the task complexity and showed that the agent learns different policies depending on the intrinsic dimension. The agent, on average, pruned off 78.48%, 77.9%, and 83.12% filters from all the layers of VGG-16 network for CIFAR-10 full, 5-class, and 2-class subsets respectively.</description><identifier>ISSN: 1559-6915</identifier><identifier>EISSN: 1931-0161</identifier><identifier>DOI: 10.1145/3663652.3663654</identifier><language>eng</language><publisher>New York, NY, USA: ACM</publisher><subject>Artificial intelligence ; Computer vision ; Computing methodologies ; Machine learning ; Machine learning approaches ; Neural networks</subject><ispartof>Applied computing review : a publication of the Special Interest Group on Applied Computing, 2024-05, Vol.24 (1), p.14-23</ispartof><rights>Copyright is held by the owner/author(s)</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Hadi, Abir Mohammad</creatorcontrib><creatorcontrib>Won, Kwanghee</creatorcontrib><title>Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data</title><title>Applied computing review : a publication of the Special Interest Group on Applied Computing</title><addtitle>ACM SIGAPP</addtitle><description>Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning, and fine-tuning steps. In this study, we introduce an approach utilizing a deep reinforcement learning-based agent to dynamically compress a deep convolutional neural network throughout its training process. We integrate the concept of the intrinsic dimension of the training data to provide the agent with insights into the task's complexity. The agent employs two distinct ranking criteria, L1-norm-based and attention-based measures, to selectively prune filters from each layer as it determines necessary. In the experiments, we used the CIFAR-10 dataset and its subsets (2-class and 5-class subsets) to model the task complexity and showed that the agent learns different policies depending on the intrinsic dimension. The agent, on average, pruned off 78.48%, 77.9%, and 83.12% filters from all the layers of VGG-16 network for CIFAR-10 full, 5-class, and 2-class subsets respectively.</description><subject>Artificial intelligence</subject><subject>Computer vision</subject><subject>Computing methodologies</subject><subject>Machine learning</subject><subject>Machine learning approaches</subject><subject>Neural networks</subject><issn>1559-6915</issn><issn>1931-0161</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNo9kM1OwzAQhC0EEqVwRuLkF0jr_8RH1EKpVMGl92jrrMHQOJWdgnh7Qls4fSPNzGo1hNxyNuFc6ak0RhotJkeqMzLiVvKCccPPB621LYzl-pJc5fzOGK-4ESPi5og7OuviZ7fd96GLsKXPuE8H9F9d-hjMdpcw58GkG8jY0EH0b0iXsU8h5uDoPLQYD4HOH6x1ghBDfKVz6OGaXHjYZrw5cUzWjw_r2VOxelksZ_erArRVRSlE5X0prHUcASxIabw2w_vSKCus2RjXlKLEBsuKyQatAgdMNa6y0juQYzI9nnWpyzmhr3cptJC-a87q34nq00QnqqFxd2yAa__Df-YP3uViQQ</recordid><startdate>20240502</startdate><enddate>20240502</enddate><creator>Hadi, Abir Mohammad</creator><creator>Won, Kwanghee</creator><general>ACM</general><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20240502</creationdate><title>Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data</title><author>Hadi, Abir Mohammad ; Won, Kwanghee</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Artificial intelligence</topic><topic>Computer vision</topic><topic>Computing methodologies</topic><topic>Machine learning</topic><topic>Machine learning approaches</topic><topic>Neural networks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Hadi, Abir Mohammad</creatorcontrib><creatorcontrib>Won, Kwanghee</creatorcontrib><collection>CrossRef</collection><jtitle>Applied computing review : a publication of the Special Interest Group on Applied Computing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Hadi, Abir Mohammad</au><au>Won, Kwanghee</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data</atitle><jtitle>Applied computing review : a publication of the Special Interest Group on Applied Computing</jtitle><stitle>ACM SIGAPP</stitle><date>2024-05-02</date><risdate>2024</risdate><volume>24</volume><issue>1</issue><spage>14</spage><epage>23</epage><pages>14-23</pages><issn>1559-6915</issn><eissn>1931-0161</eissn><abstract>Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning, and fine-tuning steps. In this study, we introduce an approach utilizing a deep reinforcement learning-based agent to dynamically compress a deep convolutional neural network throughout its training process. We integrate the concept of the intrinsic dimension of the training data to provide the agent with insights into the task's complexity. The agent employs two distinct ranking criteria, L1-norm-based and attention-based measures, to selectively prune filters from each layer as it determines necessary. In the experiments, we used the CIFAR-10 dataset and its subsets (2-class and 5-class subsets) to model the task complexity and showed that the agent learns different policies depending on the intrinsic dimension. The agent, on average, pruned off 78.48%, 77.9%, and 83.12% filters from all the layers of VGG-16 network for CIFAR-10 full, 5-class, and 2-class subsets respectively.</abstract><cop>New York, NY, USA</cop><pub>ACM</pub><doi>10.1145/3663652.3663654</doi><tpages>10</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1559-6915
ispartof Applied computing review : a publication of the Special Interest Group on Applied Computing, 2024-05, Vol.24 (1), p.14-23
issn 1559-6915
1931-0161
language eng
recordid cdi_crossref_primary_10_1145_3663652_3663654
source Association for Computing Machinery:Jisc Collections:ACM OPEN Journals 2023-2025 (reading list)
subjects Artificial intelligence
Computer vision
Computing methodologies
Machine learning
Machine learning approaches
Neural networks
title Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T12%3A11%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-acm_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20Convolutional%20Neural%20Network%20Compression%20based%20on%20the%20Intrinsic%20Dimension%20of%20the%20Training%20Data&rft.jtitle=Applied%20computing%20review%20:%20a%20publication%20of%20the%20Special%20Interest%20Group%20on%20Applied%20Computing&rft.au=Hadi,%20Abir%20Mohammad&rft.date=2024-05-02&rft.volume=24&rft.issue=1&rft.spage=14&rft.epage=23&rft.pages=14-23&rft.issn=1559-6915&rft.eissn=1931-0161&rft_id=info:doi/10.1145/3663652.3663654&rft_dat=%3Cacm_cross%3E3663654%3C/acm_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true