Loading…
Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data
Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning,...
Saved in:
Published in: | Applied computing review : a publication of the Special Interest Group on Applied Computing 2024-05, Vol.24 (1), p.14-23 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | cdi_FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3 |
container_end_page | 23 |
container_issue | 1 |
container_start_page | 14 |
container_title | Applied computing review : a publication of the Special Interest Group on Applied Computing |
container_volume | 24 |
creator | Hadi, Abir Mohammad Won, Kwanghee |
description | Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning, and fine-tuning steps. In this study, we introduce an approach utilizing a deep reinforcement learning-based agent to dynamically compress a deep convolutional neural network throughout its training process. We integrate the concept of the intrinsic dimension of the training data to provide the agent with insights into the task's complexity. The agent employs two distinct ranking criteria, L1-norm-based and attention-based measures, to selectively prune filters from each layer as it determines necessary. In the experiments, we used the CIFAR-10 dataset and its subsets (2-class and 5-class subsets) to model the task complexity and showed that the agent learns different policies depending on the intrinsic dimension. The agent, on average, pruned off 78.48%, 77.9%, and 83.12% filters from all the layers of VGG-16 network for CIFAR-10 full, 5-class, and 2-class subsets respectively. |
doi_str_mv | 10.1145/3663652.3663654 |
format | article |
fullrecord | <record><control><sourceid>acm_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1145_3663652_3663654</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3663654</sourcerecordid><originalsourceid>FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3</originalsourceid><addsrcrecordid>eNo9kM1OwzAQhC0EEqVwRuLkF0jr_8RH1EKpVMGl92jrrMHQOJWdgnh7Qls4fSPNzGo1hNxyNuFc6ak0RhotJkeqMzLiVvKCccPPB621LYzl-pJc5fzOGK-4ESPi5og7OuviZ7fd96GLsKXPuE8H9F9d-hjMdpcw58GkG8jY0EH0b0iXsU8h5uDoPLQYD4HOH6x1ghBDfKVz6OGaXHjYZrw5cUzWjw_r2VOxelksZ_erArRVRSlE5X0prHUcASxIabw2w_vSKCus2RjXlKLEBsuKyQatAgdMNa6y0juQYzI9nnWpyzmhr3cptJC-a87q34nq00QnqqFxd2yAa__Df-YP3uViQQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data</title><source>Association for Computing Machinery:Jisc Collections:ACM OPEN Journals 2023-2025 (reading list)</source><creator>Hadi, Abir Mohammad ; Won, Kwanghee</creator><creatorcontrib>Hadi, Abir Mohammad ; Won, Kwanghee</creatorcontrib><description>Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning, and fine-tuning steps. In this study, we introduce an approach utilizing a deep reinforcement learning-based agent to dynamically compress a deep convolutional neural network throughout its training process. We integrate the concept of the intrinsic dimension of the training data to provide the agent with insights into the task's complexity. The agent employs two distinct ranking criteria, L1-norm-based and attention-based measures, to selectively prune filters from each layer as it determines necessary. In the experiments, we used the CIFAR-10 dataset and its subsets (2-class and 5-class subsets) to model the task complexity and showed that the agent learns different policies depending on the intrinsic dimension. The agent, on average, pruned off 78.48%, 77.9%, and 83.12% filters from all the layers of VGG-16 network for CIFAR-10 full, 5-class, and 2-class subsets respectively.</description><identifier>ISSN: 1559-6915</identifier><identifier>EISSN: 1931-0161</identifier><identifier>DOI: 10.1145/3663652.3663654</identifier><language>eng</language><publisher>New York, NY, USA: ACM</publisher><subject>Artificial intelligence ; Computer vision ; Computing methodologies ; Machine learning ; Machine learning approaches ; Neural networks</subject><ispartof>Applied computing review : a publication of the Special Interest Group on Applied Computing, 2024-05, Vol.24 (1), p.14-23</ispartof><rights>Copyright is held by the owner/author(s)</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Hadi, Abir Mohammad</creatorcontrib><creatorcontrib>Won, Kwanghee</creatorcontrib><title>Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data</title><title>Applied computing review : a publication of the Special Interest Group on Applied Computing</title><addtitle>ACM SIGAPP</addtitle><description>Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning, and fine-tuning steps. In this study, we introduce an approach utilizing a deep reinforcement learning-based agent to dynamically compress a deep convolutional neural network throughout its training process. We integrate the concept of the intrinsic dimension of the training data to provide the agent with insights into the task's complexity. The agent employs two distinct ranking criteria, L1-norm-based and attention-based measures, to selectively prune filters from each layer as it determines necessary. In the experiments, we used the CIFAR-10 dataset and its subsets (2-class and 5-class subsets) to model the task complexity and showed that the agent learns different policies depending on the intrinsic dimension. The agent, on average, pruned off 78.48%, 77.9%, and 83.12% filters from all the layers of VGG-16 network for CIFAR-10 full, 5-class, and 2-class subsets respectively.</description><subject>Artificial intelligence</subject><subject>Computer vision</subject><subject>Computing methodologies</subject><subject>Machine learning</subject><subject>Machine learning approaches</subject><subject>Neural networks</subject><issn>1559-6915</issn><issn>1931-0161</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNo9kM1OwzAQhC0EEqVwRuLkF0jr_8RH1EKpVMGl92jrrMHQOJWdgnh7Qls4fSPNzGo1hNxyNuFc6ak0RhotJkeqMzLiVvKCccPPB621LYzl-pJc5fzOGK-4ESPi5og7OuviZ7fd96GLsKXPuE8H9F9d-hjMdpcw58GkG8jY0EH0b0iXsU8h5uDoPLQYD4HOH6x1ghBDfKVz6OGaXHjYZrw5cUzWjw_r2VOxelksZ_erArRVRSlE5X0prHUcASxIabw2w_vSKCus2RjXlKLEBsuKyQatAgdMNa6y0juQYzI9nnWpyzmhr3cptJC-a87q34nq00QnqqFxd2yAa__Df-YP3uViQQ</recordid><startdate>20240502</startdate><enddate>20240502</enddate><creator>Hadi, Abir Mohammad</creator><creator>Won, Kwanghee</creator><general>ACM</general><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20240502</creationdate><title>Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data</title><author>Hadi, Abir Mohammad ; Won, Kwanghee</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Artificial intelligence</topic><topic>Computer vision</topic><topic>Computing methodologies</topic><topic>Machine learning</topic><topic>Machine learning approaches</topic><topic>Neural networks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Hadi, Abir Mohammad</creatorcontrib><creatorcontrib>Won, Kwanghee</creatorcontrib><collection>CrossRef</collection><jtitle>Applied computing review : a publication of the Special Interest Group on Applied Computing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Hadi, Abir Mohammad</au><au>Won, Kwanghee</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data</atitle><jtitle>Applied computing review : a publication of the Special Interest Group on Applied Computing</jtitle><stitle>ACM SIGAPP</stitle><date>2024-05-02</date><risdate>2024</risdate><volume>24</volume><issue>1</issue><spage>14</spage><epage>23</epage><pages>14-23</pages><issn>1559-6915</issn><eissn>1931-0161</eissn><abstract>Selecting the optimal deep learning architecture for a particular task and dataset remains an ongoing challenge. Typically, this decision-making process involves exhaustive searches for neural network architectures or multi-phase optimization, which includes initial training, compression or pruning, and fine-tuning steps. In this study, we introduce an approach utilizing a deep reinforcement learning-based agent to dynamically compress a deep convolutional neural network throughout its training process. We integrate the concept of the intrinsic dimension of the training data to provide the agent with insights into the task's complexity. The agent employs two distinct ranking criteria, L1-norm-based and attention-based measures, to selectively prune filters from each layer as it determines necessary. In the experiments, we used the CIFAR-10 dataset and its subsets (2-class and 5-class subsets) to model the task complexity and showed that the agent learns different policies depending on the intrinsic dimension. The agent, on average, pruned off 78.48%, 77.9%, and 83.12% filters from all the layers of VGG-16 network for CIFAR-10 full, 5-class, and 2-class subsets respectively.</abstract><cop>New York, NY, USA</cop><pub>ACM</pub><doi>10.1145/3663652.3663654</doi><tpages>10</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1559-6915 |
ispartof | Applied computing review : a publication of the Special Interest Group on Applied Computing, 2024-05, Vol.24 (1), p.14-23 |
issn | 1559-6915 1931-0161 |
language | eng |
recordid | cdi_crossref_primary_10_1145_3663652_3663654 |
source | Association for Computing Machinery:Jisc Collections:ACM OPEN Journals 2023-2025 (reading list) |
subjects | Artificial intelligence Computer vision Computing methodologies Machine learning Machine learning approaches Neural networks |
title | Deep Convolutional Neural Network Compression based on the Intrinsic Dimension of the Training Data |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T12%3A11%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-acm_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20Convolutional%20Neural%20Network%20Compression%20based%20on%20the%20Intrinsic%20Dimension%20of%20the%20Training%20Data&rft.jtitle=Applied%20computing%20review%20:%20a%20publication%20of%20the%20Special%20Interest%20Group%20on%20Applied%20Computing&rft.au=Hadi,%20Abir%20Mohammad&rft.date=2024-05-02&rft.volume=24&rft.issue=1&rft.spage=14&rft.epage=23&rft.pages=14-23&rft.issn=1559-6915&rft.eissn=1931-0161&rft_id=info:doi/10.1145/3663652.3663654&rft_dat=%3Cacm_cross%3E3663654%3C/acm_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a594-7228ff7299c1eaa9a336f561613649296b6cd727ede7803de94aca04dc893fca3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |