Loading…

Accelerating one-shot neural architecture search via constructing a sparse search space

Neural Architecture Search (NAS) has garnered significant attention for its ability to automatically design high-quality deep neural networks (DNNs) tailored to various hardware platforms. The major challenge for NAS is the time-consuming network estimation process required to select optimal network...

Full description

Saved in:
Bibliographic Details
Published in:Knowledge-based systems 2024-12, Vol.305, p.112620, Article 112620
Main Authors: Huang, Hongtao, Chang, Xiaojun, Yao, Lina
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c185t-604f9e959e59a098329e948c0ea3e65975ce4f279b7e82aa07d7e4ee0c9b8e0f3
container_end_page
container_issue
container_start_page 112620
container_title Knowledge-based systems
container_volume 305
creator Huang, Hongtao
Chang, Xiaojun
Yao, Lina
description Neural Architecture Search (NAS) has garnered significant attention for its ability to automatically design high-quality deep neural networks (DNNs) tailored to various hardware platforms. The major challenge for NAS is the time-consuming network estimation process required to select optimal networks from a large pool of candidates. Rather than training each candidate from scratch, recent one-shot NAS methods accelerate the estimation process by only training a supernet and sampling sub-networks from it, inheriting partial network architectures and weights. Despite significant acceleration, the supernet training with a large search space (i.e., the number of candidate sub-networks) still requires thousands of GPU hours to support high-quality sub-network sampling. In this work, we propose SparseNAS, an approach for one-shot NAS acceleration by reducing the redundancy of the search space. We observe that many sub-networks in the space are underperforming, with significant performance disparity to high-performance sub-networks. Crucially, this disparity can be observed early in the beginning of the supernet training. Therefore, we train an early predictor to learn this disparity and filter out high-quality networks in advance. Then, the supernet training will be conducted in this space sub-space. Compared to the state-of-the-art one-shot NAS, our SparseNAS reports a 3.1× training speedup with comparable network performance on the ImageNet dataset. Compared to the state-of-the-art acceleration method, SparseNAS reports a maximum of 1.5% higher Top-1 accuracy and 28% training cost reduction with a 7× bigger search space. Extensive experiment results demonstrated that SparseNAS achieves better trade-offs between efficiency and performance than state-of-the-art one-shot NAS. [Display omitted] •Introduce efficient one-shot Neural Architecture Search (NAS) for network design.•Investigate and analyze the substantial costs of one-shot NAS searching.•Propose a sub-space building approach for efficient NAS.•Report on-device performance evaluation.
doi_str_mv 10.1016/j.knosys.2024.112620
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_knosys_2024_112620</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0950705124012541</els_id><sourcerecordid>S0950705124012541</sourcerecordid><originalsourceid>FETCH-LOGICAL-c185t-604f9e959e59a098329e948c0ea3e65975ce4f279b7e82aa07d7e4ee0c9b8e0f3</originalsourceid><addsrcrecordid>eNp9kE1Lw0AQhvegYK3-Aw_7BxJnt9kkexFK8QsKXhSPy3Y6sRtjUnaSQv-9qRGPnoYX5nmZeYS4UZAqUPltnX62HR851aCzVCmdazgTM7AGkgKMuhCXzDUAaK3KmXhfIlJD0feh_ZBdSwnvul62NETfSB9xF3rCfogkmU5RHoKX2LXcxwF_IC957yP_LYwJ6UqcV75huv6dc_H2cP-6ekrWL4_Pq-U6QVWaPskhqyxZY8lYD7Zc6DFlJQL5BeXGFgYpq3RhNwWV2nsotgVlRIB2UxJUi7nIpl6MHXOkyu1j-PLx6BS4kxBXu0mIOwlxk5ARu5swGm87BIqOMVCLtA1xfNdtu_B_wTeEEm9x</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Accelerating one-shot neural architecture search via constructing a sparse search space</title><source>ScienceDirect Freedom Collection</source><creator>Huang, Hongtao ; Chang, Xiaojun ; Yao, Lina</creator><creatorcontrib>Huang, Hongtao ; Chang, Xiaojun ; Yao, Lina</creatorcontrib><description>Neural Architecture Search (NAS) has garnered significant attention for its ability to automatically design high-quality deep neural networks (DNNs) tailored to various hardware platforms. The major challenge for NAS is the time-consuming network estimation process required to select optimal networks from a large pool of candidates. Rather than training each candidate from scratch, recent one-shot NAS methods accelerate the estimation process by only training a supernet and sampling sub-networks from it, inheriting partial network architectures and weights. Despite significant acceleration, the supernet training with a large search space (i.e., the number of candidate sub-networks) still requires thousands of GPU hours to support high-quality sub-network sampling. In this work, we propose SparseNAS, an approach for one-shot NAS acceleration by reducing the redundancy of the search space. We observe that many sub-networks in the space are underperforming, with significant performance disparity to high-performance sub-networks. Crucially, this disparity can be observed early in the beginning of the supernet training. Therefore, we train an early predictor to learn this disparity and filter out high-quality networks in advance. Then, the supernet training will be conducted in this space sub-space. Compared to the state-of-the-art one-shot NAS, our SparseNAS reports a 3.1× training speedup with comparable network performance on the ImageNet dataset. Compared to the state-of-the-art acceleration method, SparseNAS reports a maximum of 1.5% higher Top-1 accuracy and 28% training cost reduction with a 7× bigger search space. Extensive experiment results demonstrated that SparseNAS achieves better trade-offs between efficiency and performance than state-of-the-art one-shot NAS. [Display omitted] •Introduce efficient one-shot Neural Architecture Search (NAS) for network design.•Investigate and analyze the substantial costs of one-shot NAS searching.•Propose a sub-space building approach for efficient NAS.•Report on-device performance evaluation.</description><identifier>ISSN: 0950-7051</identifier><identifier>DOI: 10.1016/j.knosys.2024.112620</identifier><language>eng</language><publisher>Elsevier B.V</publisher><subject>Automated machine learning (AutoML) ; Multi-platform network deployment ; Neural architecture search (NAS)</subject><ispartof>Knowledge-based systems, 2024-12, Vol.305, p.112620, Article 112620</ispartof><rights>2024 The Authors</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c185t-604f9e959e59a098329e948c0ea3e65975ce4f279b7e82aa07d7e4ee0c9b8e0f3</cites><orcidid>0009-0001-1239-1286</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Huang, Hongtao</creatorcontrib><creatorcontrib>Chang, Xiaojun</creatorcontrib><creatorcontrib>Yao, Lina</creatorcontrib><title>Accelerating one-shot neural architecture search via constructing a sparse search space</title><title>Knowledge-based systems</title><description>Neural Architecture Search (NAS) has garnered significant attention for its ability to automatically design high-quality deep neural networks (DNNs) tailored to various hardware platforms. The major challenge for NAS is the time-consuming network estimation process required to select optimal networks from a large pool of candidates. Rather than training each candidate from scratch, recent one-shot NAS methods accelerate the estimation process by only training a supernet and sampling sub-networks from it, inheriting partial network architectures and weights. Despite significant acceleration, the supernet training with a large search space (i.e., the number of candidate sub-networks) still requires thousands of GPU hours to support high-quality sub-network sampling. In this work, we propose SparseNAS, an approach for one-shot NAS acceleration by reducing the redundancy of the search space. We observe that many sub-networks in the space are underperforming, with significant performance disparity to high-performance sub-networks. Crucially, this disparity can be observed early in the beginning of the supernet training. Therefore, we train an early predictor to learn this disparity and filter out high-quality networks in advance. Then, the supernet training will be conducted in this space sub-space. Compared to the state-of-the-art one-shot NAS, our SparseNAS reports a 3.1× training speedup with comparable network performance on the ImageNet dataset. Compared to the state-of-the-art acceleration method, SparseNAS reports a maximum of 1.5% higher Top-1 accuracy and 28% training cost reduction with a 7× bigger search space. Extensive experiment results demonstrated that SparseNAS achieves better trade-offs between efficiency and performance than state-of-the-art one-shot NAS. [Display omitted] •Introduce efficient one-shot Neural Architecture Search (NAS) for network design.•Investigate and analyze the substantial costs of one-shot NAS searching.•Propose a sub-space building approach for efficient NAS.•Report on-device performance evaluation.</description><subject>Automated machine learning (AutoML)</subject><subject>Multi-platform network deployment</subject><subject>Neural architecture search (NAS)</subject><issn>0950-7051</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp9kE1Lw0AQhvegYK3-Aw_7BxJnt9kkexFK8QsKXhSPy3Y6sRtjUnaSQv-9qRGPnoYX5nmZeYS4UZAqUPltnX62HR851aCzVCmdazgTM7AGkgKMuhCXzDUAaK3KmXhfIlJD0feh_ZBdSwnvul62NETfSB9xF3rCfogkmU5RHoKX2LXcxwF_IC957yP_LYwJ6UqcV75huv6dc_H2cP-6ekrWL4_Pq-U6QVWaPskhqyxZY8lYD7Zc6DFlJQL5BeXGFgYpq3RhNwWV2nsotgVlRIB2UxJUi7nIpl6MHXOkyu1j-PLx6BS4kxBXu0mIOwlxk5ARu5swGm87BIqOMVCLtA1xfNdtu_B_wTeEEm9x</recordid><startdate>20241203</startdate><enddate>20241203</enddate><creator>Huang, Hongtao</creator><creator>Chang, Xiaojun</creator><creator>Yao, Lina</creator><general>Elsevier B.V</general><scope>6I.</scope><scope>AAFTH</scope><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0009-0001-1239-1286</orcidid></search><sort><creationdate>20241203</creationdate><title>Accelerating one-shot neural architecture search via constructing a sparse search space</title><author>Huang, Hongtao ; Chang, Xiaojun ; Yao, Lina</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c185t-604f9e959e59a098329e948c0ea3e65975ce4f279b7e82aa07d7e4ee0c9b8e0f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Automated machine learning (AutoML)</topic><topic>Multi-platform network deployment</topic><topic>Neural architecture search (NAS)</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Huang, Hongtao</creatorcontrib><creatorcontrib>Chang, Xiaojun</creatorcontrib><creatorcontrib>Yao, Lina</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>CrossRef</collection><jtitle>Knowledge-based systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Huang, Hongtao</au><au>Chang, Xiaojun</au><au>Yao, Lina</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Accelerating one-shot neural architecture search via constructing a sparse search space</atitle><jtitle>Knowledge-based systems</jtitle><date>2024-12-03</date><risdate>2024</risdate><volume>305</volume><spage>112620</spage><pages>112620-</pages><artnum>112620</artnum><issn>0950-7051</issn><abstract>Neural Architecture Search (NAS) has garnered significant attention for its ability to automatically design high-quality deep neural networks (DNNs) tailored to various hardware platforms. The major challenge for NAS is the time-consuming network estimation process required to select optimal networks from a large pool of candidates. Rather than training each candidate from scratch, recent one-shot NAS methods accelerate the estimation process by only training a supernet and sampling sub-networks from it, inheriting partial network architectures and weights. Despite significant acceleration, the supernet training with a large search space (i.e., the number of candidate sub-networks) still requires thousands of GPU hours to support high-quality sub-network sampling. In this work, we propose SparseNAS, an approach for one-shot NAS acceleration by reducing the redundancy of the search space. We observe that many sub-networks in the space are underperforming, with significant performance disparity to high-performance sub-networks. Crucially, this disparity can be observed early in the beginning of the supernet training. Therefore, we train an early predictor to learn this disparity and filter out high-quality networks in advance. Then, the supernet training will be conducted in this space sub-space. Compared to the state-of-the-art one-shot NAS, our SparseNAS reports a 3.1× training speedup with comparable network performance on the ImageNet dataset. Compared to the state-of-the-art acceleration method, SparseNAS reports a maximum of 1.5% higher Top-1 accuracy and 28% training cost reduction with a 7× bigger search space. Extensive experiment results demonstrated that SparseNAS achieves better trade-offs between efficiency and performance than state-of-the-art one-shot NAS. [Display omitted] •Introduce efficient one-shot Neural Architecture Search (NAS) for network design.•Investigate and analyze the substantial costs of one-shot NAS searching.•Propose a sub-space building approach for efficient NAS.•Report on-device performance evaluation.</abstract><pub>Elsevier B.V</pub><doi>10.1016/j.knosys.2024.112620</doi><orcidid>https://orcid.org/0009-0001-1239-1286</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0950-7051
ispartof Knowledge-based systems, 2024-12, Vol.305, p.112620, Article 112620
issn 0950-7051
language eng
recordid cdi_crossref_primary_10_1016_j_knosys_2024_112620
source ScienceDirect Freedom Collection
subjects Automated machine learning (AutoML)
Multi-platform network deployment
Neural architecture search (NAS)
title Accelerating one-shot neural architecture search via constructing a sparse search space
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-01T01%3A14%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Accelerating%20one-shot%20neural%20architecture%20search%20via%20constructing%20a%20sparse%20search%20space&rft.jtitle=Knowledge-based%20systems&rft.au=Huang,%20Hongtao&rft.date=2024-12-03&rft.volume=305&rft.spage=112620&rft.pages=112620-&rft.artnum=112620&rft.issn=0950-7051&rft_id=info:doi/10.1016/j.knosys.2024.112620&rft_dat=%3Celsevier_cross%3ES0950705124012541%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c185t-604f9e959e59a098329e948c0ea3e65975ce4f279b7e82aa07d7e4ee0c9b8e0f3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true