Loading…
Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better
Deep Learning has revolutionized the fields of computer vision, natural language understanding, speech recognition, information retrieval and more. However, with the progressive improvements in deep learning models, their number of parameters, latency, resources required to train, etc. have all have...
Saved in:
Published in: | arXiv.org 2021-06 |
---|---|
Main Author: | |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-a1432-dd854bb8a63068a4ad6612004f8025a9e74d442f427c2b5f784c72c175efd8b83 |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Menghani, Gaurav |
description | Deep Learning has revolutionized the fields of computer vision, natural language understanding, speech recognition, information retrieval and more. However, with the progressive improvements in deep learning models, their number of parameters, latency, resources required to train, etc. have all have increased significantly. Consequently, it has become important to pay attention to these footprint metrics of a model as well, not just its quality. We present and motivate the problem of efficiency in deep learning, followed by a thorough survey of the five core areas of model efficiency (spanning modeling techniques, infrastructure, and hardware) and the seminal work there. We also present an experiment-based guide along with code, for practitioners to optimize their model training and deployment. We believe this is the first comprehensive survey in the efficient deep learning space that covers the landscape of model efficiency from modeling techniques to hardware support. Our hope is that this survey would provide the reader with the mental model and the necessary understanding of the field to apply generic efficiency techniques to immediately get significant improvements, and also equip them with ideas for further research and experimentation to achieve additional gains. |
doi_str_mv | 10.48550/arxiv.2106.08962 |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2543475295</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2543475295</sourcerecordid><originalsourceid>FETCH-LOGICAL-a1432-dd854bb8a63068a4ad6612004f8025a9e74d442f427c2b5f784c72c175efd8b83</originalsourceid><addsrcrecordid>eNpVjU1Lw0AURQdBsNT-AHcDbk2cvHnzEXe1tiqkuKjuhDLJzEhqnNRMUuq_N6IbV-dyOdxLyEXGUtRCsGvTHetDChmTKdO5hBMyAc6zRCPAGZnFuGOMgVQgBJ-Q16X3dVW70NM75_a0cKYLdXi7oXO6GbqD-6JtoGvzPnb_DbpurWsi3XyYpnHdFV2Z2P_QBEtvXT_mc3LqTRPd7I9T8rJaPi8ekuLp_nExLxKTIYfEWi2wLLWRnElt0FgpM2AMvWYgTO4UWkTwCKqCUnilsVJQZUo4b3Wp-ZRc_u7uu_ZzcLHf7tqhC-PlFgRyVAJywb8BWQFS4w</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2543475295</pqid></control><display><type>article</type><title>Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better</title><source>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</source><creator>Menghani, Gaurav</creator><creatorcontrib>Menghani, Gaurav</creatorcontrib><description>Deep Learning has revolutionized the fields of computer vision, natural language understanding, speech recognition, information retrieval and more. However, with the progressive improvements in deep learning models, their number of parameters, latency, resources required to train, etc. have all have increased significantly. Consequently, it has become important to pay attention to these footprint metrics of a model as well, not just its quality. We present and motivate the problem of efficiency in deep learning, followed by a thorough survey of the five core areas of model efficiency (spanning modeling techniques, infrastructure, and hardware) and the seminal work there. We also present an experiment-based guide along with code, for practitioners to optimize their model training and deployment. We believe this is the first comprehensive survey in the efficient deep learning space that covers the landscape of model efficiency from modeling techniques to hardware support. Our hope is that this survey would provide the reader with the mental model and the necessary understanding of the field to apply generic efficiency techniques to immediately get significant improvements, and also equip them with ideas for further research and experimentation to achieve additional gains.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2106.08962</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Computer vision ; Deep learning ; Efficiency ; Experimentation ; Hardware ; Information retrieval ; Machine learning ; Modelling ; Natural language (computers) ; Speech recognition</subject><ispartof>arXiv.org, 2021-06</ispartof><rights>2021. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-a1432-dd854bb8a63068a4ad6612004f8025a9e74d442f427c2b5f784c72c175efd8b83</citedby></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2543475295?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,27925,37012,44590</link.rule.ids></links><search><creatorcontrib>Menghani, Gaurav</creatorcontrib><title>Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better</title><title>arXiv.org</title><description>Deep Learning has revolutionized the fields of computer vision, natural language understanding, speech recognition, information retrieval and more. However, with the progressive improvements in deep learning models, their number of parameters, latency, resources required to train, etc. have all have increased significantly. Consequently, it has become important to pay attention to these footprint metrics of a model as well, not just its quality. We present and motivate the problem of efficiency in deep learning, followed by a thorough survey of the five core areas of model efficiency (spanning modeling techniques, infrastructure, and hardware) and the seminal work there. We also present an experiment-based guide along with code, for practitioners to optimize their model training and deployment. We believe this is the first comprehensive survey in the efficient deep learning space that covers the landscape of model efficiency from modeling techniques to hardware support. Our hope is that this survey would provide the reader with the mental model and the necessary understanding of the field to apply generic efficiency techniques to immediately get significant improvements, and also equip them with ideas for further research and experimentation to achieve additional gains.</description><subject>Computer vision</subject><subject>Deep learning</subject><subject>Efficiency</subject><subject>Experimentation</subject><subject>Hardware</subject><subject>Information retrieval</subject><subject>Machine learning</subject><subject>Modelling</subject><subject>Natural language (computers)</subject><subject>Speech recognition</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNpVjU1Lw0AURQdBsNT-AHcDbk2cvHnzEXe1tiqkuKjuhDLJzEhqnNRMUuq_N6IbV-dyOdxLyEXGUtRCsGvTHetDChmTKdO5hBMyAc6zRCPAGZnFuGOMgVQgBJ-Q16X3dVW70NM75_a0cKYLdXi7oXO6GbqD-6JtoGvzPnb_DbpurWsi3XyYpnHdFV2Z2P_QBEtvXT_mc3LqTRPd7I9T8rJaPi8ekuLp_nExLxKTIYfEWi2wLLWRnElt0FgpM2AMvWYgTO4UWkTwCKqCUnilsVJQZUo4b3Wp-ZRc_u7uu_ZzcLHf7tqhC-PlFgRyVAJywb8BWQFS4w</recordid><startdate>20210621</startdate><enddate>20210621</enddate><creator>Menghani, Gaurav</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20210621</creationdate><title>Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better</title><author>Menghani, Gaurav</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a1432-dd854bb8a63068a4ad6612004f8025a9e74d442f427c2b5f784c72c175efd8b83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer vision</topic><topic>Deep learning</topic><topic>Efficiency</topic><topic>Experimentation</topic><topic>Hardware</topic><topic>Information retrieval</topic><topic>Machine learning</topic><topic>Modelling</topic><topic>Natural language (computers)</topic><topic>Speech recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Menghani, Gaurav</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database (Proquest) (PQ_SDU_P3)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Menghani, Gaurav</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better</atitle><jtitle>arXiv.org</jtitle><date>2021-06-21</date><risdate>2021</risdate><eissn>2331-8422</eissn><abstract>Deep Learning has revolutionized the fields of computer vision, natural language understanding, speech recognition, information retrieval and more. However, with the progressive improvements in deep learning models, their number of parameters, latency, resources required to train, etc. have all have increased significantly. Consequently, it has become important to pay attention to these footprint metrics of a model as well, not just its quality. We present and motivate the problem of efficiency in deep learning, followed by a thorough survey of the five core areas of model efficiency (spanning modeling techniques, infrastructure, and hardware) and the seminal work there. We also present an experiment-based guide along with code, for practitioners to optimize their model training and deployment. We believe this is the first comprehensive survey in the efficient deep learning space that covers the landscape of model efficiency from modeling techniques to hardware support. Our hope is that this survey would provide the reader with the mental model and the necessary understanding of the field to apply generic efficiency techniques to immediately get significant improvements, and also equip them with ideas for further research and experimentation to achieve additional gains.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2106.08962</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2021-06 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2543475295 |
source | Publicly Available Content Database (Proquest) (PQ_SDU_P3) |
subjects | Computer vision Deep learning Efficiency Experimentation Hardware Information retrieval Machine learning Modelling Natural language (computers) Speech recognition |
title | Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-02T16%3A56%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Efficient%20Deep%20Learning:%20A%20Survey%20on%20Making%20Deep%20Learning%20Models%20Smaller,%20Faster,%20and%20Better&rft.jtitle=arXiv.org&rft.au=Menghani,%20Gaurav&rft.date=2021-06-21&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2106.08962&rft_dat=%3Cproquest%3E2543475295%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a1432-dd854bb8a63068a4ad6612004f8025a9e74d442f427c2b5f784c72c175efd8b83%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2543475295&rft_id=info:pmid/&rfr_iscdi=true |