Loading…

BranchyNet: Fast Inference via Early Exiting from Deep Neural Networks

Deep neural networks are state of the art methods for many learning tasks due to their ability to extract increasingly better features at each network layer. However, the improved performance of additional layers in a deep network comes at the cost of added latency and energy usage in feedforward in...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2017-09
Main Authors: Teerapittayanon, Surat, McDanel, Bradley, Kung, H T
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Teerapittayanon, Surat
McDanel, Bradley
Kung, H T
description Deep neural networks are state of the art methods for many learning tasks due to their ability to extract increasingly better features at each network layer. However, the improved performance of additional layers in a deep network comes at the cost of added latency and energy usage in feedforward inference. As networks continue to get deeper and larger, these costs become more prohibitive for real-time and energy-sensitive applications. To address this issue, we present BranchyNet, a novel deep network architecture that is augmented with additional side branch classifiers. The architecture allows prediction results for a large portion of test samples to exit the network early via these branches when samples can already be inferred with high confidence. BranchyNet exploits the observation that features learned at an early layer of a network may often be sufficient for the classification of many data points. For more difficult samples, which are expected less frequently, BranchyNet will use further or all network layers to provide the best likelihood of correct prediction. We study the BranchyNet architecture using several well-known networks (LeNet, AlexNet, ResNet) and datasets (MNIST, CIFAR10) and show that it can both improve accuracy and significantly reduce the inference time of the network.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2075875980</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2075875980</sourcerecordid><originalsourceid>FETCH-proquest_journals_20758759803</originalsourceid><addsrcrecordid>eNqNyrEOgjAUQNHGxESi_MNLnElqawUdVYguTO6kIQ8FscXXovL3MvgBTme4d8ICIeUqStZCzFjoXMM5F5tYKCUDlu1Jm_I25Oh3kGnn4WwqJDQlwqvWkGpqB0g_ta_NFSqyDzgidpBjT7od8W9Ld7dg00q3DsOfc7bM0svhFHVknz06XzS2JzOmQvBYJbHaJlz-d30BWvY6oQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2075875980</pqid></control><display><type>article</type><title>BranchyNet: Fast Inference via Early Exiting from Deep Neural Networks</title><source>Publicly Available Content (ProQuest)</source><creator>Teerapittayanon, Surat ; McDanel, Bradley ; Kung, H T</creator><creatorcontrib>Teerapittayanon, Surat ; McDanel, Bradley ; Kung, H T</creatorcontrib><description>Deep neural networks are state of the art methods for many learning tasks due to their ability to extract increasingly better features at each network layer. However, the improved performance of additional layers in a deep network comes at the cost of added latency and energy usage in feedforward inference. As networks continue to get deeper and larger, these costs become more prohibitive for real-time and energy-sensitive applications. To address this issue, we present BranchyNet, a novel deep network architecture that is augmented with additional side branch classifiers. The architecture allows prediction results for a large portion of test samples to exit the network early via these branches when samples can already be inferred with high confidence. BranchyNet exploits the observation that features learned at an early layer of a network may often be sufficient for the classification of many data points. For more difficult samples, which are expected less frequently, BranchyNet will use further or all network layers to provide the best likelihood of correct prediction. We study the BranchyNet architecture using several well-known networks (LeNet, AlexNet, ResNet) and datasets (MNIST, CIFAR10) and show that it can both improve accuracy and significantly reduce the inference time of the network.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Data points ; Energy consumption ; Feature extraction ; Inference ; Neural networks</subject><ispartof>arXiv.org, 2017-09</ispartof><rights>2017. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2075875980?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25752,37011,44589</link.rule.ids></links><search><creatorcontrib>Teerapittayanon, Surat</creatorcontrib><creatorcontrib>McDanel, Bradley</creatorcontrib><creatorcontrib>Kung, H T</creatorcontrib><title>BranchyNet: Fast Inference via Early Exiting from Deep Neural Networks</title><title>arXiv.org</title><description>Deep neural networks are state of the art methods for many learning tasks due to their ability to extract increasingly better features at each network layer. However, the improved performance of additional layers in a deep network comes at the cost of added latency and energy usage in feedforward inference. As networks continue to get deeper and larger, these costs become more prohibitive for real-time and energy-sensitive applications. To address this issue, we present BranchyNet, a novel deep network architecture that is augmented with additional side branch classifiers. The architecture allows prediction results for a large portion of test samples to exit the network early via these branches when samples can already be inferred with high confidence. BranchyNet exploits the observation that features learned at an early layer of a network may often be sufficient for the classification of many data points. For more difficult samples, which are expected less frequently, BranchyNet will use further or all network layers to provide the best likelihood of correct prediction. We study the BranchyNet architecture using several well-known networks (LeNet, AlexNet, ResNet) and datasets (MNIST, CIFAR10) and show that it can both improve accuracy and significantly reduce the inference time of the network.</description><subject>Data points</subject><subject>Energy consumption</subject><subject>Feature extraction</subject><subject>Inference</subject><subject>Neural networks</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2017</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNyrEOgjAUQNHGxESi_MNLnElqawUdVYguTO6kIQ8FscXXovL3MvgBTme4d8ICIeUqStZCzFjoXMM5F5tYKCUDlu1Jm_I25Oh3kGnn4WwqJDQlwqvWkGpqB0g_ta_NFSqyDzgidpBjT7od8W9Ld7dg00q3DsOfc7bM0svhFHVknz06XzS2JzOmQvBYJbHaJlz-d30BWvY6oQ</recordid><startdate>20170906</startdate><enddate>20170906</enddate><creator>Teerapittayanon, Surat</creator><creator>McDanel, Bradley</creator><creator>Kung, H T</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20170906</creationdate><title>BranchyNet: Fast Inference via Early Exiting from Deep Neural Networks</title><author>Teerapittayanon, Surat ; McDanel, Bradley ; Kung, H T</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_20758759803</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2017</creationdate><topic>Data points</topic><topic>Energy consumption</topic><topic>Feature extraction</topic><topic>Inference</topic><topic>Neural networks</topic><toplevel>online_resources</toplevel><creatorcontrib>Teerapittayanon, Surat</creatorcontrib><creatorcontrib>McDanel, Bradley</creatorcontrib><creatorcontrib>Kung, H T</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content (ProQuest)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Teerapittayanon, Surat</au><au>McDanel, Bradley</au><au>Kung, H T</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>BranchyNet: Fast Inference via Early Exiting from Deep Neural Networks</atitle><jtitle>arXiv.org</jtitle><date>2017-09-06</date><risdate>2017</risdate><eissn>2331-8422</eissn><abstract>Deep neural networks are state of the art methods for many learning tasks due to their ability to extract increasingly better features at each network layer. However, the improved performance of additional layers in a deep network comes at the cost of added latency and energy usage in feedforward inference. As networks continue to get deeper and larger, these costs become more prohibitive for real-time and energy-sensitive applications. To address this issue, we present BranchyNet, a novel deep network architecture that is augmented with additional side branch classifiers. The architecture allows prediction results for a large portion of test samples to exit the network early via these branches when samples can already be inferred with high confidence. BranchyNet exploits the observation that features learned at an early layer of a network may often be sufficient for the classification of many data points. For more difficult samples, which are expected less frequently, BranchyNet will use further or all network layers to provide the best likelihood of correct prediction. We study the BranchyNet architecture using several well-known networks (LeNet, AlexNet, ResNet) and datasets (MNIST, CIFAR10) and show that it can both improve accuracy and significantly reduce the inference time of the network.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2017-09
issn 2331-8422
language eng
recordid cdi_proquest_journals_2075875980
source Publicly Available Content (ProQuest)
subjects Data points
Energy consumption
Feature extraction
Inference
Neural networks
title BranchyNet: Fast Inference via Early Exiting from Deep Neural Networks
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T17%3A02%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=BranchyNet:%20Fast%20Inference%20via%20Early%20Exiting%20from%20Deep%20Neural%20Networks&rft.jtitle=arXiv.org&rft.au=Teerapittayanon,%20Surat&rft.date=2017-09-06&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2075875980%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_20758759803%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2075875980&rft_id=info:pmid/&rfr_iscdi=true