Loading…

On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks

Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weigh...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2016-07
Main Authors: Julius, Gopinath Mahale, Sumana, T, Adityakrishna, C S
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Julius
Gopinath Mahale
Sumana, T
Adityakrishna, C S
description Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error function may conform to general results of the statistics of large systems. To this end we apply some results from Random Matrix Theory to analyse these functions. We model the error function in terms of a Hamiltonian in N-dimensions and derive some theoretical results about its general behavior. These results are further used to make better initial guesses of weights for the learning algorithm.
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2079939416</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2079939416</sourcerecordid><originalsourceid>FETCH-proquest_journals_20799394163</originalsourceid><addsrcrecordid>eNqNjcsKwjAQRYMgKOo_DLgWYqKtXftAwcdGcCmhnWq0TmomRfDrjeAHuLpwOefelugqrcej2USpjhgw36SUKknVdKq7oj4QhCvCzhVYWbqAK2HpvfOwaigP1hGDYVjbyxUW9oHEsTIVbA0VnJsaGcoInzACATZkgzWVfZuvCZZgi8bTd3eP4eX8nfuiXZqKcfDLnhiulsf5elR792yQw_nmGh8v-KxkmmU6m4wT_R_1AYeqSp4</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2079939416</pqid></control><display><type>article</type><title>On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks</title><source>Publicly Available Content Database</source><creator>Julius ; Gopinath Mahale ; Sumana, T ; Adityakrishna, C S</creator><creatorcontrib>Julius ; Gopinath Mahale ; Sumana, T ; Adityakrishna, C S</creatorcontrib><description>Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error function may conform to general results of the statistics of large systems. To this end we apply some results from Random Matrix Theory to analyse these functions. We model the error function in terms of a Hamiltonian in N-dimensions and derive some theoretical results about its general behavior. These results are further used to make better initial guesses of weights for the learning algorithm.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Classification ; Error functions ; Landscape ; Machine learning ; Mathematical models ; Matrix theory ; Modelling ; Neural networks ; Weight</subject><ispartof>arXiv.org, 2016-07</ispartof><rights>2016. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2079939416?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Julius</creatorcontrib><creatorcontrib>Gopinath Mahale</creatorcontrib><creatorcontrib>Sumana, T</creatorcontrib><creatorcontrib>Adityakrishna, C S</creatorcontrib><title>On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks</title><title>arXiv.org</title><description>Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error function may conform to general results of the statistics of large systems. To this end we apply some results from Random Matrix Theory to analyse these functions. We model the error function in terms of a Hamiltonian in N-dimensions and derive some theoretical results about its general behavior. These results are further used to make better initial guesses of weights for the learning algorithm.</description><subject>Algorithms</subject><subject>Classification</subject><subject>Error functions</subject><subject>Landscape</subject><subject>Machine learning</subject><subject>Mathematical models</subject><subject>Matrix theory</subject><subject>Modelling</subject><subject>Neural networks</subject><subject>Weight</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2016</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNjcsKwjAQRYMgKOo_DLgWYqKtXftAwcdGcCmhnWq0TmomRfDrjeAHuLpwOefelugqrcej2USpjhgw36SUKknVdKq7oj4QhCvCzhVYWbqAK2HpvfOwaigP1hGDYVjbyxUW9oHEsTIVbA0VnJsaGcoInzACATZkgzWVfZuvCZZgi8bTd3eP4eX8nfuiXZqKcfDLnhiulsf5elR792yQw_nmGh8v-KxkmmU6m4wT_R_1AYeqSp4</recordid><startdate>20160720</startdate><enddate>20160720</enddate><creator>Julius</creator><creator>Gopinath Mahale</creator><creator>Sumana, T</creator><creator>Adityakrishna, C S</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20160720</creationdate><title>On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks</title><author>Julius ; Gopinath Mahale ; Sumana, T ; Adityakrishna, C S</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_20799394163</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2016</creationdate><topic>Algorithms</topic><topic>Classification</topic><topic>Error functions</topic><topic>Landscape</topic><topic>Machine learning</topic><topic>Mathematical models</topic><topic>Matrix theory</topic><topic>Modelling</topic><topic>Neural networks</topic><topic>Weight</topic><toplevel>online_resources</toplevel><creatorcontrib>Julius</creatorcontrib><creatorcontrib>Gopinath Mahale</creatorcontrib><creatorcontrib>Sumana, T</creatorcontrib><creatorcontrib>Adityakrishna, C S</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Julius</au><au>Gopinath Mahale</au><au>Sumana, T</au><au>Adityakrishna, C S</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks</atitle><jtitle>arXiv.org</jtitle><date>2016-07-20</date><risdate>2016</risdate><eissn>2331-8422</eissn><abstract>Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error function may conform to general results of the statistics of large systems. To this end we apply some results from Random Matrix Theory to analyse these functions. We model the error function in terms of a Hamiltonian in N-dimensions and derive some theoretical results about its general behavior. These results are further used to make better initial guesses of weights for the learning algorithm.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2016-07
issn 2331-8422
language eng
recordid cdi_proquest_journals_2079939416
source Publicly Available Content Database
subjects Algorithms
Classification
Error functions
Landscape
Machine learning
Mathematical models
Matrix theory
Modelling
Neural networks
Weight
title On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T21%3A35%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=On%20the%20Modeling%20of%20Error%20Functions%20as%20High%20Dimensional%20Landscapes%20for%20Weight%20Initialization%20in%20Learning%20Networks&rft.jtitle=arXiv.org&rft.au=Julius&rft.date=2016-07-20&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2079939416%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_20799394163%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2079939416&rft_id=info:pmid/&rfr_iscdi=true