Loading…
On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks
Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weigh...
Saved in:
Published in: | arXiv.org 2016-07 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Julius Gopinath Mahale Sumana, T Adityakrishna, C S |
description | Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error function may conform to general results of the statistics of large systems. To this end we apply some results from Random Matrix Theory to analyse these functions. We model the error function in terms of a Hamiltonian in N-dimensions and derive some theoretical results about its general behavior. These results are further used to make better initial guesses of weights for the learning algorithm. |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2079939416</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2079939416</sourcerecordid><originalsourceid>FETCH-proquest_journals_20799394163</originalsourceid><addsrcrecordid>eNqNjcsKwjAQRYMgKOo_DLgWYqKtXftAwcdGcCmhnWq0TmomRfDrjeAHuLpwOefelugqrcej2USpjhgw36SUKknVdKq7oj4QhCvCzhVYWbqAK2HpvfOwaigP1hGDYVjbyxUW9oHEsTIVbA0VnJsaGcoInzACATZkgzWVfZuvCZZgi8bTd3eP4eX8nfuiXZqKcfDLnhiulsf5elR792yQw_nmGh8v-KxkmmU6m4wT_R_1AYeqSp4</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2079939416</pqid></control><display><type>article</type><title>On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks</title><source>Publicly Available Content Database</source><creator>Julius ; Gopinath Mahale ; Sumana, T ; Adityakrishna, C S</creator><creatorcontrib>Julius ; Gopinath Mahale ; Sumana, T ; Adityakrishna, C S</creatorcontrib><description>Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error function may conform to general results of the statistics of large systems. To this end we apply some results from Random Matrix Theory to analyse these functions. We model the error function in terms of a Hamiltonian in N-dimensions and derive some theoretical results about its general behavior. These results are further used to make better initial guesses of weights for the learning algorithm.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Classification ; Error functions ; Landscape ; Machine learning ; Mathematical models ; Matrix theory ; Modelling ; Neural networks ; Weight</subject><ispartof>arXiv.org, 2016-07</ispartof><rights>2016. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2079939416?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25753,37012,44590</link.rule.ids></links><search><creatorcontrib>Julius</creatorcontrib><creatorcontrib>Gopinath Mahale</creatorcontrib><creatorcontrib>Sumana, T</creatorcontrib><creatorcontrib>Adityakrishna, C S</creatorcontrib><title>On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks</title><title>arXiv.org</title><description>Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error function may conform to general results of the statistics of large systems. To this end we apply some results from Random Matrix Theory to analyse these functions. We model the error function in terms of a Hamiltonian in N-dimensions and derive some theoretical results about its general behavior. These results are further used to make better initial guesses of weights for the learning algorithm.</description><subject>Algorithms</subject><subject>Classification</subject><subject>Error functions</subject><subject>Landscape</subject><subject>Machine learning</subject><subject>Mathematical models</subject><subject>Matrix theory</subject><subject>Modelling</subject><subject>Neural networks</subject><subject>Weight</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2016</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNqNjcsKwjAQRYMgKOo_DLgWYqKtXftAwcdGcCmhnWq0TmomRfDrjeAHuLpwOefelugqrcej2USpjhgw36SUKknVdKq7oj4QhCvCzhVYWbqAK2HpvfOwaigP1hGDYVjbyxUW9oHEsTIVbA0VnJsaGcoInzACATZkgzWVfZuvCZZgi8bTd3eP4eX8nfuiXZqKcfDLnhiulsf5elR792yQw_nmGh8v-KxkmmU6m4wT_R_1AYeqSp4</recordid><startdate>20160720</startdate><enddate>20160720</enddate><creator>Julius</creator><creator>Gopinath Mahale</creator><creator>Sumana, T</creator><creator>Adityakrishna, C S</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20160720</creationdate><title>On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks</title><author>Julius ; Gopinath Mahale ; Sumana, T ; Adityakrishna, C S</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_20799394163</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2016</creationdate><topic>Algorithms</topic><topic>Classification</topic><topic>Error functions</topic><topic>Landscape</topic><topic>Machine learning</topic><topic>Mathematical models</topic><topic>Matrix theory</topic><topic>Modelling</topic><topic>Neural networks</topic><topic>Weight</topic><toplevel>online_resources</toplevel><creatorcontrib>Julius</creatorcontrib><creatorcontrib>Gopinath Mahale</creatorcontrib><creatorcontrib>Sumana, T</creatorcontrib><creatorcontrib>Adityakrishna, C S</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>AUTh Library subscriptions: ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Julius</au><au>Gopinath Mahale</au><au>Sumana, T</au><au>Adityakrishna, C S</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks</atitle><jtitle>arXiv.org</jtitle><date>2016-07-20</date><risdate>2016</risdate><eissn>2331-8422</eissn><abstract>Next generation deep neural networks for classification hosted on embedded platforms will rely on fast, efficient, and accurate learning algorithms. Initialization of weights in learning networks has a great impact on the classification accuracy. In this paper we focus on deriving good initial weights by modeling the error function of a deep neural network as a high-dimensional landscape. We observe that due to the inherent complexity in its algebraic structure, such an error function may conform to general results of the statistics of large systems. To this end we apply some results from Random Matrix Theory to analyse these functions. We model the error function in terms of a Hamiltonian in N-dimensions and derive some theoretical results about its general behavior. These results are further used to make better initial guesses of weights for the learning algorithm.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2016-07 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2079939416 |
source | Publicly Available Content Database |
subjects | Algorithms Classification Error functions Landscape Machine learning Mathematical models Matrix theory Modelling Neural networks Weight |
title | On the Modeling of Error Functions as High Dimensional Landscapes for Weight Initialization in Learning Networks |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T21%3A35%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=On%20the%20Modeling%20of%20Error%20Functions%20as%20High%20Dimensional%20Landscapes%20for%20Weight%20Initialization%20in%20Learning%20Networks&rft.jtitle=arXiv.org&rft.au=Julius&rft.date=2016-07-20&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2079939416%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-proquest_journals_20799394163%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2079939416&rft_id=info:pmid/&rfr_iscdi=true |