Loading…
Optimization of the weights and asymmetric activation function family of neural network for time series forecasting
•We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series. The use of neural network models for time series forecasting has been motivated by e...
Saved in:
Published in: | Expert systems with applications 2013-11, Vol.40 (16), p.6438-6446 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | cdi_FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923 |
---|---|
cites | cdi_FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923 |
container_end_page | 6446 |
container_issue | 16 |
container_start_page | 6438 |
container_title | Expert systems with applications |
container_volume | 40 |
creator | Gomes, Gecynalda S. da S. Ludermir, Teresa B. |
description | •We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series.
The use of neural network models for time series forecasting has been motivated by experimental results that indicate high capacity for function approximation with good accuracy. Generally, these models use activation functions with fixed parameters. However, it is known that the choice of activation function strongly influences the complexity and neural network performance and that a limited number of activation functions has been used in general. We describe the use of an asymmetric activation functions family with free parameter for neural networks. We prove that the activation functions family defined, satisfies the requirements of the universal approximation theorem We present a methodology for global optimization of the activation functions family with free parameter and the connections between the processing units of the neural network. The main idea is to optimize, simultaneously, the weights and activation function used in a Multilayer Perceptron (MLP), through an approach that combines the advantages of simulated annealing, tabu search and a local learning algorithm. We have chosen two local learning algorithms: the backpropagation with momentum (BPM) and Levenberg–Marquardt (LM). The overall purpose is to improve performance in time series forecasting. |
doi_str_mv | 10.1016/j.eswa.2013.05.053 |
format | article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_1701071117</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417413003515</els_id><sourcerecordid>1701071117</sourcerecordid><originalsourceid>FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923</originalsourceid><addsrcrecordid>eNqFkU1r3DAQhkVpodu0f6AnXQq9eKuxZcuCXkroFwRyyV2M5XGirT-2Gm2W7a-vjEOPDQyMhJ55B72vEO9B7UFB8-mwJz7jvlRQ7VWdq3ohdtCaqmiMrV6KnbK1KTQY_Vq8YT4oBUYpsxN8e0xhCn8whWWWyyDTA8kzhfuHxBLnXiJfpolSDF6iT-FxA4fT7LcDTmG8rIMznSKOuaXzEn_JYYkyK5NkioF4vZNHTmG-fyteDTgyvXvqV-Lu29e76x_Fze33n9dfbgqvTZWKqgPjqTNmMFoN0Nn8K7AeO9shtsoC9KXt-rqn1jaYHy1l0nTaQ0O2rK7Ex032GJffJ-LkpsCexhFnWk7ssgOgDACY59HGQF0qrdrnUa1bU7Z12WS03FAfF-ZIgzvGMGG8OFBujc0d3BqbW2Nzqs5V5aEPT_rIHsch4uwD_5ssTQM1WJ25zxtH2cHHQNGxDzR76kM2Orl-Cf9b8xdFR69i</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1448728526</pqid></control><display><type>article</type><title>Optimization of the weights and asymmetric activation function family of neural network for time series forecasting</title><source>ScienceDirect Freedom Collection 2022-2024</source><creator>Gomes, Gecynalda S. da S. ; Ludermir, Teresa B.</creator><creatorcontrib>Gomes, Gecynalda S. da S. ; Ludermir, Teresa B.</creatorcontrib><description>•We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series.
The use of neural network models for time series forecasting has been motivated by experimental results that indicate high capacity for function approximation with good accuracy. Generally, these models use activation functions with fixed parameters. However, it is known that the choice of activation function strongly influences the complexity and neural network performance and that a limited number of activation functions has been used in general. We describe the use of an asymmetric activation functions family with free parameter for neural networks. We prove that the activation functions family defined, satisfies the requirements of the universal approximation theorem We present a methodology for global optimization of the activation functions family with free parameter and the connections between the processing units of the neural network. The main idea is to optimize, simultaneously, the weights and activation function used in a Multilayer Perceptron (MLP), through an approach that combines the advantages of simulated annealing, tabu search and a local learning algorithm. We have chosen two local learning algorithms: the backpropagation with momentum (BPM) and Levenberg–Marquardt (LM). The overall purpose is to improve performance in time series forecasting.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2013.05.053</identifier><language>eng</language><publisher>Amsterdam: Elsevier Ltd</publisher><subject>Activation ; Algorithmics. Computability. Computer arithmetics ; Algorithms ; Applied sciences ; Artificial intelligence ; Asymmetric activation function ; BPM algorithm ; Computer science; control theory; systems ; Computer simulation ; Connectionism. Neural networks ; Exact sciences and technology ; Forecasting ; Free parameter ; Inference from stochastic processes; time series analysis ; LM algorithm ; Mathematical analysis ; Mathematical models ; Mathematics ; Neural networks ; Probability and statistics ; Sciences and techniques of general use ; Simulated annealing ; Statistics ; Tabu search ; Theoretical computing ; Time series</subject><ispartof>Expert systems with applications, 2013-11, Vol.40 (16), p.6438-6446</ispartof><rights>2013 Elsevier Ltd</rights><rights>2014 INIST-CNRS</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923</citedby><cites>FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&idt=27615194$$DView record in Pascal Francis$$Hfree_for_read</backlink></links><search><creatorcontrib>Gomes, Gecynalda S. da S.</creatorcontrib><creatorcontrib>Ludermir, Teresa B.</creatorcontrib><title>Optimization of the weights and asymmetric activation function family of neural network for time series forecasting</title><title>Expert systems with applications</title><description>•We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series.
The use of neural network models for time series forecasting has been motivated by experimental results that indicate high capacity for function approximation with good accuracy. Generally, these models use activation functions with fixed parameters. However, it is known that the choice of activation function strongly influences the complexity and neural network performance and that a limited number of activation functions has been used in general. We describe the use of an asymmetric activation functions family with free parameter for neural networks. We prove that the activation functions family defined, satisfies the requirements of the universal approximation theorem We present a methodology for global optimization of the activation functions family with free parameter and the connections between the processing units of the neural network. The main idea is to optimize, simultaneously, the weights and activation function used in a Multilayer Perceptron (MLP), through an approach that combines the advantages of simulated annealing, tabu search and a local learning algorithm. We have chosen two local learning algorithms: the backpropagation with momentum (BPM) and Levenberg–Marquardt (LM). The overall purpose is to improve performance in time series forecasting.</description><subject>Activation</subject><subject>Algorithmics. Computability. Computer arithmetics</subject><subject>Algorithms</subject><subject>Applied sciences</subject><subject>Artificial intelligence</subject><subject>Asymmetric activation function</subject><subject>BPM algorithm</subject><subject>Computer science; control theory; systems</subject><subject>Computer simulation</subject><subject>Connectionism. Neural networks</subject><subject>Exact sciences and technology</subject><subject>Forecasting</subject><subject>Free parameter</subject><subject>Inference from stochastic processes; time series analysis</subject><subject>LM algorithm</subject><subject>Mathematical analysis</subject><subject>Mathematical models</subject><subject>Mathematics</subject><subject>Neural networks</subject><subject>Probability and statistics</subject><subject>Sciences and techniques of general use</subject><subject>Simulated annealing</subject><subject>Statistics</subject><subject>Tabu search</subject><subject>Theoretical computing</subject><subject>Time series</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2013</creationdate><recordtype>article</recordtype><recordid>eNqFkU1r3DAQhkVpodu0f6AnXQq9eKuxZcuCXkroFwRyyV2M5XGirT-2Gm2W7a-vjEOPDQyMhJ55B72vEO9B7UFB8-mwJz7jvlRQ7VWdq3ohdtCaqmiMrV6KnbK1KTQY_Vq8YT4oBUYpsxN8e0xhCn8whWWWyyDTA8kzhfuHxBLnXiJfpolSDF6iT-FxA4fT7LcDTmG8rIMznSKOuaXzEn_JYYkyK5NkioF4vZNHTmG-fyteDTgyvXvqV-Lu29e76x_Fze33n9dfbgqvTZWKqgPjqTNmMFoN0Nn8K7AeO9shtsoC9KXt-rqn1jaYHy1l0nTaQ0O2rK7Ex032GJffJ-LkpsCexhFnWk7ssgOgDACY59HGQF0qrdrnUa1bU7Z12WS03FAfF-ZIgzvGMGG8OFBujc0d3BqbW2Nzqs5V5aEPT_rIHsch4uwD_5ssTQM1WJ25zxtH2cHHQNGxDzR76kM2Orl-Cf9b8xdFR69i</recordid><startdate>20131115</startdate><enddate>20131115</enddate><creator>Gomes, Gecynalda S. da S.</creator><creator>Ludermir, Teresa B.</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>IQODW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20131115</creationdate><title>Optimization of the weights and asymmetric activation function family of neural network for time series forecasting</title><author>Gomes, Gecynalda S. da S. ; Ludermir, Teresa B.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2013</creationdate><topic>Activation</topic><topic>Algorithmics. Computability. Computer arithmetics</topic><topic>Algorithms</topic><topic>Applied sciences</topic><topic>Artificial intelligence</topic><topic>Asymmetric activation function</topic><topic>BPM algorithm</topic><topic>Computer science; control theory; systems</topic><topic>Computer simulation</topic><topic>Connectionism. Neural networks</topic><topic>Exact sciences and technology</topic><topic>Forecasting</topic><topic>Free parameter</topic><topic>Inference from stochastic processes; time series analysis</topic><topic>LM algorithm</topic><topic>Mathematical analysis</topic><topic>Mathematical models</topic><topic>Mathematics</topic><topic>Neural networks</topic><topic>Probability and statistics</topic><topic>Sciences and techniques of general use</topic><topic>Simulated annealing</topic><topic>Statistics</topic><topic>Tabu search</topic><topic>Theoretical computing</topic><topic>Time series</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Gomes, Gecynalda S. da S.</creatorcontrib><creatorcontrib>Ludermir, Teresa B.</creatorcontrib><collection>Pascal-Francis</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Gomes, Gecynalda S. da S.</au><au>Ludermir, Teresa B.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Optimization of the weights and asymmetric activation function family of neural network for time series forecasting</atitle><jtitle>Expert systems with applications</jtitle><date>2013-11-15</date><risdate>2013</risdate><volume>40</volume><issue>16</issue><spage>6438</spage><epage>6446</epage><pages>6438-6446</pages><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>•We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series.
The use of neural network models for time series forecasting has been motivated by experimental results that indicate high capacity for function approximation with good accuracy. Generally, these models use activation functions with fixed parameters. However, it is known that the choice of activation function strongly influences the complexity and neural network performance and that a limited number of activation functions has been used in general. We describe the use of an asymmetric activation functions family with free parameter for neural networks. We prove that the activation functions family defined, satisfies the requirements of the universal approximation theorem We present a methodology for global optimization of the activation functions family with free parameter and the connections between the processing units of the neural network. The main idea is to optimize, simultaneously, the weights and activation function used in a Multilayer Perceptron (MLP), through an approach that combines the advantages of simulated annealing, tabu search and a local learning algorithm. We have chosen two local learning algorithms: the backpropagation with momentum (BPM) and Levenberg–Marquardt (LM). The overall purpose is to improve performance in time series forecasting.</abstract><cop>Amsterdam</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2013.05.053</doi><tpages>9</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0957-4174 |
ispartof | Expert systems with applications, 2013-11, Vol.40 (16), p.6438-6446 |
issn | 0957-4174 1873-6793 |
language | eng |
recordid | cdi_proquest_miscellaneous_1701071117 |
source | ScienceDirect Freedom Collection 2022-2024 |
subjects | Activation Algorithmics. Computability. Computer arithmetics Algorithms Applied sciences Artificial intelligence Asymmetric activation function BPM algorithm Computer science control theory systems Computer simulation Connectionism. Neural networks Exact sciences and technology Forecasting Free parameter Inference from stochastic processes time series analysis LM algorithm Mathematical analysis Mathematical models Mathematics Neural networks Probability and statistics Sciences and techniques of general use Simulated annealing Statistics Tabu search Theoretical computing Time series |
title | Optimization of the weights and asymmetric activation function family of neural network for time series forecasting |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T06%3A57%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Optimization%20of%20the%20weights%20and%20asymmetric%20activation%20function%20family%20of%20neural%20network%20for%20time%20series%20forecasting&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Gomes,%20Gecynalda%20S.%20da%20S.&rft.date=2013-11-15&rft.volume=40&rft.issue=16&rft.spage=6438&rft.epage=6446&rft.pages=6438-6446&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2013.05.053&rft_dat=%3Cproquest_cross%3E1701071117%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1448728526&rft_id=info:pmid/&rfr_iscdi=true |