Loading…

Optimization of the weights and asymmetric activation function family of neural network for time series forecasting

•We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series. The use of neural network models for time series forecasting has been motivated by e...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems with applications 2013-11, Vol.40 (16), p.6438-6446
Main Authors: Gomes, Gecynalda S. da S., Ludermir, Teresa B.
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923
cites cdi_FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923
container_end_page 6446
container_issue 16
container_start_page 6438
container_title Expert systems with applications
container_volume 40
creator Gomes, Gecynalda S. da S.
Ludermir, Teresa B.
description •We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series. The use of neural network models for time series forecasting has been motivated by experimental results that indicate high capacity for function approximation with good accuracy. Generally, these models use activation functions with fixed parameters. However, it is known that the choice of activation function strongly influences the complexity and neural network performance and that a limited number of activation functions has been used in general. We describe the use of an asymmetric activation functions family with free parameter for neural networks. We prove that the activation functions family defined, satisfies the requirements of the universal approximation theorem We present a methodology for global optimization of the activation functions family with free parameter and the connections between the processing units of the neural network. The main idea is to optimize, simultaneously, the weights and activation function used in a Multilayer Perceptron (MLP), through an approach that combines the advantages of simulated annealing, tabu search and a local learning algorithm. We have chosen two local learning algorithms: the backpropagation with momentum (BPM) and Levenberg–Marquardt (LM). The overall purpose is to improve performance in time series forecasting.
doi_str_mv 10.1016/j.eswa.2013.05.053
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_1701071117</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417413003515</els_id><sourcerecordid>1701071117</sourcerecordid><originalsourceid>FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923</originalsourceid><addsrcrecordid>eNqFkU1r3DAQhkVpodu0f6AnXQq9eKuxZcuCXkroFwRyyV2M5XGirT-2Gm2W7a-vjEOPDQyMhJ55B72vEO9B7UFB8-mwJz7jvlRQ7VWdq3ohdtCaqmiMrV6KnbK1KTQY_Vq8YT4oBUYpsxN8e0xhCn8whWWWyyDTA8kzhfuHxBLnXiJfpolSDF6iT-FxA4fT7LcDTmG8rIMznSKOuaXzEn_JYYkyK5NkioF4vZNHTmG-fyteDTgyvXvqV-Lu29e76x_Fze33n9dfbgqvTZWKqgPjqTNmMFoN0Nn8K7AeO9shtsoC9KXt-rqn1jaYHy1l0nTaQ0O2rK7Ex032GJffJ-LkpsCexhFnWk7ssgOgDACY59HGQF0qrdrnUa1bU7Z12WS03FAfF-ZIgzvGMGG8OFBujc0d3BqbW2Nzqs5V5aEPT_rIHsch4uwD_5ssTQM1WJ25zxtH2cHHQNGxDzR76kM2Orl-Cf9b8xdFR69i</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1448728526</pqid></control><display><type>article</type><title>Optimization of the weights and asymmetric activation function family of neural network for time series forecasting</title><source>ScienceDirect Freedom Collection 2022-2024</source><creator>Gomes, Gecynalda S. da S. ; Ludermir, Teresa B.</creator><creatorcontrib>Gomes, Gecynalda S. da S. ; Ludermir, Teresa B.</creatorcontrib><description>•We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series. The use of neural network models for time series forecasting has been motivated by experimental results that indicate high capacity for function approximation with good accuracy. Generally, these models use activation functions with fixed parameters. However, it is known that the choice of activation function strongly influences the complexity and neural network performance and that a limited number of activation functions has been used in general. We describe the use of an asymmetric activation functions family with free parameter for neural networks. We prove that the activation functions family defined, satisfies the requirements of the universal approximation theorem We present a methodology for global optimization of the activation functions family with free parameter and the connections between the processing units of the neural network. The main idea is to optimize, simultaneously, the weights and activation function used in a Multilayer Perceptron (MLP), through an approach that combines the advantages of simulated annealing, tabu search and a local learning algorithm. We have chosen two local learning algorithms: the backpropagation with momentum (BPM) and Levenberg–Marquardt (LM). The overall purpose is to improve performance in time series forecasting.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2013.05.053</identifier><language>eng</language><publisher>Amsterdam: Elsevier Ltd</publisher><subject>Activation ; Algorithmics. Computability. Computer arithmetics ; Algorithms ; Applied sciences ; Artificial intelligence ; Asymmetric activation function ; BPM algorithm ; Computer science; control theory; systems ; Computer simulation ; Connectionism. Neural networks ; Exact sciences and technology ; Forecasting ; Free parameter ; Inference from stochastic processes; time series analysis ; LM algorithm ; Mathematical analysis ; Mathematical models ; Mathematics ; Neural networks ; Probability and statistics ; Sciences and techniques of general use ; Simulated annealing ; Statistics ; Tabu search ; Theoretical computing ; Time series</subject><ispartof>Expert systems with applications, 2013-11, Vol.40 (16), p.6438-6446</ispartof><rights>2013 Elsevier Ltd</rights><rights>2014 INIST-CNRS</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923</citedby><cites>FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&amp;idt=27615194$$DView record in Pascal Francis$$Hfree_for_read</backlink></links><search><creatorcontrib>Gomes, Gecynalda S. da S.</creatorcontrib><creatorcontrib>Ludermir, Teresa B.</creatorcontrib><title>Optimization of the weights and asymmetric activation function family of neural network for time series forecasting</title><title>Expert systems with applications</title><description>•We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series. The use of neural network models for time series forecasting has been motivated by experimental results that indicate high capacity for function approximation with good accuracy. Generally, these models use activation functions with fixed parameters. However, it is known that the choice of activation function strongly influences the complexity and neural network performance and that a limited number of activation functions has been used in general. We describe the use of an asymmetric activation functions family with free parameter for neural networks. We prove that the activation functions family defined, satisfies the requirements of the universal approximation theorem We present a methodology for global optimization of the activation functions family with free parameter and the connections between the processing units of the neural network. The main idea is to optimize, simultaneously, the weights and activation function used in a Multilayer Perceptron (MLP), through an approach that combines the advantages of simulated annealing, tabu search and a local learning algorithm. We have chosen two local learning algorithms: the backpropagation with momentum (BPM) and Levenberg–Marquardt (LM). The overall purpose is to improve performance in time series forecasting.</description><subject>Activation</subject><subject>Algorithmics. Computability. Computer arithmetics</subject><subject>Algorithms</subject><subject>Applied sciences</subject><subject>Artificial intelligence</subject><subject>Asymmetric activation function</subject><subject>BPM algorithm</subject><subject>Computer science; control theory; systems</subject><subject>Computer simulation</subject><subject>Connectionism. Neural networks</subject><subject>Exact sciences and technology</subject><subject>Forecasting</subject><subject>Free parameter</subject><subject>Inference from stochastic processes; time series analysis</subject><subject>LM algorithm</subject><subject>Mathematical analysis</subject><subject>Mathematical models</subject><subject>Mathematics</subject><subject>Neural networks</subject><subject>Probability and statistics</subject><subject>Sciences and techniques of general use</subject><subject>Simulated annealing</subject><subject>Statistics</subject><subject>Tabu search</subject><subject>Theoretical computing</subject><subject>Time series</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2013</creationdate><recordtype>article</recordtype><recordid>eNqFkU1r3DAQhkVpodu0f6AnXQq9eKuxZcuCXkroFwRyyV2M5XGirT-2Gm2W7a-vjEOPDQyMhJ55B72vEO9B7UFB8-mwJz7jvlRQ7VWdq3ohdtCaqmiMrV6KnbK1KTQY_Vq8YT4oBUYpsxN8e0xhCn8whWWWyyDTA8kzhfuHxBLnXiJfpolSDF6iT-FxA4fT7LcDTmG8rIMznSKOuaXzEn_JYYkyK5NkioF4vZNHTmG-fyteDTgyvXvqV-Lu29e76x_Fze33n9dfbgqvTZWKqgPjqTNmMFoN0Nn8K7AeO9shtsoC9KXt-rqn1jaYHy1l0nTaQ0O2rK7Ex032GJffJ-LkpsCexhFnWk7ssgOgDACY59HGQF0qrdrnUa1bU7Z12WS03FAfF-ZIgzvGMGG8OFBujc0d3BqbW2Nzqs5V5aEPT_rIHsch4uwD_5ssTQM1WJ25zxtH2cHHQNGxDzR76kM2Orl-Cf9b8xdFR69i</recordid><startdate>20131115</startdate><enddate>20131115</enddate><creator>Gomes, Gecynalda S. da S.</creator><creator>Ludermir, Teresa B.</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>IQODW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20131115</creationdate><title>Optimization of the weights and asymmetric activation function family of neural network for time series forecasting</title><author>Gomes, Gecynalda S. da S. ; Ludermir, Teresa B.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2013</creationdate><topic>Activation</topic><topic>Algorithmics. Computability. Computer arithmetics</topic><topic>Algorithms</topic><topic>Applied sciences</topic><topic>Artificial intelligence</topic><topic>Asymmetric activation function</topic><topic>BPM algorithm</topic><topic>Computer science; control theory; systems</topic><topic>Computer simulation</topic><topic>Connectionism. Neural networks</topic><topic>Exact sciences and technology</topic><topic>Forecasting</topic><topic>Free parameter</topic><topic>Inference from stochastic processes; time series analysis</topic><topic>LM algorithm</topic><topic>Mathematical analysis</topic><topic>Mathematical models</topic><topic>Mathematics</topic><topic>Neural networks</topic><topic>Probability and statistics</topic><topic>Sciences and techniques of general use</topic><topic>Simulated annealing</topic><topic>Statistics</topic><topic>Tabu search</topic><topic>Theoretical computing</topic><topic>Time series</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Gomes, Gecynalda S. da S.</creatorcontrib><creatorcontrib>Ludermir, Teresa B.</creatorcontrib><collection>Pascal-Francis</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Gomes, Gecynalda S. da S.</au><au>Ludermir, Teresa B.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Optimization of the weights and asymmetric activation function family of neural network for time series forecasting</atitle><jtitle>Expert systems with applications</jtitle><date>2013-11-15</date><risdate>2013</risdate><volume>40</volume><issue>16</issue><spage>6438</spage><epage>6446</epage><pages>6438-6446</pages><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>•We present a method for optimization of the activation functions for ANN.•The proposed optimization method uses Simulated Annealing and Tabu Search.•The proposed method is good for forecasting distinct time series. The use of neural network models for time series forecasting has been motivated by experimental results that indicate high capacity for function approximation with good accuracy. Generally, these models use activation functions with fixed parameters. However, it is known that the choice of activation function strongly influences the complexity and neural network performance and that a limited number of activation functions has been used in general. We describe the use of an asymmetric activation functions family with free parameter for neural networks. We prove that the activation functions family defined, satisfies the requirements of the universal approximation theorem We present a methodology for global optimization of the activation functions family with free parameter and the connections between the processing units of the neural network. The main idea is to optimize, simultaneously, the weights and activation function used in a Multilayer Perceptron (MLP), through an approach that combines the advantages of simulated annealing, tabu search and a local learning algorithm. We have chosen two local learning algorithms: the backpropagation with momentum (BPM) and Levenberg–Marquardt (LM). The overall purpose is to improve performance in time series forecasting.</abstract><cop>Amsterdam</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2013.05.053</doi><tpages>9</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0957-4174
ispartof Expert systems with applications, 2013-11, Vol.40 (16), p.6438-6446
issn 0957-4174
1873-6793
language eng
recordid cdi_proquest_miscellaneous_1701071117
source ScienceDirect Freedom Collection 2022-2024
subjects Activation
Algorithmics. Computability. Computer arithmetics
Algorithms
Applied sciences
Artificial intelligence
Asymmetric activation function
BPM algorithm
Computer science
control theory
systems
Computer simulation
Connectionism. Neural networks
Exact sciences and technology
Forecasting
Free parameter
Inference from stochastic processes
time series analysis
LM algorithm
Mathematical analysis
Mathematical models
Mathematics
Neural networks
Probability and statistics
Sciences and techniques of general use
Simulated annealing
Statistics
Tabu search
Theoretical computing
Time series
title Optimization of the weights and asymmetric activation function family of neural network for time series forecasting
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T06%3A57%3A02IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Optimization%20of%20the%20weights%20and%20asymmetric%20activation%20function%20family%20of%20neural%20network%20for%20time%20series%20forecasting&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Gomes,%20Gecynalda%20S.%20da%20S.&rft.date=2013-11-15&rft.volume=40&rft.issue=16&rft.spage=6438&rft.epage=6446&rft.pages=6438-6446&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2013.05.053&rft_dat=%3Cproquest_cross%3E1701071117%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c473t-3b17ceb77f740f1b905319cab9baa80911d29bd5de896a0539ef747b4c16e923%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1448728526&rft_id=info:pmid/&rfr_iscdi=true