Loading…

The q-G method: A q-version of the Steepest Descent method for global optimization

In this work, the q -Gradient ( q -G) method, a q -version of the Steepest Descent method, is presented. The main idea behind the q -G method is the use of the negative of the q -gradient vector of the objective function as the search direction. The q -gradient vector, or simply the q -gradient, is...

Full description

Saved in:
Bibliographic Details
Published in:SpringerPlus 2015-10, Vol.4 (1), p.647-15, Article 647
Main Authors: Soterroni, Aline C., Galski, Roberto L., Scarabello, Marluce C., Ramos, Fernando M.
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In this work, the q -Gradient ( q -G) method, a q -version of the Steepest Descent method, is presented. The main idea behind the q -G method is the use of the negative of the q -gradient vector of the objective function as the search direction. The q -gradient vector, or simply the q -gradient, is a generalization of the classical gradient vector based on the concept of Jackson’s derivative from the q -calculus. Its use provides the algorithm an effective mechanism for escaping from local minima. The q -G method reduces to the Steepest Descent method when the parameter q tends to 1. The algorithm has three free parameters and it is implemented so that the search process gradually shifts from global exploration in the beginning to local exploitation in the end. We evaluated the q -G method on 34 test functions, and compared its performance with 34 optimization algorithms, including derivative-free algorithms and the Steepest Descent method. Our results show that the q -G method is competitive and has a great potential for solving multimodal optimization problems.
ISSN:2193-1801
2193-1801
DOI:10.1186/s40064-015-1434-4