Loading…
The q-G method: A q-version of the Steepest Descent method for global optimization
In this work, the q -Gradient ( q -G) method, a q -version of the Steepest Descent method, is presented. The main idea behind the q -G method is the use of the negative of the q -gradient vector of the objective function as the search direction. The q -gradient vector, or simply the q -gradient, is...
Saved in:
Published in: | SpringerPlus 2015-10, Vol.4 (1), p.647-15, Article 647 |
---|---|
Main Authors: | , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | In this work, the
q
-Gradient (
q
-G) method, a
q
-version of the Steepest Descent method, is presented. The main idea behind the
q
-G method is the use of the negative of the
q
-gradient vector of the objective function as the search direction. The
q
-gradient vector, or simply the
q
-gradient, is a generalization of the classical gradient vector based on the concept of Jackson’s derivative from the
q
-calculus. Its use provides the algorithm an effective mechanism for escaping from local minima. The
q
-G method reduces to the Steepest Descent method when the parameter
q
tends to 1. The algorithm has three free parameters and it is implemented so that the search process gradually shifts from global exploration in the beginning to local exploitation in the end. We evaluated the
q
-G method on 34 test functions, and compared its performance with 34 optimization algorithms, including derivative-free algorithms and the Steepest Descent method. Our results show that the
q
-G method is competitive and has a great potential for solving multimodal optimization problems. |
---|---|
ISSN: | 2193-1801 2193-1801 |
DOI: | 10.1186/s40064-015-1434-4 |