Loading…
Advanced hyperparameter optimization for improved spatial prediction of shallow landslides using extreme gradient boosting (XGBoost)
Machine learning algorithms have progressively become a part of landslide susceptibility mapping practices owing to their robustness in dealing with complicated and non-linear mechanisms of landslides. However, the internal structures of such algorithms contain a set of hyperparameter configurations...
Saved in:
Published in: | Bulletin of engineering geology and the environment 2022-05, Vol.81 (5), Article 201 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Machine learning algorithms have progressively become a part of landslide susceptibility mapping practices owing to their robustness in dealing with complicated and non-linear mechanisms of landslides. However, the internal structures of such algorithms contain a set of hyperparameter configurations whose correct setting is crucial to get the highest achievable performance. This current study investigates the effectiveness and robustness of advanced optimization algorithms, including random search (RS), Bayesian optimization with Gaussian Process (BO-GP), Bayesian optimization with Tree-structured Parzen Estimator (BO-TPE), genetic algorithm (GA), and Hyperband method, for optimizing the hyperparameters of the eXtreme Gradient Boosting (XGBoost) algorithm in the spatial prediction of landslides. 12 causative factors were considered to produce landslide susceptibility maps (LSMs) for the Trabzon province of Turkey, where translational shallow landslides are ubiquitous. Five accuracy metrics, including overall accuracy (OA), precision, recall, F1-score, area under the receiver operating characteristic curve (AUC), and a statistical significance test were employed to measure the effectiveness of the optimization strategies on XGBoost algorithm. Compared to the XGBoost model with default setting, the optimized models provided a significant improvement of up to 13% in terms of overall accuracy, which was also ascertained by McNemar’s test. AUC analysis revealed that having statistically similar performances, GA (0.942) and Hyperband (0.922) methods had the highest predictive abilities, followed by BO-GP (0.920), BO-TPE (0.899), and RS (0.894). Analysis of computational cost efficiency showed that the Hyperband approach (40.3 s) was much faster (about 13 times) than the GA in hyperparameter tuning, and thus appeared to be the best optimization algorithm for the problem under consideration. |
---|---|
ISSN: | 1435-9529 1435-9537 |
DOI: | 10.1007/s10064-022-02708-w |