Loading…
A weighted metric scalarization approach for multiobjective BOHB hyperparameter optimization in LSTM model for sentiment analysis
In recent years, hyperparameter tuning has become increasingly essential, especially for neural network models, where models with good performance generally require a long time to train. In contrast, fast-trained models often result in poor model performance. Thus, it is essential to determine the o...
Saved in:
Published in: | Information sciences 2023-10, Vol.644, p.119282, Article 119282 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | In recent years, hyperparameter tuning has become increasingly essential, especially for neural network models, where models with good performance generally require a long time to train. In contrast, fast-trained models often result in poor model performance. Thus, it is essential to determine the optimal hyperparameter configuration for a neural network model that can produce good model performance and fast training time. This paper presents a multiobjective optimization for Long-Short Term Memory (LSTM) model using Bayesian optimization-Hyperband (BOHB) with a weighted metric scalarization method. The objective functions used are the F1-score (Macro) and model training time, where both of the objective functions are considered equally important. The data used in this paper are the SmSA and EmoT datasets taken from the IndoNLU benchmark. The results are optimized model using multiobjective BOHB optimization with weighted metric scalarization method produced higher F1-score and faster model training time in both datasets compared to the baseline model and optimized model using single objective BOHB.
•This paper implements the weighted metric scalarization for the multiobjective BOHB optimization method.•We tested our proposed method to optimize the LSTM model hyperparameter for sentiment analysis.•The data used in this paper are the SmSA and EmoT datasets taken from the IndoNLU benchmark.•We compare our proposed method against the original Bayesian optimization, Hyperband, and BOHB.•The optimized LSTM model using our proposed method produced higher F1-score and faster training time, simultaneously. |
---|---|
ISSN: | 0020-0255 |
DOI: | 10.1016/j.ins.2023.119282 |