Search alternatives:
guided optimization » based optimization (Expand Search), model optimization (Expand Search)
used optimization » based optimization (Expand Search), led optimization (Expand Search), field optimization (Expand Search)
binary data » primary data (Expand Search), dietary data (Expand Search)
guided optimization » based optimization (Expand Search), model optimization (Expand Search)
used optimization » based optimization (Expand Search), led optimization (Expand Search), field optimization (Expand Search)
binary data » primary data (Expand Search), dietary data (Expand Search)
-
1
The Pseudo-Code of the IRBMO Algorithm.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
2
IRBMO vs. meta-heuristic algorithms boxplot.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
3
IRBMO vs. feature selection algorithm boxplot.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
4
-
5
MSE for ILSTM algorithm in binary classification.
Published 2023“…The proposed algorithm has two phases: phase one involves training a conventional LSTM network to get initial weights, and phase two involves using the hybrid swarm algorithms, CBOA and PSO, to optimize the weights of LSTM to improve the accuracy. …”
-
6
IRBMO vs. variant comparison adaptation data.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
7
-
8
Pseudo Code of RBMO.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
9
P-value on CEC-2017(Dim = 30).
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
10
Memory storage behavior.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
11
Elite search behavior.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
12
Description of the datasets.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
13
S and V shaped transfer functions.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
14
S- and V-Type transfer function diagrams.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
15
Collaborative hunting behavior.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
16
Friedman average rank sum test results.
Published 2025“…To adapt to the feature selection problem, we convert the continuous optimization algorithm to binary form via transfer function, which further enhances the applicability of the algorithm. …”
-
17
Algorithm for generating hyperparameter.
Published 2024“…Motivated by the above, in this proposal, we design an improved model to predict the existence of respiratory disease among patients by incorporating hyperparameter optimization and feature selection. To optimize the parameters of the machine learning algorithms, hyperparameter optimization with a genetic algorithm is proposed and to reduce the size of the feature set, feature selection is performed using binary grey wolf optimization algorithm. …”
-
18
-
19
Results of machine learning algorithm.
Published 2024“…Motivated by the above, in this proposal, we design an improved model to predict the existence of respiratory disease among patients by incorporating hyperparameter optimization and feature selection. To optimize the parameters of the machine learning algorithms, hyperparameter optimization with a genetic algorithm is proposed and to reduce the size of the feature set, feature selection is performed using binary grey wolf optimization algorithm. …”
-
20
ROC comparison of machine learning algorithm.
Published 2024“…Motivated by the above, in this proposal, we design an improved model to predict the existence of respiratory disease among patients by incorporating hyperparameter optimization and feature selection. To optimize the parameters of the machine learning algorithms, hyperparameter optimization with a genetic algorithm is proposed and to reduce the size of the feature set, feature selection is performed using binary grey wolf optimization algorithm. …”