Optimizing Machine Learning Algorithms for Landslide Susceptibility Mapping along the Karakoram Highway, Gilgit Baltistan, Pakistan: A Comparative Study of Baseline, Bayesian, and Metaheuristic Hyperparameter Optimization Techniques.
accuracy
geospatial data
machine learning
optimization techniques
practical implementation
scalability
Journal
Sensors (Basel, Switzerland)
ISSN: 1424-8220
Titre abrégé: Sensors (Basel)
Pays: Switzerland
ID NLM: 101204366
Informations de publication
Date de publication:
01 Aug 2023
01 Aug 2023
Historique:
received:
24
06
2023
revised:
19
07
2023
accepted:
25
07
2023
medline:
12
8
2023
pubmed:
12
8
2023
entrez:
12
8
2023
Statut:
epublish
Résumé
Algorithms for machine learning have found extensive use in numerous fields and applications. One important aspect of effectively utilizing these algorithms is tuning the hyperparameters to match the specific task at hand. The selection and configuration of hyperparameters directly impact the performance of machine learning models. Achieving optimal hyperparameter settings often requires a deep understanding of the underlying models and the appropriate optimization techniques. While there are many automatic optimization techniques available, each with its own advantages and disadvantages, this article focuses on hyperparameter optimization for well-known machine learning models. It explores cutting-edge optimization methods such as metaheuristic algorithms, deep learning-based optimization, Bayesian optimization, and quantum optimization, and our paper focused mainly on metaheuristic and Bayesian optimization techniques and provides guidance on applying them to different machine learning algorithms. The article also presents real-world applications of hyperparameter optimization by conducting tests on spatial data collections for landslide susceptibility mapping. Based on the experiment's results, both Bayesian optimization and metaheuristic algorithms showed promising performance compared to baseline algorithms. For instance, the metaheuristic algorithm boosted the random forest model's overall accuracy by 5% and 3%, respectively, from baseline optimization methods GS and RS, and by 4% and 2% from baseline optimization methods GA and PSO. Additionally, for models like KNN and SVM, Bayesian methods with Gaussian processes had good results. When compared to the baseline algorithms RS and GS, the accuracy of the KNN model was enhanced by BO-TPE by 1% and 11%, respectively, and by BO-GP by 2% and 12%, respectively. For SVM, BO-TPE outperformed GS and RS by 6% in terms of performance, while BO-GP improved results by 5%. The paper thoroughly discusses the reasons behind the efficiency of these algorithms. By successfully identifying appropriate hyperparameter configurations, this research paper aims to assist researchers, spatial data analysts, and industrial users in developing machine learning models more effectively. The findings and insights provided in this paper can contribute to enhancing the performance and applicability of machine learning algorithms in various domains.
Identifiants
pubmed: 37571627
pii: s23156843
doi: 10.3390/s23156843
pmc: PMC10422586
pii:
doi:
Types de publication
Journal Article
Langues
eng
Sous-ensembles de citation
IM
Références
Environ Sci Pollut Res Int. 2023 Sep 13;:
pubmed: 37702864
Entropy (Basel). 2022 Jun 20;24(6):
pubmed: 35741565
J Cheminform. 2017 Jun 28;9(1):42
pubmed: 29086090
Syst Biol. 2016 May;65(3):381-96
pubmed: 26821913
IEEE Trans Cybern. 2020 Aug;50(8):3668-3681
pubmed: 31751262