Affiliation:
1. Instituto de Computação Universidade Federal de Alagoas Maceió Brazil
2. Departamento de Matemática Universidad de Atacama Copiapo Chile
Abstract
AbstractWith computational evolution, there has been a growth in the number of machine learning algorithms and they became more complex and robust. A greater challenge is upon faster and more practical ways to find hyperparameters that will set up each algorithm individually. This article aims to use meta‐learning as a practicable solution for recommending hyperparameters from similar datasets, through their meta‐features structures, than to adopt the already trained XGBoost parameters for a new database. This reduced computational costs and also aimed to make real‐time decision‐making feasible or reduce any extra costs for companies for new information. The experimental results, adopting 198 data sets, attested to the success of the heuristics application using meta‐learning to compare datasets structure analysis. Initially, a characterization of the datasets was performed by combining three groups of meta‐features (general, statistical, and info‐theory), so that there would be a way to compare the similarity between sets and, thus, apply meta‐learning to recommend the hyperparameters. Later, the appropriate number of sets to characterize the XGBoost turning was tested. The obtained results were promising, showing an improved performance in the accuracy of the XGBoost, k = {4 − 6}, using the average of the hyperparameters values and, comparing to the standard grid‐search hyperparameters set by default, it was obtained that, in 78.28% of the datasets, the meta‐learning methodology performed better. This study, therefore, shows that the adoption of meta‐learning is a competitive alternative to generalize the XGBoost model, expecting better statistics performance (accuracy etc.) rather than adjusting to a single/particular model.
Reference75 articles.
1. Time series classifier recommendation by a meta-learning approach
2. A meta-learning approach for selecting image segmentation algorithm
3. Aha D. W. &Kibler D. F.(1989).Noise‐tolerant instance‐based learning algorithms.Proceedings of the 11th international joint conference on Artificial intelligence(794–799). Citeseer.
4. MFE: Towards reproducible meta‐feature extraction;Alcobaça E.;Journal of Machine Learning Research,2020
5. COVID-19 Outbreak Prediction with Machine Learning