Abstract
AbstractAutomated machine learning (AutoML) and artificial neural networks (ANNs) have revolutionized the field of artificial intelligence by yielding incredibly high-performing models to solve a myriad of inductive learning tasks. In spite of their successes, little guidance exists on when to use one versus the other. Furthermore, relatively few tools exist that allow the integration of both AutoML and ANNs in the same analysis to yield results combining both of their strengths. Here, we present TPOT-NN—a new extension to the tree-based AutoML software TPOT—and use it to explore the behavior of automated machine learning augmented with neural network estimators (AutoML+NN), particularly when compared to non-NN AutoML in the context of simple binary classification on a number of public benchmark datasets. Our observations suggest that TPOT-NN is an effective tool that achieves greater classification accuracy than standard tree-based AutoML on some datasets, with no loss in accuracy on others. We also provide preliminary guidelines for performing AutoML+NN analyses, and recommend possible future directions for AutoML+NN methods research, especially in the context of TPOT.
Funder
National Institute of Environmental Health Sciences
U.S. National Library of Medicine
National Institute of Allergy and Infectious Diseases
National Center for Advancing Translational Sciences
National Institute of Diabetes and Digestive and Kidney Diseases
Publisher
Springer Science and Business Media LLC
Subject
Computer Science Applications,Hardware and Architecture,Theoretical Computer Science,Software
Reference35 articles.
1. P. Auer, H. Burgsteiner, W. Maass, Reducing communication for distributed learning in neural networks, in International Conference on Artificial Neural Networks (Springer, 2002), pp. 123–128
2. W. Banzhaf, P. Nordin, R.E. Keller, F.D. Francone, Genetic Programming (Springer, Berlin, 1998)
3. M. Belkin, D. Hsu, S. Ma, S. Mandal, Reconciling modern machine-learning practice and the classical bias-variance trade-off. Proc. Nat. Acad. Sci. 116(32), 15849–15854 (2019)
4. M.B. Brown, A.B. Forsythe, Robust tests for the equality of variances. J. Am. Stat. Assoc. 69(346), 364–367 (1974)
5. T. Chen, C. Xgboost Guestrin, A scalable tree boosting system, in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (2016), pp. 785–794
Cited by
15 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献