Affiliation:
1. Brigham Young University, USA
2. University of Porto, Portugal
3. University of Houston, USA
Abstract
The application of Machine Learning (ML) and Data Mining (DM) tools to classification and regression tasks has become a standard, not only in research but also in administrative agencies, commerce and industry (e.g., finance, medicine, engineering). Unfortunately, due in part to the number of available techniques and the overall complexity of the process, users facing a new data mining task must generally either resort to trialand- error or consultation of experts. Clearly, neither solution is completely satisfactory for the non-expert end-users who wish to access the technology more directly and cost-effectively. What is needed is an informed search process to reduce the amount of experimentation with different techniques while avoiding the pitfalls of local optima that may result from low quality models. Informed search requires meta-knowledge, that is, knowledge about the performance of those techniques. Metalearning provides a robust, automatic mechanism for building such meta-knowledge. One of the underlying goals of meta-learning is to understand the interaction between the mechanism of learning and the concrete contexts in which that mechanism is applicable. Metalearning differs from base-level learning in the scope of adaptation. Whereas learning at the base-level focuses on accumulating experience on a specific learning task (e.g., credit rating, medical diagnosis, mine-rock discrimination, fraud detection, etc.), learning at the meta-level is concerned with accumulating experience on the performance of multiple applications of a learning system. The meta-knowledge induced by meta-learning provides the means to inform decisions about the precise conditions under which a given algorithm, or sequence of algorithms, is better than others for a given task. While Data Mining software packages (e.g., SAS Enterprise Miner, SPSS Clementine, Insightful Miner, PolyAnalyst, KnowledgeStudio, Weka, Yale, Xelopes) provide user-friendly access to rich collections of algorithms, they generally offer no real decision support to non-expert end-users. Similarly, tools with emphasis on advanced visualization help users understand the data (e.g., to select adequate transformations) and the models (e.g., to tweak parameters, compare results, and focus on specific parts of the model), but treat algorithm selection as a post-processing activity driven by the users rather than the system. Data mining practitioners need systems that guide them by producing explicit advice automatically. This chapter shows how meta-learning can be leveraged to provide such advice in the context of algorithm selection.
Reference40 articles.
1. Abe, H., & Yamaguchi, T. (2002). Constructing Inductive Applications by Meta-Learning with Method Repositories, in Progress in Discovery Science, Final Report of the Japanese Discovery Science Project, LNCS 2281, Springer-Verlag, 576-585.
2. Aha, D. W. (1992). Generalizing from Case Studies: A Case Study, in Proceedings of the Ninth International Workshop on Machine Learning, 1-10.
3. Bensusan, H., & Giraud-Carrier, C. (2000). Discovering Task Neighbourhoods Through Landmark Learning Performances, in Proceedings of the Fourth European Conference on Principles and Practice of Knowledge Discovery in Databases, 325,330.
4. Bensusan, H., Giraud-Carrier, C., & Kennedy, C. J. (2000). A Higher-Order Approach to Meta-Learning, in Proceedings of the ECML-2000 Workshop on Meta-Learning: Building Automatic Advice Strategies for Model Selection and Method Combination, 109-118.
5. Bensusan, H., & Kalousis, A. (2001). Estimating the Predictive Accuracy of a Classifier, in Proceedings of the Twelfth European Conference on Machine Learning, 25-36.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献