Parallel Feature Subset Selection Wrappers Using k-means Classifier
Author:
Papaioannou Nikolaos1, Tsimpiris Alkiviadis1, Talagozis Christos1, Fragidis Leonidas2, Angeioplastis Athanasios1, Tsakiridis Sotirios1, Varsamis Dimitrios1
Affiliation:
1. Department of Computer, Informatics and Telecommunications Engineering, International Hellenic University, Serres, GREECE 2. Department of Management Science and Technology, International Hellenic University, Kavala, GREECE
Abstract
In a world where the volume of data is constantly increasing, the implementation time of various processes increases significantly. Therefore, the proper management and the effort to reduce the dimensions of the datasets are considered imperative. Feature selection can reduce the size of the datasets by keeping a smaller subset, while improving the accuracy of the classification. The main purpose of this paper is to propose and examine the efficiency of parallel feature selection wrappers based on k-means classifier. The simple kmeans algorithm and a parallel version of it are used. Different parallelization variants of feature subset selection (fss) are presented and their accuracy and computation time are also evaluated on four different datasets. The comparison is performed among different parallelization variations and the serial implementation of fss with the k-means clustering algorithm. Finally, the results of the research are presented, highlighting the importance of parallelization in reducing the execution time of the proposed algorithms.
Publisher
World Scientific and Engineering Academy and Society (WSEAS)
Subject
Computer Science Applications,Information Systems
Reference33 articles.
1. S. Mittal, M. Shuja, and M. Zaman, “A review of data mining literature,” IJCSIS, vol. 14, pp. 437– 442, 2016. 2. J. A. Hartigan and M. A. Wong, “Algorithm as 136: A k-means clustering algorithm,” Journal of the Royal Statistical Society. Series C (Applied Statistics), vol. 28, no. 1, pp. 100–108, 1979. 3. M. Capo, A. Pérez, and J. Lozano, “An efficient k-means algorithm for massive data,” 2016. 4. M. Omran, A. Engelbrecht, and A. Salman, “An overview of clustering methods,” Intell. Data Anal., vol. 11, pp. 583–605, 2007. 5. X.-D. Wang, R.-C. Chen, F. Yan, Z.-Q. Zeng, and C.-Q. Hong, “Fast adaptive k-means subspace clustering for high-dimensional data,” IEEE Access, vol. 7, pp. 42639–42651, 2019.
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Hybrids of K-means and linkage algorithms;2023 International Conference on Applied Mathematics & Computer Science (ICAMCS);2023-08-08
|
|