Abstract
Metrics are used to measure the distance, similarity, or dissimilarity between two points in a metric space. Metric learning algorithms perform the finding task of data points that are closest or furthest to a query point in m-dimensional metric space. Some metrics take into account the assumption that the whole dimensions are of equal importance, and vice versa. However, this assumption does not incorporate a number of real-world problems that classification algorithms tackle. In this research, the existing information gain, the information gain ratio, and some well-known conventional metrics have been compared by each other. The 1-Nearest Neighbor algorithm taking these metrics as its meta-parameter has been applied to forty-nine benchmark datasets. Only the accuracy rate criterion has been employed in order to quantify the performance of the metrics. The experimental results show that each metric is successful on datasets corresponding to its own domain. In other words, each metric is favorable on datasets overlapping its own assumption. In addition, there also exists incompleteness in classification tasks for metrics just like there is for learning algorithms.
Publisher
Afyon Kocatepe Universitesi Fen Ve Muhendislik Bilimleri Dergisi
Reference46 articles.
1. Aha, D.W., 1998. Feature Weighting for Lazy Learning Algorithms. In: H. Liu and H. Motoda, eds. Feature Extraction, Construction and Selection. Springer, Boston, MA, 13–32.
2. Aydın, F., 2022. A class-driven approach to dimension embedding. Expert Systems with Applications, 195, 116650.
3. Bellet, A., Habrard, A., and Sebban, M., 2013. A Survey on Metric Learning for Feature Vectors and Structured Data.
4. Bellet, A., Habrard, A., and Sebban, M., 2015. Nonlinear and Local Metric Learning. In: Metric Learning. Springer, Cham., 33–42.
5. Beyer, K.S., Goldstein, J., Ramakrishnan, R., and Shaft, U., 1999. When Is ‘“Nearest Neighbor”’ Meaningful? In: ICDT ’99 Proceedings of the 7th International Conference on Database Theory. London, UK: Springer-Verlag, 217–235.