Measure of Similarity between GMMs Based on Autoencoder-Generated Gaussian Component Representations
Author:
Kalušev Vladimir1, Popović Branislav2ORCID, Janev Marko3, Brkljač Branko2, Ralević Nebojša2
Affiliation:
1. The Institute for Artificial Intelligence Research and Development of Serbia, Fruškogorska 1, 21000 Novi Sad, Serbia 2. Faculty of Technical Sciences, University of Novi Sad, Trg Dositeja Obradovića 6, 21000 Novi Sad, Serbia 3. Institute of Mathematics, Serbian Academy of Sciences and Arts, Kneza Mihaila 36, 11000 Belgrade, Serbia
Abstract
A novel similarity measure between Gaussian mixture models (GMMs), based on similarities between the low-dimensional representations of individual GMM components and obtained using deep autoencoder architectures, is proposed in this paper. Two different approaches built upon these architectures are explored and utilized to obtain low-dimensional representations of Gaussian components in GMMs. The first approach relies on a classical autoencoder, utilizing the Euclidean norm cost function. Vectorized upper-diagonal symmetric positive definite (SPD) matrices corresponding to Gaussian components in particular GMMs are used as inputs to the autoencoder. Low-dimensional Euclidean vectors obtained from the autoencoder’s middle layer are then used to calculate distances among the original GMMs. The second approach relies on a deep convolutional neural network (CNN) autoencoder, using SPD representatives to generate embeddings corresponding to multivariate GMM components given as inputs. As the autoencoder training cost function, the Frobenious norm between the input and output layers of such network is used and combined with regularizer terms in the form of various pieces of information, as well as the Riemannian manifold-based distances between SPD representatives corresponding to the computed autoencoder feature maps. This is performed assuming that the underlying probability density functions (PDFs) of feature-map observations are multivariate Gaussians. By employing the proposed method, a significantly better trade-off between the recognition accuracy and the computational complexity is achieved when compared with other measures calculating distances among the SPD representatives of the original Gaussian components. The proposed method is much more efficient in machine learning tasks employing GMMs and operating on large datasets that require a large overall number of Gaussian components.
Funder
Serbian Ministry of Education, Science and Technological Development Faculty of Technical Sciences H2020 project INCOMING
Subject
Geometry and Topology,Logic,Mathematical Physics,Algebra and Number Theory,Analysis
Reference47 articles.
1. Goldberger, J., and Aronowitz, H. (2005, January 4–8). A distance measure between GMMs based on the unscented transform and its application to speaker recognition. Proceedings of the INTERSPEECH, Lisbon, Portugal. 2. Goldberger, J., Gordon, S., and Greenspan, H. (2003, January 13–16). An efficient image similarity measure based on approximations of KL-divergence between two Gaussian mixtures. Proceedings of the International Conference on Computer Vision, Nice, France. 3. Wu, Y., Chan, K.L., and Huang, Y. (2003, January 13–16). Image texture classification based on finite Gaussian mixture models. Proceedings of the 3rd Int. Workshop on Text. Anal. and Synth., Int. Conf. on Computer Vision, Nice, France. 4. Dilokthanakul, N., Mediano, P.A., Garnelo, M., Lee, M.C., Salimbeni, H., Arulkumaran, K., and Shanahan, M. (2017). Deep unsupervised clustering with Gaussian mixture variational autoencoders. arXiv. 5. A novel image retrieval technique based on semi supervised clustering;Gangodkar;Multimed. Tools Appl.,2021
|
|