Towards Explainability of the Latent Space by Disentangled Representation Learning
-
Published:2023-11-30
Issue:
Volume:26
Page:41-48
-
ISSN:2255-9094
-
Container-title:Information Technology and Management Science
-
language:
-
Short-container-title:ITMS
Author:
Namatēvs Ivars1, Ņikuļins Artūrs2, Slaidiņa Anda3, Neimane Laura3, Radziņš Oskars3, Sudars Kaspars2
Affiliation:
1. Riga Technical university, Institute of Electronics and Computer Science 2. Institute of Electronics and Computer Science, Latvia 3. Rīga Stradiņš University, Latvia
Abstract
Deep neural networks are widely used in computer vision for image classification, segmentation and generation. They are also often criticised as “black boxes” because their decision-making process is often not interpretable by humans. However, learning explainable representations that explicitly disentangle the underlying mechanisms that structure observational data is still a challenge. To further explore the latent space and achieve generic processing, we propose a pipeline for discovering the explainable directions in the latent space of generative models. Since the latent space contains semantically meaningful directions and can be explained, we propose a pipeline to fully resolve the representation of the latent space. It consists of a Dirichlet encoder, conditional deterministic diffusion, a group-swap and a latent traversal module. We believe that this study provides an insight into the advancement of research explaining the disentanglement of neural networks in the community.
Publisher
Riga Technical University
Reference62 articles.
1. [1] J. Egger, A. Pepe, C. Gsaxner, Y. Jin, J. Li, and R. Kern, “Deep learning – a first meta-survey of selected reviews across scientific disciplines, their commonalities, challenges and research impact”, PeerJ Computer Science, vol. 7, 2021, Art. no. e773. https://doi.org/10.7717/peerj-cs.773 2. [2] Y. Bengio, A. Courville, and P. Vincent, “Representation learning: A review and new perspectives”, TPAMI, vol. 35, no. 8, pp. 1798–1828, Aug. 2013. https://doi.org/10.1109/TPAMI.2013.50 3. [3] K. Ridgeway and M.C. Mozer, “Learning deep disentangled embeddings with the F-statistic loss”, in 32nd Conference on Neural Information Processing Systems (NeurIPS2018), Montréal, Canada, 2018, pp. 1–10. https://proceedings.neurips.cc/paper_files/paper/2018/file/2b24d495052a8ce66358eb576b8912c8-Paper.pdf 4. [4] R. Suter, D. Miladinovic, B. Schölkopf, and S. Bauer, “Robustly disentangled causal mechanisms: Validating deep representations for interventional robustness,” in International Conference on Machine Learning, PMLR, vol. 97, 2019, pp. 6056–6065. https://proceedings.mlr.press/v97/suter19a/suter19a.pdf 5. [5] D. Friede, C. Reimers, H. Stuckenschmidt, and M. Niepert, “Learning disentangled discrete representations”, in Machine Learning and Knowledge Discovery in Databases: Research Track. ECML PKDD 2023. Lecture Notes in Computer Science, D. Koutra, C. Plant, M. Gomez Rodriguez, E. Baralis, and F. Bonchi, Eds., vol. 14172. Springer, Cham. https://doi.org/10.1007/978-3-031-43421-1_35
|
|