Affiliation:
1. University of California, Berkeley, USA. begus@berkeley.edu
Abstract
Abstract
This paper models unsupervised learning of an identity-based pattern (or copying) in speech called reduplication from raw continuous data with deep convolutional neural networks. We use the ciwGAN architecture (Beguš, 2021a) in which learning of meaningful representations in speech emerges from a requirement that the CNNs generate informative data. We propose a technique to wug-test CNNs trained on speech and, based on four generative tests, argue that the network learns to represent an identity-based pattern in its latent space. By manipulating only two categorical variables in the latent space, we can actively turn an unreduplicated form into a reduplicated form with no other substantial changes to the output in the majority of cases. We also argue that the network extends the identity-based pattern to unobserved data. Exploration of how meaningful representations of identity-based patterns emerge in CNNs and how the latent space variables outside of the training range correlate with identity-based patterns in the output has general implications for neural network interpretability.
Reference50 articles.
1. Investigating under and overfitting in Wasserstein Generative Adversarial Networks;Adlam,2019
2. Pre-wiring and pre-training: What does a neural network need to learn truly general identity rules?;Alhama;Journal of Artificial Intelligence Research,2018
3. Wasserstein Generative Adversarial Networks;Arjovsky,2017
4. vq-wav2vec: Self-supervised learning of discrete speech representations;Baevski,2020
5. Ciwgan and fiwgan: Encoding information in acoustic data to model lexical learning with Generative Adversarial Networks;Beguš;Neural Networks,2021
Cited by
5 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献