Author:
Peric Zoran,Denic Bojan,Savic Milan,Despotovic Vladimir
Abstract
A compression method based on non-uniform binary scalar quantization, designed for the memoryless Laplacian source with zero-mean and unit variance, is analyzed in this paper. Two quantizer design approaches are presented that investigate the effect of clipping with the aim of reducing the quantization noise, where the minimal mean-squared error distortion is used to determine the optimal clipping factor. A detailed comparison of both models is provided, and the performance evaluation in a wide dynamic range of input data variances is also performed. The observed binary scalar quantization models are applied in standard signal processing tasks, such as speech and image quantization, but also to quantization of neural network parameters. The motivation behind the binary quantization of neural network weights is the model compression by a factor of 32, which is crucial for implementation in mobile or embedded devices with limited memory and processing power. The experimental results follow well the theoretical models, confirming their applicability in real-world applications.
Reference39 articles.
1. Digital Coding of Waveforms: Principles and Applications to Speech and Video;Jayant,1984
2. Vector Quantization and Signal Compression;Gersho,1992
3. Speech Coding Algorithms: Foundation and Evolution of Standardized Coders;Chu,2003
4. Forward Adaptive Dual-Mode Quantizer Based on the First-Degree Spline Approximation and Embedded G.711 Codec
5. Lloyd–Max's Algorithm Implementation in Speech Coding Algorithm Based on Forward Adaptive Technique
Cited by
10 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献