1. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding;han;ArXiv,2015
2. An sram optimized approach for constant memory consumption and ultra-fast execution of ml classifiers on tinyml hard-ware;sudharsan;2021 IEEE International Conference on Services Computing (SCC),0
3. A review on tinyml: State-of-the-art and prospects;ray;Journal of King Saud University-Computer and Information Sciences,2021
4. Tinytl: Reduce memory, not parameters for efficient on-device learning;cai;ArXiv,2020
5. Progress & compress: A scalable framework for continual learning;schwarz;International Conference on Machine Learning,0