1. Distilling the knowledge in a neural network;Hinton;eprint arXiv 10.48550/arXiv.1503.02531,2015
2. Operator fusion in XLA: analysis and evaluation;Snider;arXiv:2301.13062,2023
3. Lost in Pruning: The effects of pruning neural networks beyond test accuracy;Liebenwein;arXiv 2103.03014v1,2021
4. Structured Pruning of Deep Convolutional Neural Networks
5. Pruning vs Quantization: Which is Better?;Kuzmin;Qualcomm AI Research,2023