1. Wu, J., Leng, C., Wang, Y., et al. (2016). Quantized convolutional neural networks for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 4820–4828).
2. Gong, Y., Liu, L., Yang, M., et al. (2014). Compressing deep convolutional networks using vector quantization. arXiv:
1412.6115
.
3. Courbariaux, M., Bengio, Y., David, J.-P. (2015). Binaryconnect: Training deep neural networks with binary weights during propagations. In Advances in Neural Information Processing Systems (pp. 3123–3131).
4. Soulié, G., Gripon, V., Robert, M. (2016). Compression of deep neural networks on the fly. In International Conference on Artificial Neural Networks (pp. 153–160). Cham: Springer.
5. Suda, N., Chandra, V., Dasika, G., et al. (2016). Throughput-optimized openCL-based FPGA accelerator for large-scale convolutional neural networks. In Proceedings of the 2016 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays (pp. 16–25): ACM.