1. Yao Fu, Ephrem Wu, Ashish Sirasao, Sedny Attia, Kamran Khan, and Ralph Wittig. 2016. Deep learning with int8 optimization on xilinx devices. White Paper (2016).
2. Song Han, Huizi Mao, and William J Dally. 2015. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015).
3. Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference
4. OctCNN: A high throughput FPGA accelerator for CNNs using octave convolution algorithm;Lou Wenqi;IEEE Trans. Comput.,2021
5. An Efficient Hardware Accelerator for Sparse Convolutional Neural Networks on FPGAs