Author:
Liu Chunlei,Chen Peng,Zhuang Bohan,Shen Chunhua,Zhang Baochang,Ding Wenrui
Abstract
Binary Neural Networks (BNNs) have received significant attention due to the memory and computation efficiency recently. However, the considerable accuracy gap between BNNs and their full-precision counterparts hinders BNNs to be deployed to resource-constrained platforms. One of the main reasons for the performance gap can be attributed to the frequent weight flip, which is caused by the misleading weight update in BNNs. To address this issue, we propose a state-aware binary neural network (SA-BNN) equipped with the well designed state-aware gradient. Our SA-BNN is inspired by the observation that the frequent weight flip is more likely to occur, when the gradient magnitude for all quantization states {-1,1} is identical. Accordingly, we propose to employ independent gradient coefficients for different states when updating the weights. Furthermore, we also analyze the effectiveness of the state-aware gradient on suppressing the frequent weight flip problem. Experiments on ImageNet show that the proposed SA-BNN outperforms the current state-of-the-arts (e.g., Bi-Real Net) by more than 3% when using a ResNet architecture. Specifically, we achieve 61.7%, 65.5% and 68.7% Top-1 accuracy with ResNet-18, ResNet-34 and ResNet-50 on ImageNet, respectively.
Publisher
Association for the Advancement of Artificial Intelligence (AAAI)
Cited by
8 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. 1-D Spatial Attention in Binarized Convolutional Neural Networks;ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP);2024-04-14
2. Partial Binarization of Neural Networks for Budget-Aware Efficient Learning;2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV);2024-01-03
3. Binary Neural Networks;Computational Intelligence Methods and Applications;2024
4. How to Train Accurate BNNs for Embedded Systems?;Embedded Machine Learning for Cyber-Physical, IoT, and Edge Computing;2023-10-10
5. Rad-Bnn: Regulating Activation Distribution for Accurate Binary Neural Network;2023