Our proposed method for systematically binarizing ANNs

From quantitative analysis to synthesis of efficient binary neural networks

Abstract

Binary Neural Networks (BNNs) offer an effective way to slash the cost of computation and memory accesses in inference. Recently, a plurality of ideas has been proposed, some of which are complementary while others are incompatible. This work presents a thorough review of state-of-the-art methods and an analysis of their computational cost based on the energy consumption of fixed-point, ternary and binary MAC vector operations. We derive an approach on how to systematically design a cost-efficient BNN. Our quantized LeNet and VGGNet architectures highlight the benefit of prudent capacity augmentation, with layer-wise ternarization providing best improvement of accuracy over μJ/classification in BNNs.

Publication
In 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA)
ANN CNN Efficiency Quantization
Tim Stadtmann
Electrical Engineer