“…The obvious comparison point of this paper would be recent efforts at training quantized networks with bit-precision greater than single bit. There have been a multitude of approaches (Li et al, 2016 ; Zhou et al, 2016 , 2017 ; Choi et al, 2018 ; Deng et al, 2018 ; Zhang et al, 2018 ) with recent efforts aimed at designing networks with hybrid precision where the bit-precision of each layer of the network can vary (Prabhu et al, 2018 ; Wu et al, 2018 ; Chakraborty et al, 2019 ; Wang et al, 2019 ). However, in order to support variable bit-precision for each layer, the underlying hardware would need to be designed accordingly to handle mixed-precision (which usually is characterized by much higher area, latency and power consumption than BNN hardware accelerators.…”