“…Several sparse CNN (s-CNN) accelerators have recently been proposed to exploit the structured [28,43] as well as the unstructured sparsity [1,3,5,7,8,15,16,19,31,39,40,41,42] in both CNN model parameters and its activations. Although this works [1,3,5,7,8,15,16,19,28,31,39,40,41,42,43] outperform dense CNN accelerators in terms of energy efficiency and performance, they use compressed vector-vector multiplication (VVM) or matrix-vector multiplication (MVM) engines based on traditional dataflow techniques such as weight stationary, input stationary, output stationary, and row stationary techniques [32,35]. When the sparsity of inputs or weights becomes excessively large (≥ 60%-90%), previous VVM and MVM accelerators suffer from low utilization issues.…”