“…To deploy deep HAR model on mobile wearable devices, model compression with small accuracy diminution is a critical and challenging mission. Universally, model compression methods include pruning [32], [33], [34], [35], [36], quantization [37], [38], low-rank approximation & sparsity [39], [40], and knowledge distillation [27], [41], [42]. Model pruning aims to prune non-significant weights in large models, and those pruned large-sparse models also have significant performance [43].…”