“…In related research, many compression methods for neural network have been proposed successively, including parameter pruning (Hu et al, 2016;Pan et al, 2020;Sui et al, 2021), parameter sharing (Wu et al, 2018), low-rank decomposition (Swaminathan et al, 2020) and knowledge distillation (Li et al, 2020;Prakosa et al, 2021), etc. But most of these methods have certain limitations, such as not being applicable to large neural network models, being applicable only to classification tasks, or some parameter settings depending on empirical and so on (Neill, 2020;Rongrong et al, 2018).…”