Knowledge distillation aims to transfer the knowledge from one model to another, which is widely used in model compression and transfer learning. Recent knowledge distillation algorithms mainly focused on the inter-sample or intra-sample knowledge, which haven't efficiently exploited supervised labels. In this work, we introduce the self-supervised tasks into conventional knowledge distillation framework to construct a new category-based knowledge distillation algorithm. To be specific, two new memory banks are constructed to store the category embedding predicted by the teacher and student models, respectively. Then, the student model can absorb the knowledge from the teacher's memory bank instead of the model itself. It utilizes the cross-batch knowledge based on the contrast of category embedding among various batches.Moreover, additional self-supervised tasks with various data augmentation strategies are are designed to combine with the knowledge distillation process, which guides the model training in a multi-task manner. In the experiments, our proposed algorithm outperforms most of the sample-based knowledge distillation on both CIFAR-100 and ImageNet datasets.