Recent feature contrastive learning (FCL) has shown promising performance in self-supervised representation learning. For domain adaptation, however, FCL cannot show overwhelming gains since the class weights are not involved during optimization, which does not guarantee the produced features to be clustered around the class weights learned from source data. To tackle this issue, we propose a novel probability contrastive learning (PCL) in this paper, which not only produces compact features but also enforces them to be distributed around the class weights. Specifically, we propose to use the output probabilities after softmax to perform contrastive learning instead of the extracted features and remove the 2 normalization in the traditional FCL. In this way, the probability will approximate the onehot form, thereby narrowing the distance between the features and the class weights. Our proposed PCL is simple and effective. We conduct extensive experiments on two domain adaptation tasks, i.e., unsupervised domain adaptation and semi-supervised domain adaptation. The results on multiple datasets demonstrate that our PCL can consistently get considerable gains and achieves the state-of-theart performance. In addition, our method also obtains considerable gains on semi-supervised tasks when labeled data is scarce.