Convolutional neural networks (CNNs) have demonstrated impressive performance in fitting data distribution. However, due to the complexity in learning intricate features from data, networks usually experience overfitting during the training. To address this issue, many data augmentation techniques have been proposed to expand the representation of the training data, thereby improving the generalization ability of CNNs. In this paper, we propose PatchMix, a novel mixup-based augmentation method that applies mixup to patches within an image rather than blending between multiple images. Despite the challenge of label mixing, PatchMix generates numerous reliable training samples. Additionally, we propose PatchMix-R to enhance the robustness of the model against perturbations by processing adjacent pixels. Easy to implement, our methods can be integrated with most CNN-based classification models and combined with varying data augmentation techniques. The experiments show that PatchMix and PatchMix-R consistently outperform other state-of-the-art methods in terms of accuracy and robustness. Class activation mappings (CAMs) of the trained model are also investigated to visualize the effectiveness of our approach.