In recent decades, the wide use of deep learningbased methods has consistently improved the performance of remote sensing images and is widely used for hyperspectral change detection (HCD) tasks. However, most of the existing HCD method is based on the convolutional neural network (CNN), which shows limitations in long-range dependencies and also cannot mine sequence features well. The CD performance still has margins for improvement. In this study, inspired by the excellent performance of transformers in computer vision and which has shown a significant ability to model global dependencies to attenuate the loss of long-range information, we built a hybrid spatial-spectral convolutional vision transformer (SSViT) for HCD. Our proposed method combines the merits of CNN and transformer to fulfill effective and efficient HCD. This study focused on highly reliable pseudo-sample data generation by selection scenario. To generate a pseudo sample, we have used different methods: (1) we predict change and no-change areas by using Euclidean distance, (2) thresholding by Chan-Vese segmentation method for determining change and no-change pixels for intensity maps, (3) sorting of change and no-change pixels, and (4) selection of the minimum value of initial no-change pixels as pseudo change sample data, in addition to, choosing the maximum intensity value for change candidate pixels as change sample data. The highly reliable change pixels were selected, and then pseudo-training data was used to train the SSViT model. At last, the change map is generated by training the SSViT network based on pseudo-training data. The performance of the SSViT model is evaluated for real-world hyperspectral (HS) datasets with different change landcover types. Furthermore, a new series of HS images is introduced for CD purposes. The results of CD show that the HS images have a high potential for detecting subtle changes. The experimental results demonstrate that the proposed SSViT could outperform the advanced HCD methods.