“…On the other hand, Transformer networks, with their self-attention mechanism, proficient in capturing long-range dependencies, enhancing feature extraction and generalization when combined with CNNs (Bai & Tahmasebi, 2022;Vaswani et al, 2017). This integrated CNN-Transformer approach achieves high accuracy with fewer parameters, proving advantageous in computational resource-limited settings and facilitating research on large-sized porous media (Meng et al, 2023). Moreover, the infusion of physical information into the network-permitting the direct assimilation of porous media's physical parameters during training-markedly elevates the model's predictive precision and its generalization capacity (Kamrava, Im, et al, 2021;Meng et al, 2023;Tang et al, 2022).…”