Low visibility always leads to serious traffic accidents worldwide, although extensive works are studied to deal with the estimation of visibility in meteorology areas, it is still a tough problem. Deep learning-based visibility estimation methods, it has low accuracy due to lacking "specific features" of the foggy images. Meanwhile, physical modelbased visibility estimation methods are only applicable to some specific scenes due to its high requirements for extra auxiliary parameters. Therefore, This paper proposes a novel end-to-end framework named STCN-Net for visibility estimation, which combined the "engineered features" and "learned features" to achieve higher accuracy. Specifically, a novel 3D multi-feature stream Matrix, named DDT, is designed for visibility estimation, which is consisted of a transmittance matrix, a dark channel matrix, and a depth matrix. Unlike traditional deep learning methods which only use convolutional neural networks (CNN) to deal with the input data or images, our method combines CNN and Transformer to process the input data or images. In STCN-Net, Swin-Transformer(Swin-T) module takes the original image as input while the CNN module takes the DDT matrix as input. Moreover, in order to integrate different feature information from the CNN and Swin-T, we embed a Coordinate Attention (CA) module in STCN-Net. Finally, two visibility datasets: Visibility Image Dataset Ⅰ (VID I) and Visibility Image Dataset Ⅱ (VID II) were constructed for evaluation where VID I is a real scene visibility dataset and VID II is a synthetic visibility dataset. The experimental results show that our method has better performance than classical methods on the two datasets. And compared with the runner-up, it has 2.1% more accuracy in VID I and 0.5% in VID II.
Low visibility always leads to serious traffic accidents worldwide, and it remains a tough problem despite the extensive work researched in the field of meteorology. Therefore, we propose a novel end-to-end network, FGS-Net, for visibility estimation, combining "engineered features" and "learned features" to achieve higher accuracy. Specifically, we propose a novel and effective fog region segmentation method, named ASRS(Auto Seed Region Segmentation), to segment the fog regions in the input image. Subsequently, two "specific features" (transmittance matrix, dark channel matrix) and three "common features" (contrast, average gradient, brightness) are extracted from the fog region. Next, the feature information of the fog region is statistically calculated as "engineering features" for visibility estimation. In addition, our approach uses Transformer, a classical model commonly used in the field of Natural Language Processing (NLP) to obtain the "learned features" for visibility estimation. Furthermore, in order to obtain more efficient "learned features", we embed the Coordinate Attention (CA) module in FGS-Net. Finally, to verify the effectiveness and superiority of our method, We evaluate our approach on two visibility datasets: Visibility Image Dataset Ⅰ (VID I) and Visibility Image Dataset Ⅱ (VID II), where VID I is a real scene visibility dataset and VID II is a synthetic visibility dataset. The experimental results show that our method has better performance than the classical one on these two datasets. And compared with the runner-up, it has 2.2% and 0.9% higher accuracy in VID I and VID II, respectively.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.