Stereo matching has been widely used in various computer vision applications and it is still a challenging problem. Adaptive support weights (ASW) methods represent the state of the art in stereo matching and have achieved outstanding performance. However, the local ASW methods fail to resolve the matching ambiguity in low texture areas because their cost aggregation is limited within local fixed or adaptive support windows. On the other hand, the non-local ASW methods perform cost aggregation along a special tree, so that these methods are often sensitive to high texture areas since some useful connectivity constrains between adjacent pixels are broken during constructing the special tree. To solve these problems, in this paper, a novel and generic fusing ASW framework are proposed for stereo matching. In this framework, we establish dual support windows for each pixel, i.e., a local window and the whole image window. As such, the primitive connectivity between each pixel and its neighboring pixels in the local window can be maintained, and then each pixel not only gets appropriate supports from neighboring pixels within its local support window but also receives more adaptive supports from the other pixels outside the local window. Furthermore, a local edge-aware filter and a non-local edge-aware filter, whose kernel windows correspond to the dual support windows, are merged in order to achieve collaborative filtering of the cost volume. The performance evaluation on the Middlebury and KITTI datasets shows that the proposed stereo matching method outperforms the current state-of-the-art methods. INDEX TERMS Stereo matching, cost aggregation, adaptive support weight, edge-aware filtering.
a faculty of automation and information engineering, Xi'an university of technology, Xi'an, china; b faculty of electronics and information engineering, ankang university, ankang, china ABSTRACT Outdoor images captured during sand-dust weather condition typically yield poor contrast and colour shift. A novel method for single sand-dust images restoration is introduced in this paper, which relies on the atmospheric scattering model and information loss constraint. To compensate the colour shift and achieve proper luminance, the proposed atmospheric light is changing with the content of the local scenes, which is initially estimated on the basis of the general scattering model and the grey-world assumption. Then, the initial atmospheric light is updated and the coarse transmission is estimated under the information loss constraint. Next, the fast guide filter is exploited in the post-refinement process to inhibit the halo artefacts. Comparison experiments demonstrate that the proposed algorithm is straightforward and efficient, the contrast and colour shift of different kinds of sand-dust images can be well compensated, especially, nice colour fidelity and proper luminance can be maintained.
Visually sensitive regions in the scene are thought to be important for scene categorization. In this paper, we propose to utilize the important visually sensitive information represented by deep features for scene categorization. Specifically, the context relationship between the objects and the surroundings is fully utilized as the main basis for judging the content of the scene, and combining with the deep convolution neural networks (CNNs), a scene categorization model based on deep visually sensitive features is constructed. First, the saliency regions of the scene images are marked according to the context-based saliency detection algorithm. Then, the original images and the corresponding visually sensitive region detection images are superimposed to obtain the visually sensitive region enhancement images. Second, the deep convolution features of the original images, the visually sensitive region detection images, and the visually sensitive region enhancement images are extracted through the deep CNNs pre-trained on the large-scale scene dataset Places. Finally, considering that the deep features extracted by different layers of the convolution network have different capabilities of discrimination, the fusion features are generated from multiple convolution layers to construct visually sensitive CNN model (VS-CNN). In order to verify the effectiveness of the proposed model, the experiments are conducted on the five standard scene datasets, i.e., LabelMe, UIUC-Sports, Scene-15, MIT67, and SUN. The experimental results show that the proposed model is effective and has good adaptability. Especially, our categorization performance is superior to many state-of the-art methods for a complex indoor scene.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.