2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020
DOI: 10.1109/cvpr42600.2020.00370
|View full text |Cite
|
Sign up to set email alerts
|

Assessing Image Quality Issues for Real-World Problems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 47 publications
(14 citation statements)
references
References 48 publications
0
14
0
Order By: Relevance
“…Datasets need to be large enough to account for the variability between questions, images and concepts that occur in real world scenarios (Kafle & Kanan, 2017). Our work complements prior work (Chiu, Zhao, & Gurari, 2020) by offering rich, user-centered guidance in how to construct future large-scale datasets to meet the real needs of real users of VQA services, specifically people with visual impairments.…”
Section: Visual Question Answering (Vqa)mentioning
confidence: 93%
“…Datasets need to be large enough to account for the variability between questions, images and concepts that occur in real world scenarios (Kafle & Kanan, 2017). Our work complements prior work (Chiu, Zhao, & Gurari, 2020) by offering rich, user-centered guidance in how to construct future large-scale datasets to meet the real needs of real users of VQA services, specifically people with visual impairments.…”
Section: Visual Question Answering (Vqa)mentioning
confidence: 93%
“…In addition, we also utilize a relevant unrecognizability prediction task, which predicts the unrecognizable degree of an image. This task is trained on the VizWiz-QualityIssues dataset (Chiu et al, 2020 ), containing images with labels of the unrecognizable degree. Even if intelligibility features of heavily distorted images cannot obtain desired results in original tasks, they can still be distinguished from features of high-quality images, which is beneficial to the IQA task.…”
Section: Proposed Methodsmentioning
confidence: 99%
“…For example, Kang et al [7][8] proposed a multi-task shallow CNN to learn both the distortion type and the quality score; Kim and Lee [9] applied state-of-the-art FR-IQA methods to provide proxy quality scores for each image patch as the ground truth label in the pre-training stage, and the proposed network was fine-tuned by the Subjective annotations. Similarly, Da Pan et al [10] employed the U-Net to learn the local quality predicting scores previously calculated by Full-Reference IQA methods, several Dense layers were then incorporated to pool the local quality predicting scores into an overall perceptual quality score; Liang et al [11] tried to utilize similar scene as reference to provide more prior information for the IQA model; Liu et al [12] proposed to use RankNet to learn the quality rank information of image pairs in the training set, and then used the output of the second last layer to predict the quality score; Yee et al [13] tried to learn the corresponding unknown reference image from the distorted one by resorting the Generative Adversarial Networks, and to assess the perceptual quality by comparing the hallucinated reference image and the distorted image; Chiu et al [1] proposed a new IQA framework and corresponding dataset that links the IQA issue to two practical vision tasks which are image captioning and visual question answering respectively; Su et al [14] employed self-adaptive hyper network whose parameters could adjust according to image contents; Zhu et al [15] leveraged meta-learning to learn a general-purpose BIQA model from training set of several specific distortion types.…”
Section: Related Workmentioning
confidence: 99%
“…As described above, recent state-of-the-art BIQA methods focus on predicting the distorted image solely, and take less consideration on how to make their proposed models incorporated into other downstream vision tasks. Amongst the IQA works above, only [1] tries to link IQA issues with other image vision tasks. Our work exploits how to leverage IQA models to optimize the perceptual quality of multi-channel transmitting systems, which is of much importance because such work could not only optimize the transmitting system both for servers and clients but also represents a beneficial attempt for linking the gap between IQA and other image vision tasks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation