2015 IEEE International Conference on Computer Vision (ICCV) 2015
DOI: 10.1109/iccv.2015.203
|View full text |Cite
|
Sign up to set email alerts
|

Weakly-and Semi-Supervised Learning of a Deep Convolutional Network for Semantic Image Segmentation

Abstract: Deep convolutional neural networks (DCNNs) trained on a large number of images with strong pixel-level annotations have recently significantly pushed the state-of-art in semantic image segmentation. We study the more challenging problem of learning DCNNs for semantic image segmentation from either (1) weakly annotated training data such as bounding boxes or image-level labels or (2) a combination of few strongly labeled and many weakly labeled images, sourced from one or multiple datasets. We develop Expectati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

7
1,065
0
2

Year Published

2017
2017
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,043 publications
(1,074 citation statements)
references
References 41 publications
7
1,065
0
2
Order By: Relevance
“…Another important WSL application is segmentation. Many methods are based on MIL framework: MIL-FCN [49] extends MIL to multi-class segmentation, MIL-Base [50] introduces a soft extension of MIL, EM-Adapt [45] includes an adaptive bias into the MIL framework, and Constrained CNN (CCNN) [48] uses a loss function optimized for any set of linear constraints on the output space of a CNN.…”
Section: Related Workmentioning
confidence: 99%
“…Another important WSL application is segmentation. Many methods are based on MIL framework: MIL-FCN [49] extends MIL to multi-class segmentation, MIL-Base [50] introduces a soft extension of MIL, EM-Adapt [45] includes an adaptive bias into the MIL framework, and Constrained CNN (CCNN) [48] uses a loss function optimized for any set of linear constraints on the output space of a CNN.…”
Section: Related Workmentioning
confidence: 99%
“…Specifically, the validation set consists of 3 tiles (areas: 26,28,30) and the testing set consists of 3 tiles (areas: 32, 34, 37). The remaining 10 tiles represent the training set, and each tile is used for training an individual RF classifier.…”
Section: Experiments Setup and Detailsmentioning
confidence: 99%
“…At the same time, progress in graphic processing unit (GPU) and parallel computing technology has significantly increased the computing capability, such that learning a more complicated classifier with a larger amount of training data has becomes accessible to more researchers. Specifically, one of the most successful practices in this direction was the launch of deep CNN (convolutional neural networks) [19][20][21][22] in the computer vision community, which has become the dominant method for visual recognition and semantic classification [13,[23][24][25][26]. Furthermore, one of the most distinct characteristics of CNN is its ability to automatically learn the most suitable features, which has made the manual feature extraction process that is used in the traditional supervised-based classification methods unnecessary.…”
Section: Introductionmentioning
confidence: 99%
“…Subsampling layers, which are also called pooling layers, adjust outputs from convolutional layer to get translation invariance. CNN is mainly applied in computer vision field for big data, for example, image classification [124,125] and image segmentation [126]. Document (or textual) representation, also part of NLP, is the basic method for information retrieval and important to understand natural language.…”
Section: Big Data Deep Learningmentioning
confidence: 99%