2018
DOI: 10.1007/978-3-030-01258-8_23
|View full text |Cite
|
Sign up to set email alerts
|

Modeling Visual Context Is Key to Augmenting Object Detection Datasets

Abstract: Performing data augmentation for learning deep neural networks is well known to be important for training visual recognition systems. By artificially increasing the number of training examples, it helps reducing overfitting and improves generalization. For object detection, classical approaches for data augmentation consist of generating images obtained by basic geometrical transformations and color changes of original training images. In this work, we go one step further and leverage segmentation annotations … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
200
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 213 publications
(203 citation statements)
references
References 40 publications
3
200
0
Order By: Relevance
“…Alternatively, artificial images can be produced to augment the dataset [63], [64], [90], [91] by generating composite images in which multiple crops and/or images are blended. A straightforward approach is to randomly place cropped objects onto images as done by [90].…”
Section: Generative Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Alternatively, artificial images can be produced to augment the dataset [63], [64], [90], [91] by generating composite images in which multiple crops and/or images are blended. A straightforward approach is to randomly place cropped objects onto images as done by [90].…”
Section: Generative Methodsmentioning
confidence: 99%
“…However, the produced images may look unrealistic. This problem is alleviated by determining where to paste and the size of the pasted region according to the visual context [91]. In a similar vein, the objects can be swapped between images: Progressive and Selective Instance-Switching (PSIS) [64] swaps single objects belonging to the same class between a pair of images considering also the scales and shapes of the candidate instances.…”
Section: Generative Methodsmentioning
confidence: 99%
“…Previous method [15] investigated applying context model to explicitly model the consistency of the object and background in semantic space. Different from their approach, our appearance consistency map does not consider the semantic consistency explicitly but enforces the object to be pasted at places with similar background pattern on the original image.…”
Section: Discussionmentioning
confidence: 99%
“…Instance-level augmentation. One branch of recent work has emerged with more precise instance-level image augmentation, laying potential to fully exploit the supervised information in the existing dataset [16,26,14,15,18,25,43]. Dwibedi et al [16] improved instance detection by simple cut-and-paste strategy with extra instances that have annotated masks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation