Proceedings of the 2017 ACM on Conference on Information and Knowledge Management 2017
DOI: 10.1145/3132847.3132920
|View full text |Cite
|
Sign up to set email alerts
|

Dual Learning for Cross-domain Image Captioning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
27
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
2
2

Relationship

1
8

Authors

Journals

citations
Cited by 48 publications
(27 citation statements)
references
References 17 publications
0
27
0
Order By: Relevance
“…One captioner adapts the sentence style from source to target domain, whereas two critics, namely domain critic and multi-modal critic, aim at distinguishing them. Zhao et al [141] fine-tuned the pre-trained source model on limited data in the target domain via a dual learning mechanism.…”
Section: G Image Captioningmentioning
confidence: 99%
“…One captioner adapts the sentence style from source to target domain, whereas two critics, namely domain critic and multi-modal critic, aim at distinguishing them. Zhao et al [141] fine-tuned the pre-trained source model on limited data in the target domain via a dual learning mechanism.…”
Section: G Image Captioningmentioning
confidence: 99%
“…It can learn the distribution of the real dataset and generate synthetic samples conforming to that distribution. GAN have been successfully applied in image generation, image inpainting [48], image captioning [49][50][51], object detection [52], semantic segmentation [53,54], natural language processing [55,56], speech enhancement [57], credit card fraud detection [58] and supervised learning with insufficient training data [59]. From the experiments and results of these studies, it is evident that GAN conforms to the distribution of the original data samples and can generate realistic synthetic data.…”
Section: Related Workmentioning
confidence: 99%
“…Thus, methods developed on such datasets might not be easily adopted in the wild. Nevertheless, great efforts have been made to extend captioning to out-of-domain data [3,9,69] or different styles beyond mere factual descriptions [22,55]. In this work we explore unsupervised captioning, where image and language sources are independent.…”
Section: Language Domainmentioning
confidence: 99%
“…Chen et al [9] address cross-domain captioning, where the source domain consists of image-caption pairs and the goal is to leverage unpaired data from a target domain through a critic. In [69], the cross-domain problem is addressed with a cycle objective. Similarly, unpaired data can be used to generate stylized descriptions [22,46].…”
Section: Related Workmentioning
confidence: 99%