The World Wide Web Conference 2019
DOI: 10.1145/3308558.3313691
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic Deep Multi-modal Fusion for Image Privacy Prediction

Abstract: With millions of images that are shared online on social networking sites, e ective methods for image privacy prediction are highly needed. In this paper, we propose an approach for fusing object, scene context, and image tags modalities derived from convolutional neural networks for accurately predicting the privacy of images shared online. Speci cally, our approach identi es the set of most competent modalities on the y, according to each new target image whose privacy has to be predicted. e approach conside… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
16
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 24 publications
(16 citation statements)
references
References 52 publications
0
16
0
Order By: Relevance
“…Buschek et al (2015) proposed a multi-modal method that assigns privacy labels to the images based on visual features and metadata like location and publication time. Tonge, Caragea, and Squicciarini (2018) utilized another kind of metadata, tag, and Tonge and Caragea (2019) further derived features of the object, scene, and tags for privacy-leaking image detection. Yang et al (2020) extracted a knowledge graph from the images and identified private images based on object detection and graph neural networks.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Buschek et al (2015) proposed a multi-modal method that assigns privacy labels to the images based on visual features and metadata like location and publication time. Tonge, Caragea, and Squicciarini (2018) utilized another kind of metadata, tag, and Tonge and Caragea (2019) further derived features of the object, scene, and tags for privacy-leaking image detection. Yang et al (2020) extracted a knowledge graph from the images and identified private images based on object detection and graph neural networks.…”
Section: Related Workmentioning
confidence: 99%
“…Yang et al focuses on objects and their correlation to identify private images based on object detection. However, they neglect other important elements like scenes (Tonge and Caragea 2019), textures, and objects beyond the capacity of pre-trained object detectors. Furthermore, the correlation among objects is fixed, but the elements vary in different images, making the fixed correlation inappropriate.…”
Section: Introductionmentioning
confidence: 99%
“…In particular, Tonge and Caragea [55] use PicAlert image dataset at one-shot to train an SVM classifier, which achieves an accuracy of 83.14% by using object tags created via ImageNet. In a more recent work [57], they propose an approach for fusing object, scene context, and image tags modalities. The model identifies the set of most competent modalities on the fly and obtains an accuracy of 86.36% .…”
Section: Performance Of the Estimation Functionmentioning
confidence: 99%
“…Although the problem of privacy setting recommendation for images have been extensively studied before, none of the approaches address all of these requirements. An important set of approaches [47,55,57,62] train various machine learning approaches to predict the privacy labels of images. The size of the data they need changes based on the trained model but many of them require a large amount of data to train accurate classifiers because of the model complexity.…”
Section: Introductionmentioning
confidence: 99%
“…Interestingly, the resulting model provided more consistent predictions compared to users' judgments, indicating that users might fail to follow their own privacy-related preferences. A multimodal prediction model which mixes visual content and tags is introduced in [52]. Performance is improved by exploiting predictions from neighboring photos from the user's stream.…”
Section: Related Workmentioning
confidence: 99%