Proceedings of the 28th ACM International Conference on Multimedia 2020
DOI: 10.1145/3394171.3414047
|View full text |Cite
|
Sign up to set email alerts
|

Privacy-Preserving Visual Content Tagging using Graph Transformer Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 19 publications
(10 citation statements)
references
References 25 publications
0
10
0
Order By: Relevance
“…In addition, when DA-GAT model exploits the ResNeXt-50 [53] as the backbone, the performance of DA-GAT model is 87.1%, and it is 0.5%, 2.3% and 6.3% higher than SGTN [42] model on mAP, C-F1 (All) and O-F1 (All), respectively. Therefore, the experimental results conirm the efectiveness and superiority of DA-GAT model compared with current state-of-the-art methods.…”
Section: Methodsmentioning
confidence: 97%
See 1 more Smart Citation
“…In addition, when DA-GAT model exploits the ResNeXt-50 [53] as the backbone, the performance of DA-GAT model is 87.1%, and it is 0.5%, 2.3% and 6.3% higher than SGTN [42] model on mAP, C-F1 (All) and O-F1 (All), respectively. Therefore, the experimental results conirm the efectiveness and superiority of DA-GAT model compared with current state-of-the-art methods.…”
Section: Methodsmentioning
confidence: 97%
“…In addition, in order to compare with the recently published literatures [3,4,55], we also exploit the image scale of 576 × 576 to train and test DA-GAT model. Meanwhile, to compare with the recently published literatures [34,42], we exploit ResNeXt-50 [53] network for image feature extraction with a semi-weakly supervised pre-trained model on ImageNet [8]. For label representations, following previous works [4,6], we leverage the 300-dim GloVe [29] model trained on the Wikipedia dataset.…”
Section: Datasets and Experimental Setingsmentioning
confidence: 99%
“…Singh et al [35] investigated human attributes (emotion, age, and gender) prediction under various de-identification privacy scenarios by body parts obfuscation. Vu et al [36] designed a framework for social media content tagging by constructing a global knowledge graph to avoid sensitive local data such as faces, passport numbers, and vehicle plates. However, these works primarily focused on static images, and none of them considered privacy issues in human motion evaluation, such as movement synchrony.…”
Section: B Privacy-preserving Machine Learningmentioning
confidence: 99%
“…To name a few, BERT (Devlin et al 2018), RoBERTa (Liu et al 2019) are helping many language related tasks to achieve new state-of-the-art results. FastText and GloVe were tested by Chen et al (2019b), Char2Vec and BERT were experimented by Vu et al (2020). Here we investigate more into these following language models: • Char2Vec (Kim et al 2015) is a deep language model that learns at character-level inputs.…”
Section: Language Embeddingsmentioning
confidence: 99%
“…Real-world images generally embodies rich and diverse semantic information with multiple objects or actions; therefore, multi-label classification has attracted a large number of recent studies in the artificial intelligence (AI) community (Wang et al 2020a;Yeh et al 2017;Zhu et al 2017). Recognising object labels in images has many applications, ranging from social tag recommendation (Nam et al 2019;Vu et al 2020) and fashion trend analysis (Inoue et al 2017) to functional genomics (Bi and Kwok 2011). The core challenge in multi-label learning is to understand and model object dependencies to exploit attributive knowledge.…”
Section: Introductionmentioning
confidence: 99%