2022
DOI: 10.1007/978-3-031-20065-6_23
|View full text |Cite
|
Sign up to set email alerts
|

UIA-ViT: Unsupervised Inconsistency-Aware Method Based on Vision Transformer for Face Forgery Detection

Abstract: Intra-frame inconsistency has been proved to be effective for the generalization of face forgery detection. However, learning to focus on these inconsistency requires extra pixel-level forged location annotations. Acquiring such annotations is non-trivial. Some existing methods generate large-scale synthesized data with location annotations, which is only composed of real images and cannot capture the properties of forgery regions. Others generate forgery location labels by subtracting paired real and fake ima… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
17
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 48 publications
(17 citation statements)
references
References 29 publications
0
17
0
Order By: Relevance
“…To tackle the malicious use of face swapping, numerous detection methods have been proposed. Most existing works model face swapping detection as a binary classification problem and focus on designing better the features classified on [8], [10], [11], [13], [16], [36], the classifier network [7], [9], [12], [15], [37], [38] or the training policies [14], [39], [40] to improve the accuracy and generalization. For example, low-level artifacts are used in early detection methods, such as face warping artifacts [8].…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…To tackle the malicious use of face swapping, numerous detection methods have been proposed. Most existing works model face swapping detection as a binary classification problem and focus on designing better the features classified on [8], [10], [11], [13], [16], [36], the classifier network [7], [9], [12], [15], [37], [38] or the training policies [14], [39], [40] to improve the accuracy and generalization. For example, low-level artifacts are used in early detection methods, such as face warping artifacts [8].…”
Section: Related Workmentioning
confidence: 99%
“…As for the network, powerful backbones like Xception [9] and EfficientNet [12] are usually used. Recent works like UIA-VIT [37] utilize Transformer in face forgery detection and achieve high performance. Besides, different training policies such as contrastive learning [39] are applied to make the classifier generalize better on unseen face swapping methods.…”
Section: Related Workmentioning
confidence: 99%
“…We adopt the HQ version of FF for both training and testing, and only use one frame every video for testing. We compare our results with state-of-the-art image-based methods Multi-Attention [45], DCL [40], RECCE [3] and UIA-ViT [47]. We ran the public code of RECCE and UIA-ViT to produce results under the same setting.…”
Section: Cross-domain Performance Evaluationmentioning
confidence: 99%
“…We adopt the HQ version of FF for training, and only use one frame every video for testing. Under the same setting, we ran the public code of RECCE [3], UIA-ViT [47] and SBI [38] to produce corresponding results. In Table 2, we show a competitive performance with existing image-based methods, signaling satisfying adaptability of RFFR to different datasets, especially high quality datasets like Celeb-DF.…”
Section: Cross-domain Performance Evaluationmentioning
confidence: 99%
See 1 more Smart Citation