Abstract
In recent years, significant progress has been made in the field of face forgery and face forgery detection. However, the performance of the detection methods in the unknown environment is far beyond satisfactory due to the feature distribution deviation of different fake face generators. In this paper, we adopt the domain generalization theory to improve the generality of fake face detection. The utilized method augments the original image samples by introducing gradient noise yielded during back-propagation, simulating the forgery features in unknown domains. In the construction of the detection network, we propose a multi-scale synthetic artifact trace tracker (MSATT) to enhance the manipulation traces through multi-scale content suppression. Meanwhile, we observed that the synthesized images present a noticeable color abnormality after going through the proposed MSATT module. Therefore, we designed a color difference perception network (CDPNet) to capture this unique feature. Experimental results demonstrate that both the domain augmentation and the proposed CDPNet can effectively improve the performance of the detection network. The proposed method is competitive with the state-of-the-art face forgery detection methods on both intra- and inter-dataset evaluations.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
The highly realistic fake human faces in digital images or videos synthesized by machine-learning systems (deepfakes) have caused broad concerns. Abuse of these generated faces poses considerable threats to social security, such as fake news and video scams. Therefore, it is urgent to develop effective detection techniques to ensure the credibility of multimedia information.
Significant achievements have been made in the field of face forgery detection in recent years. Some of these methods [1,2,3,4,5] were designed based on Convolution Neural Networks (CNN). Afchar et al. [1] used CNN to extract mesoscopic features for face forgery detection. Masi et al. [2] proposed a two-branch network structure. One branch propagates the original information, while the other branch suppresses the face content yet amplifies multi-band frequencies using a Laplacian of Gaussian (LoG) as a bottleneck layer. Nguyen et al.[3] adopted a novel capsule network for detecting forged images and videos. Liu et al. [4] combined spatial image and phase spectrum to capture the up-sampling artifacts of face forgery to improve the transferability for face forgery detection. Zhou et al. [6] proposed a two-stream network structure. One stream to detect tampering artifacts for classification. The other stream trains a patch-based triplet network to leverage features capturing local noise residuals and camera characteristics as a second stream. Zhao et al. [6] defined face forgery detection as a fine-grained classification task for the first time and proposed a multi-attentional deepfake detection network. Guo et al. [7] suppressed the content features to make the model pay more attention to the artifact regions. Kohli et al. [8] proposed a light weight 3DCNN for face forgery detection to increase efficiency. What’s more, Kohli et al. [9] is used for facial forgery detection by searching for operational traces in the frequency. These methods can reach high accuracy on hold-out test sets, but their performance usually drops significantly on unseen domains.
A pressing need for deepfake forensics is improving the detectors’ generality. In this paper, we incorporate the domain generalization theory to address this issue. To the best of our knowledge, the application of domain generalization theory in face forgery detection has yet to be fully discussed. Few works use domain generalization theory in deepfake detection. For example, the authors of [10] enrich the diversity of fake samples by giving a pool of the forgery configuration. In contrast, our method improves the generality of the detection model by introducing adaptive gradient noise to the training data. The additional gradient noise is generated during backpropagation, which effectively uses the information learned by the network itself. It is worth noting that the training phase only uses forged samples from a single domain (single domain refers to the feature space of samples generated using a single forgery method) and training in a single domain is challenging. Since the enhanced domain is generated under the premise that the worst-case scenario is satisfied, it will be constrained by semantic consistency [11], resulting in poor domain transmission, we adopt semantic constraints relaxing module used to relax semantics constraints during domain augmentation.
The statistically based deepfake artifacts usually exist in the form of subtle traces, which may be obscured by the media content. Accordingly, we design an MSATT module to suppress content features and enhance subtle manipulation traces in a multi-scale fashion. Furthermore, we observed that there exists color difference in images before and after the MSATT pre-processing, and the color difference is significantly more evident in the synthesized samples than in the real samples. In light of this phenomenon, we propose a CDPNet to capture this color-based feature to enhance the discriminability of the detection method.
The key contributions of this paper are threefold as below:
-
1.
We introduce the domain augmentation theory into fake face detections. A domain augmentation module automatically simulates the unknown feature domains through perturbations on the source domain, therefore enhancing the generality of the detection model.
-
2.
We design an MSATT module to extract robust face forgery traces. This module enhances the subtle forgery traces by employing multi-scale feature fusion as well as content feature suppression.
-
3.
We propose a novel color-difference feature for fake face detection. We observe that the MSATT also reveals a distinct color-behavior difference between the original and the synthesized images. We construct a CDPNet to capture this unique feature.
2 Proposed method
In this section, we introduce the proposed method in detail. We first give the overall of the proposed method in Section 2.1. Section 2.2 introduce the task module, which includes the MSATT and the CDPNet. Then, the domain augmentation strategy will be discussed in Section 2.3.
2.1 Overview
The overall architecture of the proposed method is shown in Fig. 1.
The framework mainly includes a task module and a domain augmentation module. In the task module, the RGB images to be tested are sent to the MSATT to capture the enhanced forgery traces. In addition, the high-frequency images are extracted from the RGB sequence using the SRM filter [12] to explore the noise information. We proposed a CDPNet to deal with the MSATT exposed color abnormality feature. Following are the backbone layers and a binary classifier. The domain augmentation module (DA) uses the method in [13] to enhance the source domain by simulating the distributions in the unknown environment under the worst-case constraint [11]:
Where D is the similarity measurement between the original domain and the generalized target domain, and \(\delta \) represents the largest domain boundary between Source Domain(S) and Target Domain(T). \(\theta \) is a parameter optimized according to the objective function L of a specific task. In addition, we maximize domain transmission capacity with training with the assistance of the SCR module. The SCR module is a Wasserstein Auto-Encoders (WAEs) [14], which is used to relax the semantics constraints to generate more challenging samples. It is worth noting that the input images of the task module come from S and \(S^+\). \(S^+\) means the enhanced domain generated by the domain augmentation module.
2.2 Task module
The task module extracts discriminative forgery features as well as trains a classifier. In each iteration, the samples from the original and enhancement domains are fed into the SRM filter and the MSATT module. The following backbone layers and the CDPNet extract discriminative features to train a classifier.
The pre-processing stage The task module starts with a two-branched pre-processing. The first branch adopts the SRM filter to extract the noise features. The other branch is the MSATT. In this paper, we mainly utilize the low-level statistical-based deepfake features to expose fake face forgery. According to [7], CNNs are more inclined to extract content representations due to their relatively fixed structures. Moreover, the statistically based deepfake traces are often fragile and could be easily diminished. Therefore, we propose the MSATT to enhance the delicate forgery traces while suppressing the content features.
The MSATT is divided into three stages. In the first stage, a convolution operation Conv(*) is used to calculate a feature map \(F_1\):
Where \(\theta _1\) denotes the parameters of the convolution layers in the first stage. \(x\in \mathbb {R}^{W\times H\times C}\) denotes the input tensor. W denotes the height, H denotes the width, and C denotes the number of tensor channels.
The manipulation traces extracted using only one layer are fragile and could easily disappear in subsequent convolutions. Therefore in the second stage, a multi-scale convolution module is used to calculate multi-level feature maps \(F_2\) from different perception fields.
Where \(\theta _2\) denotes the parameters of multi-scale convolution layers in the second stage. \(<*>\) indicates the different size of the convolution layer’ filter.
In the third stage of the MSATT module, we subtract the input tensor x from \(F_1\) and \(F_2\), respectively, to suppress the content representations while exposing the subtle manipulation traces. For further exposing as many discriminative cues for face forgery detection, we use a convolution layer in the third stage to perform convolution on the output \(F_2\) of the second stage to fully utilize the remaining information. Meanwhile, we add the output of the high-frequency branch Q to get the final output tensor x’ of the pre-processing stage:
where \(\theta _3\) denotes the convolutional parameters of the third stage that preserves the additional discriminative information.
Then we use the backbone layers to obtain discriminative features and train a classifier. We use the cross-entropy loss function for the classification task:
where \(\hat{y_i}\) is the softmax output of the task model, y denotes the label, and N is the number of inputs.
Color difference perception network (CDPNet) After extensive observations, we found that the output image of MSATT often has an overall hue difference compared to the input image. The hue difference is more significant and stable in synthesized images than in pristine images. We speculate that this phenomenon stems from the inherent color distribution abnormalities present in deepfake images and videos. However, these abnormalities are often subtle and imperceptible to the human eye. After the MSATT module, the deepfake anomalies are amplified, making such hue differences apparent. We illustrate this phenomenon in Fig. 2.
Motivated by this observation, we designed a CDPNet to capture the color-based feature. We first use (6) to measure the color difference in the image before and after the MSATT pre-processing.
where \(I_1\) and \(I_2\) represent the pixel value matrix of the input and output images of the MSATT module, respectively. R, G and B represent three color channels. After obtaining the color-difference measurement Diff, we fed it into the self-designed CDPNet. We only designed a few layers of convolutional structure and combined it with the short connections of ResNet to obtain more robust features. The structure of the CDPNet is listed in Table 1.
2.3 Domain augmentation
The purpose of domain augmentation is to make the detection model learn features from unknown distributions. The concept is shown in Fig. 3.
The domain augmentation simulates cross-domain distributions by adding gradient noise to the source domain. Due to the unpredictable nature of the noise signal, the augmentation expands in near-random directions based on the source domain. According to our observations and understanding of Deepfake datasets, we have found that the feature spaces of different Deepfake datasets tend to cluster together. Therefore, despite the random nature of the domain augmentation, it still has the ability to extend to adjacent unknown domains.
We hope that the distribution of the enhancement domain should deviate from the source domain as much as possible to simulate a broader distribution of the unknown environment. This paper implements domain augmentation by satisfying the worst-case scenario under the constraint of ensuring semantic consistency, but the constraint of semantic consistency will limit the transmission capacity (the expansion capability from the source domain to the target domain), shown in Fig. 3(b), red spots generated from blue spots will be limited to a certain range. So we alleviate the constraints of semantic consistency while keeping it across the line between real and fake.
To achieve this goal, we exploit a specific domain augmentation strategy that simulates the unknown domain by introducing gradient noise into the source domain. The samples that need to be enhanced are considered as part of the trainable parameters of the task model and then use objective function \(L_{total}\) to calculate the gradients of the input layer to update the input samples, similar to the network’s backpropagation process. \(L_{total}\) is the total loss required for the sample enhancement stage, as shown in (7).
\(L_{total}\) consists of three losses: \(L_{cls}\), \(L_{const}\), and \(L_{relax}\). \(\alpha \) and \(\beta \) are hyperparameters to balance \(L_{const}\) and \(L_{relax}\). We will detail these losses in the following.
\(L_{cls}\) is the classification loss defined in (5), it is the optimization function in the worst-case scenario.
\(L_{const}\) is used to maximize the difference between the source domain and the target domain while satisfying semantic consistency. Ensure that high-level semantic features are related to class labels. The definition is as follows:
where \(z^{+}\) represents the discriminative features extracted from enhanced samples. 1• indicate 0-1indicator function and it will be \(\infty \) if the class label of \(x^{+}\) is different from x. \(L_{const}\) can achieve a certain degree of out-of-domain generalization in the embedded space, but its out-of-domain transmission ability is limited due to semantic consistency constraints. In order to enhance the transmission capability outside the domain and increase the diversity of samples, we adopted \(L_{relax}\) to alleviate the constraint of semantic consistency which is defined as follows.
\(L_{relax}\) is for mitigating constraints on semantic consistency by limiting the encoding capabilities of WAEs which are defined in (9).
Where E and D represent the Encoder and Decoder, respectively. In this paper, we use WAEs [14] to implement \(L_{relax}\). By limiting the encoding ability of WAE, the reconstruction error increases, and more disturbances are generated to enhance the domain transmission capability. The structure of the encoder and decoder is described in Table 2. The encoder and decoder need to be trained in advance to better capture the distribution of the source domain. Then limit the encoding ability of WAEs to maximize the difference between the enhancement domain and the source domain.
After obtaining the objective function \(L_{total}\) that needs to be optimized, we can use the iterative method to perturb the original sample along the direction of gradient change to generate more samples \(x^+\):
where \(\zeta \) represents the scale factor. \(\theta \) and \(\varphi \) represent the convolutional layers’ parameters of the task module and SCR module respectively. Our main idea is to add disturbance to the sample to obtain new ones. The feature vector z is obtained from the samples in the source domain through the task model. In order to make the difference between z and the feature vector \(z^+\) of the samples corresponding to the augment domain larger, we augment the samples along the gradient change direction in the back propagation process through adaptive learning. In simple terms, we update the input samples in a way similar to the network parameter update. It is also worth noting that during the non-data augmentation training phase, we only use \(L_{cls}\) for optimization of the parameters \(\theta \) of the task module:
Where \(\eta \) is the learning rate. Training task module on the original and enhanced domains to achieve better generalization performance.
Our work mainly includes the following two points. First, in order to better capture manipulation traces, MSATT is used to suppress content features to expose manipulation traces. Second, simulate the distribution outside the domain through the domain enhancement strategy. Using \(L_{const}\) and \(L_{relax}\) to maximize the expansion outside the domain by relaxing the constraints of semantic consistency. Our method implementation is summarized in Algorithm 1.
3 Experiment
In this section, we conduct several experiments to verify the effectiveness of the proposed method. Section 3.1 provides the details of the experimental setup. Section 3.2 reports the ablation experiment. Section 3.3 provides visualization of the generalization of our method. Section 3.4 reports the experimental results with recent works. Section 3.5 verifies the model generalization performance.
3.1 Experimental setup
Datasets We performed experiments on several of the most popular deepfake datasets: FaceForensics++ (FF++) [15], Celeb-DF [16], Deepfake Detection Challenge (DFDC) [17] and FaceShifter [18]. The FF++ contains four forgery patterns: DeepFake (DF) [19], Face2Face (F2F) [20], FaceSwap (FS) [21], and NeuralTexture (NT) [22]. A total of 4,000 forged videos were generated based on 1,000 pristine videos. In addition, according to different compression rates, FF++ also provides three different levels of compressed video: pristine quality (raw), high quality (HQ), and low quality (LQ). In this paper, We regard DeepFake, Face2Face, FaceSwap, and NeuralTexture as four datasets. Note that we divide the dataset according to the official ratio of 720:140:140. Celeb-DF [16] is another widely used deep forged dataset. It improves the visual quality of the video samples and is more challenging for face forgery detection tasks. DeepFake Detection Challenge (DFDC) [17] is another more challenging dataset which contains 1,000 pristine videos and over 4,000 fake videos manipulated by multiple DeepFake, GAN-based and non-learned methods. FaceShifter [18] is a more challenging face forgery detection that is not only considerably more perceptually appealing, but also better identity preserving in comparison.
Metrics The metrics in our experiments are Accuracy (ACC) and Aera under the curve (AUC), which are most commonly used for evaluating face forgery detection methods.
Implementation details The backbone of the proposed architecture is the Xception [23] which is pre-trained on imagenet. We use MTCNN [24] to exact the face areas, and align and resize them to 256\(\times \)256 pixels. The hyper-parameter used in (7) are \(\alpha = 0.0001\) and \(\beta = 1e9\). We set the batch size to 32, and use the Adam optimizer. The learning rate of the task module and WAE is set to 0.00002 and 0.0005, respectively. Our experiments run on an NVIDIA GTX GeForce 1080Ti GPU.
3.2 Ablation study
In this section, we carry out several ablation experiments to verify the effectiveness of the proposed MSATT, CDPNet, and Data Augmentation (DA) strategy. All ablation experiments were trained on DF and tested on each of the four sub-datasets in FF++. The results are shown in Table 3. The results show that the main modules we propose can improve the detection performance on both intra- and inter-datasets. The first row is the detection results of the backbone. In the second row, we tested the efficacy of the MSATT. Compared to the backbone, the performance slightly drops on the same domain but improves on the cross-domain.
In the third row, we evaluated the CDPNet. Since there is a dependency between the CDPNet and the MSATT, we tested the two modules jointly. Compared to the backbone, it has increased by 0.23%, 1.14%, 1.40%, and 0.57% on DF, F2F, FS, and NT, respectively. We analyzed the DA strategy in the fourth row. The results on cross-domain datasets have significantly improved in comparison to the baseline, which is evident that the DA enhances the generality of the detection algorithm. In the last row, we evaluated the performance after integrating all proposed modules. It demonstrates the optimal results both within and across datasets. Compared to the backbone, it has increased by 0.59%, 3.25%, 2.13%, and 1.11% on DF, F2F, FS and NT, respectively. We can see that the improvement in the intra-domain performance is not substantial, whereas there are notable improvements in cross-domain performances. The reason is that the baseline performance of the intra-domain (tested on DF) has already reached 99.14%, leaving little room for significant improvement. In contrast, there is ample room for improvement in cross-domain performances compared to the baseline. Moreover, the significant improvement in cross-dataset performance demonstrates the effectiveness of domain generalization.
3.3 Visualization of improved generalization
The purpose of the proposed method is to improve the cross-domain generalization of the face forgery detection model.
We perturb the original samples to let the model learn more diverse representations under unknown distributions. Figure 4 demonstrates the visualization results of the feature domain augmentation. It realizes dimension reduction by using t-sne [25]. By comparing Fig.4(a) and (b), we find that the feature distributions deviate further from the centroid after adding the DA. It means that our method does expand both the real and fake data domains. By observing (c) and (d), we find that after domain generalization, the model is better able to distinguish the real and fake data in the cross-dataset evaluation, which proves the DA can expand the feature domains along the correct directions to a certain extent.
3.4 In-Domain evaluation
In this section, we evaluate the in-domain performance of the proposed method. We compare our method with five SOTA deepfake forensic methods. All the methods are trained and tested on each of the four sub-datasets in FF++. Table 4 shows the detection results. We can see that the proposed method achieves competitive performances within the same domain. Our method’s AUC metric is 0.72% higher than that of Yang et al. [27] in the DeepFake subset. Our method’s AUC metric is 0.89% higher than that of Yang et al. [27] in the Face2Face. Testing on FaceSwap, our method’s AUC is 1.40% higher than that of Qian et al. [26]. Similarly, Testing on NeuralTexture, our method’s AUC is 3.99% higher than that of Yang et al. [27].
We show the comprehensive performance of all methods by averaging the AUC on four sub-datasets. We can see that our method is superior to other methods. Meanwhile, Fig. 5 visualize the ROC curve to show it more intuitively.
3.5 Out-of-domain evaluation
Cross-Manipulation Evaluation In this part, we evaluate the generalization of the proposed method to unseen manipulations. The datasets and results are shown in Table 5. For a fair comparison, the methods compared were all re-implemented so that the training sets could be kept the same. We can see that our method shows overall superior performance in comparison to the rest. Especially, on the training set F2F, our method is superior to all other methods on all the test sets. However, the proposed method does not perform adequately on the training set NT. Our analysis suggests that our method has a relatively ideal effect on the expression transfer method based on computer graphics, while the effect on deep learning expression transfer is not very satisfactory. We calculated the average AUC in the last column. Our results are respectively 1.74%, 3.53%, 4.36%, and 0.96% higher than the second-highest. We also provide histograms to intuitively demonstrate the generalization of the proposed method against the comparison methods, as in Fig. 6.
Cross-dataset evaluation In this part, we assess the cross-dataset generalization of the proposed method. The test datasets are Celeb-DF, DFDC, and FaceShifter. We train each model on DF, F2F, FS, and NT, respectively. We compared several SOTA deep face detection methods with good generalization performance. Li et al. [30] identify authenticity by detecting mixed boundaries. Qian. et al. [26] obtained identification clues through the frequency domain. The comparison results are shown in Table 6. Analyzing the experiment data can clearly find that our method can achieve the best or second-best results. Especially the results of training on DeepFake(DF), our method achieves the best test results on Celeb-DF, DFDC, and FaceShifter. This indicates that our method performs well in terms of generalization for forgery types of DeepFake. The training results on F2F and NT show that our method has satisfactory generalization performance on Celeb-DF and FaceShifter, and can also achieve the second best performance on DFDC. The effectiveness of our method on FS is not very satisfactory, possibly due to insufficient learning of this type of forgery in our method. In addition, we also calculated the average AUC of the four different manipulation test datasets. It can be intuitively seen that our method can achieve the best result on F2F, it is 2.71% higher than that of the second-best result method. The other three results can achieve the second-best effect.
4 Conclusion
In this paper, we provide a new feasible method for improving the generality of face forgery detection. We adopt the domain generalization theory to simulate the real and fake face feature distributions in the unknown environment. The employed technique introduces gradient disturbances to the source domain in an automatic fashion. We demonstrate the improved generalization through visualizations and quantitative results. Since the domain extension is uncontrollable in direction, it is actually generalizing in a nearly-random manner. In the future, we will investigate methods that can control domain augmentation under meaningful ranges.
We find that when trying to enhance the subtle manipulation traces through specifically designed CNN structures, the outputs reveal chromatic anomalies, such as hue-shifts after the proposed MSATT module. This result supports that the colors of deepfake images are actually unnatural, which will become evident with magnification. In future research, we will continue to study these hidden features, such as color features of other attributions.
Data Availability
The datasets generated during and analyzed during the current study are available from the corresponding author upon reasonable request.
References
Afchar D, Nozick V, Yamagishi J, Echizen I (2018) Mesonet: a compact facial video forgery detection network. In: 2018 IEEE international workshop on information forensics and security, pp 1–7. IEEE
Masi I, Killekar A, Mascarenhas, RM, Gurudatt, SP, AbdAlmageed W (2020) Two-branch recurrent network for isolating deepfakes in videos. In: European conference on computer vision, pp 667–684. Springer
Nguyen HH, Yamagishi J, Echizen I (2019) Capsule-forensics: using capsule networks to detect forged images and videos. In: ICASSP 2019-2019 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp 2307–2311. IEEE
Liu H, Li X, Zhou W, Chen Y, He Y, Xue H, Zhang W, Yu N (2021) Spatial-phase shallow learning: rethinking face forgery detection in frequency domain. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 772–781
Zhou P, Han X, Morariu VI, Davis LS (2017) Two-stream neural networks for tampered face detection. In: 2017 IEEE conference on computer vision and pattern recognition workshops (CVPRW), pp 1831–1839. IEEE
Zhao H, Zhou W, Chen D, Wei T, Zhang W, Yu N (2021) Multi-attentional deepfake detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2185–2194
Guo Z, Yang G, Chen J, Sun X (2021) Fake face detection via adaptive manipulation traces extraction network. Comput Vis Image Underst 204:103170
Kohli A, Gupta A (2022) Light-weight 3dcnn for deepfakes, faceswap and face2face facial forgery detection. Multimed Tool Appl 81(22):31391–31403
Kohli A, Gupta A (2021) Detecting deepfake, faceswap and face2face facial forgeries using frequency cnn. Multimed Tool Appl 80:18461–18478
Chen L, Zhang Y, Song Y, Liu L, Wang J (2022) Self-supervised learning of adversarial example: towards good generalizations for deepfake detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 18710–18719
Volpi R, Namkoong H, Sener O, Duchi JC, Murino V, Savarese S (2018) Generalizing to unseen domains via adversarial data augmentation. Advan Neural Inform Process Syst 31
Luo Y, Zhang Y, Yan J, Liu W (2021) Generalizing face forgery detection with high-frequency features. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 16317–16326
Qiao F, Zhao L, Peng X (2020) Learning to learn single domain generalization. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 12556–12565
Tolstikhin I, Bousquet O, Gelly S, Schoelkopf B (2017) Wasserstein auto-encoders. arXiv:1711.01558
Rossler A, Cozzolino D, Verdoliva L, Riess C, Thies J, Nießner M (2019) Faceforensics++: learning to detect manipulated facial images. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1–11
Li Y, Yang X, Sun P, Qi H, Lyu S (2020) Celeb-df: a large-scale challenging dataset for deepfake forensics. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3207–3216
Dolhansky B, Bitton J, Pflaum B, Lu J, Howes R, Wang M, Ferrer CC (2020) The deepfake detection challenge (dfdc) dataset. arXiv:2006.07397
Jiang L, Li R, Wu W, Qian C, Loy CC (2020) Deeperforensics-1.0: a large-scale dataset for real-world face forgery detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2889–2898
Deepfakes github. website, https://github.com/deepfakes/faceswap. Accessed: 24 March 2022
Thies J, Zollhofer M, Stamminger M, Theobalt C, Nießner M (2016) Face2face: real-time face capture and reenactment of rgb videos. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2387–2395
Faceswap (2018) https://github.com/marekkowalski/faceswap/. Accessed: 29 October 2018
Thies J, Zollhöfer M, Nießner M (2019) Deferred neural rendering: Image synthesis using neural textures. Acm Transactions on Graphics (TOG) 38(4):1–12
Chollet F (2017) Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1251–1258
Xiang J, Zhu G (2017) Joint face detection and facial expression recognition with mtcnn. In: 2017 4th International conference on information science and control engineering (ICISCE), pp 424–427. IEEE
Van der Maaten L, Hinton G (2008) Visualizing data using t-sne. J Mach Learn Res 9(11)
Qian Y, Yin G, Sheng L, Chen Z, Shao J (2020) Thinking in frequency: face forgery detection by mining frequency-aware clues. In: European conference on computer vision, pp 86–103. Springer
Yang J, Xiao S, Li A, Lu W, Gao X, Li Y (2021) Msta-net: forgery detection by generating manipulation trace based on multi-scale self-texture attention. IEEE Trans Circ Syst Video Technol
Tan M, Le Q (2019) Efficientnet: rethinking model scaling for convolutional neural networks. In: International conference on machine learning, pp 6105–6114. PMLR
Yu C-M, Chen K-C, Chang C-T, Ti Y-W (2022) Segnet: a network for detecting deepfake facial videos. Multimedia Syst 28(3):793–814
Li L, Bao J, Zhang T, Yang H, Chen D, Wen F, Guo B (2020) Face x-ray for more general face forgery detection. In: Proceedings of the IEEE conference on CVPR, pp 5001–5010
Acknowledgements
This work was supported by the National Natural Science Foundation of China under Grant 61802064.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interests
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Li, W., Feng, C., Wei, L. et al. Improving the generalization of face forgery detection via single domain augmentation. Multimed Tools Appl 83, 63975–63992 (2024). https://doi.org/10.1007/s11042-023-17840-2
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-023-17840-2