Keywords

1 Introduction

Copy–move forged [1] image is obtained by imitating a part of image and gluing it at different locations upon the same image as shown in Fig. 1. The principal motive behind such kind of forgery technique is to replicate the objects of image or to conceal the selective information rendered through image. Copy–move forgery follows above perception, because of which forged portion has a constitutional resemblance to rest of the image. As forged section has implicit resemblance to rest portion of the image which made detection of such kind of forgery a complex process. Copy–move forged images can be fabricated using image processing software without leaving any trace of forgery. Tampered region of the image may go through several geometrical and post-processing attacks to make detection of forgery a complicated task. Detection of copy–move forged images is integrative part of image forensics [2]. Significant amount of research work is practiced in this field. Fridrich et al. [3] designed pioneer technique for sleuthing copy–move forgery. Their method extracts discrete cosine transform (DCT) features of overlapping blocks. The features are sorted and resemblance among feature vectors is computed to detect fiddled areas. Alkawaz et al. [4] suggested a colligated algorithm that also uses DCT coefficients as features for distinct blocks of varying sizes. Both of the techniques possess high computational complexity and inappropriate identification of altered areas when manipulated areas of image suffer from post-processing operations. Zandi et al. [5] proposed a method utilizing adjustive similarity threshold. Their method utilized standard deviation of the image blocks for computation of threshold proportions. In [6], Lee et al. suggested a strategy employing histogram of orientated gradients (HOG) to spot tampered areas. Their algorithm performed well when altered image suffer from small degree rotations, blurring, brightness adjustment, and color reduction. Silva et al. [7] applied point of interest as well as blocks of pixels for forged region detection. They utilized voting process in multiscale space.

Fig. 1
figure 1

An instance of region duplication image forgery (tampered region is outlined with red color rectangle)

From the literature survey, we observed that copy–move forgery detection results may sustain false matches (which are incorrectly detected as forged even if primitively they are not forged region of image). To increase the detection accuracy and reduce false matches while forged image enduring several post-processing attacks (blurring, contrast modification, luminance change, color diminution, etc.) is an open- research subject to cultivate. To minimize the presence of fictitious matches, we have used lexicographical sorting with Euclidean distance calculation accompanied with computation of shift vector. For precise localization of forged region, we have used hybridized feature extraction technique in which we have combined singular values (obtained from decomposition over GLCM) with LBP features. Proposed approach is invariant to post-processing attacks and obtained improved detection accuracy as well as reduced false matches than former techniques.

GLCM [8] is a texture feature extraction technique, which represents the frequency of occurrence of one gray tone, in a delimitated spatial linear relationship with other gray tones in image. SVD [9] possess algebraic and geometric invariant properties. For a given matrix, unique singular values [10] are obtained which is useful in firm representation for image blocks. Most important data of image block is possessed by largest singular values (LSV) while small singular values are sensitive towards noise. LSV possess effective stability even when an image endure minor distortions. LBP [11] operator facilitates integrated description for a texture patch with structural and statistical features. LBP is highly discriminative texture operator [12] which analyze numerous patterns corresponding to each pixel with its neighborhood. The signed difference between center and neighboring pixel value is invariant to variation in mean luminance as well as change in grayscale.

This paper is structured as follows. Section 2 illustrates the proposed algorithm. Section 3 shows the experimental outcome validating the legitimacy of the proposed approach by utilizing evaluation metrics. Finally, the proposed study is concluded in Sect. 4.

2 Proposed Methodology

Figure 2 presents the fundamental framework followed for copy–move forgery detection. The proposed technique utilize following steps:

2.1 Preprocessing

  1. 1.

    Conversion of color channel:

    Initially, if the input image is RGB then it is converted to grayscale image.

  2. 2.

    Division of image in overlapping blocks:

    After preprocessing step, image is parted into partially overlapping blocks. If the gray scale image is of dimension \(M \times N\) and block size used for division of image is \(B \times B\) then total number of blocks are \((M-B+1) \times (N-B+1)\).

Fig. 2
figure 2

Framework of the proposed methodology

2.2 Feature Extraction

  1. 1.

    Computing gray-level co-occurrence matrix:

    For each block, GLCM [8] is obtained to represent spatial relationship between the gray tones within each image block. The size of GLCM depends on maximum variation in gray level values of pixel within a block. For a block of dimension \(B \times B\), co-occurrence matrix of size \(B \times B\) is incurred.

  2. 2.

    Calculation of singular values:

    SVD [9] is applied over co-occurrence matrix. As result of decomposition, GLCM decomposed in three components: left unitary, right unitary, and diagonal matrix. Diagonal matrix contains singular values [10] on diagonal positions. Singular values represent feature vectors for all blocks of image. For a \(B \times B\) dimensional matrix feature vector obtained is of length B.

  3. 3.

    Extraction of LBP features and fusion with singular values:

    Blocks are processed using LBP [11] as shown in Fig. 3. LBP features [12] are stored for each block. Singular values prevailed using SVD, and LBP features combinedly formulate feature vector to represent each block of image.

Fig. 3
figure 3

Working rationale of local binary pattern operator

2.3 Comparison of Feature Vectors

Feature vectors for overlapping blocks are placed in matrix FM. The dimension of feature vector matrix FM is \((M-B+1)(N-B+1)\times len \), where ‘len’ represents length of feature vector. For robust localization of similar feature vectors of duplicated blocks, Lexicographical sorting is applied over feature matrix. As a resultant, we achieve similar feature vector settled at neighboring locations.

$$FM=\left\{ \begin{array}{c} FV_{1} \\ FV_{2} \\ FV_{3} \\ \vdots \\ FV_{(M-B+1)(N-B+1)} \end{array} \right\} $$

2.4 Similarity Detection and False Match Removal

  1. 1.

    Duplicate region detection:

    Euclidean distance is measured to find similarity between feature vectors as indicated in Eq. 2.

    $$\begin{aligned} D(F,F')=(\sum \limits _{i=1}^{n_f} (F_i-{F_i}')^2)^\frac{1}{2} \end{aligned}$$
    (2)

    where F and F\('\) shows feature vectors obtained from blocks and \({n_f}\) represent the length of feature vector. To detect similar feature vectors from sorted feature matrix threshold value is used, such that \( D(F, {F}^{'})\le {T_{dis}}\). To detect similar regions of image, shift vectors between similar blocks are measured. The top left corner location of a block is deliberated as its emplacement within image. Let, \((i_{1},j_{1})\) and \((i_{2},j_{2})\) be the coordinates of blocks \(\bar{FM}\). Shift vector between two co-ordinates can be obtained as in Eq. 3.

    $$\begin{aligned} S=(s_{1}, s_{2})=(i_{1}-i_{2},j_{1}-j_{2}) \end{aligned}$$
    (3)

    Both Shift vectors \(-S\) and S symbolize same order of magnitude for shifting. So, displacement vectors are normalized by multiplying with \(-1\) to maintain \(S\ge 0\). A counter value C is initialized with zero. For similar shift vector between neighboring blocks, counter value increased by 1 as in Eq. 4.

    $$\begin{aligned} C(s_{1},s_{2})=C(s_{1},s_{2})+1 \end{aligned}$$
    (4)
  2. 2.

    Removal of false matches:

    Counter value corresponding to different shift vector represents the frequency of similar shifting between block pairs. The proposed method detects all normalized shift vectors \(S1, S2, S3,\ldots , Sr\). Shift vectors whose occurrence is higher than user-defined threshold \(T_{sh}\) shows forged blocks. \(C(S_{k})>T_{sh}\) for \(k= 1, 2,\ldots , R\). High value of \(T_{sh}\) may leave altered regions undetected, while low value of \(T_{sh}\) can give rise to false matches. We employ a color map for localizing forged blocks of image.

3 Experimental Results and Discussion

3.1 Dataset

The source images for our experiments accumulated from CoMoFoD dataset [13]. Experiments are carried out on images with dimension \(512 \times 512\). We selected 60 images suffering from color reduction, 60 images bearing contrast adjustment, 60 images with blurring attack, and 60 images with brightness change attack. In CoMoFoD dataset, color reduction images are divided into three categories with color channels contracted from 256 to 128, 64, and 32. Images with contrast adjustment also have three ranges [0.01, 0.95], [0.01, 0.9], and [0.01, 0.8]. Image with blurring attacks also have three categories based on average filter size as \(3 \times 3\), \(5 \times 5\), \(7\times 7\), and images with illumination change attack classified in three categories [0.01, 0.95], [0.01, 0.9], and [0.01, 0.8].

3.2 Experimental Setup

All experiments are executed over MATLAB 2016a, installed upon a platform equipped with 64-bit Windows (8 GB RAM) and Intel core i7 processor. We set every parameter as \(B=8\), \(t_{dis}=50\), and \(t_{sh}=50\). Various size of forged images can be used as input to proposed method, so values of M and N may vary. As here we consider all input images with size \(512 \times 512\) so, \(M=512\) and \(N=512\). From experimental results, we found that when \(B=4\), too many false matches appear. when B increased to 16 then detection accuracy is compromised. \(T_{sh}\) is used for locating groups of blocks with similar shifting and meaningful forged region detection. \(t_{dis}\) is used for spotting similar feature vectors. Higher value of \(t_{dis}\) results in detection of different feature vectors as similar whereas smaller values of \(t_{dis}\) perform rigorous detection of similar feature vectors which may stipulate slightly dissident feature vectors as different. Here, length of extracted feature vector ‘len’ is 67.

3.3 Performance Evaluation

To demonstrate the outcomes of proposed technique, we have computed DA, FPR, and F-Measure by comparing the forgery detection outcomes with ground truth. DA, FPR, and F-Measure can be calculated as in Eqs. 5, 6, and 7 respectively.

$$\begin{aligned} DA=\frac{TP+TN}{TP+FP+TN+FN} \end{aligned}$$
(5)
$$\begin{aligned} FPR=\frac{FP}{TN+FP} \end{aligned}$$
(6)
$$\begin{aligned} F{\text {-}}Measure=\frac{2\times TP}{2\times TP+FP+FN} \end{aligned}$$
(7)

True Positive (TP) symbolizes the number of pixels correctly perceived as spoofed. False Positive (FP) represents the number of pixels which are falsely identified as faked but primitively they are not forged. False Negative (FN) shows the number of pixels which are manipulated but not detected. True Negative (TN) indicates number of pixels correctly detected as not forged. DA symbolizes the measure to which forgery detection algorithm can detect the altered region of image correctly, whereas FPR represents the percentage of pixels falsely detected as altered. F-Measure expresses the alliance of number of accurately detected forged pixels with respect to number of forged pixels perceived, and relevant forged region detection (i.e., proportion of correct detection of tampered pixels with respect to number of altered pixels acquainted by ground truth of forged image.).

3.3.1 Contrast Adjustment and Brightness Change Attacks

Table 1 illustrates average quantitative results obtained through the proposed method for forgery detection when images are enduring contrast adjustment and brightness change alterations. Qualitative outcomes incurred applying the proposed method are pictured in Fig. 4.

Table 1 Average DA and FPR results for contrast adjustment and brightness change attack
Fig. 4
figure 4

a, f shows untampered image. b, g shows corresponding forged images. c, d, and e shows results for forgery detection in case of contrast modification category range as \([0.01{-}0.8]\), \([0.01{-}0.9]\), and \([0.01{-}0.95]\) respectively. h, i, and j shows results for forgery detection in case of tampered images with luminance change range as \([0.01{-}0.8]\), \([0.01{-}0.9]\), and \([0.01{-}0.95]\) respectively

3.3.2 Image Blurring and Color Reduction Attacks

Table 2 shows average quantitative results obtained using the proposed method for forgery detection when images are suffering from image blurring and color reduction attack. Qualitative results incurred using proposed method are shown in Fig. 5.

Table 2 Average DA and FPR results for image blurring and color reduction attack
Fig. 5
figure 5

a, f displays original image. b, g represents corresponding forged images. c, d, and e show results for forgery detection in case of image blurring with filter size \(3\times 3\), \(5\times 5\), and \(7\times 7\), respectively. h, i, and j shows results for forgery detection in case of color reduction with level 32, 64, and 128 respectively

Fig. 6
figure 6

a, b, c, and d portray comparative results for F-measure when image endures contrast adjustment, image blurring, brightness change, and color reduction attacks, respectively

Fig. 7
figure 7

a, c, e, and g show comparative results for DA, whereas b, d, f, and h represent comparative outcomes for FPR when forged image bears contrast adjustment, blurring, luminance change, and color diminution attacks, respectively

3.4 Comparative Analysis

Experimental outcomes establish the fact that our method is an efficacious technique for copy–move forgery detection. In comparison to other features based on DCT [3], SVD [11], HOG [6], and discrete wavelet transform (DWT) [14], our method achieved highest DA and FPR rates for forgery detection, when altered image is suffering from contrast adjustment, image blurring, brightness change, and color reduction attacks. For color reduction attack, HoG-based features obtained highest F-measure for all color reduction levels. SVD features achieved better results for F-Measure than proposed method toward color reduction levels 32 and 64. Figure 6 displays comparative results obtained for F-measure when image endures post-processing attacks, whereas Fig. 7 shows comparative results for DA and FPR.

The complexity of copy–move forgery detection algorithms primarily depends on number of blocks accessed and length of feature vector generated, which are represented by \(n_b\) and \(n_f\) respectively. For input image size \(512 \times 512\), the proposed algorithm obtains \(n_f=67\) and \(n_b=255{,}055\). DCT-based feature extraction method possess \(n_f=64\) and \(n_b= 255{,}055\). For SVD-based method \(n_f=8\) and \(n_b= 255{,}055\). Forgery detection scheme using HoG features obtains \(n_f=4\) and \(n_b= 247{,}009\). DWT-based method acquires \(n_b= 62{,}201\) and \(n_f= 64\). For recording forgery detection time, we performed forgery detection operation over image size \(256 \times 256\). The detection time for proposed algorithm is 98.65 s. For HoG, DWT, DCT, and SVD based methods detection times are 17.55 s, 1.69 s, 46.91 s, and 10.99 s respectively. Due to fusion of features obtained through singular value decomposition over GLCM and LBP, the proposed algorithm takes high computational time as compared to other cited methods, but facilitates high detection accuracy with low rate of occurrence of false matches.

4 Conclusion

Images have their applications in various fields like criminal investigation, courts of law, medical imaging, document analysis, etc. On account of speedy growth of technology, lot of potent computer applications are available which have made forging process easier. Forgery over images are practiced for denigration, blackmailing, harassment, political disputation, fun-making, etc. Digital images are not adequate in courts of law for witness, without forensic investigation over evidences. Such concerns related to multimedia security and forensic probes resulted in evolution of various advancements in image tampering detection techniques. As a little contribution in field of image forgery detection, our method obtained substantially high DA and low FPR when forged image suffer from blurring, color reduction, contrast adjustment, and brightness change attack. As future work, we will search for methods invariant to geometrical attacks with greater robustness than state-of-the-art methods of copy–move image forgery detection.