1 Introduction

The growth of advanced image editing tools forced us to think about sophisticated authentication mechanisms [22]. Distinguishing original images from fake ones and localizing the tampered area is a challenging issue for industries and academia. Recently, an image hashing-based approach has been widely used for image authentication as well as tampered area localization. In this technique, an image is represented by an image hash or a digital signature, which is a visual representation of image contents. An image hash should be robust to content-preserving operations (CPOs) such as compression, rotation, translation, etc., and reactive to change in content manipulations [31]. The image hashing approach has been used in different applications like image authentication [24, 38], image retrieval [4, 37], tampering detection [3, 12,13,14, 25], and some other applications [16, 17]. An image hashing technique was first introduced by Venkatesan et al. [33]. In this method, wavelet coefficients are extracted to form an image signature, which is robust in compression, geometric distortions, but sensitive to some CPOs. To authenticate and localize the area of tampering, Roy and Sun [27] first introduced a block-based image hashing approach. This method is robust to compression, rotation, but translation has not been explored. Ahmed et al. [2] considered the LL band of DWT from the 16 × 16 sizes of image blocks to construct hash. This method can detect tampering and robust against filtering operation as well as compression, but the hash length is very high. To achieve better authentication and tampering localization approach, Pun et al. [25] combined global and local features to construct a hash. This method is robust to compression, filtering, and rotation up to 5 degrees, but limited in the case of combined RST transformation.

There are some image alignment techniques [3, 20, 21, 35] to eliminate the impact of geometric transformations, which have been employed to authenticate images and to localize tampering. These image alignment methods require too long hash to reconstruct the original image from the geometric translated one. Lu et al. [21] introduced the theory of scale space and the Radon transform to obtain the parameters of geometric distortion and reconstructed the image. In another work, Lu and Wu [20] improved [21] using SIFT features. Battiato et al. [3] presented a new approach using a voting procedure in the space model. The performance of the geometric correction in [3] is improved compared to [20, 21], but the major limitation is too long hash length, i.e., 1000 digits. Yan et al. [35] represented an image alignment approach based on a quaternion Fourier-Mellin transform. In this method, the estimation of geometric parameters has been affected by tampering operations. In another study, Karsh et al. [14] introduced image alignment based on a furthest non-zero pixel, but this method is limited only to the positive angle of rotations.

Yan et al. [36] represented an image hashing based on multi-scale, where an image is divided into multiple rings. This method is robust to most CPOs, but sensitive to composite RST attacks. In another work, Yan et al. [34] presented tamper detection based on multi-scale difference map fusion. This method is robust against some CPOs, but sensitive to translation. The hashing techniques [8, 9, 11, 23, 26, 28,29,30] are applied for image authentication, but fail to locate tampering, if composite RST and tampering occur simultaneously.

It can be observed from the literature that most of the existing image authentication methods are sensitive to composite RST transformation. Also, the tampered area may not be localized, if tampering and composite RST transformation occur simultaneously. Based on the limitation of the existing methods, the main contribution of the proposed work can be encapsulated as follows.

The main contributions of the work are as follows.

  • It has been observed from the literature that most of the existing image authentication methods are sensitive to composite RST transformation. Also, the tampered area may not be localized, if tampering and composite RST transformation occur simultaneously. Hence, a combination of LWT and a modified image compression approach based on DCT has been proposed to construct a short image hash.

  • The image map has been constructed using the proposed modified image de-compression via inverse DCT of hash. The difference of image map from the received hash and received image yields tamper localization. As per our literature survey, this is the first-time hashing based on LWT and modified image compression based on DCT, used for image authentication and tampering area localization, in the proposed work.

  • Besides, a modified blind geometric distortion correction approach based on inherent geometric characteristics has been proposed. Due to this, the proposed method is robust to composite RST transformation. Also, the proposed system obtained the tampering location, where tampering and geometric transformation occur simultaneously, which is a major limitation in the state-of-the-art methods.

The arrangement of the paper is as follows. Section 2 represents a detailed description of the proposed image hashing methodology. Section 3 demonstrates the proposed blind geometric distortion correction approach, image authentication, and tampering localization. The experimental results and analysis have been discussed in Section 4. Finally, the conclusions and future scopes are mentioned in Section 5.

2 Proposed image hashing

The proposed image hashing method is shown in Fig. 1. Brief descriptions of each step of the proposed image hashing are drawn in the subsequent subsections.

2.1 Preprocessing

An arbitrary size of the source image is resized into \( p\times p \) using bilinear interpolation. Here, image resizing is necessary to keep the hash length fixed and maintain robustness against scaling operation. If the source image is an RGB image, it is mapped into CIE \({\mathbf{L}}^{\mathbf{*}}{\mathbf{a}}^{\mathbf{*}}{\mathbf{b}}^{\mathbf{*}}\)color space [6], and the intensity component (\({\mathbf{L}}^{\text{*}}\)) has been considered for further processing. The reason behind choosing the CIE \({\mathbf{L}}^{\mathbf{*}}{\mathbf{a}}^{\mathbf{*}}{\mathbf{b}}^{\mathbf{*}}\)color space is that it is perceptually uniform and reasonably related to human perception space [31]. Hence, features extracted from the \({\mathbf{L}}^{\text{*}}\) component are stable compared to other color spaces.

Fig. 1
figure 1

Depicts a procedure to generate a hash code from an image

2.2 LWT

The pre-processed image is decomposed using LWT up to the third level [7], as shown in Fig. 2. By experiment, to maintain trade-off between the length of hash and discrimination performance, the LL3 sub-band, let it \(\mathbf{F}\) a square matrix sized \(q\times q\), has been selected for further processing, which approximates the pre-processed image. The reason to use LWT after pre-processing is to reduce the dimension, while keeping most information intact. Furthermore, DCT is applied on \(\mathbf{F}\) to get the compressed hash, discussed in the following subsection.

Fig. 2
figure 2

Sketch map, LWT decomposition of an image. H and L indicate high and low, respectively. Number represents level

2.3 DCT (Discrete Cosine Transform)

The mathematical expressions to obtain DCT of \(\mathbf{F}\) is discussed as follows. Firstly, \(\mathbf{F}\) is divided into non-overlapping blocks of sized \(m\times m\). For simplicity, let \(q\) be an integral multiple of \(m\). Hence, the total number of blocks is \(\delta ={(q/m)}^{2}\). Let \({\mathbf{M}}_{k}\left(x,y\right); 0\le x,y\le m-1,\) be the \(k\)-th block indexed from left to right and top to bottom (\(1\le k\le \delta\)). Then, 2-D discrete cosine transform of \({\mathbf{M}}_{k}\left(x,y\right)\) has been obtained as follows:

$${\mathbf{B}}_{k}(u,v)={\mathbf{T}}(u,v)\times {\mathbf{M}}_{k}(x,y)\times {\mathbf{T}}^{\prime}(u,v)$$
(1)

Here, three matrices \(\mathbf{T}\), \(\mathbf{M},\) and \({\mathbf{T}}^{\prime}\) are multiplied (\(1\le k\le \delta\)), where \({\mathbf{T}}\left(u,v\right)\) is a discrete cosine transform matrix [1, 18], shown in Appendix 1. \({\mathbf{T}}^{{\prime }}(u,v)\) represents a transpose of \({\mathbf{T}}(u,v)\). \({\mathbf{B}}_{k}(u,v)\) is the DCT of \({k}^{th}\) (\(1\le k\le \delta\)) non-overlapping pixel blocks. The reason to use DCT after LWT is its energy compaction property. Most of the contents of an image may be represented using only a few low-frequency components in the transform domain (in this paper, \(n\) represents the number of low-frequency DCT coefficients from each \(k\)-th block). From these low-frequency coefficients, using inverse DCT (IDCT), the approximate image contents may be reconstructed (in this manuscript is known as an image map, which has been used for tampered area localization, discussed in sub-Section 3.3). The \(n\) low-frequency DCT coefficients from each \(k\)-th block are selected and concatenated to generate a short image hash, discussed in the following subsection.

2.4 Generation of an image hash

Let \(n\) low-frequency DCT coefficients are selected from each \({k}^{th}\) (\(1\le k\le \delta\)) matrices, \({\mathbf{B}}_{k}(u,v)\), using zigzag ordering, as shown in Appendix 2. It has been observed that low-frequency components are represented in the front side of the zigzag sequence, while high-frequency coefficients are in the later parts. Here, we select the first \(n\) coefficients in the zigzag sequence to form a vector \({\mathbf{h}}^{k}\) sized \(1\times n\) Now, concatenate \(n\) low-frequency DCT coefficients from \({k}^{th}\)blocks yields image hash, \(\mathbf{h}\), of \(n\times \delta =r\) digits as follows.

$${\mathbf{h}}=[{h}^{1}\left(1\right),{h}^{1}\left(2\right),\dots ,{h}^{1}\left(n\right), {h}^{2}\left(1\right), {h}^{2}\left(2\right),\dots ,{h}^{2}\left(n\right),\dots ,{h}^{\delta }\left(1\right),{h}^{\delta }\left(2\right),\dots ,{h}^{\delta }\left(n\right)]$$
(2)

The length of the final image hash is \(r\) digits. Also, the details of the hash generation, in implementation form, are shown in algorithm 1.

figure b

2.5 Metric of performance

The metric of performance comparison is the L2 norm. Let, \(\mathbf{h}\) and \({\mathbf{h}}^{\prime}\) are the hash for transmitted and received images, respectively. The L2 norm (or Hash Distance) is given by:

$$Hash \ Distance\, \left(d\right)=\sqrt{{\sum} _{i=1}^{r}{\left|h\left(i\right)-{h}^{\prime}\left(i\right)\right|}^{2}}$$
(3)

where, \(h\left(i\right)\) and \({h}^{\prime}\left(i\right)\) show \({i}^{th}\)elements of \(\mathbf{h}\) and \({\mathbf{h}}^{\prime}\), respectively. When the hash distance is less than a threshold \(\tau 1\) (\(d< \tau 1\)), the image pair is considered to be “similar image pairs”, if \(\tau 1< d < \tau 2\) then “tampered image pairs”, otherwise “different content image pairs”. FPR (False Positive Rate) and TPR (True Positive Rate) are two other performance comparison metrics, based on hash distances, discussed as follows.

$$FPR = {\zeta }_{1}/{\eta }_{1}$$
(4)
$$TPR = {\zeta }_{2}/{\eta }_{2}$$
(5)

where ζ1 reveals the total number of the different content image pairs considered similar ones, and ζ2 reveals the total number of similar content image pairs considered similar ones. η1 and η2 represent visually different and similar content image pairs, respectively. For a better image authentication approach, TPR and FPR should be high and low, respectively. The receiver operating characteristics (ROC) is drowned using TPR in the ordinate and FPR in abscissa, which is used to compare the overall performance of different image authentication methods.

3 Proposed image authentication system

The proposed image authentication system is shown in Fig. 3. The process of image authentication is discussed in detail in the following subsections.

3.1 Blind geometric correction approach

Let us have the received image and hash. For image authentication and tampered area localization, the effect of the geometric transformations is eliminated in the received image. In the literature, methods [3, 14, 20, 21, 35] are geometric correction approaches. The methods [3, 20, 21, 35] are affected by tampering as well as the length of a hash is too large. In [14], it needs additional information from the transmitter side. In this paper, we have proposed a blind geometric correction method that does not require any information from the transmitter side, which discussed as follows.

Let a received image is processed through combined geometric correction such as RST, shown in Fig. 4a. First, the rotation angle is estimated as follows.

Fig. 3
figure 3

Proposed image authentication system

$$\begin{array}{*{20}c}\theta =\text{arctan}(\Delta Y/\Delta X),\ {\theta }^{\prime}=\text{arctan}(\Delta {Y}^{\prime}/\Delta {X}^{\prime})\end{array}$$
(6)

where, \(\Delta Y={Y}_{b}-{Y}_{r}\), \(\Delta X={X}_{r}-{X}_{b}\), \({\Delta Y}^{\prime}={Y}_{l}-{Y}_{t}\), and \({\Delta X}^{\prime}={X}_{t}-{X}_{l}\). \(({X}_{r},{Y}_{r})\), \(({X}_{l}, {Y}_{l})\), \(({X}_{t}, {Y}_{t})\), and \(({X}_{b},{Y}_{b})\) are indexes of non-zero-pixel values of the rightmost, leftmost, top, and bottom points, respectively. If \(\theta \cong {\theta }^{\prime}\) and \({X}_{t}\)>\({X}_{b}\), then the image is considered to be rotated anticlockwise (i.e., angle of rotation is positive, \({{\theta }_{p}=\theta }^{\prime}\)), as shown in Fig. 4a, otherwise clockwise (i.e., angle of rotation is negative, \({{\theta }_{n}=\theta }^{\prime}\)), as shown in Fig. 4b. Then, the rotated image is anti-rotated to either \({\theta }_{p}\) or \({\theta }_{n}\) to restore the image, shown in Fig. 4c. Finally, the region of interest is extracted, as shown in Fig. 4d. The proposed geometric correction is limited in the range (\(-45^\circ+45^\circ\)). The Pseudocode for the blind geometric transformation correction algorithm is shown in algorithm 2.

figure c
Fig. 4
figure 4

Correction in geometric transformation (a) Composite RST, rotated in anticlockwise direction, [250, 0.5, [100 100] (b) Composite RST, rotated in clockwise direction, [-250, 0.5, [100 100]) (c) Anti-rotated image (d) Restored image

3.2 Image authentication

After eliminating the geometric transformation effect in the received image (if any), the image hash, \({\mathbf{h}}^{\prime}\) is generated as discussed in Section 2. Now, estimate the hash distance (\(d\)) between \(\mathbf{h}\) and \({\mathbf{h}}^{\prime}\). If\(d\) lies between threshold \({\uptau }1\) and \({\uptau }2\) (both the thresholds are selected based on an experiment in Section 4), then the received image has been manipulated by fraudulent during the transmission, i.e., tampered version of transmitted one. Else if \(d\) is less than \({\uptau }1\), the received image is considered a similar one, otherwise a different one. From the tampered image, the tampering area has been localized, as discussed in the following subsection.

3.3 Tampering localization

The received image has been passed through blind geometric correction. The image hash \({\mathbf{h}}^{\prime}\) has also been obtained, as shown in Fig. 5. Now, the image maps are generated for both the received hash, \(\mathbf{h}\), and generated hash, \({\mathbf{h}}^{\prime}\), as follows.

First, dis-concatenate \({\mathbf{h}}^{\prime}\) into \(k\)-th (\(1\le k\le \delta\)) number of arrays, i.e., \({z}^{k}\left(i\right)\) where \(i=\text{1,2},\dots ,n\), each consisting \(n\) low-frequency DCT coefficients. Next, padding 49 zeroes on \({z}^{k}\left(i\right)\) yields \({c}^{k}\left(j\right)\) where (\(j=\text{1,2},\dots ,n+49\)). Apply inverse zigzag coding (shown in Appendices 2) on \({c}_{k}\), which generates \(m\times m\) size \(k\)-th \({B}_{k}^{\prime}(u,v)\) sub-matrices. This process extracts the original sequence of \(n\) high energy frequency coefficients for each \(k\)-th sub-matrices. After that, find the inverse discrete cosine transform (IDCT) of \(k\)-th sub-matrices \({\mathbf{B}}_{k}^{\prime}(u,v)\) as follows

$$\begin{array}{*{20}c}{f}_{k}^{\prime}(x,y)={\mathbf{T}}^{\prime}(u,v)\times {\mathbf{B}}_{k}^{\prime}(u,v) \times {\mathbf{T}}(u,v); \ 1\le k\le \delta\end{array}$$
(7)

where \(\mathbf{T}\) is DCT matrix as discussed in Appendix 1 and \({\mathbf{T}}^{\prime}\) is the transpose of \(\mathbf{T}\). Rearrange \(k\)-th sub-matrices \({f}_{k}^{\prime}(x,y)\) from left to right, top to bottom non-overlapping pixel blocks yields an image map \({f}^{\prime}(x,y)\) of size\(p/m\times p/m\). Similarly, find an image map \({f}^{\prime\prime}(x,y)\) from the received image hash, \(\mathbf{h}\). The two image maps are subtracted and normalized. Next, it has been converted into a binary image, multiplied with the restored received image to detect the tampered regions. The details in the implementation form are shown in algorithm 3.

figure d

4 Analysis of experimental results

In this section, the proposed method has been evaluated using a large number of image pairs, and the selected optimal value of the model parameters are as follows: \(p\times p=128\times 128\), \(m\times m=8\times 8\), \(\delta =4\), \(n=15\), \(r=60\), \(\tau 1=0.88\), and \(\tau 1=27\). The proposed model has been analyzed in three categories: the performance of blind geometric transformation correction, the performance of hashing technique for robustness and discriminative capability, and detection of tampered images and their localization, shown in Sections 4.1, 4.2, and 4.3, respectively.

Fig. 5
figure 5

Flow chart of the proposed tampering localization method

4.1 Performance of blind geometric transformation correction

To examine the efficacy of the proposed blind geometric transformation correction method, we have selected 200 similar images and 400 tampered images from Ground Truth Database [10] and CASIA V2.0 database [5], respectively. These selected images are gone through the geometric transformation (i.e., composite RST) with the parameters shown in Table 1. The geometric transformation parameters, i.e., rotation angles for clockwise and anticlockwise, are estimated using the proposed method. The estimation error (mean and standard deviation of the error), is shown in Tables 2 and 3. It can be observed that for both similar and tampered images, the mean and standard deviation of estimation error is too low. It can also be seen that the mean and standard deviation estimation error is invariant to rotation angles. In the case of tampered images, the estimation error is highly affected in the existing methods [3, 20, 21, 35]. The geometric distortions are eliminated due to the very low estimation error, as discussed in Section 3.1. However, the proposed blind geometric transformation correction approach is limited for the rotation angle from − 45 to + 45 degrees.

Table 1 Content preserving operations specifications
Table 2 Performance of the proposed blind geometric transformation correction, the error of the rotation angle estimation in case of clockwise rotation
Table 3 Performance of the proposed blind geometric transformation correction, the error of the rotation angle estimation in case of anticlockwise rotation

4.2 Performance of hashing for robustness and discriminative capability

In this section, the proposed model is used to segregate the received image as “perceptually similar image pairs”, “tampered image pairs”, or “different image pairs”. The experiment has been carried out on 3,948 “perceptually similar image pairs”, generated as shown in Table 1. Here, 42 source images are selected from the USC-SIPI database [32], such as 37 and 5 from the “Aerial” and “Miscellaneous” categories, respectively, the size varies from 512 x 512 to 2250 x 2250. Also, 19,900 “different image pairs” are generated from different combinations of 200 source images. Where, the source images are as follows: 75 from Ground Truth Database [10] of size 756 x 504, 75 using Nikon D3200, of size from 3008 x 2000 to 4512 x 3000; and 50 from the Internet, of size from 256 x 256 to 1024 x 768. 400 “tampered image pairs” are selected from the CASIA V2.0 database [5], where size varies from 240 x 160 to 900 x 600.

The hash distances are calculated for the above three categories of image pairs and estimated the true positive rate via varying the threshold, drawn in Fig. 6. Figure 6a shows the true positive rate for “perceptually similar image pairs” and “tampered image pairs”. Here, the overlapping curve shown in the red rectangle in the right upper corner is expanded and presented in the right bottom corner to view the details. It can be observed that an optimal threshold \(\tau 1=0.88\) may be selected to segregate between “perceptually similar image pairs” and “tampered image pairs”. Similarly, Fig. 6b shows the true positive rate for “tampered image pairs” and “different image pairs”. The overlapping in the center portion is enlarged and presented at the bottom to view details. It can be seen that \(\tau 2=27\) may be selected to differentiate between “tampered image pairs” and “different image pairs”. Both thresholds are determined based on a trade-off between robustness and discriminations. It can be observed from Fig. 6 that the proposed method has better robustness (TPR is high at the optimal threshold), as well as most of the “tampered image pairs” and “different image pairs”, are truly detected.

Fig. 6
figure 6

True positive rate (performance of image authentication) with the varying threshold for (a) perceptually similar and tampered image pairs and (b) tampered and different image pairs

4.3 Detection of tampered images and its localization

This section discussed the experiment on 400 tampered image pairs taken from the CASIA V2.0 database [5], which varies from 240 x 160 to 900 x 600. The hash distances of tampered image pairs are shown in Fig. 7, where the red and green line shows \(\tau 1=0.88\) and \(\tau 2=27\), respectively. It can be seen that most of the tampered images are detected by the proposed method. In case the tampering area is large, the hash distances are larger than \(\tau 2=27\), shown above the green line. The large area tampering may be considered different image pairs. For the detected tampered images, the tampering area can be localized using the proposed method, as shown in Table 4. Here, the first and second row shows original and tampered pairs, respectively. The tampering is localized in the third row. It can be observed that the tampered areas are indeed detected in the second and third columns, respectively. But, in the fourth column, some small regions which are not tampered, along with tampered ones, are detected as tampered areas, which is a limitation of the proposed method. Whereas, if some portions of the tampered regions are similar to the original image, that portion has not been detected, as shown in the fifth and sixth, respectively. However, most of the tampered objects are detected using the proposed method. Due to the page limitation, few samples are presented. Besides, the proposed method can detect tampering, even if the composite RST and tampering occur simultaneously, as shown in Table 5. Here, the first column shows the original image, and the second one reflects tampering and the composite RST transformation. In the third column, the tampered regions (i.e., some objects) are identified using the proposed method.

Fig. 7
figure 7

Hash distance distribution of tampered image pairs, where red and green line shows \(\tau 1=0.88\) and \(\tau 2=27\), respectively

Table 4 Localization of tampered regions
Table 5 Localization of tampered regions, even if tampering and composite RST occur simultaneously

5 Performance comparison with state-of-the-art methods

The proposed method is compared with state-of-the-art techniques such as Radon Transform, and Discrete Fourier Transform (RT-DFT) based hashing [15], Binary multi-view based hashing [9], SIFT based hashing [19], Zernike Moments (ZM) based hashing [38], and Ring invariant vector distance (Ring-IVD) based hashing [31]. Tables 6, 7, and 8 demonstrate the benefits and drawbacks of comparative approaches. The values of parameters required for the implementation of the compared methods have been reserved from corresponding papers. But, the input images are resized to 128 × 128, and the threshold values are selected based on our image database, discussed in Section 4. The Euclidean distance that produces TPR and FPR has only been chosen as a performance metric for fair comparison among all the methods. A ROC curve consists of several coordinate points (TPR, FPR), with the x-axis being FPR and the y-axis being TPR. If compared algorithms have the same FPR in the ROC curve, high TPR methods are better than low TPRs. Likewise, if two algorithms with the same TPR, the low FPR approach outperforms the high FPR method. All compared methods are evaluated with the same database, discussed in Section 4.

The proposed method has been compared with existing techniques via four phases: Firstly, the overall robustness and discriminative capability (considering only different image pairs). Secondly, individual robustness against digital operations. Next, sensitiveness towards content-changing operations. Finally, some more performance parameters such as length of hash, time of computation in MATLAB, etc.

The overall robustness and discriminative capability performance are represented using the receiver operating characteristic (ROC) curve, as shown in Fig. 8. Here, the curve near the upper left portion (defined within the red rectangle) has been zoomed and kept in the right lower part to view details. Each compared method estimates TPR and FPR via varying thresholds and generates the ROC curve, as shown in Fig. 8. TPR and FPR indicate the performance of robustness and discriminative capability, respectively. A ROC curve followed towards the upper left corner is a better technique. It can be observed that the proposed image authentication system ROC curve is followed in the leftmost corner, hence better than the compared ones, for robustness and discriminative capability. The methods [9, 15, 31] successively followed the proposed method.

Fig. 8
figure 8

Performance comparison of the proposed technique with some existing techniques

The performances of compared methods against digital operations are shown in Table 6. A better approach should have high TPR (i.e., close to one) and low FPR (i.e., close to zero). It can be observed that the performance of the proposed method for robustness is better than the existing methods, particularly in the case of geometric transformations such as rotation, translation, and composite RST, etc. The techniques [31, 38] are robust against many digital manipulations, but sensitive to translation and composite RST. Whereas the method [15] robustness is better in translation, but susceptible to scaling and compression. The robustness of the method [9] against geometric operations followed [15].

Table 6 Comparison of robustness against digital operations

The sensitivity of the compared methods towards content-changing operations at an optimal threshold is shown in Table 7. It can be observed that the proposed method sensitiveness is higher (false acceptance is lower, i.e., FPR = 0.0843) than that of compared methods. The methods [9, 15, 31, 38] successively followed the proposed method. It has been experimentally observed that in the case tampered color is similar to the original one; then the proposed method may not detect that part. Hence, there is a small misclassification. However, FPR is the lowest among the compared methods due to better image map construction.

Table 7 Performance comparison of image hashing methods for content-changing operations

The performance compared with some more parameters is shown in Table 8. It can be observed that the proposed method TPR = 0.9901, which is higher than the compared ones. Whereas the FPR = 2.512×10− 4 is the lowest among the compared methods. Hence, the proposed method has a better trade-off between robustness and discrimination. The robustness against an arbitrary rotation and composite RST is a major finding, severely limiting the state-of-the-art methods. The robustness against these geometric operations has been achieved due to the proposed blind geometric correction approach. The method [31] is robust to an arbitrary rotation, but there is much information loss. All compared methods are implemented using a desktop computer with an Intel i7 processor of 8 GB RAM having a windows 8 operating system using MATLAB 2015a. The image hash has been generated for 200 images and considers the average time. It can be seen that computational cost is the lowest among compared methods. But, the hash length is slightly larger than some of the techniques. However, the proposed method can locate the tampering region, even if tampering and composite RST occur simultaneously, which is the main focus of the proposed method.

Table 8 Performance comparison for discrimination with different existing techniques

6 Conclusion and future works

In this work, a blind geometric transformation correction has been proposed. An image hash has been generated based on LWT and DCT. The proposed image hashing technique is applied for content authentication and tampered area localization. An image map has been generated from the short hash. Based on the differences in image maps, the tampered areas have been localized. The main focus of this work is to keep the hash length short, maintain robustness against digital operations, and localize the area of tampering, even if the tampering and composite RST occur simultaneously. The experiment has been carried out on an extensive database, and the results demonstrated the effectiveness of the proposed method against digital operations. Besides, good discriminative capability and localize tampered regions, even if tampering and composite RST occur simultaneously. ROC curve shows that the proposed trade-off between robustness and discriminative capacity is better than some state-of-the-art techniques.

In future work, the accuracy of tampering area localization may be improved. The proposed method may be extended for video hashing.