1 Introduction

The practice of sharing information secretly has been going on for ages. One of the earliest methods was to etch the message on a messenger’s head, wait for the hair to grow again, and then send him to the intended recipient. The recipient had to shave off the messenger’s head to read the secret message. This was very tedious and time-taking process and could not be made use of much successfully during wars, except for few instances. A better approach was to use wooden tablets covered in wax. The technique was to remove the wax, write the message on the wood, and re-cover it with wax. Advancement led to development of better techniques such as the invisible ink and null ciphers, but the latter mostly raised suspicion as they sounded strange. Hence, almost all the methods used for secrecy were either detectable or at least, suspicious [16].

With the arrival of computers, cryptography took over all the conventional secret communication methods. It aimed at encrypting–replacing each letter with a combination of symbols–the data that is to be kept safe, thus making it incomprehensible for an adversary. However, encryption was not concerned with hiding the existence of secret communication. This approach leads to suspicion which in turn might make data transmission unsuccessful.

Information hiding techniques resolved all the above mentioned issues and are categorized in two branches: steganography and watermarking. Attributes of the information hiding techniques include (i) imperceptibility, (ii) survivability, (iii) capacity, and (iv) security. Steganography, in general, is the process of hiding information in the cover medium such that it is undetectable, which makes it a prime candidate for hidden communication. Image steganography is carried out by manipulating and altering image pixels to hide the data [6, 30]. In addition to spatial domain, image steganography is also carried out in several transform domains [19].

A generic description of the steganographic process is as follows:

$$\label{eq1} \textit{cover data} + \textit{secret data} \xrightarrow{\text{embedding algorithm}} \textit{stego data } $$
(1)

Nowadays, steganography is often used in conjunction with cryptography for dual protection of data. As the encrypted information is hidden using steganography, the adversary has to first find the hidden information–which itself is a difficult task–and then decrypt it to find the secret data.

With the rapid expansion in information technology, problems such as piracy, counterfeit and violation of copyrights increased dramatically. One of the proposed solutions was to have a digital signature in the data, so that the rightful owner may prove his ownership when required [17]. Watermarking - a technique to embed digital signature in data–served the purpose. However, in case of watermarking, non-visibility was not a necessary criterion for every scenario. High capacity was also not needed as signatures are usually small. The only essential requirement was to embed the signature in such a way that it could be robust against removal or destruction attempts; hence robustness and security were the attributes of prime importance. In general, watermarking secures digital data against copyright violation and counterfeit issues, by embedding a digital signature with minimal distortion in cover data. But, in case of military and medical images, this minute distortion was not permissible.

In case of steganography, survivability has not been an important attribute so far, rather more important is the imperceptibility and capacity with secure delivery. These requirements could be better understood by the visual requirement model given in Fig. 1 [15]. While watermarking algorithms are primarily concerned with the robustness of the embedded watermark, steganography algorithms are concerned with imperceptibility and capacity of the algorithm.

Fig. 1
figure 1

Requirement model

Cao et al. proposed a digital image zero-watermarking method based on Discrete Wavelet Transform (DWT) and chaotic modulation [5]. Zero-watermarking, as the name suggests, made no changes to the cover and solved the problems caused due to distortion. The employed technique was the development of a key, based on a relationship between watermark–digital signature–and cover image. Zero-watermarking is a relatively new field and is being explored to solve problems that were not addressable with former watermarking techniques. Inspired by zero-watermarking, the proposed work presents Zero-steganography–a steganography scheme that is totally imperceptible and hence, undetectable.

The proposed algorithm makes use of logistic map to generate a chaotic matrix. Stego-keys are developed based on certain relationships between cover image, payload and chaotic matrix. Payload is retrieved by means of a relationship between cover image, stego-keys and chaotic matrix. Cover image is not modified during the whole steganographic process, therefore the name Zero-steganography. Hence, instead of directly embedding payload into the cover image that leaves signs of steganography [7, 20, 23], the proposed algorithm performs the steganographic process without any modifications in the cover image.

Rest of the paper is organized as follows: Section 2 gives a brief introduction of chaos and chaotic map used in the proposed algorithm. The proposed algorithm is presented in Section 3. In Section 4, the algorithm is analyzed for survivability, imperceptibility and security. Finally, conclusions are drawn in Section 5.

2 Chaos and one-dimentional chaotic maps

Chaos represents a state of disorder. In scientific sense, there is much more to chaos than just disarray [37]. Instead of random disorder, the chaos that is useful for scientific applications is deterministic. Contrary to the intuition, it means that such kind of chaos can be generated using mathematical equations. However, the initial parameters required to reproduce the same chaotic sequence must be exact. Any slight variation in the initial parameters results in a totally different chaotic sequence.

It is rather difficult to define chaos completely. Williams Garnett defines it as, “Chaos is sustained and orderly-looking long term evolution that satisfies a certain special mathematical criteria and that occurs in deterministic nonlinear systems” [37].

Logistic map is arguably the simplest nonlinear difference equation and appears in many contexts [24]. It was introduced by P.F. Verhulst in order to simulate the growth of population in a closed area [32]. Logistic map belongs to the family of first order difference equations and can be represented mathematically as:

$$\label{lmap} X_{k+1} = \mu X_k (1 - X_k) $$
(2)

where the system parameter μ ∈ [0,4] and initial condition X 0 ∈ (0,1).

The logistic map behaves chaotic with μ ∈ (3.5699456, 4] [27]. The chaotic sequences {X k ; k = 1, 2, 3, ...} generated by this one-dimensional logistic map are unique for every system parameter within this range.

Logistic map is considered as one of the simplest chaotic maps. As it is highly sensitive to system parameter μ, any two chaotic sequences produced using slightly different system parameters are uncorrelated statistically. The proposed algorithm makes use of logistic map to generate a chaotic matrix. Both the system parameter and initial condition are to be known to the intended recipient, in order to regenerate the same chaotic sequence on the receiver side. Early works in digital watermarking which used logistic map are [39, 40].

More complex chaotic maps like the Hénon map [12] and Lorenz attractor [22] can be used in order to directly generate two-dimensional and three-dimensional chaotic sequences respectively. We will now present the proposed algorithm in detail.

3 Proposed algorithm

Given the advantage of steganography over cryptography, and the lack of imperceptible yet robust steganography algorithms in the spatial domain, the proposed algorithm adopts a different approach to achieve steganography. Figure 2 shows an overview of the procedure involved in the proposed algorithm. The technique is to operate characteristics extraction procedure on the cover image. A pseudo random binary key to denote mathematical operations is generated. These mathematical operations are performed between payload and cover image characteristics, and the obtained data is used to construct an intermediate key. Then, the same mathematical operations are performed between intermediate key and chaotic sequence and resultant stego-key is stored as data key. Since the secret information resides in the relationship between cover image and keys, no change is made to the cover image during the whole procedure. Hence, making the proposed algorithm imperceptible and undetectable. We will now describe the procedure for stego-key generation and payload extraction in detail.

Fig. 2
figure 2

Block diagram of the steganographic process

3.1 Stego-key generation

For a secret payload P of dimensions m×n×o and a cover image M with i rows, j columns and k depth, the stego-key generation procedure is as follows:

  1. 1.

    Characteristics of cover image M are extracted and stored as \(\overline{M}\).

  2. 2.

    A binary key K b of dimensions m×n×o–same as that of payload–is generated using a pseudo random process. The key K b denotes mathematical operations that are to be performed between corresponding payload pixel and \(\overline{M}\). The resultant data is stored as intermediate key K i .

    $$\label{eq:Ki} K_i (x,y,z) = \left\{\begin{array}{ll} P(x,y,z) + \overline{M}(x,y,z), & \quad \text{if $K_b(x,y,z)$ = 1}\\ P(x,y,z) - \overline{M}(x,y,z), & \quad \text{if $K_b(x,y,z)$ = 0}\\ \end{array} \right. $$
    (3)

    where x = 1, 2, 3,...,m ; y = 1, 2, 3,...,n and z = 1, 2, 3,...,o.

    The procedure of generating K i , (3) is repeated till all the locations of P have been catered for.

  3. 3.

    A chaotic sequence C is generated using (2), having the same dimensions as \(\overline{M}\). The logistic map is provided with system parameter μ and initial condition X 0 to generate the sequence.

    System parameter μ and initial condition X 0 are to be known only to the sender and intended recipient, so only they may be able to generate the same chaotic sequence.

  4. 4.

    The chaotic sequence is then scaled and rounded off, such that a sequence of uniformly distributed integers in range [a,b] is obtained.

  5. 5.

    Now, K i is operated with C in accordance with the operation defined by K b as:

    $$\label{eq:bi1} K_d (x,y,z) = \left\{ \begin{array}{ll} K_i(x,y,z) + C(x,y,z), & \quad \text{if $K_b(x,y,z)$ = 1}\\ K_i(x,y,z) - C(x,y,z), & \quad \text{if $K_b(x,y,z)$ = 0}\\ \end{array} \right. $$
    (4)

    where K d denotes the data key.

At the end of step 5, we will have two keys–K b and K d –and cover image that are to be transmitted. The keys are to be transmitted through a secure channel.

3.2 Extraction

For successful retrieval of payload on the receiving side: cover image, secret keys, system parameter and initial condition for generating chaotic sequence are needed. Chaos is extremely sensitive to the system parameter μ; any minor difference in its value generates an entirely different chaotic sequence. Therefore, for accurate retrieval of the payload, exact system parameter μ is required. Steps involved in extraction process are:

  1. 1.

    Characteristics of cover image \(\overline{M}\) are obtained by using the same characteristic extraction method as in stego-key generation procedure. Also, C is generated with system parameter μ and initial condition using (2).

  2. 2.

    Now, information of payload and cover image relationship–K i –is retrieved from K d using the binary key, K b .

    $$\label{eq5} K_i (x,y,z) = \left\{\begin{array}{l l} K_d(x,y,z) - C(x,y,z), & \quad \text{if $K_b(x,y,z)$ = 1}\\ K_d(x,y,z) + C(x,y,z), & \quad \text{if $K_b(x,y,z)$ = 0}\\ \end{array} \right. $$
    (5)

    Step 2 is repeated till all the indices of K b are catered for and stored in K i .

  3. 3.

    Finally, characteristics of cover image are separated from secret data present in K i using K b to recover the payload P.

    $$\label{eq6} P (x,y,z) = \left\{\begin{array}{l l} K_i(x,y,z) - \overline{M}(x,y,z), & \quad \text{if $K_b(x,y,z)$ = 1}\\ K_i(x,y,z) + \overline{M}(x,y,z), & \quad \text{if $K_b(x,y,z)$ = 0}\\ \end{array} \right. $$
    (6)

    Step 3 is repeated till the payload matrix is completely retrieved.

We will now analyze the survivability, imperceptibility and security of the proposed algorithm in different scenarios.

4 Results and discussion

As imperceptibility and undetectability are the key issues that need to be addressed in steganography, steganography algorithms are evaluated for survivability against steganalysis and statistical attacks [9, 10, 14, 21, 36, 38]. Steganalysis generally uses following techniques to identify steganography: i) changes between color values, ii) analysis of exaggerated noise, iii) stego-image size analysis for detecting abnormal changes in stego-image, and iv) existence of obvious and repetitive patterns [14]. However, in our case, since no changes have been made to the cover image, the proposed algorithm is undetectable by steganalysis. For our experiments, averaging has been used as a characteristic extraction method. Averaging is carried out on M using an averaging window of size p ×q and is stored as \(\overline{M}\):

$$\label{eq7} \overline{M}\left(\frac{x}{p},\frac{y}{q},z\right)=\left\lceil \frac{\sum\limits_{a=x-(p-1)}^{x} \sum\limits_{b=y-(q-1)}^{y}M(a,b,c)}{p\times q} \right\rceil $$
(7)

where x = p, 2p, 3p,...,i ; y = q, 2q, 3q, ..., j and z = c = 1,2,...,k. In our analysis, we have selected p = q = 8.

As far as payload capacity is concerned for the tested scenario, the proposed algorithm can take a maximum payload up to the size of averaged cover image given by:

$$\label{eq:maxcapacity} Cap_{\max} = \frac{1}{(S_{AW})^2} * {S_M} $$
(8)

Where Cap max is the maximum payload capacity, S M is the size of cover image in bits and S AW is the size of averaging window matrix, S AW ∈ {1, 2, 3, ...}.

Comparison between capacity of proposed scheme with several other steganography schemes is presented in Table 1. Though comparison between capacity of conventional steganography and proposed scheme is hard to do as in conventional steganography algorithms capacity is directly linked to the imperceptibility of the algorithm. We have presented a comparison table based on the capacity and imperceptibility trade-off. The results presented here can be found in [8]. All of the conventional steganography methods have been subjected to blind steganalysis [9] and detection of presence of steganography has been evaluated using the detection reliability ρ. Detection reliability is normalized so that ρ = 1 means perfect detection and ρ = 0 means no detection at all.

Table 1 Comparison based on capacity and imperceptibility trade offs between: F5, F5 without matrix embedding (1,1,1) [36], OutGuess 0.2 (OG) [28], Model based Steganography without and with deblocking (MB1 and MB2 respectively) [31], Perturbed Quantization [8] and Chaos based Zero-steganography scheme (U = unachievable capacity)

The algorithms under consideration are F5, F5 without matrix embedding (1,1,1) [36], OutGuess 0.2 (OG) [28], Model based Steganography without and with deblocking (MB1 and MB2, respectively) [31], Perturbed Quantization scheme [8] and proposed Zero-steganography scheme. The capacity of all of the algorithms except the proposed scheme is measured in bits per non-zero DCT coefficient (bpc), while the metric of capacity for the proposed scheme is bit per pixel (bpp). The maximum capacity for the proposed algorithm would be 0.125 bpp for an 8 × 8 window characteristic extraction method. All of the other algorithms under consideration have maximum capacity that is higher than the maximum capacity of the proposed algorithm. However, the increased capacity comes at the cost of imperceptibility as the blind steganalysis technique can detect these algorithms if high embedding rate is used. As shown in Table 1 even for capacity as low as 0.1 bpc or 0.05 bpc the detection reliability of the blind steganalysis technique stays well above 0. The only exception is the Perturbed Quantization method, for which ρ ≈ 0 for capacity of 0.05 bpc, meaning that for 0.05 bpc capacity, it is completely imperceptible against this particular steganalysis technique.

Even though the maximum capacity of proposed scheme is lower than many conventional stegnography algorithms, the immunity of the proposed scheme against any steganalysis makes it sufficiently capacious while totally imperceptible. Note that using any other characteristics extraction procedure will result in different capacity and survivability, depending upon the methodology used for characteristics extraction and the type of attack [1]. The analyses of the proposed algorithm for survivability, imperceptibility, and security are given in the following Sections:

4.1 Survivability

Survivability is the measure that indicates the robustness of steganography algorithm against an attack by the adversary to limit the usability of the steganography algorithm. There are several attacks in the literature that are used to render the steganography algorithms unusable to some extent [15, 26, 35]. Survivability or robustness is not a primary concern in steganography algorithms, however, a great deal of work has been done in order to improve robustness in digital watermarking algorithms [3, 13, 18]. In order to evaluate the survivability of the proposed algorithm, we tested it against common attacks such as low-pass filtering, JPEG compression and noise. Survivability is measured in terms of Bit Error Rate (BER), defined as:

$$\label{eq9} \mbox{BER} = \frac{B_c}{B_t} $$
(9)

Where B c is the number of corrupted payload bits upon extraction and B t is the total number of payload bits.

For the purpose of our tests a set of 512 ×512 and 256 ×256 color images–shown in Figs. 3 and 4, respectively–were used as cover images. Whereas, a 64 ×64 and 32 ×32 color image was used as payload for the former and the latter set, respectively–shown in Fig. 5.

Fig. 3
figure 3

Test images of size 512 × 512

Fig. 4
figure 4

Test images of size 256 × 256

Fig. 5
figure 5

Payload image (Coin)

At first, respective keys were generated for each of the cover images, then the cover images were attacked and extraction was carried out from the attacked cover images. BER for extracted payload was computed by comparison between original payload bits and extracted payload bits. To further reduce the BER produced as a result of attacks, Anderson and Petitcolas [2] suggested use of redundancy. We used the simple form of redundancy–data repetition–with a repetition factor of 24 for error correction and as a result, improved, lower BER was obtained. It is to be noted here that the payload capacity will decrease by a factor of 24 when we use repetition technique.

Combined results for survivability, with and without repetition, are plotted against different attack strengths for selected attacks. The following plots represent averaged results of all five cover images for selected payload image in terms of BER. The results for scenarios with data repetition use a random payload matrix instead of the payload image. Detailed analysis of the results is presented in the Sections 4.1.1, 4.1.2, and 4.1.3.

4.1.1 Low-pass filtering attack

For the low-pass filtering attack, we used the averaging filter technique. Cover images were attacked with averaging filters of sizes 3 ×3, 5 ×5, 7 ×7 and 9 ×9–where 9 ×9 represents the highest low-pass filtering attack strength. BER for retrieved payload was calculated for each attack strength. Figure 6 illustrates the average BER obtained with all the cover images for low-pass filtering attack for both cases–with and without repetition of payload. Subsequent results were computed accordingly for each attack strength.

Fig. 6
figure 6

Low-pass filtering attack

Without repetition, in case of the highest attack strength, less than 14 % bits were corrupted and rest of the payload was extracted accurately. However, use of repetition reduced BER appreciably, as it dropped down to 2.2 %, for a filter size of 3 ×3. Similarly, for the worst attack case of filter size 9 ×9, it remained up to about 9 %, which was 6 % less than the previous case. As a whole, the proposed algorithm is fairly robust to low-pass filtering attack, promising an accurate payload retrieval rate of more than 90 %, even in the worst scenario using error-correction.

Table 2 provides the comparison between survivability of a Chaos based DCT stegnography algorithm [34] and the proposed Zero-steganography scheme for Gaussian low-pass filter attack. The results show that for both values of variance the survivability of the proposed algorithm is better than the Chaos based DCT steganography scheme [34]. As the attack strength increases the difference between the corresponding BER values for each algorithm also increase. The proposed algorithm performs much better under high attack strength as compared to the DCT algorithm.

Table 2 Comparison between survivability of Chaos based DCT steganography [34] and Chaos based Zero-steganography (CBZS) scheme for Gaussian low-pass filter attack

4.1.2 Noise attack

The proposed algorithm is analyzed for two noise attacks, Additive White Gaussian Noise (AWGN) and salt and pepper noise. AWGN damages the data by linear addition of white noise with a constant spectral density and a Gaussian distribution of amplitude. The attack makes use of two parameters, mean m and variance σ 2. Mean m is 0 and σ 2 varies from 0 to 1, with σ 2 = 1 giving highest level of noise. For our tests, we used σ 2 in the range 0.1 − 1 with a step size of 0.1. As illustrated in Fig. 7, without repetition, the proposed algorithm gives a BER of 18.4 % and 22 % for σ 2 = 0.1 and σ 2 = 1 respectively. This represents a minimum of 78 % accurate payload retrieval even with highest attack strength. Repetition technique with a repetition factor of 24 for a pseudo random payload matrix, improved the BER by almost 5 % for each attacked case, resulting in 13 % BER for σ 2 = 0.1 and 17 % BER for σ 2 = 1, collectively denoting a minimum of 83 % accurate payload retrieval for any AWGN attack strength.

Fig. 7
figure 7

AWGN attack

Table 3 provides the comparison between survivability of Chaos based DCT stegnography algorithm [34] and the proposed Zero-steganography scheme for salt and pepper noise attack. The proposed algorithm has a much higher survivability as compared to the other technique, for every value of noise density tested.

Table 3 Comparison between survivability of Chaos based DCT steganography [34] and Chaos based Zero-steganography (CBZS) scheme for salt and pepper noise attack

4.1.3 JPEG compression attack

JPEG provides a compression method for compressing continuous-tone image data with a pixel depth of 6 to 24 bits. JPEG is primarily a lossy method of compression, having reasonable speed and efficiency. It may be adjusted to produce very small, compressed images that are of relatively poor quality in appearance but still suitable for many applications. At the same time, JPEG is also capable of producing very high-quality compressed images that are still far smaller than the original uncompressed data [25].

As JPEG is lossy, the information in a cover image is most likely to be damaged if this type of compression is applied. Therefore, we tested our algorithm for survivability against JPEG compression attack for various quality settings–Q–ranging from 1 to 96 with a step size of 5. Here, Q = 1 represents lowest image quality retained after compression and Q = 96 represents highest quality retained after compression in the selected range. Figure 8 shows the results of BER for JPEG compression attack. For the proposed algorithm, worst quality JPEG compression gives a BER of about 20 % which keeps decreasing as we improve the compression quality and finally drops down to 3 % for Q = 96. Using repetition, a BER of 15 % was calculated for worst quality compression, while for the highest quality compression, BER was 0.1 % denoting a 99.9 % accurate payload retrieval rate.

Fig. 8
figure 8

JPEG compression attack

Table 4 provides the comparison between survivability of a Chaos based DCT stegnography algorithm [34], PN sequence based DCT algorithm [33] and the proposed Zero-steganography scheme for JPEG compression attack. The results show that for Q = 50 the BER for the proposed algorithm is quite lower than that of other schemes. However, for higher values of Q the BER of the Chaos based DCT algorithm remains better than the proposed algorithm. Whereas, BER of PN sequence based DCT algorithm is better than the proposed Zero-steganography scheme in only intermediate cases. For both higher and lower attack strength the survivability of proposed algorithm is much better than that of PN sequence based DCT algorithm. While BER for the proposed algorithm remains quite high for high quality JPEG compression, the proposed algorithm performs much better under high attack strength as compared to other algorithms.

Table 4 Comparison between survivability of Chaos based DCT steganography [34], PN sequence based Data hiding scheme [33] and Chaos based Zero-steganography (CBSZ) scheme for JPEG compression attack

4.1.4 Summary

Table 5 summarizes the survivability analysis of the proposed algorithm for all three types of attacks discussed in Sections 4.1.1, 4.1.2 and 4.1.3. Attacked images of Lena in case of highest attack strengths, with corresponding payload retrieval rate and Peak Signal-to-Noise Ratio (PSNR) are displayed in Table 5. PSNR is given by the equation:

$$\label{eq10} \mbox{PSNR} = 20 \times \log \Big(\frac{1}{\mbox{MSE}}\Big) db $$
(10)

where, Mean Square Error (MSE) is given as:

$$\label{eq11} {\mbox{MSE}} = \sqrt{\left(\frac{1}{N^2}\right) \sum\limits_{i=1}^{N} \sum\limits_{j=1}^{N} \left(x_{ij} - \widehat{x_{ij}}\right)^2} $$
(11)

where x ij represent the pixels of the original image, \(\widehat{x_{ij}}\) represent the pixels of distorted image and N is the total number of pixels.

Survivability of the proposed algorithm is highest against low-pass filtering as the least retrieval rate is above 86 %, obtained for filter size of 9 ×9. Statistically, the most distortion occurs in case of AWGN, where the retrieval rate obtained is 78.38 % for σ 2 = 1. However, in case of JPEG compression, the least retrieval rate is above 80 %, obtained in case of compression quality Q = 1. Despite the extremely distorted attacked cover images, for the proposed algorithm, the minimum retrieval rate stays above 78 %. We will now analyze the imperceptibility of the proposed Zero-steganography algorithm.

Table 5 Attacked images, PSNR and retrieval rate

4.2 Imperceptibility

For our proposed algorithm, the stego-image quality is independent of the payload length and gives the maximum PSNR with no artifacts or change in cover image. This implies that, for a 512 ×512 color image using averaging window of size 8×8, we can use a maximum payload length L = 98,304 bits–as calculated by (8)–and still get the highest quality stego-image with a PSNR value of ∞. This is because the proposed algorithm makes no change to the cover image; and stego-image is the same as cover image in our case. Section 4.3 provides a detailed analysis of the security perspective of the proposed algorithm in several compromised scenarios.

4.3 Security

One of the key features of the proposed algorithm is that it provides high security by using chaos and stego keys. Security is compromised when a steganographic system is broken. This process is carried out in two stages [41]:

  1. 1.

    An attacker detects the use of steganography.

  2. 2.

    He is able to read the hidden message.

It is not possible to detect the use of steganography in the proposed Zero-stegano-graphic system as no change is made to the cover image. An adversary can not detect the use of steganography due to absence of any distortion, artifacts and patterns in the stego image–it is the same as cover image. Hence, the proposed technique is secure and not vulnerable to visual and statistical attacks for detection.

Some of the types of attacks on steganographic systems include [14]: stego-only, known cover, and chosen stego. For the stego-only attack, the attacker is assumed to have stego medium only. In our case, it is the stego image–which is the same as cover image. From this information alone, an attacker can not determine if steganography is used and hence can not break the steganographic system. Similarly, for known cover attack, an attacker can not detect if steganography is used as the cover image and stego image are exactly the same. However, in case of chosen stego attack, the cover image is available along with the algorithm. Even then, it is computationally intensive for the attacker to extract the payload as he would require exact parameters for chaos and also the two keys.

Chaos has statistically infinite combinations of parameters–system parameter and initial condition–and along with two keys K b and K d , it is computationally intensive to figure out the right combination of the three elements to successfully retrieve the payload.

As stated in Section 3.2, an attacker would require i) cover image, ii) binary key K b , iii) data key K d , and iv) exact combination of parameters (μ and X 0) for generating chaotic sequence, for successful retrieval of payload. We have analyzed two special cases where three of the four necessary elements for extraction are compromised and only one is unknown to the attacker:

  • Case I–Unknown data key K d : In this case, the attacker has cover image, binary key K b and exact parameters for generating chaos. Now he may use a pseudo randomly generated data key K d to extract the payload. To demonstrate the effect of using a randomly generated data key for extraction, we have considered two cases with Coin and some text as payload in the other. Tables 6 and 7 show the results obtained in case of text and image payload respectively. Our tests revealed that extraction in case of payload image is not successful as the retrieved payload is extremely distorted. Similarly, in case of payload text, the retrieved text is completely destroyed.

  • Case II–Unknown chaos parameters–μ and X 0: In this case, the attacker has all the information required for payload extraction except the chaos system parameter μ and initial condition X 0. An attacker will now generate chaotic sequence by random selection of system parameter and initial condition. Again, we used Coin as payload image, and payload text to demonstrate the effect of this compromised scenario on payload extraction. The corresponding results for text and payload image are shown in Tables 6 and 7 respectively. As shown in the Table 7, extraction in case of payload image is not successful as a distorted and noisy image is obtained. Similarly, in case of payload text, the retrieved text was completely destroyed as shown in Table 6.

Table 6 Extracted payload text in highly compromised scenarios
Table 7 Extracted payload images in highly compromised scenarios using Coin as payload

In addition, a more sophisticated adversary could try to reconstruct the chaotic sequence C by analyzing the difference between first elements of \(\overline{M}\) and K d , assuming that he has complete knowledge of the data key and first elements of the payload. The adversay can try to figure out the initial parameters for the logistic map by calculating the initial values of the chaotic sequence, using (12).

$$\label{back_tracing} C = K_{d} \mp P \mp \overline{M} $$
(12)

Such a technique would present several problems to the adversary in order to regenerate the chaotic sequence. The resultant value for the initial guess for the first elements of payload could lead to a large number of possible choices of system parameter for the logistic map and those possible choices would vary depending upon the scaling factor used in the generation of chaotic sequence. If the scaling factor changes the possible choices for system parameter to choose from also change and this adds another parameter for the adversary to have knowledge of, if they want to break the algorithm. Based on these results, we conclude that the proposed algorithm is secure in these highly-compromised scenarios. Following Section concludes the proposed work based on analysis and the presented results.

5 Conclusion

Zero-steganography implies no direct embedding which makes it totally undetectable. Conventional steganography methods are evaluated against steganalysis as the key concern is undetectability and imperceptibility. However, in case of Zero-steganography, no steganalysis is required as the cover image retains its original form after stego-key generation procedure. Along with imperceptibility and undetectability, the proposed algorithm also provides security and sufficient capacity while at the same time providing survivability of more than 75 % against highest attack strengths for selected attacks.

Imperceptibility analysis shows that the PSNR of stego image remains ∞ regardless of the payload length. Survivability analysis of the proposed algorithm provides us with a maximum BER of about 14 % for low-pass filtering attack, occurring in case of 9 × 9 filter size. For AWGN attack, maximum BER was 22 %, occurring in case of σ 2 = 1. For JPEG compression attack, maximum BER was 20 %, observed in case of compression quality Q = 1. Our test results show that by using error correcting techniques such as repetition, the BER is reduced by almost 6 % for all cases.

The major limitation of the proposed algorithm is the higher bandwidth required for the data key to be transmitted through secure means. However, even if the data key is discovered, the presence of a chaos based sequence limits any advantage an adversary can take out of the knowledge of data key. The proposed algorithm is not currently robust against geometric attacks, however, geometric transformation resistant characteristics extraction methods can be investigated to make the algorithm robust against such attacks. In addition, better error correcting techniques such as Reed Solomon codes [29], Bose, Chaudhuri, and Hocquenghem (BCH) codes [4] and Hamming codes [11] can be used to improve the survivability results [1].

The proposed algorithm satisfies the merit of imperceptibility due to absence of any artifacts and patterns as no change is made to the cover image. Moreover, using chaos provides extended security as not only the two secret keys, but also the accurate system parameter and initial condition for generating chaotic sequence are required for successful payload retrieval. As discussed in Section 2, even the slightest of change in the system parameter for a chaotic sequence leads to a chaotic map that is significantly different from the original one. Results in Tables 67 show that even in highly compromised scenarios, the payload can not be retrieved successfully. Hence, we conclude that the proposed Zero-steganography is totally imperceptible, undetectable, sufficiently survivable and highly secure.