Keywords

1 Introduction

Quantum key distribution (QKD) protocols allow for the establishment of a shared secret key between two parties, referred to as Alice (A) and Bob (B), which is secure against an all-powerful adversary, customarily referred to as Eve (E). Such a task is impossible to achieve when using only classical communication; indeed, when parties have access only to classical resources, key-distribution is only secure if certain computational assumptions are made on the power of the adversary. With QKD protocols, however, the only required assumption is that the adversary is bounded by the laws of physics. Furthermore, QKD is a practical technology today with several experimental and commercial demonstrations. For a general survey of QKD protocols, the reader is referred to [1].

In general, most QKD protocols are designed, and their security proven, within a standard adversarial model of security. In this case, parties A and B run the protocol with the goal of establishing a shared secret key. An all-powerful adversary sits in the middle of the channel, intercepting, and probing, each quantum bit (or qubit) sent from A to B. As is standard in this usual model of cryptography, it is assumed that E is simply malicious and has no motivation to attack, nor does E “care” about the cost of attacking.

In this paper, we investigate the use of game theory to study the security of QKD protocols. While we are not the first to propose a game theoretic analysis of cryptographic protocols (quantum or otherwise - see the “Related Work” section below for a summary), we propose a more general-purpose model which can be applied to arbitrary QKD protocols. Compared with prior work, our new approach is more general and, most importantly, allows for meaningful key-rate and noise tolerance computations to be performed which are vital when considering QKD security and comparing benefits of distinct protocols.

Beyond introducing our model, we also apply it to analyze certain important QKD protocols against both all-powerful, quantum, attacks and also more practical attacks based on current-day technology. For each, we compute the critical noise tolerance values and compare with the standard adversarial model. We also discuss the efficiency of the resulting protocols in our model. Such computations were not possible in prior work, applying game theory to QKD thus showing the significance of our new methods. We stress that this work’s prime contribution is to develop a general framework for the modeling of QKD security, and various important computations involving these protocols (namely, key-rate computations and noise tolerances), through the use of game theory. We expect this work to be the foundation of future significant developments both in the fields of quantum key distribution, and also in game theory. Furthermore, our rational model of security may lead to more efficient secure communication systems as we discuss in the text.

1.1 Related Work

Game theory has seen great success when applied to classical cryptography (see [2] for a general survey). It has also raised a lot of interest recently in the study of Cyber-Physical System (CPS) security problems and network security [3,4,5,6].

Only recently have there been attempts and interest in applying game theory to quantum cryptography. Outside of key-distribution (our subject of interest in this paper), game theory has been used for secret sharing [7], rational state sharing [8], bit commitment [9], certain function computations [10], and secure direct communication [11].

The prior work discussed above all involve cryptographic primitives very different from QKD. However, some attempt has been made recently to apply game theory to QKD. In [12] a cooperative game was used to establish a quantum network consisting of point-wise QKD links which could relay information from one node to the other. However, QKD was only used as a tool in their work, the primary motivation for using game-theory was for the nodes to construct an optimal network topology in a vehicular network.

Closest to our work are [13, 14]. In [13], game theory was used to analyze the BB84 QKD protocol. Their model, however, only considered strategies affecting certain choices within the protocol. In their work, a three-party game was constructed (consisting of A, B, and the adversary E). The strategy space of each participant was to chose a basis (either Z or X) to send and receive quantum bits in (we will discuss quantum measurement in the next section). The goal of the parties A and B was to detect E while the goal of E was to avoid detection. There was no goal of establishing an actual secret key at the end of maximal length; furthermore, E did not have a goal of learning information on the key. Both of these goals will be incorporated in our more general model.

In the recently published work of [14], the model proposed in [13] was extended and applied to the so-called Ping-Pong protocol [15] and also the LM05 protocol [16]. Their work considered certain attacks E may perform against the system which were previously proposed in the literature against the ping-pong protocol. The strategy space for A and B (now considered one party in their work) consisted of choosing to run the protocol, or a variant of it (there was no choice to simply “abort” which is an important choice in QKD security [1]). The goal of E was to maximize her information on the final raw-key while avoiding detection; the goal of the party “AB” was to maximize their mutual information. Our model will also consider these two as goals; however we will not be concerned about probability of detection (which, in typical applications, is not a concern as there is always natural noise in the channel anyway). However, we will go beyond this by also setting a goal to maximize the efficiency of the protocol. Furthermore, the model we introduce in this work allows for critical key-rate and noise tolerance computations, not possible in prior work.

1.2 Notation and Definitions

We use H(X) to denote the Shannon entropy of random variable X. In particular, if \(P(X=x) = p_x\), then \(H(X) = -\sum _xp_x\log p_x\), where all logarithms in this paper are base two unless otherwise stated. By h(x) we mean the binary entropy function defined \(h(x) = -x\log x - (1-x)\log (1-x)\). Given two random variables X and Y, then H(XY) is the joint Shannon entropy of random variables X and Y defined in the usual way. H(X|Y) denotes the conditional entropy defined \(H(X|Y) = H(XY) - H(Y)\). By I(X : Y) we mean the mutual information between X and Y, defined to be \(I(X:Y) = H(X) + H(Y) - H(XY)\).

We assume a familiarity with game theory, and include the following definitions only for completeness. Given a tuple \(q = (q_1,\cdots , q_n)\) we write \(q_{-i}\) to mean the \(n-1\) tuple consisting of all \(q_j\) for \(j \ne i\); i.e., \(q_{-i} = (q_1, \cdots , q_{i-1},q_{i+1},\cdots ,q_n)\).

Definition 1

An n-player normal (strategic) form game G is an n-tuple\(\{(S_1,u_1),\) \(\dots ,(S_n,u_n)\}\), where for each i,

  • \(S_i\) is a nonempty set, called i’s strategy space, and

  • \(u_i\):\(S\rightarrow \mathbb {R}\) is called i’s utility function, where \(S=S_1\times \cdots \times S_n\).

Definition 2

Dominant Strategy (DS). A strategy \(s'_i\) (weakly) dominates \(s''_i\), if \(\forall s_{-i} \in S_{-i}\), \(u_i(s'_i, s_{-i}) \ge u_i(s''_i, s_{-i}),\) and \(\exists s'_{-i} \in S_{-i}\), \(u_i(s'_i, s'_{-i}) > u_i(s''_i, s'_{-i}).\)

Definition 3

Strict Nash Equilibrium (NE). \(s^* \in S\) is a strict Nash equilibrium of \(G= \{(S_1,u_1),\dots ,(S_n,u_n)\}\) if for all i and for all \(s_i\in S_i\), \(u_i(s^*_i, s^*_{-i}) > u_i (s_i, s^*_{-i}).\)

2 Quantum Communication and Cryptography

For completeness, we review some basic concepts in quantum key distribution and quantum communication. Due to length constraints, this section is necessarily short; however, the interested reader is referred to [17].

A quantum bit or qubit is modeled, mathematically, as a normalized vector in \(\mathbb {C}^2\). More generally, an arbitrary n-dimensional quantum state may be modeled as a normalized vector in \(\mathbb {C}^n\). Quantum states are typically denoted as “kets” of the form where the \(\psi \) can be replaced with any arbitrary label. The inner product of two kets and is denoted .

The measurement postulate of quantum mechanics gives rules for how quantum states may be observed. We are interested only with projective measurements in this work. Let be an orthonormal basis of \(\mathbb {C}^n\). Then, given a quantum state , after measurement in basis \(\mathcal {B}\), one observes basis state with probability . Note, therefore, that measurements are probabilistic processes and the outcome and distribution depends on the basis one performs a measurement in. For qubits, two common bases are the Z basis, denoted and the X basis, denoted where . Note that, once observed, the original quantum state is destroyed and “collapses” to the observed basis state. In theory, one may perform a measurement in any basis. Note, also, that the No Cloning Theorem prevents the exact duplication of an unknown quantum state. Thus, when used as communication resources, an adversary is forced to attack immediately (she cannot copy the qubits to attack later); furthermore, if she attempts to extract information from the qubits via a measurement, this may cause disturbances that may be detected by honest users later. (Measurements are not the only way E can attack a qubit - however, for understanding our work in this paper, measurements are sufficient.)

2.1 Quantum Key Distribution

Quantum key distribution takes advantage of certain properties unique to quantum mechanics to allow for the establishment of a shared secret key between A and B, secure against an all powerful adversary E, a task impossible to achieve with classical communication only. There are many different QKD protocols at this point, with the first being discovered in 1984 now known as the BB84 protocol [18]. Another important protocol, discovered in 1992, is the B92 protocol [19]. The basic operation of these protocols is shown in Protocols 1 and 2. It is important to note that, in addition to a quantum channel, allowing for the transmission of qubits from A to B, there is also an authenticated classical channel connecting the two users. This channel is not secret, however, so any message sent from A to B can be read by the attacker (though, the attacker cannot write on this channel). Authentication may be done in an information theoretic manner assuming the existence of an initial (small) secret key. Thus, QKD protocols are sometimes referred to as quantum key expansion protocols as, technically, they require an initial shared secret key which they will then expand through the use of quantum communication.

In general, QKD protocols consist of a quantum communication stage followed by a classical reconciliation stage. The first stage utilizes, through multiple iterations the quantum and authenticated channels to produce a raw key - a string of classical bits that is partially correlated and partially secret. If the error rate is “low enough” (which depends on the protocol and security model), the second stage is employed which consists of an error-correcting protocol (done over the authenticated channel, thus leaking information to E “for free”) and a privacy amplification protocol, yielding a secret key. The size of the secret key is directly correlated with the noise in the quantum channel and the amount of information an adversary potentially has on the raw-key. The more information the adversary has and the more noise, the smaller the secret key will be. In the standard adversarial model of security, the noise is assumed to be the product of the adversary’s attack and the two are directly correlated; in fact, one important aspect of QKD research is to determine a protocols maximally tolerated noise level, that is the value of noise for which QKD is possible against a malicious adversary.

For more details on all these concepts, the reader is referred to [1].

figure a
figure b

3 Game Theoretic Model

We now introduce our game theoretic security model for QKD. While in practice, A and B are two separate entities, in our game theoretic model, we will consider them as one party which we denote by AB. We therefore consider a two-party game consisting of player AB and player E. The goal of party AB is to establish a long, secret key shared between each other. The goal of E is to limit the length of the final secret key. Since it is trivial for E to cause a denial-of-service attack in a point-to-point communication protocol (of which QKD is one), we limit E’s strategy space to consisting of attacks which induce less noise than some maximal value Q which AB advertise as tolerating. This parameter Q can also represent certain “natural” noise in the quantum channel - AB will abort if the noise exceeds this value, thus if E attacks, she must “hide” in the natural noise. Noise for us is defined to be the average probability of a flipping to a and a flipping to a . One key interest to us will be for what values of Q, QKD is possible in our game theoretic model and compare this with the standard adversarial model.

Beyond these goals, there are costs for using certain quantum (and potentially also classical) resources. For AB sending and receiving qubits can be a costly activity. Thus, though AB wish to establish a key, if doing so is “too expensive” they may wish to simply “abort” and do nothing. On the other hand, to gain information is E’s goal (as this limits the size of the secret key), however attacking the quantum channel is a costly activity and extracting maximal information may require expensive quantum memory systems. Thus, it is the goal of our framework to construct a protocol (game strategy) where it is in AB’s interest to run the protocol (and not abort), while it is in E’s interest not to perform a complicated attack against it. Passive attacks (as opposed to more powerful quantum attacks) can greatly increase the efficiency of the protocol as we will see. Thus, if users employ the rational model of security for QKD, more efficient quantum communication may be possible.

One may consider applying our model to classical key-distribution (for instance, by using a hard problem that takes a large amount of classical resources to break); however this problem scenario, and the rewards for attacking, are very different from the quantum case. In a QKD protocol, the generated key is information theoretically secure, thus, for example, any message encrypted using the produced key is perfectly secret for all time. However, if a classical key-distribution system is used, an adversary may copy all communication sent by the protocol and attack offline; eventually if the system is broken, that adversary can learn all messages encrypted with that key. This is a very powerful motivating factor for an adversary. Contrast this with QKD: first, the adversary cannot attack offline and must attack actively. Furthermore, the adversary cannot learn the produced secret key nor any message encrypted with it.

We formulate our game-theoretic security model as follows. Let \(\varSigma _{AB}\) be a set of strategies (i.e., protocols) which party AB may choose to run and let \(\varSigma _E\) be the set of strategies (i.e., attacks) which party E may choose to employ against AB. We always assume the “do nothing” strategy (denoted \(I_{AB}\) for AB and \(I_E\) for E) is an option for either party. (We use I for “identity operation.”)

Now, in reality, player AB actually consists of two separate entities, thus it is important to ensure that our game-theoretic model can actually be employed in practice. In particular, we must ensure that A and B can agree on a strategy in a way that makes sense. There are many ways to achieve this; one in particular is they can sacrifice some of their initial shared secret key to send a constant-length message encrypted with one-time-pad (this message is the protocol to use). As mentioned earlier (see Sect. 2.1), for the authentication channel to work, A and B must begin with some shared random key already. They may use a constant amount \(c = \log _2|\varSigma _{AB}|\) to send, with perfect secrecy, the choice of protocol. So long as c is not a function of the number of iterations used in the quantum channel (which it is not), there is no contradiction to the key-expansion properties of QKD. Note that one cannot use this shared initial key to send securely a longer classical key - it can only be used for a small, constant, amount of initial communication such as picking from a small subset of strategies.

There are other ways for A and B to agree on a strategy, however, we may safely assume that party AB, though two distinct entities separated physically, may, at the start, agree on a single protocol to use from the set of allowed strategies \(\varSigma _{AB}\). Note that a mixed strategy may also be agreed on by having A choose a random protocol and sending the choice, securely, to B.

Let \(Q \in [0, .5]\) be the maximal noise level in the channel which is publicly known to both players before the game begins (alternatively, Q may be a value set by AB that is the “maximal tolerated” noise allowed in the channel, either naturally or artificially). Thus, even if E chooses not to attack (i.e., she chooses strategy \(I_E\)), she will still learn something about the raw key without incurring any costs (due to the information leaked by error correction). However, if she wishes to learn more (causing the secret key length to drop further) she must choose to attack the channel. We will assume that this attacker, if she chooses to attack, is able to replace the noisy quantum channel with an ideal one and then hide the noise her attack inevitably creates within this natural noise parameter Q. Such an operation (attacking, and setting up her equipment to hide within the natural noise) will be potentially expensive, though she will gain more information on the raw key thereby decreasing the efficiency of AB, her goal.

After running their respective protocol \(\varPi \in \varSigma _{AB}\) (which includes running a quantum communication stage for N iterations, followed by error correction and privacy amplification), with E attacking using attack \(\mathcal {A}\in \varSigma _E\), each party is given a utility for the outcome of the game. The outcome of the game for party AB is a function of the resulting secret key length (i.e., after error correction and privacy amplification), denoted M along with the cost of running the chosen strategy (denoted, \(C_{AB}(\varPi )\)). For our analysis, we will assume the utility is a simple linear function of the form:

$$\begin{aligned} u_{AB}(M, C_A(\varPi )) = w_g^{AB}M - w_c^{AB}C_{AB}(\varPi ). \end{aligned}$$

where \(w_g^{AB}\) and \(w_c^{AB}\) are non-negative weights for the “gain” and “cost” respectively of AB’s utility. We will assume that these weights are simply 1.

For E, her utility is a function of the information she learned on the error-corrected raw key (before privacy amplification but after error correction) and the cost of running her chosen attack. Let K be the information E learns on the raw-key and \(C_E(\mathcal {A})\) the cost of attack \(\mathcal {A} \in \varSigma _E\). Then, her utility will be:

$$\begin{aligned} u_{E}(K, C_E(\mathcal {A})) = w_g^EK - w_c^EC_E(\mathcal {A}). \end{aligned}$$

where \(w_g^E\) and \(w_c^E\) are non-negative weights for the “gain” and “cost” of E’s utility. As with \(u_{AB}\), we will assume that these weights are simply 1.

The reader may wonder what E’s rational motivation would be for learning information about the raw-key (before privacy amplification) when it is the secret key (after privacy amplification) that is actually used by A and B later to, for example, encrypt messages. First, note that if we define the model to be E gains utility for learning information on the secret key, by the very definition of privacy amplification, her gain would be negligible (and, in the asymptotic scenario, we may even say it would be zero); thus this could never motivate her. On the other hand, we could not give her utility for causing A and B to simply abort due to high noise levels above Q as this is a form of denial of service attack which would cost E little to nothing to execute (E can simply cut the quantum channel!) and is a weakness for any point-to-point communication system, especially QKD. Thus, since gaining information on the secret key is not possible, and since a denial of service attack is outside the scope of the model, E’s goal is to minimize the key-rate of the protocol (i.e., minimize its efficiency). Since the more information E has on the raw-key the smaller the secret key will be (after privacy amplification), it is E’s goal to increase her information (thus shrinking the size of the final secret key) while minimizing her cost and staying below the natural noise level of Q.

We will use \(U_{AB}(\varPi , \mathcal {A})\) to denote the expected utility given to player AB if that player chooses strategy \(\varPi \in \varSigma _{AB}\) and if E chooses strategy \(\mathcal {A}\in \varSigma _E\). \(U_{E}(\varPi ,\mathcal {A})\) is defined similarly for E.

The goal in our game-theoretic model of QKD security is to construct a protocol (strategy) “\(\varPi \)” such that the joint strategy \((\varPi , I_E)\) is a strict Nash equilibrium (NE). In particular AB are motivated to actually run the protocol while E is motivated to not launch a complicated quantum attack against it. If such a protocol exists, then, under the assumption of a rational adversary, that adversary will choose not to implement a powerful quantum attack as it will be too expensive. This security model guarantees that if AB and E are rational, then, assuming the protocol is a strict NE, the resulting key is information theoretically secure. In the standard adversarial model the key is also information theoretically secure, however the effective key-rate will be lower after privacy amplification as one must “remove” E’s additional information from her quantum attack. Thus, by assuming rational adversaries, one still maintains information theoretic security, but with greater communication efficiency.

In this work, we will consider standard QKD protocols (such as BB84 [18]) and add to these protocols additional “decoy” iterations. These decoy iterations will be, during the operation of the protocol, completely indistinguishable from standard iterations. At the end of the game (protocol run), AB will announce which iterations were “real” and which were decoys. Decoy iterations, which are useless to both parties, cost AB resources as they must still prepare and measure qubits (if they do not send qubits, this is distinguishable to E and she will know it is a decoy). However, since E cannot tell which are the decoy iterations, she is forced to attack them all the same, thus costing her resources also. If E’s attack is very expensive (e.g., requires an expensive quantum memory to operate), then the more decoy iterations there are, the less incentive she will have to attack at all. Of course, the more decoy iterations there are, the less incentive AB will have to run the protocol as it will become too expensive for too little reward.

To incorporate this decoy method, we will introduce a parameter \(\alpha \in [0,1]\) which may be set by AB. On any iteration of a protocol, during the quantum communication stage, AB (in practice, just party A) will decide whether this iteration is a real iteration (with probability \(\alpha \)) or a decoy iteration (with probability \(1-\alpha \)), however they run the iteration normally regardless so that E cannot distinguish the two cases. At the conclusion of the protocol, all decoy iterations are discarded (to achieve this in practice, A will transmit, at the conclusion of the protocol, through the authenticated classical channel, which iterations were decoys - thus E also learns this at the end of the game, but at that point, she already used resources to attack; furthermore, properties such as the No-Cloning Theorem, prevent her from making copies of qubits and later changing her attack based on this new knowledge). A protocol strategy, therefore, will be denoted \(\varPi ^{(\alpha )}\). Ultimately, the goal within this game-theoretic model is to find a value for \(\alpha \) such that the joint strategy \((\varPi ^{(\alpha )}, I_E)\) is a strict NE. Furthermore, we wish to determine what values of Q allow for an \(\alpha \) to exist and to determine the efficiency of the resulting protocol.

3.1 All-Powerful Attacks Against BB84

In this section, we apply our framework to model security of the BB84 protocol allowing E the ability to launch all-powerful attacks (e.g., attacks requiring quantum memories). We will prove that the noise tolerance of the BB84 protocol in our game theoretic framework remains \(11\%\), the same as in the standard adversarial model [20]. However, we will show that, for noise levels less than \(11\%\), the efficiency of the protocol can be substantially higher in our game theoretic model than in the standard adversarial model.

We will consider the BB84 protocol parameterized by \(\alpha \), denoted here as \(\varPi ^{(\alpha )}_{BB84}\). We will consider what is required for \((\varPi ^{(\alpha )}_{BB84}, I_E)\) to be a strict Nash equilibrium. First, consider AB’s utility for this strategy; we assume N is the number of iterations they run the protocol for. In this case, since E is not attacking, after error correction and privacy amplification, the secret key will be of expected length \(\frac{N\alpha }{2}(1-h(Q))\) (recall, in BB84, only half the iterations are expected to be kept - see Protocol 1). Thus:

$$\begin{aligned} U_{AB}(\varPi ^{(\alpha )}_{BB84}, I_E) = \frac{N}{2}\alpha (1-h(Q)) - C_{AB}, \end{aligned}$$
(1)

were we use \(C_{AB}\) to mean \(C_{AB}(\varPi ^{(\alpha )}_{BB84})\) (a value that AB must decide on, though its actual numerical value will not be important to us in this section). On the other hand, we have \(U_{AB}(I_{AB}, I_E) = 0\). Thus, for a strict NE to exist, we require:

$$\begin{aligned} \alpha > \frac{2C_{AB}}{N(1-h(Q))}. \end{aligned}$$

Naturally, this requires \(1-h(Q) > \frac{2}{N}C_{AB}\). Thus, if this expression cannot be satisfied, then the natural noise in the quantum channel (denoted Q) is too great and AB cannot justify the cost of running the protocol. In the following analysis, we will assume this inequality is satisfied.

Let us now consider E’s expected utility. If E does not attack (i.e., she chooses to play strategy \(I_E\)), then, since we are also considering “natural noise” in the channel at a rate of Q, party E will gain \(\frac{N\alpha }{2}h(Q)\) bits of information on AB’s raw key “for free” simply by listening in to the authenticated classical channel (we are assuming optimal error-correcting). Thus her expected utility is: \(U_E(\varPi ^{(\alpha )}_{BB84}, I_E) = \alpha \frac{N}{2} h(Q).\)

Now, assume that E chooses an optimal quantum attack strategy \(\mathcal {A} \in \varSigma _E\). From this, she will gain more information on the raw key (thus shrinking the final secret key size, her ultimate goal), though it also will cost something to implement. Furthermore, she will waste resources on attacking decoy states. It is known that \(I(A:E) = \frac{\alpha N}{2}h(Q)\) when E performs an optimal attack [1]. Thus, her utility (based on I(A : E) and also the information learned from error correction) is:

$$\begin{aligned} U_E(\varPi ^{(\alpha )}_{BB84}, \mathcal {A}) = I(A:E) + \alpha \frac{N}{2}h(Q) - C_E(\mathcal {A}) = \alpha Nh(Q) - C_E(\mathcal {A}) \end{aligned}$$

Thus, to be a strict NE, we require \(U_E(\varPi ^{(\alpha )}_{BB84}, I_E) > U_E(\varPi ^{(\alpha )}_{BB84}, \mathcal {A})\). For this inequality to hold it must be that: \(\alpha < \frac{2C_E(\mathcal {A})}{Nh(Q)}.\) Thus, for the strategy \((\varPi ^{(\alpha )}_{BB84}, I)\) to be a strict NE, we require an \(\alpha \) to exist that satisfies the following inequalities:

$$\begin{aligned} \frac{2C_A}{N(1-h(Q))}< \alpha < \frac{2C_E(\mathcal {A})}{Nh(Q)}. \end{aligned}$$
(2)

If such an \(\alpha \) exists, and if AB choose that for their decoy state probability, they can be assured, in our rational model of security, that E will prefer to not attack the quantum channel but instead, simply eavesdrop on the authenticated channel. Furthermore, with such an \(\alpha \), rational AB are also motivated to run the protocol, as opposed to simply aborting.

To determine suitable values for \(\alpha \) we require values for \(C_{AB}\) and \(C_E(\mathcal {A})\). Let’s assume a worst-case scenario in that \(C_{AB} = C_E(\mathcal {A})\). Note that, to implement \(\mathcal {A}\) in practice, E must somehow cut into the quantum channel, replace the natural noise with a more precise channel, setup attack equipment, and, in this scenario where \(I(A:E) = h(Q)\), construct and operate a perfect quantum memory. In reality, it seems reasonable to expect that \(C_E(\mathcal {A}) > C_{AB}\). Thus, making these equal models a “worst-case” scenario of benefit to E.

Now, by assumption, we have \(\frac{2}{N}C_{AB} < 1-h(Q)\) (i.e., the cost per-bit for AB is less than \((1-h(Q))/2\); if this assumption is not made, then AB have no motivation to run the protocol). Thus, the left-hand-side of Eq. 2 is strictly less than 1 and, so, a solution for \(\alpha \) exists only if the following inequality is satisfied:

$$\begin{aligned} \frac{C_A}{1-h(Q)} < \frac{C_E(\mathcal {A})}{h(Q)}. \end{aligned}$$

Since we are assuming in this section that \(C_{AB} = C_E(\mathcal {A})\), then \((\varPi ^{(\alpha )}_{BB84}, I_E)\) is a strict NE only if the noise in the channel Q satisfies the following inequality:

$$\begin{aligned} 1-2h(Q) > 0. \end{aligned}$$
(3)

This is exactly the same noise tolerance bound as is derived in the standard adversarial model for BB84 as reported in [1, 20, 21]! In particular, a solution for \(\alpha \) exists only if \(Q \le 11\%\).

However, despite the noise tolerance threshold being the same in our new game-theoretic model and the standard adversarial model, our game theoretic model may be used to gain a significantly improved key-rate as we now demonstrate. Assume that \(Q \le 11\%\) (and so \(1-2h(Q) > 0\) and thus an \(\alpha \) exists). Let \(\alpha \) be the largest allowed by Eq. 2 (the higher \(\alpha \) is, the better for AB as the more “real” iterations are being used on average). We may thus set:

$$\begin{aligned} \alpha = \min \left( \frac{2c_E(\mathcal {A})}{Nh(Q)} - \epsilon , 1-\epsilon \right) , \end{aligned}$$

for some small \(\epsilon >0\). Since we are assuming \(C_E(\mathcal {A}) = C_{AB}\) and we also require \(\frac{2}{N}C_A < 1-h(Q)\), we may write \(C_E = \frac{\gamma }{2}\cdot N(1-h(Q))\) for some constant \(\gamma < 1\) and thus we have:

$$\begin{aligned} \alpha = \min \left( \gamma \frac{1-h(Q)}{h(Q)} - \epsilon , 1-\epsilon \right) . \end{aligned}$$
(4)

With \(\alpha \) chosen as this, it is in E’s interest to not attack, but to instead only gain the free information from the error-correction due to the natural noise level Q. In this case, the Csiszar-Korner bound [22] applies (as E no longer has a quantum system, but a classical one) which gives us a secret key size, after privacy amplification and error correction, of:

$$\begin{aligned} \ell _{GT}(N) = \alpha \frac{N}{2}(1-h(Q)) = \frac{N}{2}\min \left( \gamma \cdot \frac{(1-h(Q))^2}{h(Q)}, (1-\epsilon )(1-h(Q))\right) . \end{aligned}$$
(5)

On the other hand, in the standard adversarial model for a noise level of Q, the secret key size would be: \(\ell _{SAM}(N) = \frac{N}{2}(1-2h(Q)).\) Discounting the \(\epsilon \) term (which may be made arbitrarily close to 0), we plot the conditional key-rate of the BB84 protocol in both our new game theoretic model and the standard adversarial model (i.e., we plot \(2\ell _{GT}(N)/N\) and \(2\ell _{SAM}(N)/N\) respectively) in Fig. 1. Note that, even though the noise tolerance is the same in both security models, our game-theoretic security model may provide a much higher key-rate (i.e., efficiency) depending on the cost \(C_{AB}\) (i.e., \(\gamma \)). Thus, by using a game-theoretic model of security, more efficient quantum secure communication systems may be employed!

Fig. 1.
figure 1

Showing the key-rate of the BB84 protocol in the Standard Adversarial Model (SAM) compared with our game-theoretic model at high noise levels (x axis) for various values of \(\gamma \). Higher means more efficient communication.

3.2 Intercept/Resend Attacks

In the previous section, we considered \(\varSigma _{AB} = \{I_{AB}, \varPi ^{(\alpha )}_{BB84}\}\) while E’s strategy space was \(\varSigma _E = \{I_E, \mathcal {A}\}\) where \(\mathcal {A}\) was an optimal attack against the BB84 protocol utilizing a quantum memory system. We also assumed that the cost of performing attack \(\mathcal {A}\) was similar to the cost of AB running the actual protocol (a very strong assumption in favor of the adversary). In practice, such an attack would be very difficult to launch against the protocol (and, with current technology, impossible as it would require a perfect quantum memory to perform successfully). In this section, we consider practical, so-called Intercept-Resend (IR) attacks. These attacks can be performed using today’s technology; they also require hardware similar to that used by A and B, allowing us to more accurately compute the cost of an attack compared with the cost of running the actual protocol.

For this attack, on each iteration of the quantum communication stage, E will, with probability p, choose to attack and with probability \(1-p\) choose to ignore the incoming qubit. This value p will control how much noise E’s IR attack actually creates (which, as before, must be kept below the natural noise level Q). This choice to attack or not is part of the strategy and is made independently for each iteration of the quantum communication stage. This is also different from the \(I_E\) strategy which chooses to not attack every iteration.

Should E decide to attack a particular iteration (with probability p), she will first measure the incoming qubit in a basis (this is fixed for each iteration and part of the strategy) causing the qubit to collapse to one of the basis states or . If E observes , she will “guess” that the key-bit for this iteration is \(i\in \{0,1\}\). She will then send a fresh qubit in the state to B.

There are two important parameters for an IR attack; first the value p and, second, the basis choice. We consider three common bases choices for IR attacks: , (see Sect. 2, and the Breidbart basis , where: and .

The value of p will be fixed to be the maximum value so that the induced noise is equal to Q. This makes sense, since the larger the value of p, the more information E may learn (since she is attacking more often), and since we cannot have p so large that the induced noise is higher than Q, the allowed maximum. Thus, once Q is given, the set \(\varSigma _E\) will consist of four distinct strategies: \(I_E\) (the “do nothing” attack); along with three strategies, one for each basis choice (we denote these attack strategies simply as ZX, and B).

As for AB, we will consider three possible strategies: \(I_{AB}\) (i.e., “do nothing”); \(\varPi ^{(\alpha )}_{BB84}\) the BB84 protocol as analyzed previously (see Protocol 1); and \(\varPi ^{(\alpha )}_{B92}\), the B92 protocol [19] (see Protocol 2). Both BB84 and B92 are common protocols used in practical implementations of QKD [1]; B92 has the advantage that it requires less quantum resources to implement (and, so, is cheaper). However, at least in the standard adversarial model, B92 has a lower noise tolerance [20]. In this section, we will show that, so long as Q satisfies certain bounds, the joint strategy \((\varPi ^{(\alpha )}_{BB84}, I_E)\) is a strict NE (for suitably chosen \(\alpha \)); we will also show that \(\varPi ^{(\alpha )}_{BB84}\) is a dominate strategy for player AB and \(I_E\) is a DS for E for certain critical values of noise levels Q.

We begin by computing the utility of each possible action pair \((\varPi ^{(\alpha )}, \mathcal {A})\). First, we must compute the cost associated to each strategy. To do so, we will define the following cost values for certain, basic, functionalities needed to implement the QKD protocol, and the IR attack:

$$\begin{aligned} C_S:&\text { The initial cost for}\, E \,\text {to setup her attack equipment}\\&\text { (e.g., splicing into the quantum channel)}\\ C_M:&\text { The cost to perform a measurement in a single basis}\\ C_P:&\text { The cost to prepare a qubit basis state}\\ C_R(\delta ):&\text { The cost to produce a} \,\delta \, \text {-biased bit}\\&\text { We assume that}\, C_R(\delta ) = h(\delta )C_R\, \text {for some cost}\, C_R\\ C_{auth}:&\text { The cost for}\, AB \, \text {to use the authenticated channel} \end{aligned}$$

We will assume that, if one requires an apparatus that is capable of producing a qubit in x different states, the cost is \(\gamma _x C_P\) for some function \(\gamma _x\). Similarly, for an apparatus capable of measuring a qubit in x different states, the cost is \(\gamma _x C_M\). Our analysis below will be suitable for any non-decreasing \(\gamma _x\); however when we evaluate our results, we will consider two cases: first \(\gamma _x = 1\) for all x (i.e., there is no increase in cost) and, second, \(\gamma _x = x\) (the cost increases linearly in the number of required states). Note that we will assume \(C_P \le C_M\) which is a reasonable assumption since measurement devices are generally more complicated (and sensitive) than preparation devices [1]. These cost values may take into account such practical issues as device energy consumption over time for example (thus running the devices for longer, or having devices capable of performing additional measurements, will potentially cost users more).

From this, we can compute the following costs after N iterations of each protocol:

$$\begin{aligned} C_{AB}(\varPi ^{(\alpha )}_{BB84})&= N[(3 + h(\alpha ))C_R + \gamma _4 C_M + \gamma _4 C_P] + C_{auth}\\ C_{AB}(\varPi ^{(\alpha )}_{B92})&= N[(2+h(\alpha ))C_R + \gamma _4C_M + \gamma _2C_P] + C_{auth}.\nonumber \end{aligned}$$
(6)

For BB84, AB must choose, each iteration, whether the iteration is a decoy or not (costing \(h(\alpha )C_R\)); what basis A should send in (with probability 1/2 each, thus costing \(C_R\)); what basis to measure in (costing \(C_R)\); and, finally, A must choose a random key bit (again, costing \(C_R\)). For B92, only one basis choice is required (from B). Finally, note that, BB84 is a four-state protocol in that A must prepare one of four possible qubit states each iteration. B92, however, is a two-state protocol - A must only be capable of preparing a state of the form or . In both cases, however, B must be able to measure one of four states (from two bases). It is clear that the cost of running B92 is no greater than the cost of running BB84.

The cost for E to operate attack \(I_E\) is zero (i.e., \(C_E(I_E) = 0\)). The cost for the other strategies is the same: first, she must choose to attack or not, costing \(h(p)C_R\); then she must measure and prepare a qubit in one basis. Those operations are performed for all N iterations of the quantum communication stage. Furthermore, she must also spend resources costing \(C_S\) to setup her attack initially (this is a one-time cost). The total cost for any attack \(\mathcal {A} = Z, X, B\) is:

$$\begin{aligned} C_E(\mathcal {A}) = N[h(p)C_R + p(\gamma _2C_M + \gamma _2C_P)] + C_S \text {, for any } \mathcal {A} \in \{Z,X,B\}. \end{aligned}$$
(7)

To complete our utility computation, we must also compute the secret key length for each protocol under each attack. Since an IR attack results in three classical random variables (one for Alice, Bob, and Eve), we may use the Csiszar-Korner bound [22] to compute the number of secret bits that may be distilled from these sources. Let \(\ell (N, \varPi ^{(\alpha )}, \mathcal {A})\) be the amount of secret key bits that may be distilled after N iterations of protocol \(\varPi ^{(\alpha )}\) given that E used attack \(\mathcal {A}\). Then from this bound, we have: \(\ell (N, \varPi ^{(\alpha )}, \mathcal {A}) = \eta N\alpha [I(A:B) - I(A:E)],\) where \(\eta \) is the proportion of non-discarded iterations; namely \(\eta = 1/2\) for BB84 and \(\eta = 1/4\) for B92 (see Protocols 1 and 2).

Note that the information computations above are dependent on only a single iteration of the protocol when faced with the specified attack since we are assuming iid attacks. Let \(\mathcal {I}(\varPi ^{(\alpha )},\mathcal {A})\) be equal to I(A : E) for the specified protocol and attack; then, the utility functions, for a fixed N, will be:

$$\begin{aligned} U_{AB}(\varPi ^{(\alpha )}, \mathcal {A})&= \eta N\alpha [I(A:B) - \mathcal {I}(\varPi ^{(\alpha )}, \mathcal {A})] - C_{AB}(\varPi ^{(\alpha )})\end{aligned}$$
(8)
$$\begin{aligned} U_{E}(\varPi ^{(\alpha )}, \mathcal {A})&= \eta N\alpha [\mathcal {I}(\varPi ^{(\alpha )},\mathcal {A}) + h(\widetilde{Q})] - C_E(\mathcal {A}), \end{aligned}$$
(9)

where we use \(\widetilde{Q}\) to denote the raw-key error rate; i.e., the error of the actual raw key which undergoes error correction (which, in the case of B92, is actually greater than the noise in the channel Q). The value \(\eta N\alpha h(\widetilde{Q})\) denotes the information leaked to E “for free” during error correction.

To complete the utility computation, we require I(A : B) and I(A : E) for all possible protocols and strategy pairs. It is not difficult to show that \(I(A:B) = 1-h(\widetilde{Q})\). For BB84, a raw-key error occurs when a flips to a (for \(i=0,1\)) or when a flips to a . By definition, this is exactly the channel noise level Q. Thus, for \(\varPi ^{(\alpha )}_{BB84}\), we have \(I(A:B) = 1-h(Q)\). For B92 it can be shown (see, for example, [23]) that the raw-key error is in fact: \(\widetilde{Q}= 2Q/(1-2Q)\). Next, we must compute \(\mathcal {I}(\varPi ^{(\alpha )}, \mathcal {A})\). Clearly, \(\mathcal {I}(\varPi ^{(\alpha )}, I_E) = 0\) for any protocol. Consider, now, an IR attack where E measures and resends in a basis (in our case, either Z, X, or B, however the equations we derive here may be applied to other attack bases). By the measurement postulate, if A sends a qubit of the form (for \(i=0,1,+,-\)), E will observe with probability . To compute \(\mathcal {I}(\varPi ^{(\alpha )}, \mathcal {A})\) we will need the joint distribution held between A and E. This is straight-forward arithmetic: one must simply trace the execution of each protocol and use the measurement postulate. We summarize this distribution in Table 1.

Table 1. Showing the joint probability distribution for A’s raw key bit and E’s “guess” based on her attack (conditioning on the event she chooses to attack). For B92, we require a normalization term, denoted M which is: \(M = v_{0,0}(v_{-,0} + v_{1,0}) + v_{0,1}(v_{-,1} + v_{1,1}) + v_{+,0}(v_{-,0} + v_{1,0}) + v_{+,1}(v_{-,1} + v_{1,1})\). The values here are found by tracing the protocol and using the measurement postulate.

By definition, we have \(\mathcal {I}(\varPi ^{(\alpha )}, \mathcal {A}) = p(H(A) + H(E) - H(AE))\) where the Shannon entropies may be computed easily from data in Table 1 and substituting in for the appropriate basis state depending on the attack E uses (note that when E chooses to not attack, which occurs with probability \(1-p\), she learns nothing, thus the need for the factor p in this expression). In summary, these are found to be:

$$\begin{aligned}&\mathcal {I}(\varPi ^{(\alpha )}_{BB84}, Z) \approx .189p&\mathcal {I}(\varPi ^{(\alpha )}_{BB84}, X) \approx .189p&\mathcal {I}(\varPi ^{(\alpha )}_{BB84}, B) \approx .399p\\&\mathcal {I}(\varPi ^{(\alpha )}_{B92}, Z) \approx .459p&\mathcal {I}(\varPi ^{(\alpha )}_{B92}, X) \approx .459p&\mathcal {I}(\varPi ^{(\alpha )}_{B92}, B) = 0. \end{aligned}$$

What remains is to find a value for p. As stated, we will assume that p is chosen to maximize E’s information while keeping the induced noise from her attack equal to Q. The natural noise in the channel is the average of the Z basis noise (which, in turn, is the average error of a flipping to a when it arrives at B’s lab) and X basis noise (the average of a flipping to a ); that is: \(Q = \frac{p}{4}(v_{0,0}v_{1,0} + v_{0,1}v_{1,1} + v_{1,0}v_{0,0} + v_{1,1}v_{0,1}+ v_{+,0}v_{-,0} + v_{+,1}v_{-,1} + v_{-,0}v_{+,0} + v_{-,1}v_{+,1}),\) from which it easily follows that \(p=2Q\) for \(\mathcal {A} = Z,X\) and \(p=4Q\) for \(\mathcal {A} = B\). Note that E may attack more often with the B basis as it induces less noise, on average, than the Z or X based IR attacks. From this analysis, we are now able to prove our two main results in this section involving sufficient conditions of the noise level for \((\varPi ^{(\alpha )}_{BB84},I_E)\) to be a strict NE and for each to be a DS.

Theorem 1

Assume classical resources are free for both parties AB and E (that is, let \(C_R = C_{auth} = C_S = 0\)) and let \(C_P \le C_M\) (as discussed in the text). Define \(A_1\) and \(A_2\) as follows:

$$\begin{aligned} A_1 = \frac{(\gamma _4-\gamma _2)C_P}{\frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q)}&A_2 = \frac{2\gamma _4(C_M+C_P)}{1-h(Q)}. \end{aligned}$$

If \(\max (A_1,A_2) < 1\) and Q, the noise in the channel is less than 0.232 and satisfies the following inequality:

$$\begin{aligned} \left\{ \begin{array}{ll} 10.025\left( \frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q)\right) - \left( \frac{\gamma _4}{\gamma _2}-1\right)> 0, &{} \text { If } A_1 \ge A_2\\ 2.506(1-h(Q)) - \frac{\gamma _4}{\gamma _2} > 0, &{} \text { Otherwise} \end{array}\right. \end{aligned}$$
(10)

Then there exists an \(\alpha \in [0,1]\) such that \((\varPi ^{(\alpha )}_{BB84}, I_E)\) is a strict NE.

Proof

Since \(C_{auth} = C_S = 0\), the factor of N may be divided out of the utility functions (we are only interested in relations between them and the factor N appears in both \(U_{AB}\) and \(U_E\). This allows us to construct the function table shown in Table 2. From this table, we see that, for \((\varPi ^{(\alpha )}_{BB84}, I_E)\) to be a strict NE, the following inequalities must be satisfied:

Table 2. Function table for utility functions \(U_{AB}\) and \(U_E\) assuming \(C_{auth} = C_S = 0\) and dividing out the factor of N on both functions.
$$\begin{aligned} \alpha&> \frac{2\gamma _4(C_M+C_P)}{1-h(Q)}\\ \alpha&> \frac{(\gamma _4-\gamma _2)C_P}{\frac{1}{4} + \frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q)} \text { } \left[ \text {If } \frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q) > 0 \right] \\ \alpha&< \frac{4\gamma _2(C_M+C_P)}{0.378} \approx 10.582\gamma _2(C_M+C_P)\\ \alpha&< \frac{8\gamma _2(C_M+C_P)}{1.596} \approx 5.013\gamma _2(C_M+C_P). \end{aligned}$$

Note that, if \(Q < .232\) (as assumed in the hypothesis), then \(\frac{1}{4}+\frac{1}{4}h(2Q/(1-2Q)) - \frac{1}{2}h(Q) > 0\). From this, it is clear that if we can find an \(\alpha \) that satisfies:

$$\begin{aligned} \max (A_1,A_2)< \alpha < \frac{8\gamma _2(C_M+C_P)}{1.596}, \end{aligned}$$

the resulting joint strategy will be a strict NE (recall, by hypothesis, \(\max (A_1,A_2)\) \(<1\)). For such a value to exist, it must be that \(\max (A_1,A_2)\) is strictly less than the right-hand side of the above expression.

We show this in two cases. First, assume \(A_2 > A_1\). Then, by our assumptions on the channel noise Q, we have:

$$\begin{aligned} \frac{\gamma _4}{\gamma _2}&< 2.506(1-h(Q))\\ \Longrightarrow&\frac{\gamma _4(C_M+C_P)}{1-h(Q)}< \frac{4\gamma _2(C_M+C_P)}{1.596} \Longrightarrow \frac{2\gamma _4(C_M+C_P)}{1-h(Q)} < \frac{8\gamma _2(C_M+C_P)}{1.596}, \end{aligned}$$

as desired.

For the second case, assume \(A_1 \ge A_2\). Then, by assumption on the channel noise Q, we have:

$$\begin{aligned}&\frac{\gamma _4}{\gamma _2}-1< 10.025\left( \frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q)\right) \\\\ \Longrightarrow&\frac{2(\gamma _4-\gamma _2)C_P}{\frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) -\frac{1}{2}h(Q)} < \frac{8\gamma _2(2C_P)}{1.596}. \end{aligned}$$

Noting that \(C_P \le C_M\) completes the proof.

Table 3. Showing the allowed noise tolerance for which \((\varPi ^{(\alpha )}_{BB84}, I_E)\) is a strict NE. When \(\gamma _4=\gamma _2\) then it is always true that \(A_2 \ge A_1\) (since \(A_1 = 0\) and \(A_2\) is always non-negative) and so we do not need to evaluate the case for \(A_1 > A_2\). When \(\gamma _4=2\gamma _2\), we must evaluate both cases. See text for explanation.

Theorem 1 gives conditions on the noise parameter Q for which \((\varPi ^{(\alpha )}_{BB84}, I_E)\) becomes a strict NE. The restrictions on \(\max (A_i) < 1\) may be satisfied if the cost \(C_P\) and \(C_M\) are low enough. The restrictions on Q depend only on the value \(\gamma _4\) and \(\gamma _2\). So long as Q satisfies Eq. 10, then AB are motivated to run the BB84 protocol and E is motivated to not perform an intercept/resend attack (but, instead, to simply “listen” on the authenticated channel). We evaluate the noise tolerance in Table 3. Surprisingly, if \(\gamma _2 = \gamma _4\), the noise tolerance is \(14.6\%\) also the maximal noise tolerance of BB84 in the standard adversarial model against optimal individual attacks (which are more general/powerful than IR attacks). Note, however, while the noise tolerance may be lower in our game theoretic model, as before, the efficiency in our game theoretic model may improve as E is not motivated to attack.

Theorem 2

Assume classical resources are free for both parties (i.e., let \(C_R = C_{auth} = C_S = 0\)) and let \(C_P \le C_M\) (as discussed in the text). Define \(A_1\) and \(A_2\) as follows:

$$\begin{aligned} A_1 = \frac{(\gamma _4-\gamma _2)C_P}{\frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q) - 0.798Q}&A_2 = \frac{2\gamma _4(C_M+C_P)}{1-h(Q)-1.596Q}. \end{aligned}$$
(11)

If \(\max (A_1,A_2)< 1\) and if Q, the noise in the channel, is strictly less than 0.185 and if it satisfies the following inequality:

$$\begin{aligned} \left\{ \begin{array}{ll} 10.025\left( \frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q) - 0.798Q\right) - \left( \frac{\gamma _4}{\gamma _2}-1\right)> 0, &{} \text { If } A_1 \ge A_2\\ 2.506(1-h(Q)-1.596Q) - \frac{\gamma _4}{\gamma _2} > 0, &{} \text { Otherwise} \end{array}\right. \end{aligned}$$
(12)

then there exists a value for \(\alpha \) such that \(\varPi ^{(\alpha )}_{BB84}\) is a dominate strategy (DS) for AB and \(I_E\) is a DS for E.

Proof

Fix \(\alpha \). For \(\varPi ^{(\alpha )}_{BB84}\) to be a DS for AB, we must show that, for every strategy \(\mathcal {E} \in \varSigma _E\), it holds that \(U_{AB}(\varPi ^{(\alpha )}_{BB84}, \mathcal {E}) \ge U_{AB}(\varPi ^{(\alpha )}, \mathcal {E})\) for \(\varPi ^{(\alpha )}= \varPi ^{(\alpha )}_{B92}\) and \(\varPi ^{(\alpha )}= I_{AB}\). We see from Table 2, for this to be true, the following inequalities must be satisfied:

$$\begin{aligned} \alpha&> \frac{2\gamma _4(C_M+C_P)}{1-h(Q)}&\alpha> \frac{(\gamma _4-\gamma _2)C_P}{\frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q)}\\ \alpha&> \frac{2\gamma _4(C_M+C_P)}{1-h(Q)-0.378Q}&\alpha> \frac{(\gamma _4-\gamma _2)C_P}{\frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q) + 0.0405Q}\\ \alpha&> \frac{2\gamma _4(C_M+C_P)}{1-h(Q)-1.596Q}&\alpha > \frac{(\gamma _4-\gamma _2)C_P}{\frac{1}{4}+\frac{1}{4}h\left( \frac{2Q}{1-2Q}\right) - \frac{1}{2}h(Q) -0.798Q} \end{aligned}$$

Note that, the denominators of the above six inequalities are all positive by assumption that \(Q < 0.185\). Note also, that there are only six inequalities, and not eight, since two are repetitions.

It is not difficult to see that if we take \(\alpha \ge \max (A_1, A_2)\), where \(A_1\) and \(A_2\) are defined in Eq. 11, then all the above inequalities are automatically satisfied and, so, \(\varPi ^{(\alpha )}_{BB84}\) will be a DS for party AB.

Now, we consider E’s strategy \(I_E\). For \(I_E\) to be a DS for party E, the following inequalities must be satisfied (again, consulting Table 2):

$$\begin{aligned} \alpha&< \frac{4Q\gamma _2(C_M+C_P)}{.378Q} \approx 10.582\gamma _2(C_M+C_P)\\ \alpha&< \frac{8Q\gamma _2(C_M+C_P)}{1.596Q} \approx 5.013\gamma _2(C_M+C_P)\\ \alpha&< \frac{8Q\gamma _2(C_M+C_P)}{0.918Q} \approx 8.715\gamma _2(C_M+C_P) \end{aligned}$$

Clearly if \(\alpha < \frac{8Q\gamma _2(C_M+C_P)}{1.596Q}\), the other two are also satisfied. All that remains to be shown is that an \(\alpha \) exists allowing both \(\varPi ^{(\alpha )}_{BB84}\) to be a DS for AB and \(I_E\) to be a DS for E. In particular, we must show that: \( \max (A_1,A_2) < \frac{8\gamma _2(C_M+C_P)}{1.596}. \) However, this can be proven in a similar manner as in the proof of Theorem 1, using the new bounds on Q from Eq. 12. This completes the proof.

The allowed noise tolerances for \(\varPi ^{(\alpha )}_{BB84}\) to be a DS for AB and \(I_E\) to be a DS for E, are reported in Table 4.

Table 4. Showing the allowed noise values Q from Theorem 2.

4 Closing Remarks

In this paper, we introduced a new game-theoretic model of QKD security. Many interesting problems remain open. It would be interesting to analyze best-reply strategies under different noise values and decoy probabilities. We may also consider adding additional strategies for AB, different, non-linear, utility functions, and support for multi-user protocols [24]. One may also analyze the NE strategies based on Stackelberg game model, when the attacker E observes the strategy of party AB and chooses her strategy accordingly. One can envision a system whereby parties re-evaluate their choices after large sequences of N iterations, taking into account noise conditions, to chose new optimal strategies.