Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Introduction

Advances in computation and information analysis have expanded the capabilities of the physical plants and devices in cyber-physical systems (CPS)[4, 13]. Fostered by advances in cloud computing, CPS have garnered significant attention from both industry and academia. Access to the cloud gives administrators the opportunity to build virtual machines that provide to computational resources with precision, scalability, and accessibility.

Despite the advantages that cloud computing provides, it also has some drawbacks. They include - but are not limited to - accountability, virtualization, and security and privacy concerns. In this paper, we focus especially on providing accurate signals to a cloud-connected device and deciding whether to accept those signals in the face of security challenges.

Recently, system designers face security challenges in the form of Advanced Persistent Threats (APTs) [19]. APTs arise from sophisticated attackers who can infer a user’s cryptographic key or leverage zero-day vulnerabilities in order to completely compromise a system without detection by the system administrator [16]. This type of stealthy and complete compromise has demanded new types of models [6, 20] for prediction and design.

In this paper, we propose a model in which a device decides whether to trust commands from a cloud which is vulnerable to APTs and may fall under adversarial control. We synthesize a mathematical framework that enables devices controlled by the cloud to intelligently decide whether to obey commands from the possibly-compromised cloud or to rely on their own lower-level control.

We model the cyber layer of the cloud-based system using the recently proposed FlipIt game [6, 20]. This game is especially suited for studying systems under APTs. We model the interaction between the cloud and the connected device using a signaling game, which provides a framework for modeling dynamic interactions in which one player operates based on a belief about the private information of the other. A significant body of research has utilized this framework for security [79, 15, 21]. The signaling and FlipIt games are coupled, because the outcome of the FlipIt game determines the likelihood of benign and malicious attackers in the robotic signaling game. Because the attacker is able to compromise the cloud without detection by the defender, we consider the strategies of the attacker and defender to be chosen with prior commitment. The circular dependence in our game requires a new equilibrium concept which we call a Gestalt equilibrium Footnote 1. We specify the parameter cases under which the Gestalt equilibrium varies, and solve a case study of the game to give an idea of how the Gestalt equilibrium can be found in general. Our proposed framework has versatile applications to different cloud-connected systems such as urban traffic control, drone delivery, design of smart homes, etc. We study one particular application in this paper:ef control of an unmanned vehicle under the threat of a compromised cloud.

Our contributions are summarized as follows:

(i) :

We model the interaction of the attacker, defender/cloud administrator, and cloud-connected device by introducing a novel game consisting of two coupled games: a traditional signaling game and the recently proposed FlipIt game.

(ii) :

We provide a general framework by which a device connected to a cloud can decide whether to follow its own limited control ability or to trust the signal of a possibly-malicious cloud.

(iii) :

We propose a new equilibrium definition for this combined game: Gestalt equilibrium, which involves a fixed-point in the mappings between the two component games.

(iv) :

Finally, we apply our framework to the problem of unmanned vehicle control.

In the sections that follow, we first outline the system model, then describe the equilibrium concept. Next, we use this concept to find the equilibria of the game under selected parameter regimes. Finally, we apply our results to the control of an unmanned vehicle. In each of these sections, we first consider the signaling game, then consider the FlipIt game, and last discuss the synthesis of the two games. Finally, we conclude the paper and suggest areas for future research.

2 System Model

We model a cloud-based system in which a cloud is subject to APTs. In this model, an attacker, denoted by \(\mathcal {A}\), capable of APTs can pay an attack cost to completely compromise the cloud without knowledge of the cloud defender. The defender, or cloud administrator, denoted by \(\mathcal {D}\), does not observe these attacks, but has the capability to pay a cost to reclaim control of the cloud. The cloud transmits a message to a robot or other device, denoted by \(\mathcal {R}\). The device may follow this command, but it is also equipped with an on-board control system for autonomous operation. It may elect to use its autonomous operation system rather than obey commands from the cloud.

This scenario involves two games: the FlipIt game introduced in [20], and the well-known signaling game. The FlipIt game takes place between the attacker and cloud defender, while the signaling game takes place between the possibly-compromized cloud and the device. For brevity, denote the FlipIt game by \(\mathbf {G_{F}}\), the signaling game by \(\mathbf {G_{S}}\), and the combined game - call it CloudControl - by \(\mathbf {G_{CC}}\) as shown in Fig. 1. In the next subsections, we formalize this game model.

2.1 Cloud-Device Signaling Game

Let \(\theta \) denote the type of the cloud. Denote compromized and safe types of clouds by \(\theta _{\mathcal {A}}\) and \(\theta _{\mathcal {D}}\) in the set \(\varTheta \). Denote the probabilities that \(\theta =\theta _{\mathcal {A}}\) and that \(\theta =\theta _{\mathcal {D}}\) by p and \(1-p\). Signaling games typically give these probabilities apriori, but in CloudControl they are determined by the equilibrium of the FlipIt game \(\mathbf {G_{F}}\).

Let \(m_{H}\) and \(m_{L}\) denote messages of high and low risk, respectively, and let \(m\in M=\left\{ m_{H},m_{L}\right\} \) represent a message in general. After \(\mathcal {R}\) receives the message, it chooses an action, \(a\in A=\left\{ a_{T},a_{N}\right\} \), where \(a_{T}\) represents trusting the cloud and \(a_{N}\) represents not trusting the cloud.

For the device \(\mathcal {R}\), let \(u_{\mathcal {R}}^{S}:\,\varTheta \times M\times A\rightarrow \mathscr {U}_\mathcal {R}\), where \(\mathscr {U}_\mathcal {R}\subset \mathbb {R}\). \(u_{\mathcal {R}}^{S}\) is a utility function such that \(u_{\mathcal {R}}^{S}\left( \theta ,m,a\right) \) gives the device’s utility when the type is \(\theta \), the message is m, and the action is a. Let \(u_{\mathcal {A}}^{S}:\, M\times A\rightarrow \mathscr {U}_\mathcal {A}\subset \mathbb {R}\) and \(u_{\mathcal {D}}^{S}:\, M\times A\rightarrow \mathscr {U}_\mathcal {D}\subset \mathbb {R}\) be utility functions for the attacker and defender. Note that these players only receive utility in \(\mathbf {G_{S}}\) if their own type controls the cloud in \(\mathbf {G_{F}}\), so that type is not longer a necessary argument for \(u_{\mathcal {A}}^{S}\) and \(u_{\mathcal {D}}^{S}\).

Denote the strategy of \(\mathcal {R}\) by \(\sigma _{\mathcal {R}}^{S}:\, A\rightarrow \left[ 0,1\right] \), such that \(\sigma _{\mathcal {R}}^{S}\left( a\,|\, m\right) \) gives the mixed-strategy probability that \(\mathcal {R}\) plays action a when the message is m. The role of the sender may be played by \(\mathcal {A}\) or \(\mathcal {D}\) depending on the state of the cloud, determined by \(\mathbf {G_{F}}\). Let \(\sigma _{\mathcal {A}}^{S}:\, M\rightarrow \left[ 0,1\right] \) denote the strategy that \(\mathcal {A}\) plays when she controls the cloud, so that \(\sigma _{\mathcal {A}}^{S}\left( m\right) \) gives the probability that \(\mathcal {A}\) sends message m. (The superscript S specifies that this strategy concerns the signaling game.) Similarly, let \(\sigma _{\mathcal {D}}^{S}:\, M\rightarrow \left[ 0,1\right] \) denote the strategy played by \(\mathcal {D}\) when he controls the cloud. Then \(\sigma _{\mathcal {D}}^{S}\left( m\right) \) gives the probability that \(\mathcal {D}\) sends message m. Let \(\varGamma _{\mathcal {R}}^{S}\), \(\varGamma _{\mathcal {A}}^{S}\), and \(\varGamma _{\mathcal {D}}^{S}\) denote the sets of mixed strategies for each player.

For \(\mathcal {X}\in \left\{ \mathcal {D},\mathcal {A}\right\} \), define functions \(\bar{u}_{\mathcal {X}}^{S}:\,\varGamma _{\mathcal {R}}^{S}\times \varGamma _{\mathcal {X}}^{S}\rightarrow \mathscr {U}_\mathcal {X}\), such that \(\bar{u}_{\mathcal {X}}^{S}\left( \sigma _{\mathcal {R}}^{S},\sigma _{\mathcal {X}}^{S}\right) \) gives the expected utility to sender \(\mathcal {X}\) when he or she plays mixed-strategy \(\sigma _{\mathcal {X}}^{S}\) and the receiver plays mixed-strategy \(\sigma _{\mathcal {R}}^{S}\). Equation (1) gives \(\bar{u}_{\mathcal {X}}^{S}\).

$$\begin{aligned} \bar{u}_{\mathcal {X}}^{S}\left( \sigma _{\mathcal {R}}^{S},\sigma _{\mathcal {X}}^{S}\right) =\underset{a\in A}{\sum }\underset{m\in M}{\sum }u_{\mathcal {X}}^{S}\left( m,a\right) \sigma _{\mathcal {R}}^{S}\left( a\,|\, m\right) \sigma _{\mathcal {X}}^{S}\left( m\right) ,\,\mathcal {X}\in \left\{ \mathcal {A},\mathcal {D}\right\} \end{aligned}$$
(1)

Next, let \(\mu :\,\varTheta \rightarrow \left[ 0,1\right] \) represent the belief of \(\mathcal {R}\), such that \(\mu \left( \theta \,|\, m\right) \) gives the likelihood with which \(\mathcal {R}\) believes that a sender who issues message m is of type \(\theta \). Then define \(\bar{u}_{\mathcal {R}}^{S}:\,\varGamma _{\mathcal {R}}^{S}\rightarrow \mathscr {U}_\mathcal {R}\) such that \(\bar{u}_{\mathcal {R}}^{S}\left( \sigma _{\mathcal {R}}^{S}\,|\, m,\mu \left( \bullet \,|\, m\right) \right) \) gives the expected utility for \(\mathcal {R}\) when it has belief \(\mu \), the message is m, and it plays strategy \(\sigma _{\mathcal {R}}^{S}\). \(\bar{u}_{\mathcal {R}}^{S}\) is given by

$$\begin{aligned} \bar{u}_{\mathcal {R}}^{S}\left( \sigma _{\mathcal {R}}^{S}\,|\, m,\mu \right) =\underset{\theta \in \varTheta }{\sum }\underset{a\in A}{\sum }u_{\mathcal {R}}^{S}\left( \theta ,m,a\right) \mu _{R}\left( \theta \,|\, m\right) \sigma _{\mathcal {R}}^{S}\left( a\,|\, m\right) . \end{aligned}$$
(2)

The expected utilities to the sender and receiver will determine their incentives to control the cloud in the game \(\mathbf {G_{F}}\) described in the next subsection.

Fig. 1.
figure 1

The CloudControl game. The FlipIt game models the interaction between an attacker and a cloud administrator for control of the cloud. The outcome of this game determines the type of the cloud in a signaling game in which the cloud conveys commands to the robot or device. The device then decides whether to accept these commands or rely on its own lower-level control. The FlipIt and signaling games are played concurrently.

2.2 FlipIt Game for Cloud Control

The basic version of FlipIt [20]Footnote 2 is played in continuous time. Assume that the defender controls the resource - here, the cloud - at \(t=0\). Moves for both players obtain control of the cloud if it is under the other player’s control. In this paper, we limit our analysis to periodic strategies, in which the moves of the attacker and the moves of the defender are both spaced equally apart, and their phases are chosen randomly from a uniform distribution. Let \(f_{\mathcal {A}}\in \mathbb {R}_{+}\) and \(f_{\mathcal {D}}\in \mathbb {R}_{+}\) (where \(\mathbb {R}_{+}\) represents non-negative real numbers) denote the attack and renewal frequencies, respectively.

Players benefit from controlling the cloud, and incur costs from moving. Let \(w_{\mathcal {X}}\left( t\right) \) denote the average proportion of the time that player \(\mathcal {X}\in \left\{ \mathcal {D},\mathcal {A}\right\} \) has controlled the cloud up to time t. Denote the number of moves up to t per unit time of player \(\mathcal {X}\) by \(z_{\mathcal {X}}\left( t\right) \). Let \(\alpha _{\mathcal {D}}\) and \(\alpha _{\mathcal {A}}\) represent the costs of each defender and attacker move. In the original formulation of FlipIt, the authors consider a fixed benefit for controlling the cloud. In our formulation, the benefit depends on the equilibrium outcomes of the signaling game \(\mathbf {G_{S}}\). Denote these equilibrium utilities of \(\mathcal {D}\) and \(\mathcal {A}\) by \(\bar{u}_{\mathcal {D}}^{S*}\) and \(\bar{u}_{\mathcal {A}}^{S*}\). These give the expected benefit of controlling the cloud. Finally, let \(u_{\mathcal {D}}^{F}\left( t\right) \) and \(u_{\mathcal {A}}^{F}\left( t\right) \) denote the time-averaged benefit of \(\mathcal {D}\) and \(\mathcal {A}\) up to time t in \(\mathbf {G_F}\). Then

$$\begin{aligned} u_{\mathcal {X}}^{F}\left( t\right) =\bar{u}_{\mathcal {X}}^{S*}w_{\mathcal {X}}\left( t\right) -\alpha _{\mathcal {X}}z_{\mathcal {X}}\left( t\right) ,\,\mathcal {X}\in \left\{ \mathcal {D},\mathcal {A}\right\} , \end{aligned}$$
(3)

and, as time continues to evolve, the average benefits over all time become

$$\begin{aligned} \underset{t\rightarrow \infty }{\liminf }\,\bar{u}_{\mathcal {X}}^{S*}w_{\mathcal {X}}\left( t\right) -\alpha _{\mathcal {X}}z_{\mathcal {X}}\left( t\right) ,\,\mathcal {X}\in \left\{ \mathcal {D},\mathcal {A}\right\} . \end{aligned}$$
(4)

We next express these expected utilities over all time as a function of periodic strategies that \(\mathcal {D}\) and \(\mathcal {A}\) employ. Let \(\bar{u}_{\mathcal {X}}^{F}:\,\mathbb {R}_{+}\times \mathbb {R}_{+}\rightarrow \mathbb {R}\), \(\mathcal {X}\in \left\{ \mathcal {D},\mathcal {A}\right\} \) be expected utility functions such that \(\bar{u}_{\mathcal {D}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}\right) \) and \(\bar{u}_{\mathcal {A}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}\right) \) give the average utility to \(\mathcal {D}\) and \(\mathcal {A}\), respectively, when they play with frequencies \(f_{\mathcal {D}}\) and \(f_{\mathcal {A}}\). If \(f_{\mathcal {D}}\ge f_{\mathcal {A}} > 0\), it can be shown that

$$\begin{aligned} \bar{u}_{\mathcal {D}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}\right) =\bar{u}_{\mathcal {D}}^{S*}\left( 1-\frac{f_{\mathcal {A}}}{2f_{\mathcal {D}}}\right) -\alpha _{\mathcal {D}}f_{\mathcal {D}}, \end{aligned}$$
(5)
$$\begin{aligned} \bar{u}_{\mathcal {A}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}\right) =\bar{u}_{\mathcal {A}}^{S*}\frac{f_{\mathcal {A}}}{2f_{\mathcal {D}}}-\alpha _{\mathcal {A}}f_{\mathcal {A}}, \end{aligned}$$
(6)

while if \(0 \le f_{\mathcal {D}} < f_{\mathcal {A}}\), then

$$\begin{aligned} \bar{u}_{\mathcal {D}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}\right) =\bar{u}_{\mathcal {D}}^{S*}\frac{f_{\mathcal {D}}}{2f_{\mathcal {A}}}-\alpha _{\mathcal {D}}f_{\mathcal {D}}, \end{aligned}$$
(7)
$$\begin{aligned} \bar{u}_{\mathcal {A}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}\right) =\bar{u}_{\mathcal {A}}^{S*}\left( 1-\frac{f_{\mathcal {D}}}{2f_{\mathcal {A}}}\right) -\alpha _{\mathcal {A}}f_{\mathcal {A}}, \end{aligned}$$
(8)

and if \(f_{\mathcal {A} }=0\), we have

$$\begin{aligned} \bar{u}_{\mathcal {A}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}\right) = 0, \, \, \, \, \bar{u}_{\mathcal {D}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}\right) = \bar{u}_\mathcal {D}^{S*}-\alpha _\mathcal {D}f_\mathcal {D}. \end{aligned}$$
(9)

Equations (5)–(9) with Eq. (1) for \(\bar{u}_{\mathcal {X}}^{S}\), \(\mathcal {X\in \left\{ \mathcal {D},\mathcal {A}\right\} }\) and Eq. (2) for \(\bar{u}_{\mathcal {R}}^{S}\) will be main ingredients in our equilibrium concept in the next section.

3 Solution Concept

In this section, we develop a new equilibrium concept for our CloudControl game \(\mathbf {G{}_{CC}}\). We study the equilibria of the FlipIt and signaling games individually, and then show how they can be related through a fixed-point equation in order to obtain an overall equilibrium for \(\mathbf {G_{CC}}.\)

3.1 Signaling Game Equilibrium

Signaling games are a class of dynamic Bayesian games. Applying the concept of perfect Bayesian equilibrium (as it e.g., [10]) to \(\mathbf {G_{S}}\), we have Definition 1.

Definition 1

Let the functions \(\bar{u}_{\mathcal {X}}^{S}\left( \sigma _{\mathcal {R}}^{S},\sigma _{\mathcal {X}}^{S}\right) ,\,\mathcal {X}\in \left\{ \mathcal {D},\mathcal {A}\right\} \) and \(\bar{u}_{\mathcal {R}}^{S}\left( \sigma _{\mathcal {R}}^{S}\right) \) be formulated according to Eqs. (1) and (2), respectively. Then a perfect Bayesian equilibrium of the signaling game \(\mathbf {G_{S}}\) is a strategy profile \(\left( \sigma _{\mathcal {D}}^{S*},\sigma _{\mathcal {A}}^{S*},\sigma _{\mathcal {R}}^{S*}\right) \) and posterior beliefs \(\mu \left( \bullet \,|\, m\right) \) such that

$$\begin{aligned} \forall \mathcal {X}\in \left\{ \mathcal {D},\mathcal {A}\right\} ,\,\sigma _{\mathcal {X}}^{S*}\left( \bullet \right) \in \underset{\sigma _{\mathcal {X}}^{S}}{\arg \max }\,\bar{u}_{\mathcal {X}}^{S}\left( \sigma _{\mathcal {R}}^{S*},\sigma _{\mathcal {X}}^{S}\right) , \end{aligned}$$
(10)
$$\begin{aligned} \forall m\in M,\,\sigma _{\mathcal {R}}^{S*}\left( \bullet \,|\, m\right) \in \underset{\sigma _{\mathcal {R}}^{S}}{\arg \max }\,\bar{u}_{\mathcal {R}}^{S}\left( \sigma _{\mathcal {R}}^{S}\,|\, m,\mu \left( \bullet \,|\, m\right) \right) , \end{aligned}$$
(11)
$$\begin{aligned} \mu \left( \theta \,|\, m\right) =\frac{1\left\{ \theta =\theta _{\mathcal {A}}\right\} \sigma _{\mathcal {A}}^{S*}\left( m\right) p+1\left\{ \theta =\theta _{\mathcal {D}}\right\} \sigma _{\mathcal {D}}^{S*}\left( m\right) \left( 1-p\right) }{\sigma _{\mathcal {A}}^{S*}\left( m\right) p+\sigma _{\mathcal {D}}^{S*}\left( m\right) \left( 1-p\right) }, \end{aligned}$$
(12)

if \(\sigma _{\mathcal {A}}^{S*}\left( m\right) p+\sigma _{\mathcal {D}}^{S*}\left( m\right) \left( 1-p\right) \ne 0\), and

$$\begin{aligned} \mu \left( \theta \,|\, m\right) =\text {any distribution on }\varTheta , \end{aligned}$$
(13)

if \(\sigma _{\mathcal {A}}^{S*}\left( m\right) p+\sigma _{\mathcal {D}}^{S*}\left( m\right) \left( 1-p\right) =0\).

Next, let \(\bar{u}_{\mathcal {D}}^{S*}\), \(\bar{u}_{\mathcal {A}}^{S*}\), and \(\bar{u}_{\mathcal {R}}^{S*}\) be the utilities for the defender, attacker, and device, respectively, when they play according to a strategy profile \(\left( \sigma _{\mathcal {D}}^{S*},\sigma _{\mathcal {A}}^{S*},\sigma _{\mathcal {R}}^{S*}\right) \) and belief \(\mu \left( \bullet \,|\, m\right) \) that satisfy the conditions for a perfect Bayesian equilibrium. Define a set-valued mapping \(T^{S}:\,\left[ 0,1\right] \rightarrow 2^{\mathcal {U_{D}}\times \mathcal {U}_{A}}\) such that \(T^S\left( p;G_S\right) \) gives the set of equilibrium utilities of the defender and attacker when the prior probabilities are p and \(1-p\) and the signaling game utilities are parameterized by \(G_S\) Footnote 3. We have

$$\begin{aligned} \left\{ \left( \bar{u}_{\mathcal {D}}^{S*},\bar{u}_{\mathcal {A}}^{S*}\right) \right\} =T^S\left( p;G_S\right) . \end{aligned}$$
(14)

We will employ \(T^{S}\) as part of the definition of an overall equilibrium for \(\mathbf {G_{CC}}\) after examining the equilibrium of the FlipIt game.

3.2 FlipIt Game Equilibrium

The appropriate equilibrium concept for the FlipIt game, when \(\mathcal {A}\) and \(\mathcal {D}\) are restricted to periodic strategies, is Nash equilibrium [14]. Definition 2 applies the concept of Nash Equilibrim to \(\mathbf {G_F}\).

Definition 2

A Nash equilibrium of the game \(\mathbf {G_{F}}\) is a strategy profile \(\left( f_{\mathcal {D}}^{*},f_{\mathcal {A}}^{*}\right) \) such that

$$\begin{aligned} f_{\mathcal {D}}^{*}\in \underset{f_{\mathcal {D}}}{\arg \max }\,\bar{u}_{\mathcal {D}}^{F}\left( f_{\mathcal {D}},f_{\mathcal {A}}^{*}\right) , \end{aligned}$$
(15)
$$\begin{aligned} f_{\mathcal {A}}^{*}\in \underset{f_{\mathcal {A}}}{\arg \max }\,\bar{u}_{\mathcal {D}}^{F}\left( f_{\mathcal {D}}^{*},f_{\mathcal {A}}\right) , \end{aligned}$$
(16)

where \(\bar{u}_{\mathcal {D}}^{F}\) and \(\bar{u}_{\mathcal {A}}^{F}\) are computed by Eqs. (5) and (6) if \(f_{\mathcal {D}}\ge f_{\mathcal {A}}\) and Eqs. (7) and (8) if \(f_{\mathcal {D}}\le f_{\mathcal {A}}\).

To find an overall equilibrium of \(\mathbf {G_{CC}}\), we are interested in the proportion of time that \(\mathcal {A}\) and \(\mathcal {D}\) control the cloud. As before, denote these proportions by p and \(1-p\), respectively. These proportions (as in [6]) can be found from the equilibrium frequencies by

$$\begin{aligned} p={\left\{ \begin{array}{ll} 0,\text { if } &{} f_{\mathcal {A}}=0\\ \frac{f_{\mathcal {A}}}{2f_{\mathcal {D}}},\text { if } &{} f_{\mathcal {D}}\ge f_{\mathcal {A}}>0\\ 1-\frac{f_{\mathcal {D}}}{2f_{\mathcal {A}}},\text { if } &{} f_{\mathcal {A}}>f_{\mathcal {D}}\ge 0 \end{array}\right. } \end{aligned}$$
(17)

Let \(G_F\) parameterize the FlipIt game. Now, we can define a mapping \(T^{F}:\,\mathcal {U_{D}}\times \mathcal {U_{A}}\rightarrow \left[ 0,1\right] \) such that the expression \(T^{F}\left( \bar{u}_{\mathcal {D}}^{S*},\bar{u}_{\mathcal {A}}^{S*};G_F\right) \) gives the proportion of time that the attacker controls the cloud in equilibrium from the values of controlling the cloud for the defender and the attacker. This mapping gives

$$\begin{aligned} p=T^{F}\left( \bar{u}_{\mathcal {D}}^{S*},\bar{u}_{\mathcal {A}}^{S*};G_F\right) . \end{aligned}$$
(18)

In addition to interpreting p as the proportion of time that the attacker controls the cloud, we can view it as the likelihood that, at any random time, the cloud will be controlled by the attacker. Of course, this is precisely the value p of interest in \(\mathbf {G_{S}}\). Clearly, \(\mathbf {G_{F}}\) and \(\mathbf {G_{S}}\) are coupled by Eqs. (14) and (18). These two equations specify the overall equilibrium for the CloudControl game \(\mathbf {G_{CC}}\) through a fixed-point equation, which we describe next.

3.3 Gestalt Equilibrium of \(\mathbf {G_{CC}}\)

When the CloudControl game \(\mathbf {G_{CC}}\) is in equilibrium the mapping from the parameters of \(\mathbf {G_{S}}\) to that game’s equilibrium and the mapping from the parameters of \(\mathbf {G_{F}}\) to that game’s equilibrium are simultaneously satisfied as shown in Fig. 2. Definition 3 formalizes this equilibrium, which we call Gestalt equilibrium.

Fig. 2.
figure 2

\(\mathbf {G_S}\) and \(\mathbf {G_F}\) interact because the utilities in the FlipIt game are derived from the output of the signaling game, and the output of the FlipIt game is used to define prior probabilities in the signaling game. We call the fixed-point of the composition of these two relationships a Gestalt equilibrium.

Definition 3

(Gestalt Equilibrium). The cloud control ratio \(p^{\dagger }\in \left[ 0,1\right] \) and equilibrium signaling game utilities \(\bar{u}_{\mathcal {D}}^{S\dagger }\) and \(\bar{u}_{\mathcal {A}}^{S\dagger }\) constitute a Gestalt equilibrium of the game \(\mathbf {G_{CC}}\) composed of coupled games \(\mathbf {G_{S}}\) and \(\mathbf {G_{F}}\) if the two components of Eq. (19) are simultaneously satisfied.

$$\begin{aligned} \left( \bar{u}_{\mathcal {D}}^{S\dagger },\bar{u}_{\mathcal {A}}^{S\dagger }\right) \in T^{S}\left( p^{\dagger };G_S\right) , \, \, \, \, \, p^{\dagger }=T^{F}\left( \bar{u}_{\mathcal {D}}^{S\dagger },\bar{u}_{\mathcal {A}}^{S\dagger };G_F\right) \end{aligned}$$
(19)

In short, the signaling game utilities \(\left( \bar{u}_{\mathcal {D}}^{S\dagger },\bar{u}_{\mathcal {A}}^{S\dagger }\right) \) must satisfy the fixed-point equation

$$\begin{aligned} \left( \bar{u}_{\mathcal {D}}^{S\dagger },\bar{u}_{\mathcal {A}}^{S\dagger }\right) \in T^S\left( T^F\left( \bar{u}_{\mathcal {D}}^{S\dagger },\bar{u}_{\mathcal {A}}^{S\dagger };G_F\right) ;G_S\right) \end{aligned}$$
(20)

In this equilibrium, \(\mathcal {A}\) receives \(\bar{u}_{\mathcal {A}}^{F}\) according to Eq. (6), Eq. (8), or Eq. (9), \(\mathcal {D}\) receives \(\bar{u}_{\mathcal {D}}^{F}\) according to Eq. (5), Eq. (7), or Eq. (9), and \(\mathcal {R}\) receives \(\bar{u}_{\mathcal {R}}^{S}\) according to Eq. (2).

Solving for the equilibrium of \(\mathbf {G_{CC}}\) requires a fixed-point equation essentially because the games \(\mathbf {G_{F}}\) and \(\mathbf {G_{S}}\) are played according to prior committment. Prior commitment specifies that players in \(\mathbf {G_{S}}\) do not know the outcome of \(\mathbf {G_{F}}\). This structure prohibits us from using a sequential concept such as sub-game perfection and suggests instead a fixed-point equation.

4 Analysis

In this section, we analyze the game proposed in Sect. 2 based on our solution concept in Sect. 3. First, we analyze the signaling game and calculate the corresponding equilibria. Then, we solve the \(\mathtt {FlipIt}\) game for different values of expected payoffs resulting from signaling game. Finally, we describe the solution of the combined game.

4.1 Signaling Game Analysis

The premise of \(\mathbf {G_{CC}}\) allows us to make some basic assumptions about the utility parameters that simplifies the search for equilibria. We expect these assumptions to be true across many different contexts.

(A1) :

\(u_{\mathcal {R}}(\theta _{\mathcal {D}},m_{L},a_{T})>u_{\mathcal {R}}(\theta _{\mathcal {D}},m_{L},a_{N})\): It is beneficial for the receiver to trust a low risk message from the defender.

(A2) :

\(u_{\mathcal {R}}(\theta _{\mathcal {A}},m_{H},a_{T})<u_{\mathcal {R}}(\theta _{\mathcal {A}},m_{H},a_{N})\): It is harmful for the receiver to trust a high risk message from the attacker.

(A3) :

\(\forall {m,m'}\in {M},\; u_{\mathcal {A}}(m,a_{T})>u_{\mathcal {A}}(m',a_{N})\) and \(\forall {m,m'}\in {M}\;, u_{\mathcal {D}}(m,a_{T})>u_{\mathcal {D}}(m',a_{N})\): Both types of sender prefer that either of their messages is trusted rather than that either of their messages is rejected.

(A4) :

\(u_{\mathcal {A}}(m_{H},a_{T})>u_{\mathcal {A}}(m_{L},a_{T})\): The attacker prefers an outcome in which the receiver trusts his high risk message to an outcome in which the receiver trusts his low risk message.

Pooling equilibria of the signaling game differ depending on the prior probabilities p and \(1-p\). Specifically, the messages on which \(\mathcal {A}\) and \(\mathcal {D}\) pool and the equilibrium action of \(\mathcal {R}\) depend on quantities in Eqs. (21) and (22) which we call trust benefits.

$$\begin{aligned} TB_{H}\left( p\right) =\begin{array}{c} p\left[ u_{\mathcal {R}}\left( \theta _{\mathcal {A}},m_{H},a_{T}\right) -u_{\mathcal {R}}\left( \theta _{\mathcal {A}},m_{H},a_{N}\right) \right] \\ +\left( 1-p\right) \left[ u_{\mathcal {R}}\left( \theta _{\mathcal {D}},m_{H},a_{T}\right) -u_{\mathcal {R}}\left( \theta _{\mathcal {D}},m_{H},a_{N}\right) \right] \end{array} \end{aligned}$$
(21)
$$\begin{aligned} TB_{L}\left( p\right) =\begin{array}{c} p\left[ u_{\mathcal {R}}\left( \theta _{\mathcal {A}},m_{L},a_{T}\right) -u_{\mathcal {R}}\left( \theta _{\mathcal {A}},m_{L},a_{N}\right) \right] \\ +\left( 1-p\right) \left[ u_{\mathcal {R}}\left( \theta _{\mathcal {D}},m_{L},a_{T}\right) -u_{\mathcal {R}}\left( \theta _{\mathcal {D}},m_{L},a_{N}\right) \right] \end{array} \end{aligned}$$
(22)

\(TB_{H}\left( p\right) \) and \(TB_{L}\left( p\right) \) give the benefit of trusting (compared to not trusting) high and low messages, respectively, when the prior probability is p. These quantities specify whether \(\mathcal {R}\) will trust a message that it receives in a pooling equilibrium. If \(TB_H\left( p\right) \) (respectively, \(TB_L\left( p\right) \)) is positive, then, in equilibrium, \(\mathcal {R}\) will trust all messages when the senders pool on \(m_H\) (respectively, \(m_L\)).

We illustrate the different possible combinations of \(TB_{H}\left( p\right) \) and \(TB_{L}\left( p\right) \) in the quadrants of Fig. 3. The labeled messages and actions for the sender and receiver, respectively, in each quadrant denote these pooling equilibria. These pooling equilibria apply throughout each entire quadrant. Note that we have not listed the requirements on belief \(\mu \) here. These are addressed in the Appendix A.2, and become especially important for various equilibrium refinement procedures.

The shaded regions of Fig. 3 denote additional special equilibria which only occur under the additional parameter constraints listed within the regions. (The geometrical shapes of the shaded regions are not meaningful, but their overlap and location relative to the four quadrants are accurate.) The dotted and uniformly shaded zones contain equilibria similar to those already denoted in the equilibria for each quadrant, except that they do not require restrictions on \(\mu \). The zone with horizontal bars denotes the game’s only separating equilibrium. It is a rather unproductive one for \(\mathcal {D}\) and \(\mathcal {A}\), since their messages are not trusted. (See the derivation in Appendix A.1.) The equilibria depicted in Fig. 3 will become the basis of analyzing the mapping \(T^{S}\left( p;G_S\right) \), which will be crucial for forming our fixed-point equation that defines the Gestalt equilibrium. Before studying this mapping, however, we first analyze the equilibria of the FlipIt game on its own.

Fig. 3.
figure 3

The four quadrants represent parameter regions of \(\mathbf {G_S}\). The regions vary based on the types of pooling equilibria that they support. For instance, quadrant IV supports a pooling equilibrium in which \(\mathcal {A}\) and \(\mathcal {D}\) both send \(m_H\) and \(\mathcal {R}\) plays \(a_N\), as well as a pooling equilibrium in which \(\mathcal {A}\) and \(\mathcal {D}\) both send \(m_L\) and \(\mathcal {R}\) plays \(a_T\). The shaded regions denote special equilibria that occur under further parameter restrictions.

4.2 FlipIt Analysis

In this subsection, we calculate the Nash equilibrium in the FlipIt game. Equations (5)–(9) represent both players’ utilities in FlipIt game. The solution of this game is similar to what has presented in [6, 20], except that the reward of controlling the resource may vary. To calculate Nash equilibrium, we normalize both players’ benefit with respect to the reward of controlling the resource. For different cases, the frequencies of move at Nash equilibrium are:

\(\bullet \,\,\,\dfrac{\alpha _{\mathcal {D}}}{{\bar{u}_{\mathcal {D}}^{S*}}}<\dfrac{\alpha _{\mathcal {A}}}{{\bar{u}_{\mathcal {A}}^{S*}}}\) and \({\bar{u}_{\mathcal {A}}^{S*}},{\bar{u}_{\mathcal {D}}^{S*}}>0\):

$$\begin{aligned} f_{\mathcal {D}}^{*}=\dfrac{{\bar{u}_{\mathcal {A}}^{S*}}}{2\alpha _{\mathcal {A}}},\,\, f_{\mathcal {A}}^{*}=\dfrac{\alpha _{\mathcal {D}}}{2\alpha _{\mathcal {A}}^{2}}\times \dfrac{({\bar{u}_{\mathcal {A}}^{S*}})^{2}}{{\bar{u}_{\mathcal {D}}^{S*}}}, \end{aligned}$$
(23)

\(\bullet \,\,\,\dfrac{\alpha _{\mathcal {D}}}{{\bar{u}_{\mathcal {D}}^{S*}}}>\dfrac{\alpha _{\mathcal {A}}}{{\bar{u}_{\mathcal {A}}^{S*}}}\) and \({\bar{u}_{\mathcal {A}}^{S*}},{\bar{u}_{\mathcal {D}}^{S*}}>0\):

$$\begin{aligned} f_{\mathcal {D}}^{*}=\dfrac{\alpha _{\mathcal {A}}}{2\alpha _{\mathcal {D}}^{2}}\times \dfrac{({\bar{u}_{\mathcal {D}}^{S*}})^{2}}{{\bar{u}_{\mathcal {A}}^{S*}}},\,\, f_{\mathcal {A}}^{*}=\dfrac{{\bar{u}_{\mathcal {D}}^{S*}}}{2\alpha _{\mathcal {D}}}, \end{aligned}$$
(24)

\(\bullet \,\,\,\dfrac{\alpha _{\mathcal {D}}}{{\bar{u}_{\mathcal {D}}^{S*}}}=\dfrac{\alpha _{\mathcal {A}}}{{\bar{u}_{\mathcal {A}}^{S*}}}\) and \({\bar{u}_{\mathcal {A}}^{S*}},{\bar{u}_{\mathcal {D}}^{S*}}>0\):

$$\begin{aligned} f_{\mathcal {D}}^{*}=\dfrac{{\bar{u}_{\mathcal {A}}^{S*}}}{2\alpha _{\mathcal {A}}},\,\, f_{\mathcal {A}}^{*}=\dfrac{{\bar{u}_{\mathcal {D}}^{S*}}}{2\alpha _{\mathcal {D}}}, \end{aligned}$$
(25)

\(\bullet \,\,\,{\bar{u}_{\mathcal {A}}^{S*}}\le 0\):

$$\begin{aligned} f_{\mathcal {D}}^{*}=f_{\mathcal {A}}^{*}=0, \end{aligned}$$
(26)

\(\bullet \,\,\,{\bar{u}_{\mathcal {A}}^{S*}}>0\) and \({\bar{u}_{\mathcal {D}}^{S*}}\le 0\):

$$\begin{aligned} f_{\mathcal {D}}^{*}=0\,\,\,\,\, f_{\mathcal {A}}^{*}=0^{+}. \end{aligned}$$
(27)

In the case that \(\bar{u}_{\mathcal {A}}^{S*}\le 0\), the attacker has no incentive to attack the cloud. In this case, the defender need not move since we assume that she controls the cloud initially. In the case that \(\bar{u}_{\mathcal {A}}^{S*}>0\) and \({\bar{u}_{\mathcal {D}}^{S*}}\le 0\), only the attacker has an incentive to control the cloud. We use \(f_{\mathcal {A}}^{*}=0^{+}\) to signify that the attacker moves only once. Since the defender never moves, the attacker’s single move is enough to retain control of the cloud at all times.

Next, we put together the analysis of \(\mathbf {G_{S}}\) and \(\mathbf {G_{F}}\) in order to study the Gestalt equilibria of the entire game.

4.3 \(\mathbf {G_{CC}}\) Analysis

To identify the Gestalt Equilibrium of \(\mathbf {G_{CC}}\), it is necessary to examine the mapping \(T^S\left( p;G_S\right) \) for all \(p\in {\left[ 0,1\right] }\). As noted in Sect. 4.1, this mapping depends on \(TB_H\left( p\right) \) and \(TB_L\left( p\right) \). From assumptions A1-A4, it is possible to verify that \(\left( TB_{L}\left( 0\right) ,TB_{H}\left( 0\right) \right) \) must fall in Quadrant I or Quadrant IV and that \(\left( TB_{L}\left( 1\right) ,TB_{H}\left( 1\right) \right) \) must lie in Quadrant III or Quadrant IV. There are numerous ways in which the set \(\left( TB_{L}\left( p\right) ,TB_{H}\left( p\right) \right) ,\, p\in \left[ 0,1\right] \) can transverse different parameter regions. Rather than enumerating all of them, we consider one here.

Fig. 4.
figure 4

For the parameter values overlayed on the figure, as p ranges from 0 to 1, \(TB_H\left( p\right) \) and \(TB_L\left( p\right) \) move from Quadrant I to Quadrant IV. The equilibria supported in each of these quadrants, as well as the equilibria supported on the interface between them, are presented in Table 1.

Consider parameters such that \(TB_{L}\left( 0\right) ,TB_{H}\left( 0\right) >0\) and \(TB_{L}\left( 1\right) >0\) but \(TB_{H}\left( 1\right) <0\) Footnote 4. This leads to an \(\mathscr {L}\) that will traverse from Quadrant I to Quadrant IV. Let us also assume that \(u_{\mathcal {D}}\left( m_{L},a_{T}\right) <u_{\mathcal {D}}\left( m_{H},a_{T}\right) \), so that Equilibrium 5 is not feasible. In Fig. 4, we give specific values of parameters that satisfy these conditions, and we plot \(\left( TB_L\left( p\right) ,TB_H\left( p\right) \right) \) for \(p\in \left[ 0,1\right] \). Then, in Table 1, we give the equilibria in each region that the line segment traverses. The equilibrium numbers refer to the derivations in the Appendix A.2.

Table 1. Signaling game equilibria by region for a game that traverses between Quadrant I and Quadrant IV. Some of the equilibria are feasible only for constrained beliefs \(\mu \), specified in Appendix A.2. We argue that the equilibria in each region marked by (*) will be selected.
Fig. 5.
figure 5

\(T^F\) and \(T^S\) are combined on a single set of axis. In \(T^S\) (the solid line), the independent variable is on the horizontal axis. In \(T^F\) (the dashed line), the independent variable is on the vertical axis. The intersection of the two curves represents the Gestalt equilibrium.

If p is such that the signaling game is played in Quadrant I, then both senders prefer pooling on \(m_{H}\). By the first mover advantage, they will select Equilibrium 8. On the border between Quadrants I and IV, \(\mathcal {A}\) and \(\mathcal {D}\) both prefer an equilibrium in which \(\mathcal {R}\) plays \(a_{T}\). If they pool on \(m_{L}\), this is guaranteed. If they pool on \(m_{H}\), however, \(\mathcal {R}\) receives equal utility for playing \(a_{T}\) and \(a_{N}\); thus, the senders cannot guarantee that the receiver will play \(a_{T}\). Here, we assume that the senders maximize their worst-case utility, and thus pool on \(m_{L}\). This is Equilibrium 3. Finally, in Quadrant IV, both senders prefer to be trusted, and so select Equilibrium 3. From the table, we can see that the utilities will have a jump at the border between Quadrants I and IV. The solid line in Fig. 5 plots the ratio \(\bar{u}_{\mathcal {A}}^{S*}/\bar{u}_{\mathcal {D}}^{S*}\) of the utilities as a function of p.

Next, consider the mapping \(p=T^{F}\left( \bar{u}_{\mathcal {D}}^{S*},\bar{u}_{\mathcal {A}}^{S*}\right) \). As we have noted, p depends only on the ratio \(\bar{u}_{\mathcal {A}}^{S*}/\bar{u}_{\mathcal {D}}^{S*}\) Footnote 5. Indeed, it is continuous in that ratio when the outcome at the endpoints is appropriately defined. This mapping is represented by the dashed line in Fig. 5, with the independent variable on the vertical axis.

We seek a fixed-point, in which \(p=T^{F}\left( \bar{u}_{\mathcal {D}}^{S*},\bar{u}_{\mathcal {A}}^{S*}\right) \) and \(\left( \bar{u}_{\mathcal {D}}^{S*},\bar{u}_{\mathcal {A}}^{S*}\right) =T^{S}\left( p\right) \). This shown by the intersection of the solid and dashed curves plotted in Fig. 5. At these points, the mappings between the signaling and the FlipIt games are mutually satisfied, and we have a Gestalt equilibrium.Footnote 6

5 Cloud Control Application

In this section, we describe one possible application of our model: a cyber-physical system composed of autonomous vehicles with some on-board control but also with the ability to trust commands from the cloud. Access to the cloud can offer automated vehicles several benefits [12]. First, it allows access to massive computational resources - i.e., infrastructure as a service (IaaS). (See [5].) Second, it allows access to large datasets. These datasets can offer super-additive benefits to the sensing capabilities of the vehicle itself, as in the case of the detailed road and terrain maps that automated cars such as those created by Google and Delphi combine with data collected by lidar, radar and vision-based cameras [1, 11]. Third, interfacing with the cloud allows access to data collected or processed by humans through crowd-sourcing applications; consider, for instance, location-based services [17, 18] that feature recommendations from other users. Finally, the cloud can allow vehicles to collectively learn through experience [12].

Attackers may attempt to influence cloud control of the vehicle through several means. In one type of attack, adversaries may be able to steal or infer cryptographic keys that allow them authorization into the network. These attacks are of the complete compromise and stealth types that are studied in the FlipIt framework [6, 20] and thus are appropriate for a CloudControl game. FlipIt also provides the ability to model zero-day exploits, vulnerabilities for which a patch is not currently available. Each of these types of attacks on the cloud pose threats to unmanned vehicle security and involve the complete compromise and steathiness that motivate the FlipIt framework.

5.1 Dynamic Model for Cloud Controlled Unmanned Vehicles

In this subsection, we use a dynamic model of an autonomous car to illustrate one specific context in which a cloud-connected device could be making a decision of whether to trust the commands that it would receive or to follow its own on-board control.

Fig. 6.
figure 6

A bicycle model is a type of representation of vehicle steering control. Here, \(\delta \left( t\right) \) is used to denote the angle between the orientation of the front wheel and the heading \(\theta \left( t\right) \). The deviation of the vehicle from a straight line is given by \(z\left( t\right) \)

We consider a car moving in two-dimensional space with a fixed speed \(v_0\) but with steering that can be controlled. (See Fig. 6, which illustrates the “bicycle model” of steering control from [3].) For simplicity, assume that we are interested in the car’s deviation from a straight line. (This line might, e.g., run along the center of the proper driving lane.) Let \(z\left( t\right) \) denote the car’s vertical distance from the horizontal line, and let \(\theta \left( t\right) \) denote the heading of the car at time t. The state of the car can be represented by a two-dimensional vector \(w\left( t\right) \triangleq \left[ \begin{array}{cc} z\left( t\right)&\theta \left( t\right) \end{array}\right] ^{T}\). Let \(\delta \left( t\right) \) denote the angle between the orientation of the front wheel - which implements steering - and the orientation of the length of the car. We can consider \(\delta \left( t\right) \) to be the input to the system. Finally, let \(y\left( t\right) \) represent a vector of outputs available to the car’s control system. The self-driving cars of both Google and Delphi employ radar, lidar, and vision-based cameras for localization. Assume that these allow accurate measurement of both states, such that \(y_{1}\left( t\right) =z\left( t\right) \) and \(y_{2}\left( t\right) =\theta \left( t\right) \). If the car stays near \(w\left( t\right) =\left[ \begin{array}{cc} 0&0\end{array}\right] ^{T}\), then we can approximate the system with a linear model. Let a and b denote the distances from the rear wheel to the center of gravity and the rear wheel to the front wheel of the car, respectively. Then the linearized system is given in [3] by the equations:

$$\begin{aligned} \frac{d}{dt}\left[ \begin{array}{c} z\left( t\right) \\ \theta \left( t\right) \end{array}\right] =\left[ \begin{array}{cc} 0 &{} v_{0}\\ 0 &{} 0 \end{array}\right] \left[ \begin{array}{c} z\left( t\right) \\ \theta \left( t\right) \end{array}\right] +\left[ \begin{array}{c} \frac{av_{0}}{b}\\ \frac{v_{0}}{b} \end{array}\right] \delta \left( t\right) , \end{aligned}$$
(28)
$$\begin{aligned} \left[ \begin{array}{c} y_{1}\left( t\right) \\ y_{2}\left( t\right) \end{array}\right] =\left[ \begin{array}{cc} 1 &{} 0\\ 0 &{} 1 \end{array}\right] \left[ \begin{array}{c} z\left( t\right) \\ \theta \left( t\right) \end{array}\right] \end{aligned}$$
(29)

5.2 Control of Unmanned Vehicle

Assume that the unmanned car has some capacity for automatic control without the help of the cloud, but that the cloud typically provides more advanced navigation.

Specifically, consider a control system onboard the unmanned vehicle designed to return it to the equilibrium \(w\left( t\right) =\left[ \begin{array}{cc} 0&0\end{array}\right] ^{T}\). Because the car has access to both of the states, it can implement a state-feedback control. Consider a linear, time-invariant control of the form

$$\begin{aligned} \delta _{car}\left( t\right) =-\left[ \begin{array}{cc} k_{1}&k_{2}\end{array}\right] \left[ \begin{array}{c} z\left( t\right) \\ \theta \left( t\right) \end{array}\right] . \end{aligned}$$
(30)

This proportional control results in the closed-loop system

$$\begin{aligned} \frac{d}{dt}\left[ \begin{array}{c} z\left( t\right) \\ \theta \left( t\right) \end{array}\right] =\left( \left[ \begin{array}{cc} 0 &{} v_{0}\\ 0 &{} 0 \end{array}\right] -\left[ \begin{array}{c} \frac{av_{0}}{b}\\ \frac{v_{0}}{b} \end{array}\right] \left[ \begin{array}{cc} k_{1}&k_{2}\end{array}\right] \right) \left[ \begin{array}{c} z\left( t\right) \\ \theta \left( t\right) \end{array}\right] \end{aligned}$$
(31)

The unmanned car \(\mathcal {R}\) may also elect to obtain data or computational resources from the cloud. Typically, this additional access would improve the control of the car. The cloud administrator (defender \(\mathcal {D}\)), however, may issue faulty commands or there may be a breakdown in communication of the desired signals. In addition, the cloud may be compromised by \(\mathcal {A}\) in a way that is stealthy. Because of these factors, \(\mathcal {R}\) sometimes benefits from rejecting the cloud’s command and relying on its own navigational abilities. Denote the command issued by the cloud at time t by \(\delta _{cloud}\left( t\right) \in { \delta _{\mathcal {A}}\left( t\right) , \delta _{\mathcal {D}}\left( t\right) }\), depending on who controls the cloud. With this command, the system is given by

$$\begin{aligned} \frac{d}{dt}\left[ \begin{array}{c} z\left( t\right) \\ \theta \left( t\right) \end{array}\right] =\left[ \begin{array}{cc} 0 &{} v_{0}\\ 0 &{} 0 \end{array}\right] \left[ \begin{array}{c} z\left( t\right) \\ \theta \left( t\right) \end{array}\right] +\left[ \begin{array}{c} \frac{av_{0}}{b}\\ \frac{v_{0}}{b} \end{array}\right] \delta _{cloud}\left( t\right) . \end{aligned}$$
(32)

5.3 Filter for High Risk Cloud Commands

In cloud control of an unmanned vehicle, the self-navigation state feedback input given by \(\delta _{car}\left( t\right) \) in Eq. (30) represents the control that is expected by the vehicle given its state. If the signal from the cloud differs significantly from the signal given by the self-navigation system, then the vehicle may classify the message as “high-risk.” Specifically, define a difference threshold \(\tau \), and let

$$\begin{aligned} m={\left\{ \begin{array}{ll} m_{H},\text { if } &{} \left| \delta _{cloud}\left( t\right) -\delta _{car}\left( t\right) \right| >\tau \\ m_{L},\text { if } &{} \left| \delta _{cloud}\left( t\right) -\delta _{car}\left( t\right) \right| \le \tau \end{array}\right. } \end{aligned}$$
(33)

Equation (33) translates the actual command from the cloud (controlled by \(\mathcal {D}\) or \(\mathcal {A}\)) into a message in the cloud signaling game.

Equations (31) and (32) give the dynamics of the unmanned car electing to trust and not trust the cloud. Based on these equations, Fig. 7 illustrates the combined self-navigating and cloud controlled system for vehicle control.

Fig. 7.
figure 7

Block-diagram model for unmanned vehicle navigation control. At any time, the vehicle uses strategy \(\sigma ^S_\mathcal {R}\) to decide whether to follow its own control or the control signal from the cloud, which may be \(\delta _\mathcal {A}\) or \(\delta _\mathcal {D}\), depending on the probabilities p, \(1-p\) with which \(\mathcal {A}\) and \(\mathcal {D}\) control the cloud. Its own control signal, \(\delta _{car}\), is obtained via feedback control.

6 Conclusion and Future Work

In this paper, we have proposed a general framework for the interaction between an attacker, cloud administrator/defender, and cloud-connected device. We have described the struggle for control of the cloud using the FlipIt game and the interaction between the cloud and the connected device using a traditional signaling game. Because these two games are played by prior commitment, they are coupled. We have defined a new equilibrium concept - i.e., Gestalt equilibrium, which defines a solution to the combined game using a fixed-point equation. After illustrating various parameter regions under which the game may be played, we solved the game in a sample parameter region. Finally, we showed how the framework may be applied to unmanned vehicle control.

Several directions remain open for future work. First, the physical component of the cyber-physical system can be further examined. Tools from optimal control such as the linear-quadratic regulator could offer a rigorous framework for defining the costs associated with the physical dynamic system, which in turn would define the payoffs of the signaling game. Second, future work could search for conditions under which a Gestalt equilibrium of the CloudControl game is guaranteed to exist. Finally, devices that use this framework should be equipped to learn online. Towards that end, a learning algorithm could be developed that is guaranteed to converge to the Gestalt equilibrium. Together with the framework developed in the present paper, these directions would help to advance our ability to secure cloud-connected and cyber-physical systems.