Keywords

7.1 Introduction

Multiple Input Multiple Output (MIMO) techniques [1,2,3,4] have played an important part in the remarkable explosion of wireless communication systems in the past two decades, and MIMO is now an integral component of all recent WiFi and cellular standards e.g. 802.11ac [2] and 5G [3]. Specifically, MIMO communication involves transceivers equipped with multiple antennae. This allows directional transmission, permitting control of the interference produced by wireless transmitters, and orders of magnitude increases in energy and spectral efficiency.

Yet the applicability of MIMO is in practice severely limited by constraints such as form factors, and the size and the number of antennae that can be realistically supported. An attractive alternative, springing from the pioneering work of [5], is distributed MIMO (DMIMO), [6] where instead of deploying centralized antenna systems, groups of single antennae transceivers collaborate to form a virtual antennae system that mimics the functionality of a centralized multi-antennae system. Studied extensively by theoreticians, until recently, this concept has been dismissed by practioners as being beyond the realm of practicality for several reasons. Among these are issues engendered by uncertain geometries and the fact that unlike centralized systems, by its very definition, each constituent of a DMIMO system carries its own clock and oscillator. These suffer significant and rapid drift modeled as a second-order stochastic process, [7] that induces DMIMO units to migrate from synchrony to complete incoherence within mere hundreds of milliseconds.

Over the last decade several authors, [8,9,10,11,12,13,14,15,16,17,18,19,20,21,22], have sought to realize this decades-old concept, by addressing two salient components of DMIMO: distributed beamforming, [8,9,10,11,12,13,14,15,16] and distributed nullforming, [17,18,19,20,21,22]. In the former, groups of transmitters, collaborate to form a beam of maximum possible power using constructive interference at a receiver. In the latter their transmissions cancel each other at the designated receiver. On the other hand [21] simultaneously forms nulls at some receivers and beams at others. Physical implementations of both beam and nullforming algorithms on software defined radio (SDR) platforms, are described in [12, 13, 20]. Apart from being important constituents of the overall DMIMO architecture, beamforming is key to power efficient communication, just as nullforming has applications in interference avoidance for increased spatial spectrum reuse [23], cognitive radio [24] and cyber security [25].

This brings us to the role of feedback control in these schemes. Ultimately in all these applications all transmissions must arrive at their target receiver synchronized in frequency and with precise phase, and for nullforming, amplitude relationships. Uncertain geometries and mobility, make it impossible for the transmitters to determine the phases, and amplitudes of their transmission at the receiver. Thus all the references cited above rely on some receiver cooperation. This takes the form of feedback from the receiver to the transmitters, and possibly between the transmitters, using which the transmitters must adjust the phase, amplitude and frequency of their transmissions to achieve synchronization at the receiver. A key issue, of course, is what type of feedback is needed, the minimum information exchange required to achieve one’s objective.

In this light, among the DMIMO papers we have cited here, barring [11, 19], all assume, prior frequency synchronization, presumably through information exchange among the transmitters. The earliest among the DMIMO papers, [8], assumes prior frequency synchronization, and requires that the receiver feed back to each transmitter a separate feedback signal throughout its operation. Such an algorithm is thus not scalable.

A breakthrough idea introduced in [9], and used in several subsequent papers, is the notion of common aggregate feedback. This involves the receiver broadcasting to all the transmitters either the complex baseband version of its total received signal or some attribute there of. In either case the burden of repeatedly feeding back a separate signal to each transmitter is alleviated. To wit [9], assumes prior frequency synchronization among the nodes and executes a 1-bit feedback algorithm to achieve beamforming. The algorithm itself is in the mold of randomized ascent. Each transmitter updates its phase according to a preselected distribution. The 1-bit feedback is whether or not the net received power declines as a result of these updates. If the power declines, the updates are discarded. Else they are retained. Under mild assumptions on the distribution from which the phase updates are chosen, this algorithm is provably convergent.

In this paper we consider distributed nullforming without prior frequency synchronization with only phase and frequency updates. The algorithm we formulate is akin to Lyapunov redesign in the controls and adaptive systems literature, [26, 27]. We observe that distributed nullforming algorithms can be found in [17,18,19,20,21,22]. Each of these, however assumes prior frequency synchronization at the outset of operation, presumably through information exchange among transmitters. While this is reasonable, oscillator frequencies also undergo drift modeled as Brownian motion, albeit at orders of magnitude slower rates than oscillator phases. On the other hand drift in oscillator frequencies has a more dramatic impact on performance than has phase drift. Furthermore Doppler shift occurs at receivers, thus receiver feedback should be used to guide the adjustment of frequencies at transmitters.

Nullforming is fundamentally more difficult than is beamforming, [22], for two reasons. First, it is much more sensitive to phase drift. Because of this, a 1-bit algorithm like in [9] is unable to adjust quickly enough to achieve a decent null. Second, while beamforming only requires frequency and phase alignment at the receiver, nullforming requires the precise control of the amplitude and phase of the radio-frequency signal transmitted by each cooperating transmitter to ensure that they cancel each other. Accordingly, [17, 18], requires that in addition to the common aggregate feedback of the total baseband received signal, each transmitter must also learn the channel state information of every transmitter to the receiver. This latter requirement is substantially relaxed in [22], where each channel, at the point of setup requires only the knowledge of its channel to the receiver. Simulations in [22] show that the gradient descent algorithm it employs is robust to substantial channel estimation errors, while capable of tracking significant oscillator drift.

This paper extends [22] to the case where the transmitters even if initially frequency synchronized undergo small frequency drifts. These frequency drifts even if small, can destroy a good null very quickly. As in [22], we assume that each user knows its complex channel gain to the receiver. Unlike [17, 18] it does not have the CSI seen by the other transmitters. Like [22], the receiver feeds back the net baseband signal. In [22] this feedback is used by each transmitter to perform a distributed gradient descent minimization of the total received power. The minimization is distributed, as each transmitter can perform it using only the aggregate feedback and its own complex channel gain to the receiver. Similarly, in this paper each transmitter adjusts its phase and frequency knowing only its CSI to the receiver and the aggregate feedback signal, to asymptotically drive the received signal to zero. However, we show that the lack of frequency synchronization precludes the use of gradient descent. Instead a Lyapunov redesign is needed.

We observe that [11, 19] also eschew the assumption of prior frequency synchronization. Among these, [11] uses a 1-bit type algorithm to beamform. We have however, discovered a conceptual error in the algorithm. On the other hand the preliminary paper, without proofs, [19], has the important difference that it critically assumes that each transmitter equalizes its complex channel gain. In this paper we equalize only the phase and not the magnitude of the channel. This is in the vein of [22] and permits a key application of distributed nullforming: Namely, permitting transmission at full power, thus maximizing incoherent power pooling gains, while protecting the null target. As explained in [22] this opens up the prospect of both Space Division Multiple Access (SDMA) and cyber security.

Section 7.2 provides the algorithm. Section 7.3 has some preliminary analysis. Section 7.4 presents a stability analysis. Received power which must be minimized is a nonconvex function of the transmitter phases and frequencies. Unsurprisingly our Lyapunov redesign yields a distributed algorithm that has multiple stationery points/trajectories that may not correspond to a minimum. Yet we show that only those stationary points are locally stable that do correspond to global minima. Section 7.5 is the conclusion.

7.2 The Algorithm

We now describe a scalable gradient descent algorithm for distributed nullforming in a node. As in [22] and unlike [17, 18], we assume that at the beginning of a nullforming epoch, each transmitter has access only to its own complex channel gain to the receiver, using which it equalizes the phase rather than also the magnitude of the channel to the receiver. Assume there are N transmitter nodes.

Denote \({\theta }_i(t)\) to be the equalized phase of the i-th transmitter and \(r_i>0\) is the magnitude of the received signal. Assume that \(\omega _i(t)\), is a frequency offset of the i-th transmitter, from a nominal frequency to which each transmitter should ideally have been synchronized, but oscillator drift prevents the maintenance of such synchronization.

Then the complex baseband signal received at the cooperating receiver is:

$$\begin{aligned} s(t)=R(t)+jI(t) \end{aligned}$$
(7.1)

where

$$\begin{aligned} R(t)=\sum _{i=1}^N r_i\cos (\omega _i(t)t+{\theta }_i(t)) \end{aligned}$$
(7.2)

and

$$\begin{aligned} I(t)=\sum _{i=1}^N r_i\sin (\omega _i(t)t+{\theta }_i(t)). \end{aligned}$$
(7.3)

As in [22], the transmission process is thus, equivalent to each transmitter transmitting, the phase equalized complex baseband signal, \(e^{j((\omega _i(t)t+{\theta }_i(t))}\). The baseband signal the receiver sees is

$$\begin{aligned} s(t)=\sum _{i=1}^Nr_ie^{j((\omega _i(t)t+{\theta }_i(t))} \end{aligned}$$
(7.4)

see (7.1). Indeed it is s(t) that the receiver broadcasts to all transmitters, which they must use to adjust their frequency and phase. Define \({\theta }(t)=[{\theta }_1(t), \cdots , {\theta }_N(t)]^\top \) and \(\omega (t)=[\omega _1(t),\cdots , \omega _N(t)]^\top \). We note the key difference with [19], which apart from not providing any proofs, assumes that all \(r_i=1\). The received power is:

$$\begin{aligned} J({\theta },\omega ,t)=I^2(t)+R^2(t). \end{aligned}$$
(7.5)

The receiver feeds back s(t). The i-th transmitter uses s(t) to adjust its \(\theta _i(t)\) and \(\omega _i(t)\) to asymptotically force \(J({\theta },\omega ,t)\) to zero. In reality both the adjustment and feedback will be discrete time. However, should the continuous time algorithm be uniformly asymptotically stable (u.a.s), then averaging theory, [28] ensures that for high enough feedback rates, and small enough gains, an obvious discretized version will also be u.a.s.

As noted above, [22] where \(\omega =0\), uses a gradient descent algorithm. The resulting J is autonomous in [22]. The frequency offsets here make the cost function nonautonomous, as it may change its value even if the phases and frequencies are not adjusted. To amplify this point observe that a pure gradient descent algorithm would take the form:

$$\begin{aligned} \dot{{\theta }}(t)=-\dfrac{\partial J({\theta },\omega ,t)}{\partial \theta }; ~\dot{\omega }(t)=-\dfrac{\partial J({\theta },\omega ,t)}{\partial \omega } . \end{aligned}$$
(7.6)

Now observe that under (7.6)

$$\begin{aligned} \dot{J}= & {} \dfrac{\partial J}{\partial t}+\dfrac{\partial J}{\partial {\theta }}\dot{{\theta }}+\dfrac{\partial J}{\partial \omega }\dot{\omega }\\= & {} \dfrac{\partial J}{\partial t}-\left\| \dfrac{\partial J}{\partial {\theta }}\right\| ^2-\left\| \dfrac{\partial J}{\partial \omega }\right\| ^2. \end{aligned}$$

Should \(\dfrac{\partial J}{\partial t}\) be zero, as is the case is in [22], this guarantees a nonincreasing \(J({\theta },\omega ,t)\). However, with \(\omega _i\) potentially nonzero under (7.2), \(\dfrac{\partial J}{\partial t}\) need not be zero, preventing guaranteed decrescence of \(J({\theta },\omega ,t)\). Thus a Lyapunov redesign of the nullforming algorithm is needed.

In Sect. 7.3 we present a Lyapunov function, and show that its decrescence is guaranteed by the following algorithm.

$$\begin{aligned} \dot{{\theta }}= & {} -\frac{\partial J}{\partial {\theta }} -\frac{\omega }{2}\end{aligned}$$
(7.7)
$$\begin{aligned} \dot{\omega }= & {} -\frac{1}{2}\frac{\partial J}{\partial {\theta }} . \end{aligned}$$
(7.8)

Thus the i-th node can implement these as long as it has access to its frequency, phase, CSI to the receiver and the common feedback signals I(t) and R(t) permitting a totally distributed implementation, as

$$\begin{aligned} \frac{\partial J}{\partial {\theta }_i}=-2R(t) r_i\sin (\omega _i(t)t+{\theta }_i(t)) +2I(t) r_i \cos (\omega _i(t)t+{\theta }_i(t)), \end{aligned}$$
(7.9)

and

$$\begin{aligned} \frac{\partial J}{\partial \omega }=t \frac{\partial J}{\partial {\theta }}. \end{aligned}$$
(7.10)

Also observe that unlike (7.6) there is an additional corrective term \(\frac{\omega }{2}\) in (7.7), that accounts for the frequency offsets. In keeping with our mandate for phase and frequency only updates the gains \(r_i\) are not updated.

7.3 Preliminaries of the Stability Analysis

In this section, we present certain preliminary results that among other things show the uniform convergence of the gradient of J with respect to \({\theta }\), and explore the properties of the stationary points of (7.7, 7.8).

But first, a result used in [29].

Lemma 7.1

Suppose on a closed interval \( {\mathscr {I}} \subset \mathbb {R}\) of length T, a signal \(w: {\mathscr {I}} \rightarrow \mathbb {R}\) is twice differentiable and for some \(\varepsilon \) and \(M'\)

$$ |w(t)| \le \varepsilon _1 \text { and } |\ddot{w}(t)|\le M'\;\;\; \forall \;\;\; t\in {\mathscr {I}}. $$

Then for some M independent of \(\varepsilon _1\), \({\mathscr {I}}\) and \(M'\), and \(M'' = \max (M', 2 \varepsilon _1 T^{-2})\) one has:

$$ |\dot{w}(t)| \le M(M''\varepsilon _1)^{1/2}\;\;\; \forall \;\;\; t\in {\mathscr {I}}. $$

We begin by deriving a preliminary relation.

$$\begin{aligned} \frac{\partial J}{\partial t}= & {} -2R(t) \sum _{i=1}^N\omega _i(t)r_i \sin (\omega _i(t)t+{\theta }_i(t))+2I(t) \sum _{i=1}^Nr_i\omega _i(t) \cos (\omega _i(t)t+{\theta }_i(t)),\nonumber \\= & {} \omega ^\top (t)\frac{\partial J}{\partial {\theta }}. \end{aligned}$$
(7.11)

We now show that under (7.7, 7.8) there is a Lyapunov function that is nonincreasing and the gradient of J with respect to \({\theta }\) converges uniformly to zero.

Lemma 7.2

Consider (7.7, 7.8) with (7.27.5) initial time \(t_0\ge 0\). Then the following hold:

  1. (a)

    For all \(t\ge t_0\),

    $$\begin{aligned} V(t)=J(t) +\frac{\Vert \omega (t)\Vert ^2}{2} \end{aligned}$$
    (7.12)

    is nonincreasing.

  2. (b)

    The following occurs uniformly in \(t_0\).

    $$\begin{aligned} \lim _{t\rightarrow \infty } \frac{\partial J}{\partial {\theta }}(t)=0. \end{aligned}$$
    (7.13)

Proof

Because of (7.5), (7.97.10) and (7.7, 7.8), there holds:

$$\begin{aligned} \dot{J}+\frac{d}{dt}\left\{ \frac{\omega ^\top \omega }{2}\right\}= & {} \frac{\partial J}{\partial t} +\dot{{\theta }}^\top \frac{\partial J}{\partial {\theta }}+\dot{\omega }^\top \frac{\partial J}{\partial \omega }+\omega ^\top \dot{\omega }\nonumber \\= & {} \omega ^\top \frac{\partial J}{\partial {\theta }}- \left\| \frac{\partial J}{\partial {\theta }} \right\| ^2-\frac{\omega ^\top }{2}\frac{\partial J}{\partial {\theta }}\nonumber \\&-\frac{\omega ^\top }{2}\frac{\partial J}{\partial {\theta }} -\frac{t}{2} \left\| \frac{\partial J}{\partial {\theta }} \right\| ^2\nonumber \\= & {} -\left( 1+\frac{t}{2}\right) \left\| \frac{\partial J}{\partial {\theta }} \right\| ^2 \end{aligned}$$
(7.14)

Consequently (a) holds. Further, \(\omega \) is uniformly bounded. Consequently from (7.9) there is an \(M_1\), independent of \(t_0\), such that for all \(t\ge t_0\)

$$ \left\| \frac{d}{dt}\left\{ \frac{\partial J}{\partial {\theta }}(t)\right\} \right\| \le M_1. $$

Equally, there exists an \(M_2\), also independent of \(t_0\), such that for all \(t\ge t_0\),

$$ \left\| \frac{\partial J}{\partial {\theta }}(t)\right\| \le M_2. $$

Further, since the initial time \(t_0\ge 0\), from (7.14) and V(t) is nonnegative, one obtains that for all \(t\ge t_0\):

$$\begin{aligned} \int _{t_0}^t \left\| \frac{\partial J}{\partial {\theta }} (s)\right\| ^2 \text { d}s\le & {} \int _{t_0}^t\left( 1+\frac{s}{2}\right) \left\| \frac{\partial J}{\partial {\theta }} (s)\right\| ^2 \text { d}s\\\le & {} V(t_0). \end{aligned}$$

Thus, for every \(\varepsilon >0\), there exists a T independent of \(t_0\) such that for all \(t\ge T+t_0\),

$$ \int _{T+t_0}^t \left\| \frac{\partial J}{\partial {\theta }} (s)\right\| ^2 \text { d}s \le \varepsilon . $$

Then from Lemma 7.1, there is a K independent of \(t_0\) such that for all \(\varepsilon >0\), there exists a T independent of \(t_0\) such that for all \(t\ge T+t_0\),

$$ \left\| \frac{\partial J}{\partial {\theta }} (s)\right\| ^2 \le \varepsilon , \;\;\forall s\ge T+t_0. $$

Thus indeed (b) holds uniformly in \(t_0\).

Thus, (7.7, 7.8) converges uniformly to a trajectory where:

$$\begin{aligned} \frac{\partial J}{\partial {\theta }}=0. \end{aligned}$$
(7.15)

Because of (7.2, 7.3), (7.9) there holds:

$$\begin{aligned} \dfrac{1}{2}\frac{\partial J}{\partial {\theta }_i}= & {} -R(t) \sum _{i=1}^Nr_i \sin (\omega _i(t)t+{\theta }_i(t))+I(t) \sum _{i=1}^Nr_i \cos (\omega _i(t)t+{\theta }_i(t))\nonumber \\= & {} \sum _{i=1}^N\sum _{l=1}^Nr_ir_l\left\{ \cos (\omega _i(t)t+{\theta }_i(t))\sin (\omega _l(t)t+{\theta }_l(t))\right. \nonumber \\&-\left. \sin (\omega _i(t)t+{\theta }_i(t))\cos (\omega _l(t)t+{\theta }_l(t)) \right\} \nonumber \\= & {} \sum _{i=1}^N\sum _{l=1}^Nr_ir_l\sin ( (\omega _l(t)-\omega _i(t))t+{\theta }_l(t)-{\theta }_i(t)). \end{aligned}$$
(7.16)

Thus (7.15) also implies that for some constant \(\omega ^*\) on this trajectory the frequency offsets

$$\begin{aligned} \omega _i=\omega ^*, ~\forall i\in \{1,\cdots ,N\}. \end{aligned}$$
(7.17)

Observe from (7.5), (7.2) and (7.3),

$$\begin{aligned} J({\theta },\omega ,t)= & {} \left( \sum _{i=1}^N r_i\cos (\omega _i(t)t+{\theta }_i(t)) \right) ^2 +\left( \sum _{i=1}^N r_i\sin (\omega _i(t)t+{\theta }_i(t)) \right) ^2\nonumber \\= & {} \sum _{i=1}^Nr_i^2+2\sum _{i=1}^N \sum _{{\mathop {l\ne i}\limits ^{l=1}}}^N r_ir_l \cos ((\omega _l(t)-\omega _i(t))t +{\theta }_l(t)-{\theta }_i(t)) \qquad \qquad \end{aligned}$$
(7.18)

This shows that these are in fact manifolds rather than points.

We now turn to a major nontrivial consequence of having potentially nonunity \(r_i\), a problem absent in [19] where all \(r_i\) are 1. There are sets of positive \(r_i\) for which a null may not be possible. It is thus useful to first characterize conditions on the \(r_i\) that ensure that a null is possible. The theorem below provides this characterization. We note it is similar to a result in [22] where the \(\omega _i\) are all fixed to zero. The theorem also characterizes the global minimum value

$$\begin{aligned} J^*=\max _{t\ge 0}\min _{{\theta }, \omega }J({\theta },\omega ,t). \end{aligned}$$
(7.19)

Theorem 7.1

Consider \(J({\theta },\omega ,t)\) defined in (7.2, 7.3, 7.5) and \(J^*\) as in (7.19), with all \(r_i>0\). Without loss of generality label \(r_i\), so that \(r_i\ge r_{i+1}\). Then the following hold:

  1. (i)

    \(J^*=0\) iff

    $$\begin{aligned} r_1\le \sum _{i=2}^N r_i. \end{aligned}$$
    (7.20)
  2. (ii)

     If (7.20) is violated

    $$\begin{aligned} J^* =\left( r_1-\sum _{i=2}^N r_i\right) ^2. \end{aligned}$$
    (7.21)

Proof

Observe \(J({\theta },\omega ,t)=|s(t)|^2\), with s(t) defined in (7.4). Suppose (7.20) is violated. Clearly, under (7.4), as \(r_i>0\),

$$ J({\theta },\omega ,t)\ge \left( r_1-\sum _{i=2}^N r_i\right) ^2. $$

Thus \(J^*>0\). Further this minimum is attained by choosing \(\omega _i=\omega _l\), for all il, \({\theta }_1=0\) and \({\theta }_i=\pi , ~\forall ~ i>1\). This proves (ii) and the “only if” part of (i).

To prove the “if” part of (i), set all \(\omega _i\) to zero and assume that (7.20) holds. We use induction. Consider \(N=2\). Then \(r_1=r_2\). Thus with \({\theta }_1=0, {\theta }_2=\pi \), \(J({\theta },0,t)=\left| r_1-r_2 \right| ^2=0, ~ \forall t\). Now suppose the result holds for some \(N=n\ge 2\). Consider \(N=n+1\).

Observe with

$$\begin{aligned} J(0,0,t)=\sum _{i=1}^{n+1}r_i>0. \end{aligned}$$
(7.22)

Consider two cases.

Case I: \(r_2>\sum _{i=3}^{n+1} r_i\): In this case, by hypothesis \(0< r_2 - \sum _{i=3}^{n+1}r_i < r_1\). Thus,

$$ J(\pi [1,0,[1,\cdots , 1]]^T,0,t)=-r_1+r_2 - \sum _{i=3}^{n+1}r_i <0. $$

As for every t, \(J(\theta ,0,t)\) is continuous in \({\theta }\), and (7.22) holds, there exist a \({\theta }\) for which \(J(\theta ,0,t)=0, ~\forall t\). Thus \(J^*=0\).

Case II: \(r_2\le \sum _{i=3}^{n+1} r_i\): From the induction hypothesis, there exist \(\hat{\theta }_2,\cdots , \hat{\theta }_{n+1}\) such that \(\sum _{i=2}^{n+1}r_i e^{j\hat{{\theta }}_i}=0<r_1.\) Moreover, by assumption \( \sum _{i=2}^{n+1}r_i \ge r_1\). Since \(\left| \sum _{i=2}^{n+1}r_i e^{j{\theta }_i}\right| \) is continuous in \({\theta }\), moving continuously between \([{\theta }_2,\cdots ,{\theta }_{n+1}]\) between 0 and \([\hat{\theta }_2,\cdots , \hat{\theta }_{n+1}]\) one can find a set of phases \([{\theta }_2^*,\cdots ,{\theta }_{n+1}^*]\) such that \(\left| \sum _{i=2}^{n+1}r_i e^{j{\theta }_i^*}\right| =r_1.\) Thus, for some \(\delta \), \( \sum _{i=2}^{n+1}r_i e^{j{\theta }_i^*}=r_1e^{j\delta }\). Then \(J([\pi +\delta ,{\theta }_2^*,\cdots ,{\theta }_{n+1}^*]^T,0,t)=0 \), completing the proof.

Returning to the stationary trajectories, given by (7.15) and (7.17), some of these correspond to the desired null, or in their absence (7.19). Others do not, and will be called false stationary points. We show below that the latter are locally unstable and are thus, rarely attained, and even if attained cannot be practically maintained as noise would drive the trajectories away from them. Thus, by showing the local stability of the global minimum, we will demonstrate the practical uniform convergence of the algorithm to the global minimum.

In the stationary trajectory, (7.15) and (7.17) are not nulls, i.e., \(s(t)\ne 0\), then from (7.9) for all il, and \(t>0\), \(\tan ( \omega ^*t+{\theta }_i)=\tan ( \omega ^*t+{\theta }_l)\). Consequently on stationary trajectories that are not nulls,

$$\begin{aligned} {\theta }_i-{\theta }_l =k_{il} \pi , ~\forall \{i,l\}\subset \{1,\cdots ,N\} \end{aligned}$$
(7.23)

where \(k_{il}\) are integers.

The local analysis of these stationary trajectories, will require the examination of the Hessian with respect to \({\theta }\) on these trajectories. Consider again (7.16). Due to (7.17), along (7.15) and (7.17) the il-th element of the Hessian along the stationary trajectory is given by:

$$\begin{aligned} \left. \left[ H(\theta ) \right] _{il}\right| _{\omega _i=\omega _l}=\frac{\partial ^2 J({\theta }) }{\partial {\theta }_i \partial {\theta }_l}=\left\{ \begin{array}{ll} -2\sum _{i\ne l}r_ir_l\cos ( \theta _i-\theta _l) &{} i=l\\ 2r_{il}\cos ( \theta _i-\theta _l) &{} i\ne l \end{array} \right. \end{aligned}$$
(7.24)

7.4 Stability Analysis

Armed with the preliminary results in Sect. 7.3 we now complete our stability analysis. Lemma 7.2 shows that uniform convergence to a stationary trajectory is guaranteed. Some of these trajectories correspond to a null. Other do not. In this section, we show that only those that correspond to a null are locally stable. The others are not. Consequently, one is assured of practical uniform convergence in the sense that stationary trajectories that do not correspond to the desired nulls if at all attained, cannot be practically maintained. Thus for all practical purposes, the algorithm defined in (7.7, 7.8) achieves a desired null.

First we demonstrate the local instability of spurious stationary trajectories. To this end we present need the following lemma.

Lemma 7.3

The linear system below with scalar \(a>0\) is unstable:

$$\begin{aligned} \dot{\eta }=\left[ \begin{matrix} a &{} ta+\frac{1}{2}\\ \frac{a }{2}&{}\frac{at}{2} \end{matrix}\right] \eta \end{aligned}$$
(7.25)

Proof

Consider the initial condition \(\eta (0)=[0,1]^\top \). Then it is evident that both elements of the state are nonegative for all \(t>0\). Then the first element of the state vector is

$$ \eta _1(t) \ge e^{at}. $$

Thus the system is unstable.

Consider next the Hessian with respect to \({\theta }\) at a critical trajectory. As given in (7.24) this is identical to the Hessian studied in [22]. From [22] we have the following lemma.

Lemma 7.4

Assume all \(r_i>0\). Consider a false stationary manifold i.e., a stationary manifold on which \(J\ne 0\) and \(J\ne J^*\). Then \(\left. H(\theta )\right| _{\forall i, \omega _i=\omega ^*}\) has at least one negative eigenvalue.

We now prove that a false stationary trajectory is unstable.

Theorem 7.2

Assume all \(r_i>0\). Consider (7.7, 7.8), and a stationary manifold defined by (7.15), (7.17) such that along this trajectory \(J\ne 0\) and \(J\ne J^*\). Then this trajectory is unstable.

Proof

Observe, that for all il

$$ \frac{\partial ^2 J({\theta }) }{\partial {\theta }_i \partial \omega _l}=t\frac{\partial ^2 J({\theta }) }{\partial {\theta }_i \partial {\theta }_l} $$

(7.7, 7.8), linearized around (7.15), (7.17) is given by:

$$\begin{aligned} \dot{x}=\left[ \begin{matrix} -H({\theta }) &{} -t H({\theta })+\frac{I}{2}\\ -\frac{H(\theta )}{2}&{} -\frac{t}{2}H({\theta }) \end{matrix}\right] x. \end{aligned}$$
(7.26)

In view of Lemma 7.4 and the symmetry of \(H({\theta })\), there is an orthogonal matrix \(\varOmega \) and real \(\lambda _i\), with \(\lambda _1>0\), such that with

$$ \varLambda =\text { diag } \{- \lambda _1,\cdots , \lambda _N\}, $$
$$ H({\theta })=\varOmega \varLambda \varOmega ^T $$

Define \(\beta =\text { diag } \{\varOmega , \varOmega \} x\). Then the linearized systems is equivalent to:

$$ \dot{{\beta }}=\left[ \begin{matrix} -\varLambda &{} -t \varLambda +\frac{I}{2}\\ -\frac{\varLambda }{2}&{} -\frac{t}{2}\varLambda \end{matrix}\right] {\beta }. $$

Then instability follows from Lemma 7.3.

Thus indeed false stationary manifolds are unstable. To complete the result, it suffices to show that the global minima, nulls or otherwise, are locally stable. As the Hessian is indefinite this is a nonhyperbolic system and a linearization approach will be inconclusive. Thus we use a more direct approach to proving local stability.

Because of (7.18) and (7.23) at false stationary points the cost function takes the values

$$\begin{aligned} \sum _{i=1}^Nr_i^2+2\sum _{i=1}^N \sum _{{\mathop {l\ne i}\limits ^{l=1}}}^N r_ir_l \mu _{il},~ \mu _{il}\in \{ -1,1 \}. \end{aligned}$$
(7.27)

Assume again \(r_i\ge r_{i+1}>0\). With \(r=[r_1\cdots , r_N]^T\), define:

$$\begin{aligned} f(r)=\left\{ \begin{matrix} \min \left\{ \left\{ \sum \limits _{i=1}^Nr_i^2+2\sum \limits _{i=1}^N \sum \limits _{{\mathop {l\ne i}\limits ^{l=1}}}^N r_ir_l \mu _{il}|\mu _{il}\in \{ -1,1 \} \right\} \setminus \{0\}\right\} &{} \text { if } r_1\le \sum \limits _{i=2}^N r_i\\ \min \left\{ \left\{ \sum \limits _{i=1}^Nr_i^2+2\sum \limits _{i=1}^N \sum \limits _{{\mathop {l\ne i}\limits ^{l=1}}}^N r_ir_l \mu _{il}|\mu _{il}\in \{ -1,1 \} \right\} \setminus \{r_1- \sum \limits _{i=2}^N r_i\}\right\} &{} \text { else } \\ \end{matrix} \right. \end{aligned}$$
(7.28)

In other words f(r) is the smallest value that J can take at a false stationary point. We can now prove local stability of the null manifold.

Theorem 7.3

Suppose \(r_i\ge r_{i+1}>0\) and with f(r) defined in (7.28). Then with positive initial time \(t_0\), (7.7, 7.8) uniformly converges to a global minimum of \(J({\theta },\omega ,t)\) if \(V(t_0)<f(r)\). Further for a constant \(\omega ^*\) and all \(i\in \{1,\cdots , N\}\),

$$\begin{aligned} \lim _{t\rightarrow \infty } \omega _i(t)=\omega ^* \end{aligned}$$
(7.29)

Proof

Item (b) of Lemma 7.2 holds uniformly in \(t_0\). Further for all \(t\ge t_0\)

$$ J({\theta }(t),\omega (t),t) \le V(t)\le V(t_0)<f(r). $$

As Lemma 7.2 guarantees convergence to a stationary manifold, and only stationary manifold at which \(J({\theta }(t),\omega (t),t)<f(r)\), convergence occurs to a global minimum. Finally, (7.29) follows from (7.15) and (7.8).

As convergence to a stationary manifold is guaranteed, and all false stationary points are locally unstable, this thus proves practical uniform convergence to a global minimum. Observe in addition the transmitters attain frequency consensus.

In principle the phases need not converge to a fixed point. However, there is a subtlety. On a stationary trajectory, (7.7) and (7.8) reduces to, (7.17) and for each i,

$$ \dot{{\theta }}_i(t)=-\dfrac{\omega ^*}{2}. $$

Thus for suitable \({\alpha }_i\) the i-th transmitted signal becomes

$$ e^{j(\frac{\omega ^*}{2}t+{\alpha }_i)}. $$

thus effectively, the attained phases are constants, and de facto frequency synchronization obtains.

7.5 Conclusion

We have provided a new distributed nullforming algorithm that is guaranteed to achieve a null, through phase and frequency adapatation. Unlike [22] this paper does not assume prior frequency synchronization. Its effect though is the achievement of frequency synchronization. That all this can be achieved with no cooperation between the transmitters, a feedback of the aggregate received signal by the receiver, and the knowledge to each transmitter of only its CSI is in our opinion an intriguing fact. Also intriguing is the fact that an otherwise nonconvex cost function has the attractive property that all its local minima are global minima. Further studies that delineate the minimum information needed to achieve distributed nullforming are warranted.