Keywords

1 Introduction

The landmark systems realization theory of Mayo and Antoulas [19] shows how to construct a dynamical system that interpolates tangential frequency domain measurements of a multi-input, multi-output system. Central to this development is the matrix pencil \(z\mathbb {L}-{\mathbb {L}_s}\) composed of Loewner and shifted Loewner matrices \(\mathbb {L}\) and \({\mathbb {L}_s}\) that encode the interpolation data. When this technique is used for exact system recovery (as opposed to data-driven model reduction), the eigenvalues of this pencil match the poles of the transfer function of the original system. However other spectral properties, including the sensitivity of the eigenvalues to perturbation, can differ greatly, depending on the location of the interpolation points in the complex plane relative to the system poles, and how the interpolation points are partitioned. Since one uses \(z\mathbb {L}-{\mathbb {L}_s}\) to learn about the original system, such subtle differences matter.

Pseudospectra are sets in the complex plane that contain the eigenvalues but provide additional insight about the sensitivity of those eigenvalues to perturbation and the transient behavior of the underlying dynamical system. While most often used to analyze single matrices, pseudospectral concepts have been extended to matrix pencils (generalized eigenvalue problems).

This introductory note shows several ways to use pseudospectra to investigate spectral questions involving Loewner pencils derived from system realization problems. Using simple examples, we explore the following questions.

  • How do the locations of the interpolation points and their partition into “left” and “right” points affect the sensitivity of the eigenvalues of \(z\mathbb {L}-{\mathbb {L}_s}\)?

  • Do solutions to the dynamical system \(\mathbb {L}\dot{\mathbf{x}}(t) = {\mathbb {L}_s}\mathbf{x}(t)\) mimic solutions to the original system \(\mathbf{E}\dot{\mathbf{x}}(t) = \mathbf{A}\mathbf{x}(t)\), especially in the transient regime? Does this agreement depend on the interpolation points?

  • How do noisy measurements affect the eigenvalues of \(z\mathbb {L}-{\mathbb {L}_s}\)?

We include an algorithm for computing pseudospectra of an n-dimensional Loewner pencil in \(\mathcal{O}(n^2)\) operations, improving the \(\mathcal{O}(n^3)\) cost for generic matrix pencils; the appendix gives a MATLAB implementation.

Throughout this note, we use \(\sigma (\cdot )\) to denote the spectrum (eigenvalues) of a matrix or matrix pencil, and \(\Vert \cdot \Vert \) to denote the vector 2-norm and the matrix norm it induces. (All definitions here can readily be adapted to other norms, as needed. The algorithm, however, is designed for use with the 2-norm.)

2 Loewner Realization Theory in a Nutshell

We briefly summarize Loewner realization theory, as developed by Mayo and Antoulas [19]; see also [1]. Consider the linear, time-invariant dynamical system

$$\begin{aligned} \mathbf{E}\dot{\mathbf{x}}(t) = \mathbf{A}\mathbf{x}(t) + \mathbf{B}\mathbf{u}(t), \qquad \mathbf{y}(t) = \mathbf{C}\mathbf{x}(t) \end{aligned}$$

for \(\mathbf{A}, \mathbf{E}\in \mathbb {C}^{n\times n}\), \(\mathbf{B}\in \mathbb {C}^{n\times m}\), and \(\mathbf{C}\in \mathbb {C}^{p\times n}\), with which we associate, via the Laplace transform, the transfer function \(\mathbf{H}(z) = \mathbf{C}(z\mathbf{E}-\mathbf{A})^{-1}\mathbf{B}\).

Given tangential measurements of \(\mathbf{H}(z)\) we seek a realization of the system that interpolates the given data. More precisely, consider the right interpolation data

  • distinct interpolation points \(\lambda _1,\ldots , \lambda _\varrho \in \mathbb {C}\);

  • interpolation directions \(\mathbf{r}_1, \ldots , \mathbf{r}_\varrho \in \mathbb {C}^m\);

  • function values \(\mathbf{w}_1, \ldots , \mathbf{w}_\varrho \in \mathbb {C}^p\);

and left interpolation data

  • distinct interpolation points \(\mu _1,\ldots , \mu _\nu \in \mathbb {C}\);

  • interpolation directions \(\boldsymbol{\ell }_1, \ldots , \boldsymbol{\ell }_\nu \in \mathbb {C}^p\);

  • function values \(\mathbf{v}_1, \ldots , \mathbf{v}_\nu \in \mathbb {C}^m\).

Assume the left and right interpolation points are disjoint, \(\{\lambda _i\}_{i=1}^\varrho \cap \{\mu _j\}_{j=1}^\nu = \emptyset \); (in our examples, all \(\varrho +\nu \) points are distinct).

The interpolation problem seeks matrices \(\widehat{\mathbf{A}}, \widehat{\mathbf{E}}, \widehat{\mathbf{B}}, \widehat{\mathbf{C}}\) for which the transfer function \(\widehat{\mathbf{H}}(z) = \widehat{\mathbf{C}}(z\widehat{\mathbf{E}}-\widehat{\mathbf{A}})^{-1}\widehat{\mathbf{B}}\) interpolates the data: for \(i=1,\ldots , \varrho \) and \(j=1,\ldots , \nu \),

$$ \widehat{\mathbf{H}}(\lambda _i) \mathbf{r}_i = \mathbf{w}_i, \qquad \boldsymbol{\ell }_j^*\widehat{\mathbf{H}}(\mu _j) = \mathbf{v}_j^*. $$

Two structured matrices play a crucial role in the development of Mayo and Antoulas [19]. From the data, construct the Loewner and shifted Loewner matrices

$$\begin{aligned} \mathbb {L}= \left[ \begin{array}{ccc} {\mathbf{v}_{1}^*\mathbf{r}_{1}^{} - \boldsymbol{\ell }_{1}^* \mathbf{w}_{1}^{} \over \mu _{1}-\lambda _{1}} &{} \cdots &{} {\mathbf{v}_{1}^*\mathbf{r}_{\varrho }^{} - \boldsymbol{\ell }_{1}^* \mathbf{w}_{\varrho }^{} \over \mu _{1}-\lambda _{\varrho }} \\ \vdots &{} \ddots &{} \vdots \\ {\mathbf{v}_{\nu }^*\mathbf{r}_{1}^{} - \boldsymbol{\ell }_{\nu }^* \mathbf{w}_{1}^{} \over \mu _{\nu }-\lambda _{1}} &{} \cdots &{} {\mathbf{v}_{\nu }^*\mathbf{r}_{\varrho }^{} - \boldsymbol{\ell }_{\nu }^* \mathbf{w}_{\varrho }^{} \over \mu _{\nu }-\lambda _{\varrho }} \\ \end{array}\right] , \quad {\mathbb {L}_s}= \left[ \begin{array}{ccc} {\mu _{1}^{}\mathbf{v}_{1}^*\mathbf{r}_{1}^{} - \lambda _{1}^{}\boldsymbol{\ell }_{1}^* \mathbf{w}_{1}^{} \over \mu _{1}-\lambda _{1}} &{} \cdots &{} {\mu _{1}^{}\mathbf{v}_{1}^*\mathbf{r}_{\varrho }^{} - \lambda _{\varrho }^{}\boldsymbol{\ell }_{1}^* \mathbf{w}_{\varrho }^{} \over \mu _{1}-\lambda _{\varrho }} \\ \vdots &{} \ddots &{} \vdots \\ {\mu _{\nu }^{}\mathbf{v}_{\nu }^*\mathbf{r}_{1}^{} - \lambda _{1}^{}\boldsymbol{\ell }_{\nu }^* \mathbf{w}_{1}^{} \over \mu _{\nu }-\lambda _{1}} &{} \cdots &{} {\mu _{\nu }^{}\mathbf{v}_{\nu }^*\mathbf{r}_{\varrho }^{} - \lambda _{\varrho }^{}\boldsymbol{\ell }_{\nu }^* \mathbf{w}_{\varrho }^{} \over \mu _{\nu }-\lambda _{\varrho }} \\ \end{array}\right] , \end{aligned}$$
(1)

i.e., the (ij) entries of these \(\nu \times \varrho \) matrices have the form

$$ (\mathbb {L})_{i,j} = {\mathbf{v}_i^*\mathbf{r}_j^{} - \boldsymbol{\ell }_i^* \mathbf{w}_j^{} \over \mu _i-\lambda _j}, \qquad ({{\mathbb {L}_s}})_{i,j} = {\mu _i^{} \mathbf{v}_i^*\mathbf{r}_j^{} - \lambda _j^{} \boldsymbol{\ell }_i^* \mathbf{w}_j^{} \over \mu _i-\lambda _j}.$$

Now collect the data into matrices. The right interpolation points, directions, and data are stored in

$$ {\boldsymbol{\Lambda }}= \left[ \begin{array}{ccc} \lambda _1 \\ &{} \ddots \\ &{} &{} \lambda _\varrho \end{array}\right] \!\in \! \mathbb {C}^{\varrho \times \varrho }, \quad \mathbf{R}= \left[ \begin{array}{ccc} \\ \mathbf{r}_1 &{} \cdots \mathbf{r}_\varrho \\ &{} \end{array}\right] \!\in \! \mathbb {C}^{m\times \varrho }, \quad \mathbf{W}= \left[ \begin{array}{ccc} \\ \mathbf{w}_1 &{} \cdots \mathbf{w}_\varrho \\ &{} \end{array}\right] \!\in \! \mathbb {C}^{p\times \varrho }, $$

while the left interpolation points, directions, and data are stored in

$$ \mathbf{M}= \left[ \begin{array}{ccc} \mu _1\\ &{} \ddots \\ &{} &{} \mu _\nu \end{array}\right] \!\in \! \mathbb {C}^{\nu \times \nu }, \quad \mathbf{L}= \left[ \begin{array}{ccc} \\ \boldsymbol{\ell }_1 &{} \cdots \boldsymbol{\ell }_\nu \\ &{} \end{array}\right] \!\in \! \mathbb {C}^{p\times \nu }, \quad \mathbf{V}= \left[ \begin{array}{ccc} \\ \mathbf{v}_1 &{} \cdots \mathbf{v}_\nu \\ &{} \end{array}\right] \!\in \! \mathbb {C}^{m\times \nu }. $$

2.1 Selecting and Arranging Interpolation Points

As Mayo and Antoulas observe, Sylvester equations connect these matrices:

$$\begin{aligned} \mathbb {L}{\boldsymbol{\Lambda }}- \mathbf{M}\mathbb {L}= \mathbf{L}^*\mathbf{W}-\mathbf{V}^*\mathbf{R}, \qquad {\mathbb {L}_s}{\boldsymbol{\Lambda }}- \mathbf{M}{\mathbb {L}_s}= \mathbf{L}^*\mathbf{W}{\boldsymbol{\Lambda }}-\mathbf{M}\mathbf{V}^*\mathbf{R}. \end{aligned}$$
(2)

Just using the dimensions of the components, note that

$$\begin{aligned} \mathrm{rank}(\mathbf{L}^*\mathbf{W}),\ \mathrm{rank}(\mathbf{L}^*\mathbf{W}{\boldsymbol{\Lambda }}),\le & {} \min \{\nu , \varrho , p\} \\ \mathrm{rank}(\mathbf{V}^*\mathbf{R}),\ \mathrm{rank}(\mathbf{V}^*\mathbf{R}\mathbf{M})\le & {} \min \{\nu , \varrho , m\}. \end{aligned}$$

Thus for modest m and p, the Sylvester equations (2) must have low-rank right-hand sides.Footnote 1 This situation often implies the rapid decay of singular values of solutions to the Sylvester equation [2, 3, 21, 22, 26]. While this phenomenon is convenient for balanced truncation model reduction (enabling low-rank approximations to the controllability and observability Gramians), it is less welcome in the Loewner realization setting, where the rank of \(\mathbb {L}\) should reveal the order of the original system: fast decay of the singular values of \(\mathbb {L}\) makes this rank ambiguous. Since \({\boldsymbol{\Lambda }}\) and \(\mathbf{M}\) are diagonal, they are normal matrices, and hence Theorem 2.1 of [3] gives

$$\begin{aligned} \frac{s_{qk+1}(\mathbb {L})}{s_1(\mathbb {L})} \le \inf _{\phi \in \mathcal{R}_{k,k}} \frac{\max \{ |\phi (\lambda )|: \lambda \in \{\lambda _1, \ldots , \lambda _\varrho \}\}}{\min \{ |\phi (\mu )|: \mu \in \{\mu _1, \ldots , \mu _\nu \}\}}, \end{aligned}$$
(3)

where \(s_j(\cdot )\) denotes the jth largest singular value, \(q = \mathrm{rank}(\mathbf{L}^*\mathbf{W}-\mathbf{V}^*\mathbf{R})\), and \(\mathcal{R}_{k,k}\) denotes the set of irreducible rational functions whose numerators and denominators are polynomials of degree k or less.Footnote 2 The right hand side of (3) will be small when there exists some \(\phi \in \mathcal{R}_{k,k}\) for which all \(|\phi (\lambda _i)|\) are small, while all \(|\phi (\mu _j)|\) are large: a good separation of \(\{\lambda _i\}\) from \(\{\mu _j\}\) is thus sufficient to ensure the rapid decay of the singular values of \(\mathbb {L}\) and \({\mathbb {L}_s}\). (Beckermann and Townsend give an explicit bound for the singular values of Loewner matrices when the interpolation points fall in disjoint real intervals [3, Corollary 4.2].)

Here, if we want the singular values of \(\mathbb {L}\) and \({\mathbb {L}_s}\) to reveal the system’s order (without decay of singular values as an accident of the arrangement of interpolation points), it is necessary for one \(|\phi (\lambda _i)|\) to be about the same size as the smallest value of \(|\phi (\mu _j)|\) for all \(\phi \in \mathcal{R}_{k,k}\). Roughly speaking, we want the left and right interpolation points to be close together (even interleaved). While this arrangement is necessary for slow decay of the singular values, it does not alone prevent such decay, as will be seen in examples in Fig. 4 (since (3) is only an upper bound).

Another heuristic, based on the Cauchy-like structure of \(\mathbb {L}\) and \({\mathbb {L}_s}\), also suggests the left and right interpolation points should be close together. Namely, \(\mathbb {L}\) and \({\mathbb {L}_s}\) are a more general form of the Cauchy matrix \(({\boldsymbol{\mathcal C}})_{i,j} = 1/(\mu _i-\lambda _j)\), whose determinant has the elegant formula (e.g., [16, p. 38])

$$\begin{aligned} \mathrm{det}({\boldsymbol{\mathcal C}}) = \frac{\prod _{1\le i<j \le n}(\mu _j-\mu _i)(\lambda _i-\lambda _j)}{\prod _{1\le i\le j \le n}(\mu _i-\lambda _j)}. \end{aligned}$$
(4)

It is an open question if \(\mathrm{det}(\mathbb {L})\) and \(\mathrm{det}({\mathbb {L}_s})\) have similarly elegant formulas. Nevertheless, \(\mathrm{det}(\mathbb {L})\) and \(\mathrm{det}({\mathbb {L}_s})\) do have the same denominator as \(\mathrm{det}({\boldsymbol{\mathcal C}})\) (which can be checked by recursively subtracting the first row from all other rows when computing the determinant). This observation suggests that to avoid artificially small determinants for \(\mathbb {L}\) and \({\mathbb {L}_s}\) (which, up to sign, are the products of the singular values) it is necessary for the denominator of (4) to be small, and, thus, for the left and right interpolation points to be close together.

In practice, we often start with initial interpolation points \(x_1,\ldots ,x_{2n}\) that we want to partition into left and right interpolation points to form \(\mathbb {L}\) and \({\mathbb {L}_s}\). Our analysis of (4) suggests a simple way to arrange the interpolation points such that the denominator of \(\mathrm{det}(\mathbb {L})\) and \(\mathrm{det}({\mathbb {L}_s})\) is small: relabel the points to satisfy

$$\begin{aligned} x_k = \underset{k \le q \le 2n}{\arg \min }~|x_{k-1}-x_q|, \quad \mathrm{for}~~ k = 2,\ldots ,2n, \end{aligned}$$
(5)
$$ \{\mu _i\} = \{x_1,x_3,\ldots ,x_{2n-1}\}, \quad \{\lambda _j\} = \{x_2,x_4,\ldots ,x_{2n}\}. $$

The greedy reordering in (5) ensures that \(|x_k-x_{k-1}|\) is small and allows us to simply interleave the left and right interpolation points. Moreover, when \(x_1,\ldots ,x_{2n}\) are located on a line, the reordering in (5) simplifies to directly interleaving \(\mu _i\) and \(\lambda _j\) and, thus, it can be skipped. This ordering need not be optimal, as we do not visit all possible combinations of \(\mu _i-\lambda _j\); it simply seeks a partition that yields a large determinant (which must also depend on the interpolation data). We note its simplicity, effectiveness, and efficiency (requiring only \(\mathcal{O}(n^2)\) operations). (For SISO systems, the state-space representation is equivalent to a barycentric form [17, p. 77]. We thank the referee for observing that in barycentric rational Remez approximations [9], a similar interleaving of support points from the set of reference points produces favorable numerical conditioning. Recent work by Gosea and Antoulas gives further numerical evidence for interleaving Loewner points [13].)

2.2 Construction of Interpolants

Throughout we make the fundamental assumptions that for all \(\widehat{z} \in \{\lambda _i\}_{i=1}^\varrho \cup \{\mu _j\}_{j=1}^\nu \),

$$ \mathrm{rank}(\widehat{z}\mathbb {L}-{\mathbb {L}_s}) = \mathrm{rank}\left( \left[ \begin{array}{c}\mathbb {L}\\ {\mathbb {L}_s}\end{array}\right] \right) = \mathrm{rank}\left( \left[ \begin{array}{cc}\mathbb {L}&{\mathbb {L}_s}\end{array}\right] \right) =: r,$$

and we presume the underlying dynamical system is controllable and observable.

When \(r = \nu =\varrho \) is the order of the system, Mayo and Antoulas [19, Lemma 5.1] show that the transfer function \(\widehat{\mathbf{H}}(z) := \widehat{\mathbf{C}}(z\widehat{\mathbf{E}}-\widehat{\mathbf{A}})^{-1}\widehat{\mathbf{B}}\) defined by

$$\begin{aligned} \widehat{\mathbf{E}}= -\mathbb {L}, \quad \widehat{\mathbf{A}}= -{\mathbb {L}_s}, \quad \widehat{\mathbf{B}}= \mathbf{V}^*, \quad \widehat{\mathbf{C}}= \mathbf{W}\end{aligned}$$
(6)

interpolates the \(\varrho +\nu \) data values.

When \(r < \max (\nu ,\varrho )\), fix some \(\widehat{z}\in \{\lambda _i\}_{i=1}^\varrho \cup \{\mu _j\}_{j=1}^\nu \) and compute the (economy-sized) singular value decomposition

$$ \widehat{z}\mathbb {L}-{\mathbb {L}_s}= \mathbf{Y}\boldsymbol{\Sigma }\mathbf{X}^*, $$

with \(\mathbf{Y}\in \mathbb {C}^{\nu \times r}\), \(\boldsymbol{\Sigma }\in \mathbb {R}^{r\times r}\), and \(\mathbf{X}\in \mathbb {C}^{\varrho \times r}\). Then with

$$\begin{aligned} \widehat{\mathbf{E}}= -\mathbf{Y}^*\mathbb {L}\mathbf{X}, \quad \widehat{\mathbf{A}}= -\mathbf{Y}^*{\mathbb {L}_s}\mathbf{X}, \quad \widehat{\mathbf{B}}= \mathbf{Y}^*\mathbf{V}^*, \quad \widehat{\mathbf{C}}= \mathbf{W}\mathbf{X}, \end{aligned}$$
(7)

\(\widehat{\mathbf{H}}(z) := \widehat{\mathbf{C}}(z\widehat{\mathbf{E}}-\widehat{\mathbf{A}})^{-1}\widehat{\mathbf{B}}\) gives an order-r system interpolating the data [19, Theorem 5.1].

3 Pseudospectra for Matrix Pencils

Though introduced decades earlier, in the 1990s pseudospectra emerged as a popular tool for analyzing the behavior of dynamical systems (see, e.g., [30]), eigenvalue perturbations (see, e.g., [4]), and stability of uncertain linear time-invariant (LTI) systems (see, e.g., [15]).

Definition 1

For a matrix \(\mathbf{A}\in \mathbb {C}^{n\times n}\) and \(\varepsilon >0\), the \(\varepsilon \)-pseudospectrum of \(\mathbf{A}\) is

$$\begin{aligned} \sigma _\varepsilon (\mathbf{A}) = \{z \in \mathbb {C} \text{ is } \text{ an } \text{ eigenvalue } \text{ of } \mathbf{A}+\boldsymbol{\Gamma } \text{ for } \text{ some } \boldsymbol{\Gamma }\in \mathbb {C}^{n\times n} \text{ with } \Vert \boldsymbol{\Gamma }\Vert < \varepsilon \}. \end{aligned}$$
(8)

For all \(\varepsilon >0\), \(\sigma _\varepsilon (\mathbf{A})\) is a bounded, open subset of the complex plane that contains the eigenvalues of \(\mathbf{A}\). (A popular variation uses the weak inequality \(\Vert \boldsymbol{\Gamma }\Vert \le \varepsilon \); the strict inequality has favorable properties for operators on infinite-dimensional spaces [5].) Definition 1 motivates pseudospectra via eigenvalues of perturbed matrices. A numerical analyst studying accuracy of a backward stable eigenvalue algorithm might be concerned with \(\varepsilon \) on the order of \(n \Vert \mathbf{A}\Vert \varepsilon _\mathrm{mach}\), where \(\varepsilon _\mathrm{mach}\) denotes the machine epsilon for the floating point system [20]. An engineer or scientist might consider \(\sigma _\varepsilon (\mathbf{A})\) for much larger \(\varepsilon \) values, corresponding to uncertainty in parameters or data that contribute to the entries of \(\mathbf{A}\).

Via the singular value decomposition, one can show that (8) is equivalent to

$$\begin{aligned} \sigma _\varepsilon (\mathbf{A}) = \{z\in \mathbb {C}: \Vert (z\mathbf{I}-\mathbf{A})^{-1}\Vert > 1/\varepsilon \}; \end{aligned}$$
(9)

see, e.g., [29, Chap. 2]. The presence of the resolvent \((z\mathbf{I}-\mathbf{A})^{-1}\) in this definition suggests a connection to the transfer function \(\mathbf{H}(z) = \mathbf{C}(z\mathbf{I}-\mathbf{A})^{-1}\mathbf{B}\) for the system

$$ \dot{\mathbf{x}}(t) = \mathbf{A}\mathbf{x}(t) + \mathbf{B}\mathbf{u}(t), \qquad \mathbf{y}(t) = \mathbf{C}\mathbf{x}(t). $$

Indeed, definition (1) readily leads to bounds on \(\Vert \mathrm{e}^{t\mathbf{A}}\Vert \), and hence transient growth of solutions to \(\dot{\mathbf{x}}(t) = \mathbf{A}\mathbf{x}(t)\); see [29, Part IV].

Various extensions of pseudospectra have been proposed to handle more general eigenvalue problems and dynamical systems; see [7] for a concise survey. The first elaborations addressed the generalized eigenvalue problem \(\mathbf{A}\mathbf{x}= \lambda \mathbf{E}\mathbf{x}\) (i.e., the matrix pencil \(z\mathbf{E}-\mathbf{A}\)) [10, 23, 24]. Here we focus on the definition proposed by Frayssé, Gueury, Nicoud, and Toumazou [10], which is ideally suited to analyzing eigenvalues of nearby matrix pencils. To permit the perturbations to \(\mathbf{A}\) and \(\mathbf{E}\) to be scaled independently, this definition includes two additional parameters, \(\gamma \) and \(\delta \).

Definition 2

Let \(\gamma ,\delta > 0\). For a pair of matrices \(\mathbf{A},\mathbf{E}\in \mathbb {C}^{n\times n}\) and any \(\varepsilon >0\), the \(\varepsilon \)-\((\gamma ,\delta )\)-pseudospectrum \(\sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E})\) of the matrix pencil \(z\mathbf{E}- \mathbf{A}\) is the set

$$\begin{aligned} \sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E})= & {} \{z \in \mathbb {C} \text{ is } \text{ an } \text{ eigenvalue } \text{ of } \text{ the } \text{ pencil } z(\mathbf{E}+\boldsymbol{\Delta }) - (\mathbf{A}+\boldsymbol{\Gamma }) \\&\ \ \ \ \ \text{ for } \text{ some } \boldsymbol{\Gamma }, \boldsymbol{\Delta }\in \mathbb {C}^{n\times n} \text{ with } \Vert \boldsymbol{\Gamma }\Vert< \varepsilon \gamma \text{, } \Vert \boldsymbol{\Delta }\Vert < \varepsilon \delta \}. \nonumber \end{aligned}$$

This definition has been extended to matrix polynomials in [14, 27].

Remark 1

Note that \(\sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E})\) is an open, nonempty subset of the complex plane, but it need not be bounded.

  1. (a)

    If \(z\mathbf{E}-\mathbf{A}\) is a singular pencil (\(\mathrm{rank}(z\mathbf{E}-\mathbf{A})<n\) for all \(z\in \mathbb {C}\)), then \(\sigma (\mathbf{A},\mathbf{E}) = \mathbb {C}\).

  2. (b)

    If \(z\mathbf{E}-\mathbf{A}\) is a regular pencil but \(\mathbf{E}\) is not invertible, then \(\sigma (\mathbf{A},\mathbf{E})\) contains an infinite eigenvalue, and hence \(\sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E})\) is unbounded.

  3. (c)

    If \(\mathbf{E}\) is nonsingular but \(\varepsilon \delta \) exceeds the distance of \(\mathbf{E}\) to singularity (the smallest singular value of \(\mathbf{E}\)), then \(\sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E})\) contains the point at infinity.

Since these pseudospectra can be unbounded, Lavallée [18] and Higham and Tisseur [14] visualize \(\sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E})\) as stereographic projections on the Riemann sphere.

Remark 2

Just as the conventional pseudospectrum \(\sigma _\varepsilon (\mathbf{A})\) can be characterized using the resolvent of \(\mathbf{A}\) in (9), Frayseé et al. [10] show that Definition 2 is equivalent to

$$\begin{aligned} \sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E}) = \left\{ z\in \mathbb {C}: \Vert (z\mathbf{E}-\mathbf{A})^{-1}\Vert > {1\over \varepsilon (\gamma + |z| \delta )}\right\} . \end{aligned}$$
(10)

This formula suggests a way to compute \(\sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E})\): evaluate \( \Vert (z\mathbf{E}-\mathbf{A})^{-1}\Vert \) on a grid of points covering a relevant region of the complex plane (or the Riemann sphere) and use a contour plotting routine to draw boundaries of \(\sigma _\varepsilon ^{(\gamma ,\delta )}(\mathbf{A},\mathbf{E})\). The accuracy of the resulting pseudospectra depends on the density of the grid. Expedient algorithms for computing \(\Vert (z\mathbf{E}-\mathbf{A})^{-1}\Vert \) can be derived by computing a unitary simultaneous triangularization (generalized Schur form) of \(\mathbf{A}\) and \(\mathbf{E}\) in \(\mathcal{O}(n^3)\) operations, then using inverse iteration or inverse Lanczos, as described by Trefethen [28] and Wright [31], to compute \(\Vert (z\mathbf{E}-\mathbf{A})^{-1}\Vert \) at each grid point in \(\mathcal{O}(n^2)\) operations. For the structured Loewner pencils of interest here, one can compute \(\Vert (z\mathbf{E}-\mathbf{A})^{-1}\Vert \) in \(\mathcal{O}(n^2)\) operations without recourse to the \(\mathcal{O}(n^3)\) preprocessing step, as proposed in Sect. 4.

Remark 3

Definition 2 can be extended to \(\delta =0\) by only perturbing \(\mathbf{A}\) [23]:

$$\begin{aligned} \sigma _\varepsilon ^{(1,0)}(\mathbf{A},\mathbf{E})= & {} \{z \in \mathbb {C}\,\text {is an eigenvalue of the pencil }z\mathbf{E}-(\mathbf{A}+\boldsymbol{\Gamma }) \\&\ \ \ \ \ \text {for some }\boldsymbol{\Gamma }\in \mathbb {C}^{n\times n}~\text {with }\Vert \boldsymbol{\Gamma }\Vert < \varepsilon \}. \nonumber \\= & {} \{ z\in \mathbb {C}: \Vert (z\mathbf{E}-\mathbf{A})^{-1}\Vert > 1/\varepsilon \}. \end{aligned}$$

This definition may be more suitable for cases where \(\mathbf{E}\) is fixed and uncertainty in the system only emerges, e.g., through physical parameters that appear in \(\mathbf{A}\).

Remark 4

Since we ultimately intend to study the pseudospectra \(\sigma _\varepsilon ^{(\gamma ,\delta )}({\mathbb {L}_s},\mathbb {L})\) of Loewner matrix pencils, one might question the use of generic perturbations \(\boldsymbol{\Gamma }, \boldsymbol{\Delta }\in \mathbb {C}^{n\times n}\) in Definition 2. Should we restrict \(\boldsymbol{\Gamma }\) and \(\boldsymbol{\Delta }\) to maintain Loewner structure, i.e., so that \({\mathbb {L}_s}+\boldsymbol{\Gamma }\) and \(\mathbb {L}+\boldsymbol{\Delta }\) maintain the coupled shifted Loewner–Loewner form in (1)? Such sets are called structured pseudospectra.

Three considerations motivate the study of generic perturbations \(\boldsymbol{\Gamma }, \boldsymbol{\Delta }\in \mathbb {C}^{n\times n}\): one practical, one speculative, and one philosophical. (a) Beyond repeatedly computing the eigenvalues of Loewner pencils with randomly perturbed data, no systematic method is known to compute the coupled Loewner structured pseudospectra, i.e., no analogue of the resolvent-like definition (10) is known. (b) Rump [25] showed that in many cases, preserving structure has little effect on the standard matrix pseudospectra. For example, the structured \(\varepsilon \)-pseudospectrum of a Hankel matrix \(\mathbf{H}\) allowing only complex Hankel perturbations exactly matches the unstructured \(\varepsilon \)-pseudospectrum \(\sigma _\varepsilon (\mathbf{H})\) based on generic complex perturbations [25, Theorem 4.3]. Whether a similar results holds for Loewner structured pencils is an interesting open question. (c) If one seeks to analyze the behavior of dynamical systems (as opposed to eigenvalues of nearby matrices), then generic perturbations give much greater insight; see [29, p. 456] for an example where real-valued perturbations do not move the eigenvalues much toward the imaginary axis (hence the real structured pseudospectra are benign), yet the stable system still exhibits strong transient growth.

As we shall see in Sect. 6, Definition 2 provides a helpful tool for investigating the sensitivity of eigenvalues of matrix pencils. A different generalization of Definition 1 gives insight into the transient behavior of solutions of \(\mathbf{E}\dot{\mathbf{x}}(t)=\mathbf{A}\mathbf{x}(t)\). This approach is discussed in [29, Chap. 45], following [23, 24], and has been extended to handle singular \(\mathbf{E}\) in [8] (for differential-algebraic equations and descriptor systems). Restricting our attention here to nonsingular \(\mathbf{E}\), we analyze the conventional (single matrix) pseudospectra \(\sigma _\varepsilon (\mathbf{E}^{-1}\mathbf{A})\). From these sets one can develop various upper and lower bounds on \(\Vert \mathrm{e}^{t\mathbf{E}^{-1}\mathbf{A}}\Vert \) and \(\Vert \mathbf{x}(t)\Vert \) [29, Chap. 15]. Here we shall just state one basic result. If \(\sup \{\mathrm{Re}(z) : z\in \sigma _\varepsilon (\mathbf{E}^{-1}\mathbf{A})\} = K \varepsilon \) for some \(K\ge 1\) (where \(\mathrm{Re}(\cdot )\) denotes the real part of a complex number), then

$$\begin{aligned} \sup _{t\ge 0} \Vert \mathrm{e}^{t\mathbf{E}^{-1}\mathbf{A}}\Vert \ge K. \end{aligned}$$
(11)

This statement implies that there exists some unit-length initial condition \(\mathbf{x}(0)\) such that \(\Vert \mathbf{x}(t)\Vert \ge K\), even though \(\sigma (\mathbf{A},\mathbf{E})\) may be contained in the left half-plane. (Optimizing this bound over \(\varepsilon >0\) yields the Kreiss Matrix Theorem [29, (15.9)].)

Pseudospectra of matrix pencils provide a natural vehicle to explore that stability of the matrix pencil associated with the Loewner realization in (6). We shall thus investigate eigenvalue perturbations via \(\sigma _\varepsilon ^{(\gamma ,\delta )}(\widehat{\mathbf{A}},\widehat{\mathbf{E}}) = \sigma _\varepsilon ^{(\gamma ,\delta )}({\mathbb {L}_s},\mathbb {L})\) and transient behavior via \(\sigma _\varepsilon (\mathbb {L}^{-1}{\mathbb {L}_s})\).

4 Efficient Computation of Loewner Pseudospectra

We first present a novel technique for efficiently computing pseudospectra of large Loewner matrix pencils, \(\sigma _\varepsilon ^{(\gamma ,\delta )}({\mathbb {L}_s},\mathbb {L})\), using the equivalent definition given in (10). When the Loewner matrix \(\mathbb {L}\) is nonsingular, we employ inverse iteration to exploit the structure of the Loewner pencil to compute \(\Vert (z\mathbb {L}-{\mathbb {L}_s})^{-1}\Vert \) (in the two-norm) using only \(\mathcal{O}(n^2)\) operations. This avoids the need to compute an initial simultaneous unitary triangularization of \({\mathbb {L}_s}\) and \(\mathbb {L}\) using the QZ algorithm, an \(\mathcal{O}(n^3)\) operation.

Inverse iteration (and inverse Lanczos) for \(\Vert (z\mathbb {L}-{\mathbb {L}_s})^{-1}\Vert \) requires computing

$$\begin{aligned} (z\mathbb {L}-{\mathbb {L}_s})^{-*}(z\mathbb {L}-{\mathbb {L}_s})^{-1}\mathbf{u} \end{aligned}$$
(12)

for a series of vectors \(\mathbf{u}\in \mathbb {C}^n\) (e.g., see [29, Chap. 39]). We invoke a property observed by Mayo and Antoulas [19], related to (2): by construction, the Loewner and shifted Loewner matrices satisfy \({\mathbb {L}_s}- \mathbb {L}{\boldsymbol{\Lambda }}= \mathbf{V}^*\mathbf{R}\). Thus the resolvent can be expressed using only \(\mathbb {L}\) and not \({\mathbb {L}_s}\):

$$ (z\mathbb {L}-{\mathbb {L}_s})^{-1} = (\mathbb {L}(z\mathbf{I}-{\boldsymbol{\Lambda }})-\mathbf{V}^*\mathbf{R})^{-1}. $$

We now use the Sherman–Morrison–Woodbury formula (see, e.g., [12]) to get

$$ (z\mathbb {L}-{\mathbb {L}_s})^{-1} = \left( \mathbf{I}+ \boldsymbol{\Upsilon }(z) (\mathbf{I}- \mathbf{R}\boldsymbol{\Upsilon }(z))^{-1} \mathbf{R}\right) (z\mathbf{I}-{\boldsymbol{\Lambda }})^{-1} \mathbb {L}^{-1} = \boldsymbol{\Theta }(z)\mathbb {L}^{-1}, $$

where \(\boldsymbol{\Upsilon }(z) := (z\mathbf{I}-{\boldsymbol{\Lambda }})^{-1} \mathbb {L}^{-1}\mathbf{V}^*\) and \(\boldsymbol{\Theta }(z) := \left( \mathbf{I}+ \boldsymbol{\Upsilon }(z) (\mathbf{I}- \mathbf{R}\boldsymbol{\Upsilon }(z))^{-1} \mathbf{R}\right) (z\mathbf{I}-{\boldsymbol{\Lambda }})^{-1}\). As a result, we can compute the inverse iteration vectors in (12) as

$$\begin{aligned} (z\mathbb {L}-{\mathbb {L}_s})^{-*}(z\mathbb {L}-{\mathbb {L}_s})^{-1}\mathbf{u}= \mathbb {L}^{-*}\boldsymbol{\Theta }(z)^*\boldsymbol{\Theta }(z)\mathbb {L}^{-1}\mathbf{u}, \end{aligned}$$
(13)

which requires solving several linear systems given by the same Loewner matrix \(\mathbb {L}\), e.g., \(\mathbb {L}^{-1}\mathbf{u}\), \(\mathbb {L}^{-1}\mathbf{V}^*\).

Crucially, solving a linear system involving a Loewner matrix \(\mathbb {L}\in \mathbb {C}^{n\times n}\) can be done efficiently in only \(2(m+p+1)n^2\) operations, since \(\mathbb {L}\) has displacement rank \(m+p\). More precisely, \(\mathbb {L}\) is a Cauchy-like matrix that satisfies the Sylvester equation (2) given by diagonal generator matrices \({\boldsymbol{\Lambda }}\) and \(\mathbf{M}\) and a right-hand side of rank at most \(m+p\), i.e., \(\mathrm{rank}(\mathbf{L}^*\mathbf{W}-\mathbf{V}^*\mathbf{R}) \le m+p\). The displacement rank structure of \(\mathbb {L}\) can be exploited to compute its LU factorization in only \(2(m+p)n^2\) flops (see [11] and [12, Sect. 12.1]). Given the LU factorization of \(\mathbb {L}\), solving \(\mathbb {L}^{-1}\mathbf{u}\) in (13) via standard forward and backward substitution requires another \(2n^2\) operations.

Next, multiplying \(\boldsymbol{\Theta }(z)\) with the solution of \(\mathbb {L}^{-1}\mathbf{u}\) requires a total of \(2mn^2 + (4m^2+6m+2)n + \frac{2}{3}m^3-m^2\) operations, namely (to leading order on the factorizations):

  • \(2mn^2 + 2mn\) operations to compute \(\boldsymbol{\Upsilon }(z) \in \mathbb {C}^{n\times m}\);

  • \(m^2(2n-1) + m\) operations to compute \(\mathbf{I}- \mathbf{R}\boldsymbol{\Upsilon }(z) \in \mathbb {C}^{m\times m}\);

  • \(\frac{2}{3}m^3+2m^2n\) operations to solve \((\mathbf{I}- \mathbf{R}\boldsymbol{\Upsilon }(z))^{-1} \mathbf{R}\in \mathbb {C}^{m\times n}\) via

    an LU factorization followed by n forward and backward substitutions;

  • \(n + m(2n-1) + (2m-1)n + 2n\) operations to multiply \(\boldsymbol{\Theta }(z)\) with \(\mathbb {L}^{-1}\mathbf{u}\).

Finally, multiplying with \(\mathbb {L}^{-*}\boldsymbol{\Theta }(z)^*\) in (13) requires an additional \(2n^2 + 2mn^2 + (4m^2+6m+2)n + \frac{2}{3}m^3-m^2\) operations, bringing the total cost of computing (13) to \(2(3m+p+1)n^2 + 4(2m^2+3m+1)n + \frac{4}{3}m^3-2m^2\) operations.

In practice, the sizes of the right and left tangential directions are much smaller than the size of the Loewner pencil, i.e., \(m, p \ll n\). For example, for scalar data (associated with SISO systems), \(m = p = 1\). Therefore, in practice, computing (13) can be done in only \(\mathcal{O}(n^2)\) operations.

Partial pivoting can be included in the LU factorization of the Loewner matrix \(\mathbb {L}\) to overcome numerical difficulties. Adding partial pivoting maintains the \(\mathcal{O}(n^2)\) operation count for the LU factorization of \(\mathbb {L}\) (see [12, Sect. 12.1]), and hence computing (13) can still be done in \(\mathcal{O}(n^2)\) operations. The appendix gives a MATLAB implementation of this efficient inverse iteration.

We measure these performance gains for a Loewner pencil generated by sampling \(f(x) = \sum _{k=1}^8 (-1)^{k+1}\left( 1+100(x-k)^2\right) ^{-1/2} + (-1)^{k+1}\left( 1+100(x-k-1/2)^2\right) ^{-1/2}\) at 2n points uniformly spaced in the interval [1, 8]. We compare our new \(\mathcal{O}(n^2)\) Loewner pencil inverse iteration against a standard implementation (see [29, p. 373]) applied to a simultaneous triangularization of \({\mathbb {L}_s}\) and \(\mathbb {L}\). (The simultaneous triangularization costs \(\mathcal{O}(n^3)\) but is fast, as MATLAB’s qz routine invokes LAPACK code. For a fair comparison, we test against a C++ implementation of the fast Loewner code, compiled into a MATLAB .mex file.) Table 1 shows timings for both implementations by computing \(\Vert (z\mathbb {L}-{\mathbb {L}_s})^{-1}\Vert \) on a \(200\times 200\) grid of points. Exploiting the Loewner structure gives a significant performance improvement for large n.

We next examine two simple examples involving full-rank realization of SISO systems, to illustrate the kinds of insights one can draw from pseudospectra of Loewner pencils. (For these small examples we use the standard \(\mathcal{O}(n^3)\) algorithm.)

Table 1 Comparison of speed of computing \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\) using the fast Loewner algorithm versus a generic inverse iteration method applied to simultaneous triangularizations of \({\mathbb {L}_s}\) and \(\mathbb {L}\)
Table 2 Right and left interpolation points for the two-dimensional SISO system (14). The two right columns report the singular values of the Loewner matrix \(\mathbb {L}\)

5 Example 1: Eigenvalue Sensitivity and Transient Behavior

We first consider a simple controllable and observable SISO system with \(n=2\):

$$\begin{aligned} \mathbf{E}=\mathbf{I}, \qquad \mathbf{A}= \left[ \begin{array}{cc} -1.1 &{} 1 \\ 1 &{} -1.1 \end{array}\right] , \qquad \mathbf{B}= \left[ \begin{array}{c} 0 \\ 1 \end{array}\right] , \qquad \mathbf{C}= \left[ \begin{array}{cc} 0&1 \end{array}\right] . \end{aligned}$$
(14)

This \(\mathbf{A}\) is symmetric negative definite, with eigenvalues \(\sigma (\mathbf{A}) = \{-0.1,-2.1\}\). Since the system is SISO, the transfer function \(\mathbf{H}(s) = \mathbf{C}(s\mathbf{I}-\mathbf{A})^{-1}\mathbf{B}\) maps \(\mathbb {C}\) to \(\mathbb {C}\), and hence the choice of “interpolation directions” is trivial (though the division into “left” and “right” points matters). We take \(\varrho =\nu =2\) left and right interpolation points, with \(\mathbf{r}_1 = \mathbf{r}_2 = 1\) and \(\boldsymbol{\ell }_1 = \boldsymbol{\ell }_2 = 1\). We will study various choices of interpolation points, all of which satisfy, for each \(\widehat{z}\in \{\lambda _1, \lambda _2, \mu _1, \mu _2\}\),

$$\begin{aligned} \mathrm{rank}(\widehat{z}\, \mathbb {L}- {\mathbb {L}_s}) = \mathrm{rank}(\mathbb {L}) = \mathrm{rank}({\mathbb {L}_s}) = n=2. \end{aligned}$$
(15)

This basic set-up makes it easy to focus on the influence of the interpolation points \(\lambda _1\), \(\lambda _2\), \(\mu _1\), \(\mu _2\). We will use the pseudospectra \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\) to examine how the interpolation points affect the stability of the eigenvalues of the Loewner pencil.

Fig. 1
figure 1

Boundaries of pseudospectra \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\) for four Loewner realizations of the system (14) using the interpolation points in Table 2. All four realizations correctly give \(\sigma ({\mathbb {L}_s},\mathbb {L}) = \sigma (\mathbf{A})\), but \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\) show how the stability of the realized eigenvalues depends on the choice of interpolation points. In this and all similar plots, the colors denote \(\log _{10}(\varepsilon )\). Thus, in plot (d), there exist perturbations to \({\mathbb {L}_s}\) and \(\mathbb {L}\) of norm \(10^{-6.5}\) that move an eigenvalue into the right half-plane

Fig. 2
figure 2

The pseudospectra \(\sigma _\varepsilon (\mathbf{A})\) (top), compared to \(\sigma _\varepsilon (\mathbb {L}^{-1}{\mathbb {L}_s})\) for four Loewner realizations of the system (14) using the interpolation points in Table 2. In all cases \(\sigma (\mathbb {L}^{-1}{\mathbb {L}_s}) = \sigma (\mathbf{A})\), but the pseudospectra of \(\mathbb {L}^{-1}{\mathbb {L}_s}\) are all quite a bit larger than \(\sigma _\varepsilon (\mathbf{A})\)

Fig. 3
figure 3

Evolution of the norm of the solution operator for the original system (black dashed line) and the four interpolating Loewner models. The instability revealed by the pseudospectra in Fig. 2 corresponds to transient growth in the Loewner systems

Table 2 records four different choices of \(\{\lambda _1, \lambda _2, \mu _1, \mu _2\}\); Fig. 1 shows the corresponding pseudospectra \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\). All four Loewner realizations match the eigenvalues of \(\mathbf{A}\) and satisfy the interpolation conditions. However, the pseudospectra show how the stability of the eigenvalues \(-0.1\) and \(-2.1\) differs across these four realizations. From example (a) to (d), these eigenvalues become increasingly sensitive as the interpolation points move farther from \(\sigma (\mathbf{A})\). Table 2 also shows the singular values of the Loewner matrix \(\mathbb {L}\), demonstrating how the second singular value \(s_2(\mathbb {L})\) decreases as the eigenvalues become increasingly sensitive. (Taken to a greater extreme, it would eventually be difficult to determine if \(\mathbb {L}\) truly is rank 2.)

Remark 5

By Remark 1, note that if \(\varepsilon < s_\mathrm{min}(\mathbb {L})\), then \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\) will be unbounded. Thus the decreasing values of \(s_\mathrm{min}(\mathbb {L})\) in Table 2 suggest the enlarging pseudospectra seen in Fig. 1. For example, in case (d) the \(\varepsilon =10^{-5}\) pseudospectrum \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\) must contain the point at infinity.

Contrast these results with the standard pseudospectra of \(\mathbf{A}\) itself, \(\sigma _\varepsilon (\mathbf{A}) = \sigma _\varepsilon ^{(1,0)}(\mathbf{A},\mathbf{I})\) shown at the top of Fig. 2. Since \(\mathbf{A}\) is real symmetric (hence normal), \(\sigma _\varepsilon (\mathbf{A})\) is the union of open \(\varepsilon \)-balls surrounding the eigenvalues. Figure 2 compares these pseudospectra to \(\sigma _\varepsilon (\mathbb {L}^{-1}{\mathbb {L}_s})\), which give insight into the transient behavior of solutions to \(\mathbb {L}\dot{\mathbf{x}}(t) = {\mathbb {L}_s}\mathbf{x}(t)\), e.g., via the bound (11). (Since \({\mathbb {L}_s}-\mathbb {L}{\boldsymbol{\Lambda }}= \mathbf{V}^*\mathbf{R}\), and \(m=1\), \(\mathbb {L}^{-1}{\mathbb {L}_s}= {\boldsymbol{\Lambda }}+ \mathbb {L}^{-1}\mathbf{V}^*\mathbf{R}\) is a rank-1 update of the matrix \({\boldsymbol{\Lambda }}\) [19, p. 643].) The top plot shows \(\sigma _\varepsilon (\mathbf{A})\), whose rightmost extent in the complex plane is always \(\varepsilon -0.1\): no transient growth is possible for this system. However, in all four Loewner realizations, \(\sigma _\varepsilon (\mathbb {L}^{-1}{\mathbb {L}_s})\) extends more than \(\varepsilon \) into the right-half plane for \(\varepsilon =10^0\) (orange level curve), indicating by (11) that transient growth must occur for some initial condition. Figure 3 shows this growth for all four realizations: the more remote interpolation points lead to Loewner realizations with greater transient growth.

6 Example 2: Partitioning Interpolation Points and Noisy Data

To further investigate how the interpolation points influence eigenvalue stability for the Loewner pencil, consider the SISO system of order 10 given by

$$ \mathbf{E}= \mathbf{I}, \quad \mathbf{A}= \mathrm{diag}(-1, -2, \ldots , -10), \quad \mathbf{B}= [1, 1, \ldots , 1]^T, \quad \mathbf{C}= [1, 1, \ldots , 1].$$

Figure  4 shows \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\) for six configurations of the interpolation points. Plots (a) and (b) use the points \(\{-10.25, -9.75, -9.25, \ldots , -1.25, -0.75\}\); in plot (a) the left and right points interleave, suggesting slower decay of the singular values of \(\mathbb {L}\), as discussed in Sect. 2.1; in plot (b) the left and right points are separated, leading to faster decay of the singular values of \(\mathbb {L}\) and considerably larger pseudospectra. (The Beckermann–Townsend bound [3, Corollary 4.2] applies to this case.) Plot (c) further separates the left and right points, giving even larger pseudospectra. In plots (d) and (f), complex interpolation points \(\{-5\pm 0.5\mathrm{i}, -5\pm 1.0\mathrm{i},\ldots ,-5\pm 5\mathrm{i}\}\) are mostly interleaved (d) (keeping conjugate pairs together) and separated (f): the latter significantly enlarges the pseudospectra. Plot (f) uses the same relative arrangement that gave such nice results in plot (a) (the singular value bound (3) is the same for (a) and (e)), but their locations relative to the poles of the original system differ. The pseudospectra are now much larger, showing that a large upper bound in (3) is not alone enough to guarantee small pseudospectra. (Indeed, the pseudospectra are so large in (e) and (f) that the plots are dominated by numerical artifacts of computing \(\Vert (z\mathbb {L}-{\mathbb {L}_s})^{-1}\Vert \).) Pseudospectra reveal the great influence interpolation point location and partition can have on the stability of the realized pencils.

Fig. 4
figure 4

Pseudospectra \(\sigma _\varepsilon ^{(1,1)}({\mathbb {L}_s},\mathbb {L})\) for six Loewner realizations of a SISO system of order \(n=10\) with poles \(\sigma (\mathbf{A}) = \{-1,-2,\ldots , -10\}\). Black dots show computed eigenvalues of the pencil \(z\mathbb {L}-{\mathbb {L}_s}\) (which should agree with \(\sigma (\mathbf{A})\), but a few are off axis in plot (e)); blue squares show the right interpolation points \(\{\lambda _i\}\); red diamonds show the left points \(\{\mu _j\}\)

Pseudospectra also give insight into the consequences of inexact measurement data. (For a recent study of how noisy data affects the recovered transfer function, see [6].) Consider the following experiment. Take the scenario in Fig. 4a, the most robust of these examples. Subject each right and left measurement \(\{\mathbf{w}_1,\ldots , \mathbf{w}_\varrho \} \subset \mathbb {C}\) and \(\{\mathbf{v}_1,\ldots , \mathbf{v}_\nu \} \subset \mathbb {C}\) to random complex noise of magnitude \(10^{-1}\), then build the Loewner pencil \(z\widehat{\mathbb {L}}-\widehat{\mathbb {L}}_s\) from this noisy data. How do the badly polluted measurements affect the computed eigenvalues? Figure 5 shows the results of 1,000 random trials, which can depart from the true matrices significantly:

$$\begin{aligned} 3.06 \le \Vert \widehat{\mathbb {L}}_s - {\mathbb {L}_s}\Vert \le 5.88, \qquad 0.49 \le \Vert \widehat{\mathbb {L}}-\mathbb {L}\Vert \le 0.63. \end{aligned}$$

Despite these large perturbations, the recovered eigenvalues are remarkably accurate: in \(99.99\%\) of cases, the eigenvalues have absolute accuracy of at least \(10^{-2}\), indeed more accurate than the measurements themselves. The pseudospectra in Fig. 4a suggest good robustness (though the pseudospectral level curves are pessimistic by one or two orders of magnitude). Contrast this with the complex interleaved interpolation points used in Fig. 4d. Now we only perturb the data by a small amount, \(5\cdot 10^{-9}\), for which the perturbed Loewner matrices (over 1,000 trials) satisfy

$$ 1.52\cdot 10^{-7} \le \Vert \widehat{\mathbb {L}}_s - {\mathbb {L}_s}\Vert \le 2.03\cdot 10^{-7}, \qquad 2.57\cdot 10^{-8} \le \Vert \widehat{\mathbb {L}}-\mathbb {L}\Vert \le 3.14\cdot 10^{-8}. $$

With this mere hint of noise, the eigenvalues of the recovered system erupt: only \(36.73\%\) of the eigenvalues are correct to two digits. (Curiously, \(-4\) and \(-5\) are always computed correctly, while \(-8\), \(-9\), and \(-10\) are never computed correctly.) The pseudospectra indicate that the leftmost eigenvalues are more sensitive, and again hint at the effect of the perturbation (though off by roughly an order of magnitude in \(\varepsilon \)).Footnote 3 Measurements of real systems (or even numerical simulations of nontrivial systems) are unlikely to produce such high accuracy; pseudospectra can reveal the virtue or folly of a given interpolation point configuration.

Fig. 5
figure 5

Eigenvalues of the perturbed Loewner pencil \(\widehat{\mathbb {L}}_s-z\widehat{\mathbb {L}}\) (gray dots), constructed from measurements that have been perturbed by random complex noise of magnitude \(10^{-1}\) (left) and \(5\cdot 10^{-9}\) (right) (1,000 trials). As the pseudospectra in Fig. 4a, d indicate, the interleaved interpolation points on the left are remarkably stable, while the similarly interleaved complex interpolation points on the right give a Loewner pencil that is highly sensitive to small changes in the data

Fig. 6
figure 6

Eigenvalues of the perturbed Loewner pencil \(\widehat{\mathbb {L}}_s-z\widehat{\mathbb {L}}\) (gray dots), constructed from measurements that have been perturbed by random complex noise of magnitude \(10^{-10}\) (10,000 trials). As suggested by the pseudospectra plots, the interleaved interpolation points (left) are more robust to perturbations than the separated points (right), though the difference is not as acute as suggested by Fig. 4d, f. For example, in these 10,000 trials, the least stable pole (\(-9\)) is computed accurately (absolute error less than 0.01) in \(10.97\%\) of trials on the left, and \(0.29\%\) on the right

In these simple experiments, pseudospectra have been most helpful for indicating the sensitivity of eigenvalues when the left and right interpolation points are favorably partitioned (e.g., interleaved). They seem to be less precise at predicting the sensitivity to noise of poor left/right partitions of the interpolation points. Figure 6 gives an example, based on the two partitions of the same interpolation points in Fig. 4d, f. The pseudospectra suggest that the eigenvalues for plot (f) should be much more sensitive to noise than those for the interleaved points in plot (d). In fact, the configuration in plot (f) appears to be only marginally less stable to noise of size \(10^{-10}\), over 10,000 trials. This is a case where one could potentially glean additional insight from structured Loewner pseudospectra.

7 Conclusion

Pseudospectra provide a tool for analyzing the stability of eigenvalues of Loewner matrix pencils. Elementary examples show how pseudospectra can inform the selection and partition of interpolation points, and bound the eigenvalues of Loewner pencils in the presence of noisy data. Using a different approach to pseudospectra, we showed that while the realized Loewner pencil matches the poles of the original system, it need not replicate transient dynamics of \(\mathbf{E}\dot{\mathbf{x}}(t) = \mathbf{A}\mathbf{x}(t)\); pseudospectra can reveal potential transient growth, which varies with the interpolation points.

In this initial study we have intentionally used simple examples involving small, symmetric \(\mathbf{A}\) and \(\mathbf{E}=\mathbf{I}\). Realistic examples, e.g., with complex poles, nonnormal \(\mathbf{A}\), singular \(\mathbf{E}\), multiple inputs and outputs, and rank-deficient Loewner matrices, will add additional complexity. Moreover, we have only sought to realize a system whose order is known; we have not addressed pseudospectra of the reduced pencils (7) in the context of data-driven model reduction.

Structured Loewner pseudospectra provide another avenue for future study. Structured matrix pencil pseudospectra have not been much investigated, especially with Loewner structure. Rump’s results for standard pseudospectra [25] suggest the following problem; its positive resolution would imply that the Loewner pseudospectrum \(\sigma _\varepsilon ^{(\gamma ,\delta )}({\mathbb {L}_s},\mathbb {L})\) matches the structured Loewner matrix pseudospectrum.

Given any \(\varepsilon , \gamma , \delta > 0\), Loewner matrix \(\mathbb {L}\) and associated shifted Loewner matrix \({\mathbb {L}_s}\), suppose \(z \in \sigma _\varepsilon ^{(\gamma ,\delta )}({\mathbb {L}_s},\mathbb {L})\). Does there exist some Loewner matrix \(\widehat{\mathbb {L}}\) and associated shifted Loewner matrix \(\widehat{\mathbb {L}}_s\) such that \(z\in \sigma (\widehat{\mathbb {L}}_s,\widehat{\mathbb {L}})\) and

$$ \Vert \widehat{\mathbb {L}}_s - {\mathbb {L}_s}\Vert< \epsilon \gamma , \qquad \Vert \widehat{\mathbb {L}} - \mathbb {L}\Vert < \epsilon \delta \ \ ?$$