Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

This chapter contains a collection of useful mathematical concepts and tools, which are useful, directly or indirectly, for the subsequent development to be covered in the main portion of the book. While much of the material is standard and can be found in classical textbooks, we also present a number of useful items that are not commonly found elsewhere. Essentially, this chapter serves as a brief overview and as a convenient reference when necessary.

1 Introduction

Hybrid systems are certainly pervasive today. Recently, we have witnessed a resurgence in examining quantization effects and a heightened interest in analog computation. There has also been recent progress in analyzing switched, hierarchical, and discretely controlled continuous-variable systems. It is time to focus on developing formal modeling, analysis, and control methodologies for hybrid systems. Therefore, hybrid systems research [357359] is devoted to modeling, design, and validation of interacting systems of continuous process and computer programs. Therefore, the identifying characteristic of hybrid systems is that they incorporate both continuous components, usually called plants, which are governed by ordinary or functional differential equations, and also digital components such as digital computers, sensors, and actuators controlled by programs. Moreover, the growing demands for control systems that are capable of controlling complex nonlinear continuous plants with discrete intelligent controllers can be addressed by the method of hybrid systems.

Throughout this book, by a switched system we mean a class of hybrid dynamical systems consisting of a family of continuous-time subsystems and a rule that orchestrates the switching between them. An integral part of this book surveys recent developments in three basic problems regarding stability and design of switched systems. These problems are:

  • stability for arbitrary switching sequences,

  • stability for certain useful classes of switching sequences, and

  • construction of stabilizing switching sequences.

We also provide motivation for studying these problems within the framework of time-delay systems. In practice, many systems encountered exhibit switching between several subsystems (are inherently multimodal) that is dependent on various environmental factors. Another source of motivation for studying switched systems comes from the rapidly developing area of switching control. Control techniques based on switching between different controllers have been applied extensively in recent years, particularly in the adaptive context, where they have been shown to achieve stability and improve transient response. The importance of such control methods also stems in part from the existence of systems that cannot be asymptotically stabilized by a single continuous feedback control law. Additionally, the fact that some of intelligent control methods are based on the idea of switching between different controllers. The existence of systems that cannot be asymptotically stabilized by a single static continuous feedback controller [47] also motivates the study. A survey of basic problems in stability and design of switched systems is given in [193].

In this book, we treat switched systems as a class of hybrid systems consisting of a family of subsystems and a switching law that specifies which subsystem will be activated along the system trajectory at each instant of time. Switched systems deserve investigation for theoretical development as well as for practical applications. To switch between different system structures is an essential feature of many control systems, for example, in power systems and power electronics [47]. There have been many studies for switched systems without uncertainties, primarily on stability analysis and design [358]. But for robust stability analysis of uncertain switched systems, there has been comparatively little work. A notable exception is the study of quadratic stability and stabilization by state-based feedback for both continuous-time and discrete-time switched linear systems composed of polytopic uncertain systems in [357]. For performance analysis of switched systems, authors of [357] investigated the disturbance attenuation properties of time-controlled switched systems consisting of several linear time invariant subsystems by using an average dwell-time approach incorporated with a piecewise Lyapunov function. Reference [133] computed the \({\cal L}_2\)-induced norm of a switched linear system when the interval between consecutive switching is large. However, uncertainty is not considered in these two papers although it is ubiquitous in the system model due to the complexity of the system itself, exogenous disturbance, measurement errors, and so on. During the past decade, there have also been many papers concerning robust (or quadratic) stability, stabilization, and robust \({\cal H_\infty}\) control of uncertain systems without switchings [331, 441].

2 Basic Mathematical Concepts

Let \(x_j,\;y_j,\;j = 1,2,\dots,n \in {\Re} (or \; {\bf C})\). Then the n-dimensional vectors \(x,\;y\) are defined by \(x=[x_1\;x_2\;\dots\;x_n]^t,\; y=[y_1\;y_2\;\dots\;y_n]^t\; \in {\Re}^n,\;\) respectively.

A nonempty set \({\cal X}\) of elements \(x,\;y,\dots\) is called the real (or complex) vector space (or real (complex) linear space) by defining two algebraic operations, vector additions and scalar multiplication, in \(x=[x_1,\ x_2,\ \dots,\ x_n]^t\) [46]

2.1 Euclidean Space

The n-dimensional Euclidean space , denoted in the sequel by \({\Re}^n\) is the linear vector space \({\Re}^n\) equipped by the inner product

$$\left\langle x,y \right\rangle = x^t \;y = \sum^n_{j=1}\; x_j y_j$$

Let \({\cal X}\) be a linear space over the field F (typically F is the field of real numbers \({\Re}\) or complex numbers C). Then a function

$$||.||: {\cal X}\rightarrow {\Re}$$

that maps \({\cal X}\) into the real numbers \({\Re}\) is a norm on \({\cal X}\) iff

  1. 1.

    \(||x|| \geq 0,\; \forall x \in {\cal X}\) (nonnegativity)

  2. 2.

    \(||x|| = 0,\; \Longleftrightarrow x = 0\) (positive definiteness)

  3. 3.

    \(||\alpha \;x|| = |\alpha| ||x|| \forall x \in {\cal X}\) (homogeneity with respect to \(|\alpha|\))

  4. 4.

    \(||x + y|| \leq ||x|| + ||y||,\;\; \forall x,y \in {\cal X}\) (triangle inequality)

Given a linear space \({\cal X}\), there are many possible norms on it. For a given norm \(||.||\;on\;{\cal X}\), the pair (\({\cal X},\;||.||\)) is used to indicate \({\cal X}\) endowed with the norm \(||.||\).

2.2 Norms of Vectors

The class of \(L_{p}\)-norms is defined by

$$\begin{array}{*{20}l}||x||_p &= \Big (\sum^n_{j=1}\;|x_j|^p \Big)^{1/p},\;\;\; f\,\,{\mathrm{or}}\,\, 1 \leq p < \infty\\ ||x||_\infty &= \max{1 \leq j \leq n} \;|x_j|\end{array}$$

The three most commonly used norms are \(||x||_1,\;||x||_2\), and \(||x||_\infty\). All p-norms are equivalent in the sense that if \(||x||_{p1}\) and \(||x||_{p2}\) are two different p-norms, then there exist positive constants \(c_{1}\) and \(c_{s}\) such that

$$c_1\;||x||_{p1}\;\leq\;||x||_{p2}\; c_2\;||x||_{p1},\;\;\;\;\;\forall x \in {\Re}^n$$

2.2.1 Induced Norms of Matrices

For a matrix \(A \in {\Re}^{n \times n},\;\) the induced p-norm of A is defined by

$$||A||_p {\stackrel{\Delta}{=}} \sup_{x \neq 0}\; \frac{||A x||_p}{||x||_p} = \sup_{||x||_p = 1} \; ||Ax||_p$$

Obviously, for matrices \(A \in {\Re}^{m \times n}\) and \(A \in {\Re}^{n \times r},\;\) we have the triangle inequality:

$$||A + B||_p \;\leq\; ||A|||_p + ||B||_p$$

It is easy to show that the induced norms are also equivalent in the same sense as for the vector norms, and satisfying

$$||A B||_p \;\leq\; ||A x||_p \;||B||_p,\;\; \forall A \in {\Re}^{n \times m},\;B \in {\Re}^{m \times r}$$

which is known as the submultiplicative property. For \(p = 1,2,\ldots \infty,\;\) we have the corresponding induced norms as follows:

$$\begin{array}{*{20}l}||A||_1 &= \max_{j} \sum^n_{s=1}\;|a_{sj}|,\;\;\; ({\mathrm{column\;sum}})\\ ||A||_2 &= \max_{j} \sqrt{\lambda_j (A^t A)}\\ ||A||_\infty &= \max_{s} \sum^n_{j=1}\;|a_{sj}|,\;\;\; ({\mathrm{row\;sum}})\\ \end{array}$$

2.3 Convex Sets

A set \({\bf S} \subset {\Re}^n\;\) is said to be open if every vector \(x \in {\bf S},\;\) there is an ε-neighborhood of x

$${\cal N} (x,\epsilon) = \{z \in \; {\Re}^n | ||z - x|| \; <\; \epsilon \}$$

such that \({\cal N} (x,\epsilon) \subset {\bf S}.\)

A set is closed iff its complement in \({\Re}^n\;\) is open; bounded if there is \(r > 0\) such that \(||x|| < r,\; \forall x \in {\bf S};\;\) and compact if it is closed and bounded; convex if for every \(x, y \in {\bf S},\;\) and every real number \(\alpha,\; 0 < \alpha < 1,\;\) the point \(\alpha\;x + (1-\alpha) x \in {\bf S}.\)

A set \({\bf K} \subset {\Re}^n\;\) is said to be convex if for any two vectors x and y in \({\bf K}\) any vector of the form \( (1 - \lambda) x + \lambda y\) is also in \({\bf K}\), where \(0 \leq \lambda \leq 1\). This simply means that given two points in a convex set, the line segment between them is also in the set. Note, in particular, that subspaces and linear varieties (a linear variety is a translation of linear subspaces) are convex. Also the empty set is considered convex. The following facts provide important properties for convex sets .

  1. 1.

    Let \({\cal C}_j,\;j=1,\dots,m\) be a family of m convex sets in \({\Re}^n\). Then the intersection \({\cal C}_1 \cap {\cal C}_2 \cap {}_{\dots} \cap{\cal C}_m\).

  2. 2.

    Let \({\cal C}\) be a convex set in \({\Re}^n\) and \(x_o \in {\Re}^n\). Then the set \(\{x_o + x: x \in {\cal C}\}\) is convex.

  3. 3.

    A set \({\bf K} \subset {\Re}^n\;\) is said to be convex cone with vertex \(x_{o}\) if \({\bf K}\) is convex, and \(x \in {\bf K}\) implies that \(x_o + \lambda x \in {\bf K}\) for any \(\lambda \geq 0\).

An important class of convex cones is the one defined by the positive semidefinite ordering of matrices, that is, \(A_1 \; \geq \; A_2 \; \geq \; A_3\). Let \(P \in {\Re}^{n \times n}\) be a positive semidefinite matrix. The set of matrices \(X \in {\Re}^{n \times n}\), such that \(X \geq P\) is a convex cone in \({\Re}^{n \times n}\).

2.4 Continuous Functions

A function \(f:{\Re}^n\;\longrightarrow\;{\Re}^m\) is said to be continuous at a point x if \(f(x + \delta x) \;\longrightarrow\; f(x)\) whenever \(\delta x \;\longrightarrow\;0.\) Equivalently, f is continuous at x if, given \(\epsilon > 0,\;\) there is \(\delta > 0\) such that

$$||x - y||\;<\; \epsilon\;\Longrightarrow\;||f(x) - f(y)||\;<\; \epsilon$$

A function f is continuous on a set of S if it is continuous at every point of S, and it is uniformly continuous on S if given \(\epsilon\; > \;0,\;\) there is \(\delta(\epsilon)\;>\;0\;\) (dependent only on ε), such that the inequality holds for all \(x, y \in {\bf S}\)

A function \(f:{\Re}\;\longrightarrow\;{\Re}\) is said to be differentiable at a point x if the limit

$$\dot{f}(x) = \lim_{\delta x \rightarrow 0} \frac{f(x + \delta x) - f(x)} {\delta x}$$

exists. A function \(f:{\Re}^n\;\longrightarrow\;{\Re}^m\) is continuously differentiable at a point x (a set S) if the partial derivatives \(\partial f_{j}/\partial x_s\;\) exist and continuous at x (at every point of S) for \(1 \; \leq \; j \;\leq \; m,\;1 \; \leq \; s \;\leq \; n\) and the Jacobian matrix is defined as

$${\bf J} = \Big [\frac{\partial f}{ \partial x} \Big] = \left [ \begin {array} {ccccc} \partial f_{1}/\partial x_1 & & \cdots & & \partial f_{1}/\partial x_n \\ \vdots & & \ddots & & \vdots \\ \partial f_{m}/\partial x_1 & & \cdots & & \partial f_{m}/\partial x_n \end {array}\right ] \; \in \; {\Re}^{m \times n}$$

2.5 Function Norms

Let \(f(t):{\Re}_+\;\longrightarrow\;{\Re}\) be a continuous function or piecewise continuous function. The p-norm of f is defined by

$$\begin{array}{*{20}l}||f||_p &= \bigg ( \int^\infty_0 \; |f(t)|^p \; {\mathrm{d}}t \bigg )^{1/p},\;\;\;\; f\,{\textrm{or}}\; p\; \in \; [1,\infty)\\ ||f||_\infty &= \sup{\,\,t \in [0,\infty)} |f(t)|,\;\;\; f\,{\textrm{or}}\; p\; =\infty \end{array}$$

By letting \(p = 1, 2, \infty,\;\) the corresponding normed spaces are called \({\bf L_1},\;{\bf L_2},\;{\bf L_\infty},\;\) respectively. More precisely, let \(f(t)\) be a function on \([0,\infty)\) of the signal spaces, they are defined as

$$\begin{array}{*{20}l}{\bf L_1} &{\stackrel{\Delta}{=}} \bigg \{ f(t):{\Re}_+\;\longrightarrow\;{\Re} | ||f||_1 = \int^\infty_0 \; |f(t)| \; {\mathrm{d}}t\; < \; \infty,\;\; {\mathrm{convolution \; kernel}} \bigg \} \\ {\bf L_2} &{\stackrel{\Delta}{=}} \bigg \{ f(t):{\Re}_+\;\longrightarrow\;{\Re} | ||f||_2 = \int^\infty_0 \; |f(t)|^2 \; {\mathrm{d}}t\; < \; \infty,\;\; {\mathrm{finite \; energy}} \bigg \} \\ {\bf L_\infty} &{\stackrel{\Delta}{=}} \bigg \{ f(t):{\Re}_+\;\longrightarrow\;{\Re} | ||f||_\infty = \sup_{t\in[0,\infty)} \; |f(t)|\; < \; \infty,\;\; {\mathrm{bounded \; signal}} \bigg \} \end{array}$$

From a signal point of view, the 1-norm , \(||x||_1\) of the signal x(t) is the integral of its absolute value, the square \(||x||^2_2\) of the 2-norm is often called the energy of the signal \(x(t),\;\) and the ∞-norm is its absolute maximum amplitude or peak value. It must be emphasized that the definitions of the norms for vector functions are not unique.

In the case of \(f(t):{\Re}_+\;\longrightarrow\;{\Re}^n,\; f(t) = [f_1(t)\;\;f_2(t) \dots f_n(t)]^t\) which denote a continuous function or piecewise continuous vector function, the corresponding p-norm spaces are defined as

$$\begin{array}{*{20}l}{\bf L^n_p} &{\stackrel{\Delta}{=}} \bigg \{ f(t):{\Re}_+\;\longrightarrow\;{\Re}^n | ||f||_p = \int^\infty_0 \, ||f(t)||^p \, {\mathrm{d}}t\; < \; \infty,\;\, f\,{{or}} \; p \; \in [1,\infty) \bigg \} \\ {\bf L^n_\infty} &{\stackrel{\Delta}{=}} \bigg \{ f(t):{\Re}_+\;\longrightarrow\;{\Re}^n | ||f||_\infty = \sup_{t\in[0,\infty)} \; ||f(t)|| \; < \; \infty \bigg \} \end{array}$$

3 Calculus and Algebra of Matrices

In this section, we solicit some basic facts and useful relations from linear algebra and calculus of matrices. The materials are stated along with some hints whenever needed but without proofs unless we see the benefit of providing a proof. Reference is made to matrix M or matrix function M(t) in the form

$$M = \left [ \begin {array} {*{20}c} M_{11} & & \cdots & & M_{1n}\\ \vdots & & \ddots & & \cdots \\ M_{m1} & & \cdots & & M_{mn} \end {array} \right ],\;\;{{or}}\;\;M(t) = \left [ \begin {array} {*{20}c} M_{11}(t) & & \cdots & & M_{1n}(t) \\ \vdots & & \ddots & & \cdots \\ M_{m1}(t) & & \cdots & & M_{mn}(t) \end {array} \right ]$$

3.1 Fundamental Subspaces

A nonempty subset \({\cal G} \subset {\Re}^n\) is called a linear subspace of \({\Re}^n\) if \(x + y\) and \(\alpha x\) are in \({\cal G}\) whenever x and y are in \({\cal G} \) for any scalar α. A set of elements \(X = \{x_1,\;x_2,\;\dots,\;x_n\}\) is said to be a spanning set for a linear subspace \({\cal G}\) of \({\Re}^n\) if every element \(g \in {\cal G} \) can be written as a linear combination of the \(\{x_j\}\). That is, we have

$${\cal G} = \{ g \in {\Re}: \;g = \alpha_1 x_1 + \alpha_2 x_2+ \;\dots\;\alpha_n x_n$$

for some scalars \(\alpha_1,\;\alpha_2,\dots,\;\alpha_n\).

A spanning set X is said to be a basis for \({\cal G}\) if no element \(x_{j}\) of the spanning set X of \({\cal G}\) can written as a linear combination of the remaining elements \(x_1,\;x_2,\dots,\;x_{j-1},\;x_{j+1},\dots,\;x_n,\;\) that is, \(x_j,\; 1 \leq i \leq n\) form a linearly independent set. It is frequent to use \(x_j=[0\;0\;\dots\;0\;1\;0\;\dots\;0]^t\) the kth unit vector.

The geometric ideas of linear vector spaces had led to the concepts of spanning a space and a basis for a space. The idea now is to introduce four important subspaces which are useful. The entire linear vector space of a specific problem can be decomposed into the sum of these subspaces.

The column space of a matrix \(A \in {Re}^{n \times m}\) is the space spanned by the columns of A, also called the range space of A, denoted by \({\cal R}[A]\). Similarly, the row space of A is the space spanned by the rows of A. Since the column rank of a matrix is the dimension of the space spanned by the columns and the row rank is the dimension of the space spanned by the rows, it is clear that the spaces \({\cal R}[A]\) and \({\cal R}[A^t]\) have the same dimension \(r = {\mathrm{rank}}(A)\).

The right null space of \(A \in {Re}^{n \times m}\) is the space spanned by all vectors x that satisfy \(A\;x = 0,\) and is denoted by \({\cal N}[A]\). The right null space of A is also called the kernel of A. The left null space of A is the space spanned by all vectors y that satisfy \(y^t\;A = 0\). This space is denoted by \({\cal N}[A^t],\;\) since it is also characterized by all vectors y such that \(A^t\;y = 0\).

The dimensions of the four spaces \({\cal R}[A],\;{\cal R}[A^t],\;{\cal N}[A]\), and \({\cal N}[A^t]\) are to be determined in the sequel. Since \(A \in {\Re}^{n \times m},\;\) we have the following

r

\({\stackrel{\Delta}{=}}\)

rank\((A)=\) dimension of column space \({\cal R}[A]\)

dim\({{\cal N}[A]}\)

\({\stackrel{\Delta}{=}}\)

dimension of right null space \({\cal N}[A]\)

n

\({\stackrel{\Delta}{=}}\)

total number of columns of A

Hence the dimension of the null space dim\({{\cal N}[A]} = n - r\). Using the fact that \({\mathrm{rank}}(A) = {\mathrm{rank}}(A^t)\), we have

r

\({\stackrel{\Delta}{=}}\)

rank\((A^t)\) = dimension of row space \({\cal R}[A^t]\)

dim\({{\cal N}[A^t]}\)

\({\stackrel{\Delta}{=}}\)

dimension of left null space \({\cal N}[A^t]\)

m

\({\stackrel{\Delta}{=}}\)

total number of rows of A

Hence the dimension of the null space dim\({{\cal N}[A^t]} = m - r\). These facts are summarized below.

Note from these facts that the entire n-dimensional space can be decomposed into the sum of the two subspaces \({\cal R}[A^t]\) and \({\cal N}[A]\). Alternatively, the entire m-dimensional space can be decomposed into the sum of the two subspaces \({\cal R}[A]\) and \({\cal N}[A^t]\).

An important property is that \({\cal N}[A]\) and \({\cal R}[A^t]\) are orthogonal subspaces, that is, \({\cal R}[A^t]^{\bot}={\cal N}[A]\). This has the meaning that every vector in \({\cal N}[A]\) is orthogonal to every vector in \({\cal R}[A^t]\). In the same manner, \({\cal R}[A]\) and \({\cal N}[A^t]\) are orthogonal subspaces, that is, \({\cal R}[A]^{\bot}={\cal N}[A^t]\). The construction of the fundamental subspaces is appropriately attained by the singular value decomposition.

\({\cal R}[A^t]\)

\({\stackrel{\Delta}{=}}\)

row space of \(A\!:\) dimension r

\({\cal N}[A]\)

\({\stackrel{\Delta}{=}}\)

right null space of \(A\!:\) dimension \(n\;-\;r\)

\({\cal R}[A]\)

\({\stackrel{\Delta}{=}}\)

column space of \(A\!:\) dimension r

\({\cal N}[A^t]\)

\({\stackrel{\Delta}{=}}\)

left null space of \(A\!:\) dimension \(n\;-\;r\)

3.2 Calculus of Vector –Matrix Functions of a Scalar

The differentiation and integration of time functions involving vectors and matrices arise in solving state equations, optimal control, and so on. This section summarizes the basic definitions of differentiation and integration on vectors and matrices . A number of formulas for the derivative of vector –matrix products are also included.

The derivative of a matrix function M(t) of a scalar is the matrix of the derivatives of each element in the matrix

$$\frac{{\mathrm{d}} M(t)}{{\mathrm{d}} t} = \left [ \begin {array} {ccccc} \frac{{\mathrm{d}} M_{11}(t)}{{\mathrm{d}} t} & & \cdots & & \frac{{\mathrm{d}} M_{1n}(t)}{{\mathrm{d}} t}\\ \vdots & & \ddots & & \cdots \\ \frac{{\mathrm{d}} M_{m1}(t)}{{\mathrm{d}} t} & & \cdots & & \frac{{\mathrm{d}} M_{mn}(t)}{{\mathrm{d}} t} \end {array} \right ]$$

The integral of a matrix function M(t) of a scalar is the matrix of the integral of each element in the matrix

$$\int^b_a \;M(t) {\mathrm{d}} t = \left [ \begin {array} {ccccc} \int^b_a \;M_{11}(t) {\mathrm{d}} t & & \cdots & & \int^b_a \;M_{1n}(t) {\mathrm{d}} t \\ \vdots & & \ddots & & \cdots \\ \int^b_a \;M_{m1}(t) {\mathrm{d}} t & & \cdots & & \int^b_a \;M_{mn}(t) {\mathrm{d}} t \end {array} \right ]$$

The Laplace transform of a matrix function M(t) of a scalar is the matrix of the Laplace transform of each element in the matrix

$$\int^b_a \;M(t) e^{-st} {\mathrm{d}} t = \left [ \begin {array} {ccccc} \int^b_a \;M_{11}(t) e^{-st}{\mathrm{d}} t & & \cdots & & \int^b_a \;M_{1n}(t) e^{-st}{\mathrm{d}} t \\ \vdots & & \ddots & & \cdots \\ \int^b_a \;M_{m1}(t) e^{-st}{\mathrm{d}} t & & \cdots & & \int^b_a \;M_{mn}(t) e^{-st}{\mathrm{d}} t \end {array} \right ]$$

The scalar derivative of the product of two matrix time functions is

$$\frac{{\mathrm{d}} (A(t) B(t))}{{\mathrm{d}}t} = \frac{A(t)}{{\mathrm{d}}t} B(t) + A(t) \frac{B(t)}{{\mathrm{d}}t}$$

This result is analogous to the derivative of a product of two scalar functions of a scalar, except caution must be used in reserving the order of the product. An important special case follows:

The scalar derivative of the inverse of a matrix time function is

$$\frac{{\mathrm{d}} A^{-1}(t)}{{\mathrm{d}}t} = - A^{-1} \frac{A(t)}{{\mathrm{d}}t} A(t)$$

3.3 Derivatives of Vector –Matrix Products

The derivative of a real scalar-valued function f(x) of a real vector \(x = [x_1,\dots,x_n]^t \;\in {Re}^n\) is defined by

$$\frac{\partial f(x)}{\partial x} = \left [ \begin {array} {c} \frac{\partial f(x)}{\partial x_1} \\ \frac{\partial f(x)}{\partial x_2} \\ \vdots \\ \frac{\partial f(x)}{\partial x_n} \end {array} \right ]$$

where the partial derivative is defined by

$$\frac{\partial f(x)}{\partial x_j} {\stackrel{\Delta}{=}} \lim_{\Delta x_j \rightarrow 0} \; \frac{f(x+ \Delta x) - f(x)}{\Delta x_j},\;\;\; \Delta x = [0\dots\Delta x_j \dots0]^t$$

An important application arises in the Taylor’s series expansion of f(x) about \(x_{o}\) in terms of \(\delta x {\stackrel{\Delta}{=}} x - x_o\). The first three terms are

$$f(x) = f(x_o) + \left (\frac{\partial f(x)}{\partial x} \right )^t \;\delta x + \frac{1}{2} \delta x^t\; \bigg [\frac{\partial^2 f(x)}{\partial x^2} \bigg ]\; \delta x$$

where

$$\begin {array}{*{20}l}\frac{\partial f(x)}{\partial x} &= \left [ \begin {array} {c} \frac{\partial f(x)}{\partial x_{1}} \\ \\ \vdots \\ \\ \frac{\partial f(x)}{\partial x_{n}} \end {array} \right ]\\ \frac{\partial^2 f(x)}{\partial x^2} = \frac{\partial}{\partial x} \bigg ( \frac{\partial f(x)}{\partial x} \bigg )^t &= \left [ \begin {array} {ccccc} \frac{\partial^2 f(x)}{\partial x^2_1} & & \cdots & & \frac{\partial^2 f(x)}{\partial x_1 \partial x_n} \\ \vdots & & \ddots & & \cdots \\ \frac{\partial^2 f(x)}{\partial x_n \partial x_1} & & \cdots & & \frac{\partial^2 f(x)}{\partial x^2_n} \end {array} \right ]\end{array}$$

The derivative of a real scalar-valued function f(A) with respect to a matrix

$$A = \left [ \begin {array}{*{20}l} A_{11} & & \cdots & & A_{1n}\\ \vdots & & \ddots & & \cdots \\ A_{n1} & & \cdots & & A_{nn} \end {array} \right ] \;\in {Re}^{n \times n}$$

is given by

$$\frac{\partial f(A)}{\partial A} = \left [ \begin {array} {ccccc} \frac{\partial f(A)}{\partial A_{11}} & & \cdots & & \frac{\partial f(A)}{\partial A_{1n}}\\ \vdots & & \ddots & & \cdots \\ \frac{\partial f(A)}{\partial A_{n1}} & & \cdots & & \frac{\partial f(A)}{\partial A_{nn}} \end {array} \right ]$$

A vector function of a vector is given by

$$v(u) = \left [ \begin {array} {c} v_1(u) \\ \vdots \\ \vdots \\ v_n(u) \end {array} \right ]$$

where \(v_j(u)\) is a function of the vector u. The derivative of a vector function of a vector (the Jacobian) is defined as follows:

$$\frac{\partial v(u)}{\partial u} = \left [ \begin {array} {ccccc} \frac{\partial v_{1}(u)}{\partial u_{1}} & & \cdots & & \frac{\partial v_{1}(u)}{\partial u_{m}}\\ \vdots & & \ddots & & \cdots \\ \frac{\partial v_{n}(u)}{\partial u_{1}} & & \cdots & & \frac{\partial v_{n}(u)}{\partial u_{m}} \end {array} \right ]$$

Note that the Jacobian is sometimes defined as the transpose of the foregoing matrix. A special case is given by

$$\frac{\partial (S\;u)}{\partial u} = S,\;\;\;\frac{\partial (u^t R u)}{\partial u} = 2\;u^t R$$

for arbitrary matrix S and symmetric matrix R.

The following section includes useful relations and results from linear algebra.

3.4 The Dini Theorem

3.5 Positive Definite and Positive Semidefinite Matrices

A matrix P is positive definite if P is real, symmetric, and \(x^t P x > 0,\;\; \forall x \neq 0\). Equivalently, all the eigenvalues of P have positive real parts. A matrix S is positive semidefinite if S is real, symmetric, and \(x^t P x \geq 0,\;\; \forall x \neq 0\).

Since the definiteness of the scalar \(x^t P x\) is a property only of the matrix P, we need a test for determining definiteness of a constant matrix P. Define a principal submatrix of a square matrix P as any square submatrix sharing some diagonal elements of P. Thus the constant, real, symmetric matrix \(P \in {\Re}^{n \times n}\) is positive definite \((P > 0)\) if either of these equivalent conditions holds:

  • All eigenvalues of P are positive

  • The determinant of P is positive

  • All successive principal submatrices of P (minors of successively increasing size) have positive determinants

3.6 Trace Properties

The trace of a square matrix P, trace (P), equals the sum of its diagonal elements or equivalently the sum of its eigenvalues. A basic property of the trace is invariant under cyclic perturbations, that is,

$${\mathrm{trace}}(AB) = {\mathrm{trace}}(BA)$$

where AB is square. Successive applications of the above results yield

$${\mathrm{trace}}(ABC) = {\mathrm{trace}}(BCA) = {\mathrm{trace}}(CAB)$$

where ABC is square. In general,

$${\mathrm{trace}}(AB) = {\mathrm{trace}}(B^t A^t)$$

Another result is that

$${\mathrm{trace}}(A^t B A) = \sum^p_{k=1}\;a^t_k B a_k$$

where \(A \in {\Re}^{n \times p},\;B \in {\Re}^{n \times n}\), and \(\{a_k\}\) are the columns of A. The following identities on trace derivatives are noted:

$$\begin {array}{*{20}l}\frac{\partial({\mathrm{trace}}(A B))} {\partial A} &= \frac{\partial({\mathrm{trace}}(A^t B^t))} {\partial A} = \frac{\partial({\mathrm{trace}}(B^t A^t))} {\partial A}\\ &= \frac{\partial({\mathrm{trace}}(B A))} {\partial A} = B^t \\ \frac{\partial({\mathrm{trace}}(AB))} {\partial B} &= \frac{\partial({\mathrm{trace}}(A^t B^t))} {\partial B} = \frac{\partial({\mathrm{trace}}(B^t A^t))} {\partial B} \\ &= \frac{\partial({\mathrm{trace}}(B A))} {\partial B} = A^t \\ \frac{\partial({\mathrm{trace}}(B A C))} {\partial A} &= \frac{\partial({\mathrm{trace}}(B^t C^t A^t))} {\partial A} = \frac{\partial({\mathrm{trace}}(C^t A^t B^t))} {\partial A} \\ &= \frac{\partial({\mathrm{trace}}(A C B))} {\partial A}= \frac{\partial({\mathrm{trace}}(C B A))} {\partial A} \\ &= \frac{\partial({\mathrm{trace}}(A^t B^t C^t))} {\partial A} = B^t\;C^t \\ \frac{\partial({\mathrm{trace}}(A^t B A))} {\partial A} &= \frac{\partial({\mathrm{trace}}(B A A^t))} {\partial A} = \frac{\partial({\mathrm{trace}}(A A^t B))} {\partial A} \\ &= (B + B^t) A \end{array}$$

Using these basic ideas, a list of matrix calculus results are given below:

$$\begin {array}{*{20}l}\frac{\partial({\mathrm{trace}}(A X^t))} {\partial X} &= A,\;\;\; \frac{\partial({\mathrm{trace}}(A X B))} {\partial X} = A^t \; B^t\\ \frac{\partial({\mathrm{trace}}(A X^t B))} {\partial X} &= B\; A,\;\;\; \frac{\partial({\mathrm{trace}}(A X))} {\partial X^t} = A\\ \frac{\partial({\mathrm{trace}}(A X^t))} {\partial X^t} &= A^t,\;\;\; \frac{\partial({\mathrm{trace}}(A X B))} {\partial X^t} = B\;A\\ \frac{\partial({\mathrm{trace}}(A X^t B))} {\partial X^t} &= A^t\; B^t,\;\;\; \frac{\partial({\mathrm{trace}}(X X))} {\partial X} = 2\;X^t\\ \frac{\partial({\mathrm{trace}}(X X^t))} {\partial X} &= 2\; X\\ \frac{\partial({\mathrm{trace}}(A X^n))} {\partial X} &= \left (\sum_{j=0}^{n-1}\;X^j\;A\;X^{n-j-1} \right)^t\\ \end{array}$$
$$\begin {array}{*{20}l}\frac{\partial({\mathrm{trace}}(A X B X))} {\partial X} &= A^t X^t B^t + B^t X^t A^t\\ \frac{\partial({\mathrm{trace}}(A X B X^t))} {\partial X} &= A^t X B^t + A X B\\ \frac{\partial({\mathrm{trace}}(X^{-1}))} {\partial X} &= - \big(X^{-2} \big)^t\\ \frac{\partial({\mathrm{trace}}(A X^{-1} B))} {\partial X} &= - \bigg(X^{-1} B A X^{-1} \bigg)^t\\ \frac{\partial({\mathrm{trace}}(A B))} {\partial A} &= B^t + B - {\mathrm{diag}}(B) \end{array}$$

3.7 Partitioned Matrices

Given a partitioned matrix (matrix of matrices) of the form

$$M= \left [ \begin {array} {*{20}c} A & & B \\ & & \\C & & D \end {array} \right ]$$

where \(A,\;B,\;C\), and D are of compatible dimensions. Then

  1. (1)

    if \(A^{-1}\) exists, a Schur complement of M is defined as \(D - C A^{-1} B\), and

  2. (2)

    if \(D^{-1}\) exists, a Schur complement of M is defined as \(A - B D^{-1} C\).

When \(A,\;B,\;C\), and D are all \(n \times n\) matrices, then

$$\begin {array}{*{20}l} a) \;\;\;\;& \det \left [ \begin {array} {ccc} A & & B \\ & & \\ C & & D \end {array} \right ] = \det(A) \det(D - C A^{-1} B),\; \det(A) \neq 0 \\ b) \;\;\;\;& \det \left [ \begin {array} {ccc} A & & B \\ & & \\ C & & D \end {array} \right ] = \det(D) \det(A - B D^{-1} C),\; \det(D) \neq 0 \end{array}$$

In the special case, we have

$$\det \left [ \begin {array} {*{20}c} A & & B \\ & & \\ C & & 0 \end {array} \right ] = \det(A) \det(C)$$

where A and C are square. Since the determinant is invariant under row, it follows

$$\begin {array}{*{20}l}\det \left [ \begin {array} {*{20}c} A & & B \\ & & \\ C & & D \end {array} \right ] &= \det \left [ \begin {array} {ccc} A & & B \\ & & \\C - CA^{-1} A& & D-CA^{-1}B \end {array} \right ] \\ &= \det \left [ \begin {array} {ccc} A & & B \\ & & \\ 0 & & D-CA^{-1}B \end {array} \right ] = \det(A) \det(D - C A^{-1} B)\end{array}$$

which justifies the forgoing result.

Given matrices \(A \in {\Re}^{m \times n}\) and \(B \in {\Re}^{n \times m}\), then

$$\det(I_m - AB) = \det(I_n - BA)$$

In case that A is invertible, then \(\det(A^{-1}) = \det(A)^{-1}\).

3.8 The Matrix Inversion Lemma

Suppose that \(A \in {\Re}^{n \times n},\;B \in {\Re}^{n \times p},\;C \in {\Re}^{p \times p},\;\) and \(D \in {\Re}^{p \times n}\). Assume that \(A^{-1}\) and \(C^{-1}\) both exist, then

$$(A + B C D )^{-1} = A^{-1} - A^{-1} B (D A^{-1} B + C^{-1})^{-1} D A^{-1}$$

In the case of partitioned matrices, we have the following result

$$\begin {array}{*{20}c}\left [ \begin {array}{*{20}c} A & & B \\ & & \\ C & & D \end {array} \right ]^{-1} &= \left [ \begin {array} {ccc} A^{-1}+A^{-1}B \Xi^{-1}C A^{-1} & & - A^{-1} B \Xi^{-1} \\ & & \\ - \Xi^{-1} CA^{-1} & & \Xi^{-1} \end {array} \right ] \\ \Xi &= (D - C A^{-1} B)\end{array}$$

provided that \(A^{-1}\) exists. Alternatively,

$$\begin {array}{*{20}c}\left [ \begin {array} {ccc} A & & B \\ & & \\ C & & D \end {array} \right ]^{-1} &= \left [ \begin {array} {ccc} \Xi^{-1} & & - \Xi^{-1} B D^{-1} \\ & & \\ - D^{-1} C \Xi^{-1} & & D^{-1} + D^{-1} C \Xi^{-1} B D^{-1} \end {array} \right ] \\ \Xi&= (D - C A^{-1} B)\end{array}$$

provided that \(D^{-1}\) exists.

For a square matrix Y, the matrices Y and \((I + Y)^{-1}\) commute, that is, given that the inverse exists

$$Y\;(I + Y)^{-1} = (I + Y)^{-1} \;Y$$

Two additional inversion formulas are given below:

$$\begin {array}{*{20}l}Y\;(I + X Y)^{-1} \;&=\; (I + Y X )^{-1} \;Y\\ (I + Y X )^{-1}\;&=\; I - Y X \;(I + Y X )^{-1}\end{array}$$

The following result provides conditions for the positive definiteness of a partitioned matrix in terms of its submatrices. The following three statements are equivalent:

$$\begin {array}{*{20}l}&1)\ \left [ \begin {array} {ccc} A_o & & A_a \\ & & \\ A^t_a & & A_c \end {array} \right ] \;>\;0 \\ &2)\ A_c > 0,\;\;\;\;A_o - A_a A^{-1}_c A^t_a \;>\; 0 \\ &3)\ A_a > 0,\;\;\;\;A_c - A^t_a A^{-1}_o A_a \;>\; 0\end{array}$$

3.9 The Singular Value Decomposition

The singular value decomposition (SVD) is a matrix factorization that has found a number of applications to engineering problems. The SVD of a matrix \(M \in {Re}^{n \times m}\) is

$$M = U\;S\;V^{\dag} = \sum^{p}_{j=1} \; \sigma_j\;U_j\;V^{\dag}_j$$

where \(U \in {Re}^{\alpha \times \alpha}\) and \(V \in {Re}^{\beta \times \beta}\) are unitary matrices (\(U^{\dag}\;U =U \;U^{\dag}= I\) and \(V^{\dag}\;V = V \;V^{\dag}I\)); \(S \in {Re}^{\alpha \times \beta}\) is a real, diagonal (but not necessarily square); and \(p min(\alpha, \beta)\). The singular values \(\{\sigma_1, \sigma_2,\dots,\sigma_{\beta} \}\) of M are defined as the positive square roots of the diagonal elements of \(S^t S\), and are ordered from largest to smallest.

To proceed further, we recall a result on unitary matrices. If U is a unitary matrix (\(U^{\dag}\;U = I\)), then the transformation U preserves length, that is,

$$\begin {array}{*{20}l}||U\;x|| \;&=\; \sqrt{(Ux)^{\dag} (Ux)} = \sqrt{x^{\dag}\;U^{\dag}\;U\;x} \\ \;&= \sqrt{x^{\dag}\;x} = ||x||\end{array}$$

As a consequence, we have

$$\begin {array}{*{20}l}||M\;x|| \;&=\; \sqrt{x^{\dag}\;M^{\dag}\;M\;x} = \sqrt{x^{\dag}\;V S^t U^{\dag}\;U S V^{\dag}\;x} \\ \;&= \sqrt{x^{\dag}\;V S^t S V^{\dag}\;x}\end{array}$$

To evaluate the maximum gain of matrix M, we calculate the maximum norm of the above equation to yield

$$\max _{||x||=1}\;||M\;x|| = \max _{||x||=1}\; \sqrt{x^{\dag}\;V S^t S V^{\dag}\;x} = \max _{||\tilde{x}||=1}\; \sqrt{\tilde{x}^{\dag}\;VS^t S \;\tilde{x}}$$

Note that maximization over \(\tilde{x} = V x\) is equivalent to maximizing over x since V is invertible and preserves the norm (equals 1 in this case). Expanding the norm yields

$$\begin {array}{*{20}l}\max _{||x||=1}\;||M\;x|| \;&=\;\max _{||\tilde{x}||=1}\; \sqrt{\tilde{x}^{\dag}\;VS^t S \;\tilde{x}}\\ &=\; \max _{||\tilde{x}||=1}\; \sqrt{ \sigma^2_1 |\tilde{x}_1|^2 + \sigma^2_2 |\tilde{x}_2|^2 +\cdots+ \sigma^2_\beta |\tilde{x}_\beta|^2}\end{array}$$

The foregoing expression is maximized, given the constraint \(||\tilde{x}||=1\), when \(\tilde{x}\) is concentrated at the largest singular value; that is, \(|\tilde{x}| = [1\;0\;\dots\;0]^t\). The maximum gain is then

$$\max _{||x||=1}\;||M\;x|| = \sqrt{ \sigma^2_1 |1|^2 + \sigma^2_2 |0|^2 +\cdots+ \sigma^2_\beta |0|^2} = \sigma_1 = {\sigma}_M$$

In words, this reads the maximum gain of a matrix is given by the maximum singular value \({\sigma}_M\). Following similar lines of development, it is easy to show that

$$\begin {array}{*{20}l}\min _{||x||=1}\;||M\;x|| \;&=\; \sigma_\beta = {\sigma}_m\\ &=\; \left \{ \begin {array} {cccc} \sigma_p & & & \alpha \geq \beta \\ 0 & & & \alpha < \beta \end {array} \right\}\end{array}$$

A property of the singular values is expressed by

$$\sigma_M(M^{-1}) \; = \; \frac{1} {\sigma_m (M)}$$

4 Notes and References

The topics covered in this chapter is meant to provide the reader with a general platform containing the basic mathematical information needed for further examination of switched time-delay systems. These topics are properly selected from standard books and monographs on mathematical analysis. For further details, the reader is referred to the standard texts [29, 46, 157, 160, 443] where fundamentals are provided.