1 Introduction

Since neural networks have been extensively applied in many areas such as reconstruction of moving image, signal processing, the tasks of pattern recognition, associative memories, fixed-point computations, and so on, the stability analysis of the concerned neural networks is a very important and prerequisite job because the application of neural networks heavily depends on the dynamic behavior of equilibrium points [15]. Also, due to the finite speed of information processing in the implementation of the network, time-delay occurs in many neural networks. It is well known that time-delay often causes undesirable dynamic behaviors such as oscillation and instability of the networks. Thus, delay-dependent stability and stabilization problem for neural networks with time-delay have been paid more attention than delay-independent ones because the information on the size of time-delays is utilized in delay-dependent criteria, which lead to reduce the conservatism of stability and stabilization criteria. To confirm this, see [623] and references therein.

In practice, it should be noted that the signal propagation is sometimes instantaneous and can be modeled with discrete delays. Also, it may be distributed during a certain time period so that the distributed delays should be incorporated in the model. That is, it is often the case that the neural network model possesses both discrete and distributed delays [24]. In this regard, the stability of cellular neural networks with discrete and distributed delays has been investigated in [2528]. Furthermore, the norm-bounded parametric uncertainties, which sometimes affect the stability of systems, are considered in the works [2628].

On the other hand, in various scientific and engineering problems, stability issues are often linked to the theory of dissipative systems. It postulates that the energy dissipated inside a dynamic system is less than the energy supplied from the external source [29]. Based on the concept of energy, the passivity is the property of dynamical systems and describes the energy flow through the system. It is also an input/output characterization and related to Lyapunov method. In the field of nonlinear control, the concept of dissipativeness was firstly introduced by Willems [30] in the form of inequality including supply rate and the storage function. The main idea of passivity theory is that passive properties of a system can keep the system internally.

In this regard, in [3138], the passivity problem for the uncertain neural networks with both discrete and distributed time-varying delays was considered. Chen et al. [31] investigated the passivity problem of the neural networks by utilizing free-weighting matrices and the LMI framework. In [32], improved delay-dependent passivity criteria for the networks were proposed. Xu et al. [33] studied the problem of passivity analysis for neural networks with both time-varying delays and norm-bounded parameter uncertainties. In [34], improved passivity criteria for stochastic neural networks with interval time-varying delays and norm-bounded parameter uncertainties were proposed via an improved approximation method. In [35], for two types of time-varying delays, new delay-dependent passivity conditions of delayed neural networks were derived. Recently, by taking more information of states as augmented vectors, an augmented Lyapunov–Krasovskii functional was utilized in [36] to derive passivity criteria for uncertain neural networks with time-varying delays. Song and Cao [37] investigated the passivity problem for a class of uncertain neural networks with leakage delay and time-varying delays by employing Newton–Leibniz formulation and the free weighting matrix method. Very recently, in [38], by constructing a novel Lyapunov–Krasovskii functional including a tuning parameter in time-varying delays and introducing some proper free-weighting matrices, new passivity conditions for neural networks with both discrete and distributed time-varying delays were developed to guarantee the passivity performance of the networks. However, there are rooms for further improvement in enhancing the feasible regions of passivity criteria.

Motivated by above discussion, in this paper, the problem on delay-dependent passivity for uncertain neural networks with both discrete and distributed time-varying delays is addressed. The parameter uncertainties are assumed to be norm-bounded. The main contribution of this paper lies in two aspects:

  1. 1.

    Unlike the method of [38], no tuning parameters in a time-varying delay are utilized. Instead, by taking more information of states, a newly constructed Lyapunov–Krasovskii functional is proposed. Then, inspired by the work of [3941], a passivity condition for neural networks with both discrete and distributed time-varying delays and parameter uncertainties is derived in terms of LMIs which will be introduced in Theorem 1.

  2. 2.

    A novel approach partitioning m-interval of the range of the activation function divided by state will be proposed. Through three numerical examples, it will be shown the maximum delay bounds for guaranteeing the passivity of the considered neural networks increase when the partitioning number of the bounding of activation function gets larger.

Based on the result of Theorem 1, a passivity criterion for uncertain neural networks with only discrete time-varying delays will be proposed in Theorem 2. Finally, three numerical examples are included to show the effectiveness of the proposed methods.

Notation

\(\mathbb{R}^{n}\) is the n-dimensional Euclidean space, and \(\mathbb{R}^{m \times n}\) denotes the set of all m×n real matrices. For symmetric matrices X and Y, X>Y (respectively, XY) means that the matrix XY is positive definite (respectively, nonnegative). X denotes a basis for the null-space of X. I n , 0 n, and 0 mn denotes n×n identity matrix, n×n and m×n zero matrices, respectively. ∥⋅∥ refers to the Euclidean vector norm or the induced matrix norm. \(\operatorname{diag} \{ \cdots\}\) denotes the block diagonal matrix. For square matrix S, \(\operatorname{sym}\{ S\}\) means the sum of S and its symmetric matrix S T; i.e., \(\operatorname{sym}\{S\}=S+S^{T}\). ⋆ represents the elements below the main diagonal of a symmetric matrix.

2 Problem statements

Consider the following uncertain neural networks with both discrete and distributed time-varying delays:

(1)

where n denotes the number of neurons in a neural network, \(x(t)=[x_{1}(t),\ldots,x_{n}(t)]^{T}\in\mathbb{R}^{n}\) is the neuron state vector, \(y(t)\in\mathbb{R}^{n}\) is the output vector, \(f(x_{i}(\cdot))=[f_{1}(x_{i}(\cdot)),\ldots,f_{n}(x_{i}(\cdot))]^{T}\in\mathbb{R}^{n}\) denotes the neuron activation function vector, \(u(t)\in\mathbb{R}^{n}\) is the input vector, \(A=\operatorname{diag}\{a_{1},\ldots,a_{n}\}\in\mathbb {R}^{n\times n}\) is a positive diagonal matrix, \(W_{i}\in\mathbb{R}^{n\times n}\ (i=0,1,2)\) are the interconnection weight matrices, \(C_{i} \in\mathbb{R}^{n\times n}\ (i=1,2)\) are known constant matrices, and ΔA(t) and ΔW i (t)(i=0,1,2) are the parameter uncertainties of the form

where F(t) is the time-varying nonlinear function satisfying

$$ F^T(t)F(t)\leq I,\quad\forall t\geq0. $$
(2)

The delays h(t) and τ(t) are time-varying delays satisfying

$$ \begin{array}{l} 0\leq h(t)\leq h_U, \qquad -\infty< \dot{h}(t)\leq h_D, \\\noalign{\vspace{4pt}} 0 \leq\tau(t)\leq\tau_U, \end{array} $$

where h U , h D , and τ U are known positive scalars.

It is assumed that the neuron activation functions satisfy the following condition.

Assumption 1

[42] The neuron activation functions f i (⋅), i=1,…,n are continuous, bounded, and satisfy

(3)

where \(k^{+}_{i}\) and \(k^{-}_{i}\) are constants.

Remark 1

In Assumption 1, \(k^{+}_{i}\) and \(k^{-}_{i}\) can be allowed to be positive, negative, or zero. As mentioned in [14], Assumption 1 describes the class of globally Lipschitz continuous and monotone nondecreasing activation when \(k^{-}_{i}=0\) and \(k^{+}_{i}>0\). Also, the class of globally Lipschitz continuous and monotone increasing activation functions can be described when \(k^{+}_{i}>k^{-}_{i}>0\).

For passivity analysis, the systems (1) can be rewritten as

(4)

The objective of this paper is to investigate delay-dependent passivity conditions for system (4). Before deriving our main results, we state the following definition and lemmas.

Definition 1

The system (1) is called passive if there exists a scalar γ≥0 such that

(5)

for all t p ≥0 and for all solution of (1) with x(0)=0.

Lemma 1

[43] Let \(\zeta\in\mathbb{R}^{n}\), \(\varPhi=\varPhi^{T}\in\mathbb{R}^{n \times n}\), and \(\varUpsilon\in\mathbb{R}^{m \times n}\) such that \(\operatorname{rank}(\varUpsilon) < n\). The following statements are equivalent:

  1. (i)

    ζ T Φζ<0, ∀ϒζ=0, ζ≠0,

  2. (ii)

    ϒ T Φϒ <0.

3 Main results

In this section, new passivity criteria for network (4) will be proposed. For the sake of simplicity on matrix representation, \(e_{i}\in\mathbb{R}^{18n \times n}\ (i=1,2,\ldots,18)\) are defined as block entry matrices (for example, e 2=[0 n ,I n ,016n×n ]T). The notations of several matrices are defined as

(6)

Then the main result is given by the following theorem.

Theorem 1

For given positive scalars h U , h D , τ U , and a positive integer m, diagonal matrices \(K^{-}=\operatorname{diag}\{k_{1}^{-},\ldots,k_{n}^{-}\}\) and \(K^{+}=\operatorname{diag}\{k_{1}^{+},\ldots,k_{n}^{+}\}\), the system (4) is passive for 0≤h(t)≤h U , \(\dot{h}(t)\leq h_{D}\) and 0≤τ(t)≤τ U , if there exist positive scalars ε and γ, positive diagonal matrices \(\varLambda_{a}=\operatorname{diag}\{\lambda _{a1},\ldots,\lambda_{an}\}\) (a=1,2), \(\Delta_{a}=\operatorname{diag}\{ \delta_{a1},\ldots,\delta_{an}\}\ (a=1,2)\), \(H_{b}=\operatorname{diag}\{h_{1}^{b},\ldots,h_{n}^{b}\}\ (b=1,2,\ldots,3m)\), \(\tilde{H}_{b}= \operatorname{diag}\{\tilde{h}_{1}^{b},\ldots,\tilde {h}_{n}^{b}\}\ (b=1,2,\ldots,m)\), positive definite matrices \(R\in\mathbb{R}^{5n\times5n}\), \(N\in\mathbb {R}^{3n\times3n}\), \(M\in\mathbb{R}^{2n\times2n}\), \(G\in\mathbb {R}^{2n\times2n}\), \(Q_{1}\in\mathbb{R}^{3n\times3n}\), \(Q_{2}\in\mathbb {R}^{n\times n}\), \(Q_{3}\in\mathbb{R}^{n\times n}\), any symmetric matrices \(P_{1}\in\mathbb{R}^{n\times n}\), \(P_{2}\in\mathbb {R}^{n\times n}\), and any matrices \(S_{1}\in\mathbb{R}^{3n\times3n}\), \(S_{2}\in\mathbb{R}^{n\times n}\), satisfying the following LMIs:

(7)
(8)
(9)

where ϒ, \(\bar{P}_{1}\), \(\bar{P}_{2}\), Ψ, and Θ j are defined in (6).

Proof

Consider the following Lyapunov–Krasovskii functional candidate as

(10)

where

Time-derivative of V 1, V 2, and V 3 are calculated as

(11)
(12)
(13)

Inspired by the work of [40], by adding the following two zero equalities with any symmetric matrices P 1 and P 2:

(14)

into the time-derivative of V 4 and using Jensen’s inequality [44], we get

(15)

If the inequality (8) hold, then the two inequalities, \(Q_{1} +\bar{P}_{1} \geq 0\) and \(Q_{1} +\bar{P}_{2} \geq0\), are satisfied. Thus, \(\dot{V}_{4}\) can be estimated as

(16)

where \(\alpha(t)=1-h(t)h_{U}^{-1}\) and \(\beta(t)=1-\tau(t)\tau _{U}^{-1}\), which satisfy 0<α(t)<1 and 0<β(t)<1 when 0<h(t)<h U and 0<τ(t)<τ U , respectively. Then, by reciprocally convex approach [41], if the LMIs (8) and (9) satisfy, then the following inequalities hold for any matrices S 1 and S 2

(17)

Also, when h(t)=0, h(t)=h M and τ(t)=0, τ(t)=τ U , respectively, we get

and

(18)

respectively.

Thus, from (17) and (18), the following inequality still holds:

(19)

Here, if the inequality (8) holds, then an upper bound of the \(\dot{V}_{4}\) can be rebounded as

(20)

Lastly, an upper bound of \(\dot{V}_{5}\) can be obtained as

(21)

where Lemma 2 in [36] was utilized in above inequality.

Since the inequality p T(t)p(t)≤q T(t)q(t) holds from (2) and (4), there exists a positive scalar ε satisfying the following inequality:

(22)

Let us choose v=0 from (3) and divide its range of (3) into m interval. It should be noted that subinterval of the range of (3) can be described as

where m is positive integer, and each condition is equivalent to

From the conditions just above, the following inequalities hold for any positive diagonal matrices: \(H_{3(j-1)+l}=\operatorname{diag} \{h_{1}^{3(j-1)+l},\ldots ,h_{n}^{3(j-1)+l} \}\) and \(\tilde{H}_{j}=\operatorname{diag} \{\tilde{h}_{1}^{j},\ldots ,\tilde{h}_{n}^{j} \}\), where j=1,2,…,m and l=1,2,3.

Case (j) for Range (j) with j=1,2,…,m:

(23)

From (11)–(23) and by applying S-procedure [45], an upper bound of \(\dot {V}-2y^{T}(t)u(t)-\gamma u^{T}(t)u(t)\) can be

(24)

By Lemma 1, ζ T(t)(Ψ+Θ j )ζ(t) with ϒζ(t)=0 n×1 is equivalent to (ϒ )T(Ψ+Θ j )ϒ <0. Therefore, if LMIs (7), (8), and (9) hold, then (ϒ )T(Ψ+Θ j )ϒ <0 holds, which means

(25)

By integrating (25) with respect to t over the time period from 0 to t p , we have

(26)

for x(0)=0. Since V(x(0))=0, the inequality (5) in Definition 1 holds. This implies that the neural networks (1) is passive in the sense of Definition 1. This completes our proof. □

Remark 2

Unlike the method of [38], the utilized augmented vector ζ(t) includes the state vector such as f(x(tτ U )) and x(tτ U ). These state vectors have not been utilized as an element of augmented vector ζ(t) in any other literature, which is the main difference between Theorem 1 and the methods in other literature. Correspondingly, in (23), the terms such as \(-2\sum^{n}_{i=1}\tilde{h}_{i}^{j} [f_{i} (x_{i}(t-\tau_{U})) -(k_{i}^{-}+\frac {j-1}{m}(k_{i}^{+}-k_{i}^{-}))x_{i} (t-\tau_{U})][f_{i} (x_{i}(t-\tau _{U}))-(k_{i}^{-}+\frac{j}{m}(k_{i}^{+}-k_{i}^{-})) x_{i} (t-\tau_{U})]\) are utilized for the first time.

Remark 3

Recently, the reciprocally convex optimization technique to reduce the conservatism of stability for systems with time-varying delays was proposed in [41]. Motivated by this work, in (15)–(19), the proposed method of [41] was utilized in obtaining upper bounds of the terms such as

(27)

and

(28)

Remark 4

In (14), two zero equalities are proposed inspired by the work of [40] and utilized in Theorem 1 to reduce the conservatism of the stability condition. As presented in (14), the terms x T(t)(h U P 1)x(t)−x T(th(t))(h U P 1)x(th(t)) and \(x^{T}(t-h(t))\* (h_{U} P_{2})x(t-h(t))-x^{T}(t-h_{U})(h_{U} P_{2})x(t-h_{U})\) provide the enhanced feasible region of the passivity criterion. Furthermore, as shown in (15), the two integral terms such as \(- 2h_{U} \int^{t}_{t-h(t)}\dot{x}^{T}(s) P_{1} x(s)\,ds\) and \(- 2h_{U} \int^{t-h(t)}_{t-h_{U}}\dot{x}^{T}(s) P_{2} x(s)\,ds\) presented in (14) are merged into the integral terms \(- h_{U} \int^{t}_{t-h(t)}\mu^{T}(s)\* Q_{1} \mu(s)\,ds\) and \(- h_{U} \int^{t-h(t)}_{t-h_{U}}\mu^{T}(s) Q_{1} \mu(s)\,ds\), which cause the conservatism of the passivity criterion.

Remark 5

Inspired by the fact that the stability and performance of neural networks are related to the choice of activation functions [46], the range of the term, \(k^{-}_{i} \leq\frac{f_{i} (u)}{u} \leq k^{+}_{i}\), is divided into m subintervals such as \(k^{-}_{i} \leq\frac{f_{i} (u)}{u} \leq k^{-}_{i} + \frac{1}{m}(k^{+}_{i}-k^{-}_{i})\), \(k^{-}_{i} + \frac{1}{m}(k^{+}_{i}-k^{-}_{i}) \leq\frac{f_{i} (u)}{u} \leq k^{-}_{i} + \frac{2}{m}(k^{+}_{i}-k^{-}_{i}),\ldots, k^{-}_{i} + \frac{m-2}{m}(k^{+}_{i}-k^{-}_{i}) \leq\frac{f_{i} (u)}{u} \leq k^{-}_{i} + \frac{m-1}{m}(k^{+}_{i}-k^{-}_{i})\), and \(k^{-}_{i} + \frac{m-1}{m}(k^{+}_{i}-k^{-}_{i}) \leq\frac{f_{i} (u)}{u} \leq k^{+}_{i} \). By choosing u as x(t), x(th(t)), x(th U ), and x(tτ U ), the inequalities (23) are utilized in Theorem 1. This idea has not been proposed in passivity analysis for uncertain neural networks with mixed time-varying delays. The advantage of this approach is that the feasible region of passivity criterion can be enhanced as the partitioning number m increases. It should be pointed unlike the delay-partitioning approach, the augmented vector are not changed. However, as m increases, the number of decision variables becomes larger. Through three numerical examples, it will be shown the feasible region of passivity criterion introduced in Theorem 1 can be significantly enhanced as the partitioning number m increases.

As a special case, the networks (4) without distributed delays can be rewritten as

(29)

We can obtain a passivity criterion of the network (29) by the similar method of the proof of Theorem 1. This result will be introduced in Theorem 2. For the sake of simplicity on matrix representation, \(e_{i}\in\mathbb{R}^{14n \times n}\) (i=1,…,14) are defined as block entry matrices (for example, e 2=[0 n ,I n ,012n×n ]T). Before introducing this, the notations of several matrices are defined as:

(30)

where \(\bar{P}_{1}\), \(\bar{P}_{2}\), Π a (a=3,…,6) and Φ 2 are defined in (6).

Now, we have the following theorem.

Theorem 2

For given positive scalars h U , h D and a positive integer m, diagonal matrices \(K^{-}= \operatorname{diag}\{k_{1}^{-},\ldots,k_{n}^{-}\}\) and \(K^{+}=\operatorname{diag}\{k_{1}^{+},\ldots,k_{n}^{+}\}\), the system (4) is passive for 0≤h(t)≤h U and \(\dot{h}(t)\leq h_{D}\), if there exist positive scalars ε and γ, positive diagonal matrices \(\varLambda_{a}=\operatorname{diag}\{\lambda _{a1},\ldots,\lambda_{an}\}\) (a=1,2), \(\Delta_{a}=\operatorname{diag}\{ \delta_{a1},\ldots,\delta_{an}\}\) (a=1,2), \(H_{b}=\operatorname{diag}\{h_{1}^{b},\ldots,h_{n}^{b}\}\) (b=1,2,…,3m), positive definite matrices \(R\in\mathbb{R}^{4n\times4n}\), \(N\in\mathbb {R}^{3n\times3n}\), \(M\in\mathbb{R}^{2n\times2n}\), \(G\in\mathbb {R}^{2n\times2n}\), \(Q_{1}\in\mathbb{R}^{3n\times3n}\), \(Q_{3}\in\mathbb {R}^{n\times n}\) and any symmetric matrix \(P_{a}\in\mathbb{R}^{n\times n}\) (a=1,2), and any matrix \(S_{1}\in\mathbb{R}^{3n\times3n}\) satisfying the following LMIs with (8):

(31)

where \(\hat{\varUpsilon}\) and \(\hat{\varPsi}\) are defined in (30), and \(\bar{P}_{a}\) (a=1,2) and Θ 1j are in (6).

Proof

Consider the following Lyapunov–Krasovskii functional candidate as

(32)

where

The other procedure of proof is straightforward from the proof of Theorem 1, so it is omitted. □

4 Numerical examples

In this section, three numerical examples will be shown to illustrate the effectiveness of the proposed criteria. In examples, MATLAB, YALMIP 3.0, and SeDuMi 1.3 are used to solve LMI problems.

Example 1

Consider the neural networks (1) with

Tables 1 and 2 show the results of the upper bound of time-delay for the above system. It can be seen that Theorem 1 in this paper provides larger delay bound than the previous results. Furthermore, when the partitioning number m increases, the maximum delay bounds get larger. This indicates that the presented passivity conditions relieve the constraint of the passivity caused by time-delay. To confirm one of the obtained results in Table 2 (h D =0.1, h U =0.7174, τ U =0.5070), a simulation result when x(0)=[−1,−0.5]T, h(t)=0.7174sin2(0.13t), τ(t)=0.5070sin2(t), u(t)=0.1sin(2πt), \(\Delta A(t)=\Delta W_{0} (t)=\Delta W_{1} (t) = \Delta W_{2} (t)=0.2 \operatorname{diag} \{\sin(t), \sin(t)\}\) are given in Fig. 1. Figure 1 shows that the system (1) with above parameters is passive in the sense of Definition 1.

Fig. 1
figure 1

State trajectories with h D =0.1, h U =0.7174 and τ U =0.5070 (Example 1)

Table 1 Upper bounds of time-delay h U and τ U with different h D (Example 1)
Table 2 Upper bounds of time-delay h U with fixed τ U and different h D (Example 1)

Example 2

Consider the neural networks (29) with

In Table 3, the results of the upper bound of time-delay for guaranteeing passivity are compared with the previous results. From the results of Table 3, it can be seen that the maximum delay bounds for guaranteeing the passivity of the above neural networks are larger than those of other literature listed in Table 3. To confirm one of the obtained result in Table 3 (h D =0.5, h U =1.1590), a simulation result when x(0)=[1,−1]T, h(t)=1.1590sin2(0.43t), u(t)=1, \(\Delta A(t)=0.01 \operatorname{diag} \{\sin(t), \sin(t)\}\), \(\Delta W_{0} (t)= 0.02 \operatorname{diag} \{\sin(t), \sin (t)\}\), \(\Delta W_{1} (t) =0.03 \operatorname{diag} \{\sin(t), \sin (t)\}\) are shown in Fig. 2. From Fig. 2, it can be confirmed that the neural networks (29) with above parameters when 0≤h(t)≤1.1590 and \(\dot{h}(t) \leq0.5\) is passive in the sense of Definition 1.

Fig. 2
figure 2

State trajectories with h D =0.5 and h U =1.1590 (Example 2)

Table 3 Upper bounds of time-delay h U with different h D (Example 2)

Example 3

Consider the neural networks (29) with

When h D =0.5, the obtained upper bounds of time-delay for guaranteeing the passivity of the above neural networks in [33] and [35] were 0.7230 and 1.3752, respectively. By applying Theorem 2 with m=2, it can be obtained that the upper bound of time-delay is 35.3121, which is much larger delay bound than one in [33] and [35]. When h D is unknown, the upper bound of time-delay obtained 0.6791 and 1.3027 in [33] and [35], respectively. However, by using Theorem 2 with m=2, one can obtain the upper bound of time-delay is 3.9715. Moreover, by utilizing Theorem 2 with m=1 and m=3, the upper bounds of time-delay with different h D listed in Table 4. From the results of Table 4, it can be confirmed that Theorem 2 gives larger delay bounds than those obtained by the method of [33] and [35]. To confirm one of the obtained result in Table 4 (h D  is unknown, h U =4.7368), a simulation result when x(0)=[0.5,1]T, h(t)=4.7368|sin(t)|, u(t)=0.1sin(2πt) are given in Fig. 3. From Fig. 3, it can be also verified that the neural networks (29) with above parameters when h U =4.7368 and h D is unknown is passive in the sense of Definition 1.

Fig. 3
figure 3

State trajectories with unknown h D and h U =4.7368 (Example 3)

Table 4 Upper bounds of time-delay h U with different h D (Example 3)

5 Conclusion

In this paper, the improved passivity criteria for uncertain neural networks with both discrete and distributed time-varying delays have been proposed. In order to drive less conservative results, the suitable Lyapunov–Krasovskii functional and decomposed conditions of activation function divided by states are utilized to enhance the feasible region of passivity criteria. Three numerical examples have been illustrated to show the effectiveness of the proposed methods. Future works will focus on passivity analysis and passification of various neural networks such as fuzzy neural networks, static neural networks, and so on. Furthermore, some new passivity analysis for discrete-time neural network with time-varying delays will be investigated in the near future.