Keywords

1 Introduction

Due to the simple topological structure and rapid training speed, the radial basis function neural networks (RBFNNs) have been widely used in many applications, such as the nonlinear system estimation [1], information processing [2], pattern recognition [3], and so on. For a RBF neural network, its performance is closely related with network architecture, thus how to determine the optimal network parameters and network size is very important.

If the network parameters are selected in a random manner, the network satisfactory performance cannot be achieved. To solve this problem, many different learning algorithms are designed to adjust network parameters. As the most commonly used method, the back propagation (BP) has been used to adjust RBFNN parameters [3]. However, the BP method always results in degraded global search experience and is easily trapped into local optimal [3]. Compared with BP method, the recursive least squares (RLS) methods can result in a more optimum convergence rate and improved network performance [4]. However, the application of RLS always needs the complicated mathematical expression and high computing resources.

The network size is also closely related with network performance of RBFNN. Hence, it is important to optimize the size of RBFNN to improve network performance. For example, the adaptive growing and pruning algorithm [5] is applied to construct the recurrent neural network, and the neurons are added or removed to improve network performance. Then, the forward and backward selection methods are used for the self-organizing RBFNN [6], in which the network size and parameters are optimized simultaneously.

Recently, the evolutionary algorithms (EA) are widely used to improve network performance. There have existed many EA algorithms, such as the differential evolution (DE), simulated annealing (SA), artificial immune algorithm (IA), particle swarm optimization (PSO), ant colony algorithm (AC), and genetic algorithm (GA). Among these algorithms, the PSO offers quick convergence rate and good global searching ability. However, the traditional PSO always suffer from weak solution diversity and premature convergence for highly nonlinear optimization problem. To solve this problem, the accelerated PSO (APSO) [7] is proposed which has shown superiority in terms of algorithm convergence speed and solution accuracy.

Based on above discussion, in this paper, the APSO is proposed to optimize network parameters and network size of RBFNN. Then, the developed network is compared with some existing methods. The simulation results show that the proposed method is efficient in terms of estimation accuracy and network compactness.

The rest of paper has been presented as follows: In Sect. 2, the RBFNN is discussed. Section 3 overviews the PSO and APSO. In Sect. 4, the simulation results have been depicted. Finally, the conclusion is presented in Sect. 5.

2 RBF Neural Network Model

The RBFNN contains three layers, including the input layer, hidden layer and output layer. The basic illustration of a RBFNN is presented in Fig. 1.

Fig. 1.
figure 1

The RBFNN network illustration

The output of RBFNN, \( \varvec{y}(t) \), is specified by the given equation:

$$ \varvec{y}(t) = \sum\nolimits_{l = 1}^{K} {w_{l} (t)\phi_{l} (t)} ;l = 1,2,3, \ldots ,K $$
(1)

Where \( w_{l} (t) \) is the weight of output at the time step t, l is the hidden neuron with \( l = 1,2,3, \ldots ,K \), \( \phi_{l} (t) \) is the lth hidden neuron output as below:

$$ \phi_{l} (t){ = }e^{{\left\| {\varvec{x}(t) - \mu_{l} (t)} \right\|/\sigma_{l}^{2} (t)}} $$
(2)

where \( \varvec{x}(t) = [x_{1} (t),x_{2} (t), \ldots ,x_{n} (t)]^{\text{T}} \) are inputs of the RBFNN, n is the nodes number of input layer; \( \varvec{\mu}_{l} (t) \) indicates the center vector of the lth hidden neuron; \( \left\| {\varvec{x}(t) -\varvec{\mu}_{l} (t)} \right\| \) is the Euclidean distance among \( \varvec{x}(t) \) and \( \varvec{\mu}_{l} (t) \); \( \sigma_{l} (t) \) is the radius or width of the lth hidden neuron. The network parameters, including width, center of hidden neurons, output weights and network size are adjusted in the training phase by using the training samples.

3 The Proposed APSO-RBFNN

3.1 Overview of PSO

In the traditional PSO [7], the swam is defined as a set \( {\text{S}} = \left\{ {\varvec{s}_{ 1} ,\varvec{s}_{ 2} , \ldots ,\varvec{s}_{N} } \right\}^{\text{T}} \) of N particles. Each particle \( \varvec{s}_{i} (k) \) at the iteration k is assigned a position vector, which is represented as below,

$$ \varvec{s}_{i} (k) = [s_{i,1} (k),s_{i,2} (k), \ldots ,s_{i,D} (k)] $$
(3)

where \( i = 1,2, \ldots ,N \), and D stands for the search space dimension. Moreover, each particle has a velocity \( \varvec{v}_{i} (k) \) which is denoted by:

$$ \varvec{v}_{i} (k) = [v_{i,1} (k),v_{i,2} (k),v_{i,3} (k), \ldots ,v_{i,D} (k)] $$
(4)

In order to allow each particle visit all the searching space, the velocity of each particle is updated according to its gained memory. Now, let us define \( \varvec{p}_{i} (k) \) as the best previous position of the ith particle, which is denoted as \( \varvec{p}_{i} (k) = [p_{i,1} (k),p_{i,2} (k), \ldots ,p_{i,D} (k)] \). Then, the best position found by the whole swarm is recorded as \( \varvec{g}_{best} (t) = [g_{best,1} (k),g_{best,2} (k), \ldots ,g_{best,D} (k)] \).

At each iteration k, the velocity of each particle is updated as:

$$ \varvec{v}_{i} (k + 1) = \omega v_{i} (k) + c_{1} r_{1} (\varvec{p}_{best} (k) - \varvec{s}_{i} (k)) + c_{2} r_{2} (\varvec{g}_{best} (k) - \varvec{s}_{i} (k)) $$
(5)

where \( \omega \) is the weight of inertia; \( r_{1} \) and \( r_{2} \) are random values derived from the interval [0,1], which are used to maintain the algorithm diversity; \( c_{1} \) and \( c_{2} \) are acceleration constants, which are used to push the swarm towards the local and global best position. Finally, the position of each particle is updated as below:

$$ \varvec{s}_{i} (k + 1) = \varvec{v}_{i} (k + 1) + \varvec{s}_{i} (k) $$
(6)

3.2 Overview of APSO

Based on the traditional PSO, the accelerated particle swarm optimization (APSO) is proposed [7]. In traditional PSO, the particle best position \( \varvec{p}_{i} \) and the global best position \( \varvec{g}_{best} \) are used to update the velocity as shown in Eq. (6). While in the APSO, only the global best position \( \varvec{g}_{best} \) are used. Consequently, the particle’s velocity in APSO is updated as below:

$$ \varvec{v}_{i} (k + 1) = \varvec{v}_{i} (k) + \alpha \varvec{q} + \beta (\varvec{g}_{best} - \varvec{s}_{i} (k)) $$
(7)

where q is a random number chosen from the range [0,1]. Generally, the parameter \( \alpha \) is chosen from the range [0.1D, 0.5D], and \( \beta \) can be chosen from the range [0.1, 0.7]. Moreover, the position of each particle is simply calculated as below:

$$ \varvec{s}_{i} (k + 1) = (1 - \beta )\varvec{s}_{i} (k) + \beta \varvec{g}_{best} + \alpha \varvec{q} $$
(8)

As illustrated in [7], compared with traditional PSO, the APSO has shown good performance in terms of algorithm convergence and diversity.

3.3 Design Process of the Proposed APSO-RBFNN

In this paper, the APSO is used to optimize network parameters of RBFNN, the proposed method is denoted as APSO-RBFNN. In APSO-RBFNN, the network parameters, including width \( \varvec{\mu}_{l} \) and center \( \varvec{\sigma}_{l} \) of hidden neurons, output weights \( w_{l} \) and network size are optimized by the APSO. The operation procedures of APSO-RBFNN are given as follows:

  • Step 1: Set the original iteration as \( k = 0 \). Randomly initialize the position \( \varvec{s}_{i} \) and velocity \( v_{i} \) of each particle \( i \).

  • Step 2: Evaluate the fitness function for each particle \( i \). Among all the particles, find \( \varvec{g}_{best} \) with the best fitness value.

  • Step 3: Calculate the positions and velocities of all particles according to Eqs. (7) and (8).

  • Step 4: Start the next generation \( k = k + 1 \) and go to Step 2. This process is repeated until the maximum iteration is reached.

4 Simulation Results

In the simulation part, the root-mean square error (RMSE) is used to test the network performance, it is also treated as the objective function of APSO. The mathematical function of RMSE is given as follows:

$$ E_{i} (k) = \sqrt {\frac{1}{L}\sum\nolimits_{t = 1}^{L} {(y_{d} (t) - \hat{y}(t)^{2} )} } $$
(9)

where, \( y_{d} \) is the desired output, L represents data samples length and \( \hat{y} \) is the network’s estimated output. The smaller the RMSE value, the better the neural network estimation performance. To demonstrate efficiency of the designed APSO- RBFNN, two non-linear system modeling experiments are illustrated.

4.1 The Mackey-Glass Time-Series Estimation

As a classical benchmark task for non-linear system modeling, the Mackey-Glass time-series are generated as below [5] (Fig. 2):

Fig. 2.
figure 2

The testing results of APSO-RBFNN for Mackey-Glass time series

$$ x(t + 1) = (1 - a)x(t) + \frac{bx(t - \tau )}{{1 + x^{10} (t - \tau )}} $$
(10)

where the parameters are set as a = 0.1, b = 0.2, τ = 17, and the initial condition is x(0) = 0. In this experiment, 1000 samples are set as the training set, the first 500 data are discarded to reduce the influence of initial condition. The samples with t = 1001 to 1500 are set as the testing set. The simulation results of the proposed APSO-RBFNN are shown in Figs. 3, 4 and 5. In Fig. 3, the comparisons between the estimated outputs and target are shown. In Fig. 4, the evolution process of the hidden neurons is illustrated. While in Fig. 5, the testing RMSE curve is displayed. From Figs. 3, 4 and 5, it can be concluded that the proposed APSO-RBFNN has obtained good estimation result.

Fig. 3.
figure 3

The evolution of neurons versus iterations for APSO-RBFNN

Fig. 4.
figure 4

The RMSE value evolving process versus iterations for APSO-RBFNN

Fig. 5.
figure 5

The testing results of APSO-RBFNN for Lorenz time series prediction

In order to test the effectiveness of the proposed APSO-RBFNN, its performance is compared with other two methods, i.e., the PSO-RBFNN in which the network parameters are optimized by PSO, the DE-RBFNN whose network parameters are optimized by differential algorithm (DE). As shown in Table 1, the APSO-RBFNN owns the optimal compact network (10 hidden neurons), the lowest training RMSE of 0.0107 and the lowest testing RMSE of 0.0127.

Table 1. Performance of various methods for Mackey-Glass time-series estimation

4.2 Lorenz Time-Series Prediction

The Lorenz time-series prediction is another benchmark to evaluate the nonlinear system modeling performance for neural networks. Its differential equations are presented as below:

$$ \begin{array}{*{20}l} {\frac{dx}{dt} = \sigma (y - x),} \hfill \\ {\frac{dy}{dt} = rx - y - xz,} \hfill \\ {\frac{dz}{dt} = xy - bz} \hfill \\ \end{array} $$
(11)

where \( \sigma = 10,b = 8/3,r = 28 \), the initial conditions for each variable are set as \( x(0) = y(0) = z(0) = 0.1 \). A total of 5000 data samples are generated, the data samples with t = 1 to 3000 are set as the training set and the remaining part is set as the testing set. In the training set, the first 1000 samples are discarded.

The estimation results on the testing set of APSO-RBFNN are shown in Fig. 5, the evolving process of the network size is illustrated in Fig. 6, and the RMSE values versus algorithm iterations are given in Fig. 7. Furthermore, the performance of the proposed APSO-RBFNN is compared with PSO-RBFNN and DE-RBFNN. The comparison details are given in Table 2, it is easily found that the designed APSO-RBFNN has better training and testing RMSE as well as network size than the other two methods.

Fig. 6.
figure 6

The network size evolution process versus iterations for APSO-RBFNN

Fig. 7.
figure 7

The RMSE values versus algorithm iterations for APSO-RBFNN

Table 2. Performance of various methods for Lorenz time series estimation

5 Conclusion

To solve the nonlinear system modeling problem, an accelerated particle swarm optimization (APSO) based radial basis function neural network (RBFNN) is proposed, which is denoted as APSO-RBFNN. In APSO-RBFNN, the center, width of hidden neurons, output weights and network size are optimized by using the APSO method. As shown in the simulation results, the proposed APSO-RBFNN outperforms other existing method in terms of network compactness and estimation accuracy.