1 Introduction

With the trend toward taller building structures, undesirable vibration levels can arise from large environmental loads such as strong wind and earthquakes. State-of-the-art active control techniques for building structures were reported by Soong (1988) and Yoshida (1994), in which some active control devices have been implemented. The tuned mass damper devices were seen in stacks, observatories, bridges, communication towers, and machinery foundations (Catalogue of Mitsubishi Heavy Industries, LTD. 2019, Tokyo, Japan). Extraordinary loading episodes can occur and cause casualties and structural damage. Development of mass-damper systems, such as tuned mass-dampers and active mass-dampers (AMD), has been introduced into high-rise buildings to reduce vibration. Vibration suppression of structure systems subject to external excitation and disturbances must be intensively studied (Li et al. 2010; Liu et al. 2011; Lazar et al. 2013; Wang et al. 2018). Recent research has been focusing on the applications of active mass damper (AMD) -an auxiliary mass with sensor, actuator and controller- to vibration suppression of a building structure (Chen 2012; Song et al. 2011). The progress in peripheral technologies, such as computation speed, low friction of linear bearing, and high horsepower servo motor have made active control feasible in structure systems (Korkmaz 2011).

Conventionally, the proportional–integral–derivative (PID) controller dominates servo system applications; however, it is not sufficiently robust to accommodate external disturbances, parameter variations, and structural perturbations. Many alternate controllers have been developed to replace the PID controller, including model reference adaptive control (Ohnishi et al. 1986), the field acceleration method (Zhong et al. 1991), fuzzy logic (Tseng and Hwang 1993), sliding mode control (Utkin 1993), neural/fuzzy control (Goode and Chow 1995a, b), input–output linearization (Marino et al. 1993; Grcar et al. 1996), and internal model control (Tzou 1996). Nevertheless, the PID controllers remain the mainstream controller for AC motors because of their relatively simple implementation. Within AC motors, the self-turning controller, in particular, may be redesigned to track differing reference inputs and structural vibration suppression.

Unmodeled nonlinear friction has been known to have an adverse influence on controller performance; therefore, a model-based approach is often employed to reduce friction effects. The critical issue in use of the model-based approach is the development of a precise friction model. Johnson and Lorenz (1992) experimentally estimated the parameters within their friction model, while Yang (1993) suggested proportional–integral–derivative (PID) control to eliminate friction effects through online model parameter estimation. Haessing and Friendland (1991) and Wit et al. (1995) have also proposed alternative friction modeling approaches. The effectiveness of these model-based studies has been found to heavily rely upon the model accuracy. Because of the inherent nonlinearity of friction, an accurate model is difficult, if not impossible, to construct.

The artificial neural network technique, with its high learning and nonlinear mapping abilities, has been successfully applied to many system identification and control problems (Sastry et al. 1994; Levin and Narendra 1996; Omatu et al. 1997). Amini et al. (1997) indicated that the structural dynamics of an object subjected to earthquake loading can be modeled by neural network. Han et al. (2001) introduced a neural network model, with recurrent architecture, for earthquake response. Yu and Li (2001) applied the neural networks to identify a nonlinear system, while Yang et al. (2006) and Chen and Yang (2014) applied the neural network models for the experimental identification of system dynamics. Yang et al. (2007) have also developed a self-organized neuro-fuzzy design methodology for modeling. Chen et al. (2009a, b, 2013, 2014) proposed a self-organized neuro-fuzzy model for system identification, in which their five-layer network adaptively adjusts the membership functions and dynamically optimizes the fuzzy rules. Learning-based control methodology via the neural networks has recently emerged as an alternative to adaptive control (Jagannathan 2001). However, there are few researches on the control of nonlinear structure using neural network for tracking control and vibration suppression under shaking table conditions, especially for high-rise building. In this work, a model combining the advantages of adaptive genetic algorithm and modified Newton method, similar to BFGS algorithm, is developed for effective training in feedforward neural networks. The benchmark tests show that the model is superior to many conventional ones: steepest descent, steepest descent with adaptive learning rate, conjugate gradient, and Newton-based methods and is suitable to small network in engineering applications. In the experimental verification, a modified Newton method with genetic algorithm for neural network training is applied to system identification, tracking control, and vibration suppression of a shaking table system.

2 Neural network and genetic algorithm

The artificial neural network is a data processing system composed of numerous inter-connected artificial neurons to mimic the biological neural network. An artificial neuron (processing element) is shown in Fig. 1a. The relationship between the input and output can be represented by \( n_{j} \equiv \sum\nolimits_{i} {w_{ij} x_{i} - \theta_{j} } \) and \( y_{j} = f (n_{j} ) \) where \( n_{j} \) is the linear combination output; \( w_{ij} \) is the synaptic weight (or simply weight) imitating the biological synapse strength between the jth neuron and the ith input; \( x_{i} \) is the ith input of the processing element (neuron); \( \theta_{j} \) is the bias representing the threshold of the transfer function; \( y_{j} \) is the output of the jth neuron and \( f (\circ ) \) is the transfer function (also called the activation function) for converting the weighted summation input \( n_{j} \) to the output \( y_{j} \). The connective weight of each connection represents the relative strength between two artificial neurons. A feedforward network may have several hidden layers as shown in Fig. 1b. The hidden layers provide the interaction of the neurons that represent the system’s internal structure. The goal of learning (training) processes is to adjust the connection weights in the layers for minimizing the error function by repeatedly training. The error function \( E_{p} \) is defined as

$$ E_{p} = \frac{1}{2}\sum\limits_{i = 1}^{{n_{o} }} { (t_{i}^{p} - y_{i}^{p} )^{2} ,\;p = 1,\;2,\; \ldots ,\;n_{p} } $$
(1)

where \( n_{o} \) is the number of neurons in the output layer; \( t_{i}^{p} \) and \( y_{i}^{p} \) are the desired (target) and actual output of the ith output neuron for the pth output sample, respectively; and \( n_{p} \) is the number of training samples. A neural network contains many processing neurons such that a higher degree of robustness and ault tolerance is possible.

Fig. 1
figure 1

a The structure of an artificial neuron. b The schematic diagram of a feedforward neural network

Genetic algorithm is an optimization algorithm based on the mechanism of natural selection and genetics. It has been demonstrated to be an effective global optimization tool. Genetic algorithm evolves a population of individuals, also called strings or chromosomes. Each individual \( v_{i} \)(i = 1, …, s) of the population V represents a trial solution of the problem. Chromosomes are usually represented by string of variables, and each element of which is called a gene. Every gene controls the inheritance of one or several characters. The value of a gene is called allelic value in [0, 1]. The algorithm can be described by reproduction, crossover, and mutation.

Reproduction is a process by which the most highly rated chromosomes in the current generation are reproduced in the new generation. Selection is often in the roulette wheel process to calculate the fitness value for each chromosome \( q_{i} = \sum\nolimits_{j = 1}^{i} {p_{j} } \), i = 1, …, s, where s is population size, and to find the total fitness of the population

$$ f = \sum\limits_{i = 1}^{s} {e(v_{i} )} , $$
(2)

where \( e(v_{i} ) \) is fitness function. The probability of a selection for each chromosome is

$$ p_{i} = \frac{1}{f}e(v_{i} ). $$
(3)

and the cumulative probability is

$$ q_{i} = \sum\limits_{j = 1}^{i} {p_{j} } . $$
(4)

For a random (floating) number a within (0, 1), if a < q1, then select the first chromosome \( v_{1} \), otherwise select the ith chromosome \( v_{i} \) (2 ≦ i ≦ s) such that qi−1 < a ≦ qi.

Crossover operator provides a mechanism for chromosomes to mix and match. The operators are selected the arithmetical crossover method

$$ \left\{ \begin{aligned} \varvec{x}_{1}^{{\prime }} = a\varvec{x}_{1} + (1 - a)\varvec{x}_{2} \hfill \\ \varvec{x}_{2}^{{\prime }} = a\varvec{x}_{2} + (1 - a)\varvec{x}_{1} \, \hfill \\ \end{aligned} \right. $$
(5)

where x1, x2 are parents, \( \varvec{x}_{1}^{{\prime }} \), \( \varvec{x}_{2}^{{\prime }} \) are offsprings, and a is a random value in \( { [0,} \ldots , { 1]}. \)

Mutation operator is a random alteration of some gene value in a chromosome. The operators are selected the non-uniform mutation method. For a parent x, if the element \( x_{k} \) is selected for this mutation, the result is

$$ \varvec{x}^{{\prime }} = [x_{1} , { } \ldots , { }x_{k}^{{\prime }} , { } \ldots , { }x_{q} ] , $$

where

$$ x^{\prime}_{k} = \left\{ {\begin{array}{*{20}l} {x_{k} + \Delta (t , { }u_{k} - x_{k} ) ,} \hfill & {{\text{if}}\;{\text{a}}\;{\text{random}}\;{\text{digit}}\;{\text{is}}\; 0} \hfill \\ {x_{k} - \Delta (t , { }x_{k} - l_{k} ) ,} \hfill & {{\text{if}}\;{\text{a}}\;{\text{random}}\;{\text{digit}}\;{\text{is}}\; 1 ,\;x_{k} \in \left[ {l_{k} { ,}u_{k} } \right]} \hfill \\ \end{array} } \right. $$
(6)

The function \( \Delta (t ,y ) \) returns a value in the range [0, y] such that the probability of \( \Delta (t ,y ) \) is close to zero as the generation number t increases. The function \( \Delta (t ,y ) \) can be expressed as

$$ \Delta (t,y) = ya\left( {1 - \frac{t}{T}} \right)^{b} , $$
(7)

where a is a random number within [0, 1], T is the maximal generation number, and b is a system parameter of the degree of non-uniformity.

There remain few guidelines in providing their initial values, though they are known critical to the training effectiveness of a neural network. The initial weight and bias are usually generated randomly. Genetic algorithm performs a multi-directional search by maintaining a population of potential solutions and it can find a near optimal solution. Although the algorithm usually requires a long training time, it can be used in the early stages of training to search for good starting point and detect the region where the global minimum is likely to be found. A two-stage model for training a feedforward neural network is proposed. The first stage is to find better initial weight and bias by using floating genetic algorithm. The second stage is to train the feedforward neural network by using modified Newton method with the specific initial weight and bias. This algorithm includes the advantages of genetic algorithm in finding better initial weight and bias and modified Newton method in accelerating the convergence of the network training.

3 Accelerated convergence by modified Newton algorithm

BPN is to utilize the gradient steepest descent algorithm to minimize the error between the desired and actual outputs in an iterative manner. The back-propagation algorithm is a first-order approximation of the steepest descent technique in the sense that it depends on the gradient of the instantaneous error surface in the weight space. The update rule is thus plagued by the drawbacks of converging slowly and being stuck easily in local minimum of the error function. To accelerate the convergence of the back-propagation algorithm, many studies have presented many different training algorithms. The modified Newton method similar to BFGS algorithm includes the advantages of the steepest descent algorithm and Newton’s algorithm, and they require less memory storage and computational load than the Newton’s algorithm and the Levenberg–Marquardt algorithm in numerical analysis. Start with the initial weights, biases and an initial estimate of the inverse of the Hessian matrix, denoted as B, of the error function as in Eq. (1). The update rule can be written by

$$ \Delta w_{ij} (k + 1) = - \eta \varvec{B}(k)\nabla E(w_{ij} (k)) $$
(8)

where E is the error function, \( \upeta \) is network learning rate, wij is the connection weight, \( \nabla E \) is the gradient of E, and B is an approximate inverse of the Hessian matrix of E. The inverse of the Hessian matrix is update by

$$ \begin{aligned} B(k + 1) & = B(k) + \left( {1 + \frac{{q^{T} (k)B(k)q(k)}}{{p^{T} (k)q(k)}}} \right)\frac{{p(k)p^{T} (k)}}{{p^{T} (k)q(k)}} \\ & \quad - \frac{{p(k)q^{T} (k)B(k) + B(k)q(k)p^{T} (k)}}{{p^{T} (k)q(k)}} \\ \end{aligned} $$

where

$$ p (k )= w_{ij} (k + 1 )- w_{ij} (k ) , $$
(9)
$$ q{ (}k )= \nabla E{ (}w_{ij} { (}k + 1 ) )- \nabla E{ (}w_{ij} (k ) ). $$
(10)

or if \( \left\| {p^{T} (k)q(k)} \right\| \le \varepsilon , \) where \( \, \varepsilon \, \) is a positive constant, then \( \varvec{B}(k + 1) \) is the identity matrix. The modified Newton method is essentially by substituting the inverse Hessian matrix with an approximation from weight and gradient changes in iteration process. It combines the stability of gradient descent with the speed of Newton-based method.

4 System identification using modified Newton method with genetic algorithm

Two examples are provided to illustrate the performance of the modified Newton method with genetic algorithm. The first example is the application of the modified Newton method with genetic algorithm in the identification of a differential equation system. Consider the plant is governed by the difference equation (Narendra and Parthasarathy 1990)

$$ y (k + 1 )= F\left[ { \, y (k ) , { }y (k - 1 ) , { }y (k - 2 ) , { }u (k ) , { }u (k - 1 )} \right] $$
(11)

where the function F has the form

$$ F\left[ {x_{1} ,x_{2} ,x_{3} ,x_{4} ,x_{5} } \right] = \frac{{x_{1} x_{2} x_{3} x_{5} (x_{3} - 1) + x_{4} }}{{1 + x_{3}^{2} + x_{2}^{2} }}. $$
(12)

A [5 × 5 × 1] feedforward neural network model, representing 5 neurons in the input layer, 5 in the hidden layer, and 1 in the output layer, is constructed to simulate the plant represented by the differential equation. The input to the plant and network model is assumed to be an independent and identically distributed random signal that is uniformly ranged in the interval [− 1, 1]. The identification procedure consists of two phases: the network training phase and the validation phase. The first is to establish the appropriate connection weights of the network by a form of supervised learning from a set of network input and desired network output. At each training step, all input are passed forward through the network to yield the network output which are then compared with the desired output. If the error is small enough, then the connection weights are fixed. Conversely, the error is passed backwards through the network, and the back-propagation algorithm with the adaptive learning rate and momentum techniques by batch learning is applied to adjust the connection weights. Note that batch learning rather than recursive learning is adopted here due to the fact that the latter tends to terminate at local minima of the parameters, hence leading to a higher total error than the former. Although recursive learning is adaptive, however, batch learning method is relatively more efficient and stable. During the validation phase, the network is given many other sets of input in addition to those used for training. The network will produce the output \( \hat{y}\left( k \right) \) very close to the actual system output \( y\left( k \right) \) if the training is successful. From this point of view, the neural network can therefore be regarded as just another form of functional representation. For a single output system, define the sum-squared error function as the performance index

$$ E = \sum\limits_{r} {(\hat{y}_{r} (k) - y_{r} (k))^{2} } $$
(13)

where the index r ranges over all training samples.

Figure 2 shows the results for sum-squared error during training obtained by using the quasi-Newton algorithm, back-propagation algorithm with momentum term and adaptive learning rate (Yang and Lee 1998), conjugation gradient algorithm (Jeng et al. 1997), back-propagation algorithm (Yang and Lee 1997), and two-stage training algorithm. The two-stage training algorithm is applied as follows: First, the floating genetic algorithm is applied with the population size s = 500 and 100 training cycles to obtain the initial weight and bias. Then, the modified Newton method is applied to train the feedforward neural network. The training pattern consists of 500 input/output pairs, and the training continues until the sum-squared error is less than 0.0001. Figure 2 shows that the result of the two-stage training algorithm is better than that of other algorithms. Compared with other algorithms, the two-stage training algorithm provides better initial weights and bias. After the training process, the validation phase is carried out by using the following input,

$$ \begin{aligned} u (k )& = 0.6{\text{sin(}}\pi \, k/20 ) ,\quad \quad \quad \;\;\;{\text{k}} < 1 0 0\\ & = 0. 6 ,\quad \quad \quad \quad \quad \quad \;\;\; 1 0 0\le {\text{k}} < 1 5 0\\ & = - 0. 6 ,\quad \quad \quad \quad \quad \quad 1 5 0\le {\text{k}} < 2 0 0\\ & = 0. 6 {\text{sin(}}\pi \, k/15 ) ,\quad 2 0 0\le {\text{k}} < 3 0 0.\\ \end{aligned} $$
(14)

The outputs of the plant and the network model are shown in Fig. 3, which also shows that neural network identification is highly accurate.

Fig. 2
figure 2

The variations of sum-squared error of the differential Eq. (12) by using different algorithms

Fig. 3
figure 3

Plant output and the network model output of the differential equation

The other benchmark test involves a nonlinear, damped Duffing oscillator, and the equation of motion is

$$ \ddot{y} + 2\dot{y} + 400y + 200y^{3} = 500u (t ) { } $$
(15)

where y is the displacement of the hard spring and u(t) is the external applied force. By transforming the continuous system into discrete-time form with the sampling interval \( \Delta t \), the equation of motion can be written as

$$ y{ (}k + 1 )= F\left[ {y (k ) , { }\dot{y} (k ) , { }u (k ) , { }\Delta t} \right] $$
(16)

A [3 × 10 × 2] feedforward neural network representing 3 input nodes, 10 neurons in a hidden layer, and 2 output nodes are employed to identify the Duffing system. The input is a random signal in the range of [− 2, 2] with an effective sampling frequency of 100 Hz. The velocity is scaled by a factor of 30 to avoid neuron disparities. The training pattern consists of 700 input/output pairs, and the training proceeds until the sum-squared error is less than 0.01. Figure 4a shows the variation in the sum-squared error of the 8 algorithms during training under the same neural network parameters. The modified Newton method is more efficient than algorithms and requires 250 training iterations. By contrast, the quasi-Newton, Levenberg–Marquardt, and Gauss–Newton algorithms require 456, 478, and 319 training iterations, respectively. Moreover, the other training algorithms do not converge even after 500 iterations. The comparison of the time responses of the system output and neural network model during the validation phase is presented in Fig. 4b, which shows that neural network identification is highly accurate such that both lines coincide. The neural network model retains its excellent performance.

Fig. 4
figure 4

a The variations of sum-squared error and b the plant output and the modified Newton model with genetic algorithm output in Duffing oscillator

5 Experimental verification for tracking control and vibration suppression in a shaking table system

In the experimental verification, the neural network with modified Newton method and genetic algorithm controller is applied in the positional tracking and vibration suppression of a shaking table system. An one-story shear-building structure model of 300 × 210 × 205 mm, weights 5.4 kg combined with an active mass damper (AMD) mounted on the top floor is manufactured as shown in Fig. 5. A low-frequency accelerometer (KISTLER 8628B5) attached at the top floor acquires the acceleration. The measurement of accelerometer is then integrated to yield the velocity signal by the integrator (G-TECH 350133) which is then integrated and filtered to displacement signal. One linear wire potentiometer (HPS-M-10-5K-F) fixed at the bottom floor through a analog I/O board (National Instruments PCI-6052E) is used to measure the displacement of shaking table. The analog I/O board (National Instruments PCI-6052E) is employed to generate the control signal to one of the AC servomotor (MITSUBISHI HC-MF73) for positional trackiing of the shaking table. The specification of the AC servomotor: the rated output is 0.75 kw, the rated torque is 2.4 N m, the maximum torque is 7.2 N m, the rated speed is 3000 rpm, the maximum speed is 4500 rpm, the permissible instantaneous speed is 5175 rpm, the inertia moment J is 0.6 kg cm2 and the total weight is 3 kg. The data acquisition and control system consists of a multi-input/output board (National Instruments PCI-6052E) and the LAB-VIEW software. Table 1 lists the specification of the AC servo motors. For safety concerns, the following devices are used to ensure the reliability of the shaking table and avoid any dangerous malfunction on the building: a mechanical brake is combined with a limit switch to stop the auxiliary mass of the shaking table system when it overruns a certain position, and an automatic switching function which cuts off the motor current in case of abnormal situation or power failure. The data acquisition is based on LAB-VIEW6.1 (National Instruments) and a PCI-6052E multifunction I/O board.

Fig. 5
figure 5

Experimental model of the building structure combined with an active mass-damper system

Table 1 Specification of AC servo motor

In the experimental verification, the neural network with modified Newton method and genetic algorithm controller is applied in the positional tracking and vibration suppression of a shaking table system (Fig. 5). The system state of displacement and velocity together with the actuator input voltage at the sampling rate of 10 Hz are used as the input of the network model. Given that a continuous system has an infinite degree-of-freedom, the number of steps of the preceding information fed as the network input depends mainly on the accuracy of the identification model and system dynamics. In this study, the previous step of system state and force, \( [y(k - 1),\dot{y}(k - 1),u(k - 1)] \), are used as network input to predict the present system state \( [y(k),\dot{y}(k)] \). The schematic of the shaking table identification (Narendra and Parthasarathy 1990) test is shown in Fig. 6. Positional tracking control experiment with the sensor and the actuator is shown in Fig. 7. The signal passes through the sensors (HPS-M-10-5K-F & KISTLER 8628B5) and then estimates the displacement \( y(k) \) and velocity \( \dot{y}(k) \). Since the output voltage of the interface board is limited to ± 10 V, a voltage amplifier (Krohn–Hite 7500) is needed to drive the actuator in the range for effective control performance.

Fig. 6
figure 6

The schematic of the shaking table identification (Narendra and Parthasarathy 1990) test

Fig. 7
figure 7

The schematic diagram of the positional tracking control and vibration suppression experiment

The system displacement and velocity state variables together with the actuator input voltage at the sampling rate of 10 Hz are input into the network model, which is sufficient for the first natural frequency at approximately 1.9 Hz. The excitation force used to generate the training sets is an identically distributed (iid) random signal that is uniformly ranged in the [− 1, 1] voltage range. The training pattern consists of 500 input/output pairs. A concise [3–3–2] three-layer network with the modified Newton algorithm is used to model the input/output transfer function. In this function, the number of hidden neurons is selected to be adequately small to restrict the error function within a small percentage. Increasing the numbers of hidden nodes does not necessarily provide good results because it reduces the generalizability of the trained network while increasing training time and complexity. The sum-squared error variation during training is shown in Fig. 8. The command responses of the system and identification model are compared in Figs. 9 and 10. Neural network identification is highly accurate such that both lines coincide. The modified Newton method effectively identifies the highly nonlinear system. This [3–3–2] BPN will be used as the identification model when training the neural network with modified Newton controller in the next verification test.

Fig. 8
figure 8

The variations of sum-squared error during training phase in the [3–3–2] neural network with modified Newton method and genetic algorithm

Fig. 9
figure 9

The experimental results of system identification for the shaking table under square excitation

Fig. 10
figure 10

Model validation of a the impulsive displacement b the sweep-sine excitation by [3–3–2] BPN

The basic structure of a neural network with the modified Newton method and genetic algorithm controller design is shown in Fig. 11. This neural network is similar to the indirect model-following control system. In this neural network, a feedforward neural network architecture is first constructed to represent system dynamics. Then, the neural network is established on the basis of the identification model. All that required for controller design is a neural network with the two-stage model of the system plant and a reference model for the desired system performance. The neural network with the two-stage model controller can be self-organized on the basis of output error between the plant and reference model. The learning algorithm then adjusts the weight and bias of the neural network with the modified Newton method and genetic algorithm controller to meet the design requirement.

Fig. 11
figure 11

The control structure of indirect model-following control system a system identification and b controller training

The neural network with modified Newton method controller’s training update rule is defined as

$$ (\Delta w_{ij} )_{p} =\upeta\delta_{i} o_{j} . $$
(17)

where the delta \( \delta_{i} \) of each layer is back propagated from the neural network model output layer to the previous layer(s) through the error function. During training process, the weight and bias of neural network identification model does not change, only by updating those in the neural network with modified Newton controller. The neural network model communicates delta \( \delta_{i} \) subsequently to the output layer. \( \delta_{i} \) is defined by

$$ \left\{ {\begin{array}{*{20}ll} {\delta_{co} = \delta_{mi} [(n_{mi} - n_{co} ):n_{mi} ]{\text{ for the hidden layer}}} \\ {\delta_{ch} = \sum\limits_{k = 1}^{{n_{co} }} {(\delta_{k} w_{ki} )} f^{'} (net_{i} ){\text{ for the hidden layer}}} \\ \end{array} } \right. $$
(18)

where \( \delta_{mi} \) and \( n_{mi} \) is from the input layer of neural network identification model, \( n_{co} \) is the output layer neurons of neural controller, and \( f^{\prime} (net_{i} ) \) is the derivation of the transfer function. In this study, the neural network with modified Newton controller training is conducted off-line; i.e., it is trained after the successful neural network identification of the plant, for the following reasons: (1) hardware limitations of real-time computation, (2) the converging too slowly of neural network training process, and (3) the inherent short transient of the structural dynamics.

The governing equation of the structure is assumed to be unknown. However, the controller design of any unknown plant requires the mathematical model a priori. System identification is therefore an important step in the design of neural network controllers. The desired performance of the closed loop system with respect to a command input is specified in the reference model in the form of a linear differential equation.

The neural network with the modified Newton method and genetic algorithm is also applied in the tracking control experiment of a shaking table and the vibration suppression of a building structure. Because a continuous structure system has infinite degree-of-freedom, the number of steps of preceding information being fed as network input depends mainly on the accuracy of the identification model and system dynamics. In this study, the previous two steps of system state and force, \( [y(k - 1),\dot{y}(k - 1),u(k - 1)] \), are used as network input to predict the present system state [y(k)], respectively. The schematic of the neural network controller design is similar to that shown in Fig. 12. Experimental observation reveals that the first vibration mode of the shaking table dominates over other modes. Thus, the reference model is chosen as a second-order system:

$$ \ddot{x}_{m} (t) + 5.0\dot{x}_{m} (t) + 142.5x_{m} (t) = 142.5u (t ) $$
(19)

wherein the damping coefficient of the model is selected to be higher than that of the uncontrolled plant and is limited by the undamped natural frequency near 1.9 Hz. Note that a considerably damped reference model can always be selected and the neural controller will remain effective. However, the control input may be saturated. The damping ratio is limited by the capability limit of the AC servo motor.

Fig. 12
figure 12

The schematic diagram of the two-stage model controller design process

The previous section provided a discussion of the shaking table test. A [3–3–1] neural network with modified Newton method and genetic algorithm is used to train the controller with the data sampling rate of 10 Hz. Thus, the first mode can be controlled effectively without exciting the other modes that may complicate the control experiment. The transfer function of the hidden layer neurons is a hyperbolic tangent function. The neural controller has three input nodes for receiving system displacement, velocity, and command input and has one output node for the control input signal. The desired output set is then calculated through the fourth-order Runge–Kutta method from Eq. (1) every 0.1 s with the input set as the initial condition. A total of 1000 training sets that is uniformly distributed in the possible tracking range defined above is generated and collected in the initial phase for controller training.

Some experiments are employed to illustrate the effectiveness of the two-stage model controller. Although numerous works have focused on the tracking control and vibration suppression of structural systems, the application of neural networks in the positional tracking and vibration suppression of shaking tables remains limited. The schematic of the positional tracking experiment is shown in Fig. 7. A low-frequency accelerometer (KISTLER 8628B5) attached at the top floor acquires the acceleration. The measurement of accelerometer is then integrated to yield the velocity signal by the integrator (G-TECH 350133) which is then integrated and filtered to displacement signal. One linear wire potentiometer (HPS-M-10-5K-F) fixed at the bottom floor through a analog I/O board (National Instruments PCI-6052E) is used to measure the displacement of shaking table. The data acquisition and control system consists of LAB-VIEW6.1 (National Instruments) and a PCI-6052E multifunction I/O board.

Figure 13a, b show the reference sigmoidal track and the response trajectory with the two-stage model under the nominal condition. The shaking table position is controlled effectively. The corresponding performances of positional control under the neural network with modified Newton method and genetic algorithm control are shown in Fig. 14. Figure 15 illustrates the response of the vibration suppression of a building structure under initial displacement and resonance excitation. To further validate the effectiveness of the neural network with modified Newton method and genetic algorithm controller, positional tracking under Ji–Ji Earthquake, Taiwan on Sep. 21, 1999 of Richer scale 7.1 is conducted. Figure 16a shows the time history of ground acceleration record in which the maximum value is 862 gal (1 gal = 1 cm/s2). Because the experimental model can not sustain such earthquake magnitude, the maximum acceleration is reduced to about 20%, 172 gal in order to simulate the earthquake signal. Figure 16b indicates the displacement vibration amplitude is seen that there is very small difference between these two traces and Fig. 16c shows the vibration response of the building structure vibration suppression under Ji–Ji Earthquake can reduce largely by using the neural network with modified Newton method and genetic algorithm controller. The above experimental results show that satisfactory performance can be achieved by using the neural network with modified Newton method and genetic algorithm controller.

Fig. 13
figure 13

Position tracking control under the periodic reference input

Fig. 14
figure 14

A tracking control experiment of the shaking table by using the neural network with modified Newton method and genetic algorithm controller

Fig. 15
figure 15

A time-domain comparison of the open- and the closed-loop building structure responses under a initial displacement and b resonance excitation

Fig. 16
figure 16

a Time history of ground acceleration record of Ji–Ji Earthquake, Taiwan (Sept. 21, 1999). b, c A time-domain comparison of the neural network with modified Newton method and genetic algorithm controller under Ji–Ji Earthquake, Taiwan (Sept. 21, 1999)

6 Conclusion

In this paper, system modeling of nonlinear dynamic system is conducted by the identification using neural networks with modified Newton method and genetic algorithm. A [3–3–2] two-stage model has been employed to identify successfully the dynamics of a shaking table and a building structure. It is should be noted that no assumption is required during modeling and training phase. Accuracy of the neural network with modified Newton method and genetic algorithm identification is validated by above two cases.

A neural network with modified Newton method and genetic algorithm controller based on indirect linear model-following control (LMFC) has been presented. A three-layer [3–3–1] two-stage model controller is developed for the tracking control of a shaking table and the building structure vibration suppression by using an AMD system. The neural controller with displacement feedback from potentiometer measurements can then minimize the positional error and suppress the vibration by using modified Newton method and genetic algorithm. Experimental verifications show that the peak-to-peak amplitude of the structure response under Ji–Ji Earthquake (Sept. 21, 1999) excitation, respectively. The experimental results show that the neural network with modified Newton method and genetic algorithm controller is effective in tracking control and vibration suppression applications.