Keywords

1 Introduction

In the past decades, smart materials, such as piezoelectric materials, magnetostrictive materials, shape memory alloys, etc., have been widely used in many fields. The piezoelectric and magnetostrictive actuators have been specially used for micro-displacement systems [1, 2]. Since smart-materials-based actuators have advantage in the output force, position resolution, and response speed [3], they have been attached importance in ultra-precision positioning systems [4]. However, the inherent hysteresis nonlinearity in smart-materials-based actuators, which is non-differentiable and multi-valued mapping, frequently leads to undesired tracking errors, oscillations, and even instability [5]. The model-based scheme for hysteresis compensation is currently popular in control systems [6]. A large amount of hysteresis models have been proposed in the past decades, such as PI model [7], KP model [8], Preisach model [9], Maxwell slip model [10], Jiles-Atherton model [11], Duhem model [12], Bouc-Wen model [13, 14], and so on. However, the ultra-precision positioning systems need more accurate hysteresis models so as to meet the requirement of science and technology.

Three-layer feed-forward neural networks (NNs) have been regarded as one of the best ways to model nonlinear systems because they can implement all kinds of nonlinear mapping. However, the mapping of hysteresis consists of one-to-multiple and multiple-to-one mappings, while the NNs are incapable of identifying one-to-multiple mapping [15], so the one-to-multiple mapping has to be eliminated so that the neural approaches can be used to model hysteresis. Ma [16] proposed a hysteretic operator (HO), expanded the input space of NN from 1-dimension to 2-dimension based on the HO so that an NN-based hysteresis model was established, and named the method as expanded-space method. Zhao [17], Dong [18], Zhang [19] and Ma [20, 21] proposed respectively new HOs and constructed neural hysteresis models, thereby improving the expanded-space method.

In this paper, a new HO, which is made up of a constraint factor (CF) and an odd m-order polynomial, is proposed to expand the input space of NN. And then, based on the proposed HO, the one-to-multiple mapping of hysteresis is transformed into one-to-one mapping so that the neural approach can be used to identify the expanded mapping. Finally, a NN-based hysteresis model is developed and used to approximate a set of real data from a magnetostrictive actuator. The experimental results demonstrate that the proposed model is effective.

2 HO Construction

2.1 HO Definition

In this paper, the HO is consisted of a CF and an odd m-order polynomial with the constant term. The function, c (x) = 1–e x, is used as the CF of HO. The role of CF is to constrain the amplitude of HO curve and ensure the curve passes through the origin in every minor coordinate system. Therefore, in the ith minor coordinate system, the HO is defined as follows:

$$ f(x_{i} ) = (1 - e^{{ - x_{i} }} )(a_{0} + \sum\limits_{j = 1}^{m} {a_{j} x_{i}^{2j - 1} } ) $$
(1)

where x i and f are respectively any input and the corresponding output of HO in the ith minor coordinate system.

In the main coordinate system, the HO is described as

$$ h(x) = \left\{ \begin{aligned} h(x_{ei} ) + f(x - x_{ei} )\;\;\;\;\;\;\;x > x_{ei} \hfill \\ h(x_{ei} ) - f(x_{ei} - x)\;\;\;\;\;\;\;x < x_{ei} \hfill \\ \end{aligned} \right. $$
(2)

where [x ei , h(x ei )] are the coordinates of the origin of the ith minor coordinate system in the main coordinate system, x is any input and h is the corresponding output of HO.

2.2 Parameter Computation

As known to all, the best method of determining polynomial coefficients is the least square method. Thus, the least square method is adopted to compute the HO parameters based on the samples used for training neural network.

In terms of the least square method, the residual δ i is written as

$$ \delta_{i} = y_{i} - f(x_{i} ) $$
(3)

Consequently, the sum of square residuals is shown as follows:

$$ S = \sum\limits_{i = 1}^{n} {\delta_{i}^{2} } = \sum\limits_{i = 1}^{n} {[y_{i} - f(x_{i} )]^{2} } = \sum\limits_{i = 1}^{n} {[y_{i} - (1 - e^{{ - x_{i} }} )(a_{0} + \sum\limits_{j = 1}^{m} {a_{j} x_{i}^{2j - 1} )} ]^{2} } $$
(4)

To minimize S, the partial derivatives of S with regard to a 0, a 1, …, a m should be set to zeros. Therefore, the (m + 1) partial derivative equations are given as follows:

$$ \left\{ {\begin{array}{*{20}l} {\frac{\partial S}{{\partial a_{0} }} = - 2\sum\limits_{i = 1}^{n} {[(1 - e^{{ - x_{i} }} ) \cdot y_{i} - (1 - e^{{ - x_{i} }} )^{2} (a_{0} + \sum\limits_{j = 1}^{m} {a_{j} x_{i}^{2j - 1)} )} ]} { = }0} \hfill \\ {\frac{\partial S}{{\partial a_{k} }} = - 2\sum\limits_{i = 1}^{n} {[(1 - e^{{ - x_{i} }} ) \cdot x_{i}^{2k - 1} \cdot y_{i} - (1 - e^{{ - x_{i} }} )^{2} (a_{0} x_{i}^{2k - 1} + \sum\limits_{j = 1}^{m} {a_{j} x_{i}^{2(j + k - 1)} )} ]} = 0,\;k = 1,2, \cdots ,m} \hfill \\ \end{array} } \right. $$
(5)

Rearranging the Eq. (5) gives

$$ \left\{ {\begin{array}{*{20}l} {a_{0} \cdot \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} } + \sum\limits_{j = 1}^{m} {a_{j} \cdot \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2j - 1)} } } \; = \;\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} ) \cdot y_{i} } } \hfill \\ {a_{0} \cdot \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2k - 1} } + \sum\limits_{j = 1}^{m} {a_{j} \cdot \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2(j + k - 1)} } } = \;\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} ) \cdot x_{i}^{2k - 1} \cdot y_{i} } ,k = 1,2, \cdots ,m} \hfill \\ \end{array} } \right. $$
(6)

The expansion form of the Eq. (6) is given as follows:

$$ \left\{ {\begin{array}{*{20}l} {a_{0} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} } + a_{1} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i} } + \cdots + a_{m} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2m - 1} } = \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )y_{i} } } \hfill \\ {a_{0} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i} } + a_{1} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2} } + \cdots + a_{m} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2m} } = \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )x_{i} y_{i} } } \hfill \\ { \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots \cdots } \hfill \\ {a_{0} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2m - 1} } + a_{1} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2m} } + \cdots + a_{m} \sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2(2m - 1)} } { = }\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )x_{i}^{2m - 1} y_{i} } } \hfill \\ \end{array} } \right. $$
(7)

The Eq. (7) is written as the following matrix Eq.

$$ \left[ {\begin{array}{*{20}c} {\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} } } & \cdots & {\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2m - 1} } } \\ \vdots & \vdots & \vdots \\ {\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2m - 1} } } & \cdots & {\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )^{2} x_{i}^{2(2m - 1)} } } \\ \end{array} } \right]\left[ {\begin{array}{*{20}c} {a_{0} } \\ {} \\ \vdots \\ {a_{m} } \\ \end{array} } \right] = \left[ {\begin{array}{*{20}c} {\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )y_{i} } } \\ {} \\ \vdots \\ {\sum\limits_{i = 1}^{n} {(1 - e^{{ - x_{i} }} )x_{i}^{2m - 1} y_{i} } } \\ \end{array} } \right] $$
(8)

i.e.,

$$ XA = Y $$
(9)

The HO parameters are obtained by solving Eq. (9),

$$ A = X^{ - 1} Y $$
(10)

3 Experimental Verification

In the following, two verification experiments are implemented. In the experiments, the presented neural hysteresis model is compared with the PI model by approximating a set of real data from a smart-material-based actuator so as to validate the effectiveness of the proposed approach.

The experimental setup is comprised of a magnetostrictive actuator (MFR OTY77), a current source, a dSPACE control board with 16-bit analog-to-digital and digital-to-analog converters, and a PC. A set of data containing 1916 input-output pairs is obtained.

The data is equally divided into two groups. One group is used for training neural networks, and another group is used for model verification.

3.1 The Proposed Model

In this section, the neural hysteresis model is employed to approximate the real data. The activation function of the hidden layer is the sigmoid function and that of the output layer is the linear function. The comparison of the different orders shows that the 9-order polynomial is most suitable to the HO in this experiment. The HO parameters are listed in the Table 1.

Table 1. HO Parameters

To determine the optimal number of hidden nodes, the number from 1 to 100 is tried in this experiment. The best 3 performances are listed in the Table 2. It can be seen from the Table 2, that the neural hysteresis model has the best performance when the number of hidden nodes becomes 5. Therefore, a three-layer feed-forward neural network with two input nodes, 5 hidden nodes and one output node was employed to approximate the real data in this experiment. After 229 iterations, the training procedure finishes. The mean square error (MSE) of model prediction is 0.0611. The Figs. 1 and 2 illustrate the model prediction and absolute error respectively.

Table 2. The top 3 performances of NN with different number of hidden neurons
Fig. 1.
figure 1

Comparison of the proposed model prediction and the real data

Fig. 2.
figure 2

The absolute error of the proposed model

3.2 PI Model

In addition, to compare with the proposed model, the PI model was also applied to approximate the measured data. The model thresholds were calculated via the following formula:

$$ r_{i} = \frac{i - 1}{N}\left[ {\hbox{max} (x(k)) - \hbox{min} (x(k))} \right] $$
(11)

where N is the number of backlash operators and i = 1, 2, …, N.

The Matlab nonlinear optimization tool was used to determine the weights of backlash operators. However, the calculation time of the PI model increases along with the increase of N, so only the PI models containing 1–5000 backlash operators were tried. The top three performances are listed in the Table 3. Therefore, N = 4994 is selected in this experiment, which leads to 20-hour calculation time. The MSE of model prediction is 0.4327. Figures 3 and 4 display the model prediction and absolute errors respectively.

Table 3. Performances of different no. of backlash operators
Fig. 3.
figure 3

Comparison of the PI model prediction and the real data

Fig. 4.
figure 4

The absolute error of the PI model

3.3 Comparison

In the above experiments, the MSE of the proposed neural hysteresis model is 85.88 % smaller than that of the PI model, which demonstrates that the proposed neural model can better approximate the real data measured from the magnetostrictive actuator than the PI model.

4 Conclusions

In this paper, a new HO is proposed. The HO consists of two components: a CF and an odd m-order polynomial. And then, based on the constructed HO, the one-to-multiple mapping of hysteresis is transformed into a continuous one-to-one mapping by expanding the input space of NN. In this way, the expanded mapping only contains one-to-one and multiple-to-one mappings, which can be identified using the neural approaches. Finally, an experiment is implemented to verify the proposed hysteresis model. The verification performance approves the proposed approach.