Keywords

This paper describes the application of complex artificial neural networks (CANN) in the inverse identification problem of the elastic and dissipative properties of solids. Additional information for the inverse problem serves the components of the displacement vector measured on the body boundary, which performs harmonic oscillations at the first resonant frequency. The process of displacement measurement in this paper is simulated using calculation of finite element (FE) software ANSYS. In the shown numerical example, we focus on the accurate identification of elastic modulus and quality of material depending on the number of measurement points and their locations as well as on the architecture of neural network and time of the training process, which is conducted by using algorithms RProp, QuickProp.

1 Introduction

The artificial neural networks (ANN) [1] are widely used in different regions of science and industry. One of their applications in mechanics is solving the coefficient inverse problem of identifying elastic [26] and dissipative properties of solids [711]. In the early 1990s, the complex artificial neural networks (CANN) were proposed in the works of T. Nitta [12, 13], which then have been widely used for various applications [1416]. The CANN parameters (weights, threshold values, inputs and outputs) are complex numbers, which used in various fields of modern technology, such as optoelectronics, filtering, imaging, speech synthesis, computer vision, remote sensing, quantum devices, spatial–temporal analysis of physiological neural devices and systems. Application of CANN in mechanics is a new research area, which has been developed over the last few years.

This paper describes the application of CANN in solving coefficient inverse problem of identifying elastic (Young’s modulus) and dissipative (quality factor) properties of solids. Additional information for solving this inverse problem serves the components of the displacement vector measured at the boundary of the body (in a discrete set of points), which performs harmonic oscillations in the vicinity of the first resonant frequency. The process of displacement measurement in this paper is simulated using calculation of finite element (FE) software ANSYS. In the numerical example shown below, the accuracy of identifying mechanical properties of the material depends on the number of measurement points and their locations, as well as on the architecture of neural network and time of the training process (which is conducted by using algorithms CBP).

2 Formulation of Identification Inverse Problems for Mechanical Properties

In the plane problem of elasticity theory, the steady-state harmonic oscillations of a rectangular area (a × b) with the angular frequency ω are considered (Fig. 12.1). The rectangle is fixed on the left side boundary S u , the vertical force F 0 is applied to the lower right corner, the rest of the right boundary S t is free from mechanical stresses. The mechanical properties of the material are described by Young’s elastic modulus, E, Poisson’s ratio, ν, and quality factor, Q.

Fig. 12.1
figure 1

Statement of the problem, the designations are interpreted into text

The vibration equations of the solid have the form [18]:

$$ \sigma_{kj,j} = - \rho (\omega^{2} - i\omega \alpha )u_{k} ;\,\,\,\,k,j = 1,2; $$
$$ \sigma_{kk} = c_{kj} \left( {1 + i\omega \beta } \right)\varepsilon_{jj} ,\;\sigma_{12} = c_{44} \left( {1 + i\omega \beta } \right)\varepsilon_{12} ,\;\varepsilon_{kj} = \frac{1}{2}\left( {u_{k,j} + u_{j,k} } \right). $$
(12.1)

The boundary conditions suppose the presence of a force at the point 2 (Fig. 12.1) and similar conditions in the rest of the boundary:

$$ u_{k} \left| {_{{S_{u} }} = 0,\,\,\,\,\sigma_{kj} n_{j} } \right|_{{S_{t} }} = 0. $$
(12.2)

The additional information to solve the inverse problem are the displacements, measured at the points 2–5 (Fig. 12.1):

$$ u_{1} = U_{mr} + iU_{mi} ,\quad \;u_{2} = V_{mr} + iV_{mi} ,\quad \;m = 2,3,4,5. $$
(12.3)

Here σ kj , ɛ kj are the stress and strain tensor components, respectively, ρ is the density. The elastic coefficients correspond to the isotropic body: c 11 = c 22 = λ + 2μ, c 12 = c 21 = λ, c 44 = μ, λ, μ are the Lamé coefficients. The applied vibration frequency ω is matched with proper first resonance frequency, which does not take into account the dissipation of mechanical energy. Coefficients α, β characterize dissipation calculated according to the method [17].

3 Structure of CANN

Structure of CANN in solving coefficient inverse problem of identification of the elastic (Young’s modulus) and dissipative (quality factor) properties of solids (Fig. 12.2).

Fig. 12.2
figure 2

Structure of CANN

This CANN is formed in three layers, called the input layer, hidden layer, and output layer. Each of layers consists of one or more nodes, represented in the diagram by the small circles, as shown in Fig. 12.2. The lines between the nodes indicate the flow of information from one node to the next (with weights W, \( \overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\smile}$}}{W} \)).

When neural networks are trained, they works basically in the following way.

  1. (1)

    The first step in training a network is to initialize the network connection weights by small random values (−1.0, 1.0).

  2. (2)

    The input pattern (X 1 = X 1r + iX 1i , X 2 = X 2r + iX 2i , X n = X nr + iX ni ) on the base of which the network will be trained are presented at the input layer of the net and the net is run normally to see what output pattern it actually produces.

  3. (3)

    The actual output pattern are compared to the desired output pattern for corresponding input pattern. The differences between actual and desired patterns form the error pattern:

    $$ Y_{n} = \mathop \sum \limits_{m} W_{nm} X_{m} = X + iY = Z; $$
    (12.4)
    $$ O_{n} = F_{c} \left( z \right) = F\left( x \right) + iF\left( y \right) = \frac{1}{1 + \exp ( - X)} + i\frac{1}{1 + \exp ( - Y)}; $$
    (12.5)
    $$ Ep = \left( \frac{1}{N} \right)\sum\nolimits_{n = 1}^{N} {\left| {T_{n} - O_{n}} \right|}^{2} = \frac{1}{2}\sum\nolimits_{n = 1}^{N} {(\left| {Re\left( {T_{n} } \right) - Re\left( {O_{n} } \right)} \right|^{2} } + \left. {\left| {Im\left( {T_{n} } \right) - Im\left( {O_{n} } \right)} \right|^{2} } \right), $$
    (12.6)

    where W nm is the (complex-valued) weight connecting neuron n and m, X m is the (complex-valued) input signal from neuron m, F is the sigmoid function, Ep is the square error for the pattern p, O n are the output values of the input neuron, T n is the target pattern.

  4. (4)

    The error pattern is used to adjust the weights on the output layer so that the error pattern reduced the next time if the same pattern was presented at the inputs:

    $$ W_{nm} = W_{nm} + \Delta W_{nm} = W_{nm} + \overline{{H_{m} }} \Delta \lambda_{n} ; $$
    (12.7)
    $$ \Delta \lambda_{n} = \varepsilon \left\{ {Re\left[ {\delta^{n} } \right]\left( {1 - Re\left[ {O_{n} } \right]} \right)Re\left[ {O_{n} } \right] + i Im\left[ {\delta^{n} } \right]\left( {1 - Im\left[ {O_{n} } \right]} \right)Im\left[ {O_{n} } \right]} \right\}, $$
    (12.8)

    where λ n is the threshold value of the output neuron n, \( \delta^{n} = T_n - O_n \) is the difference (error) between the actual pattern and the target pattern, H m are the output values of the hidden neuron.

  5. (5)

    Then the weights of the hidden layer are adjusted similarly, it is stated in this time what they actually produce the neurons in the output layer to form the error pattern for the hidden layer:

    $$ \overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\smile}$}}{W}_{lm} = \overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\smile}$}}{W}_{lm} + \Delta \overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\smile}$}}{W}_{lm} = \overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\smile}$}}{W}_{lm} = \overline{{{\text{l}}_{l} }} \Delta \theta_{m} ; $$
    (12.9)
    $$ \begin{aligned} \Delta \theta_{m} =&\,\varepsilon\{ \left( {1- Re\left[ {H_{m} } \right]} \right)Re\left[ {H_{m} }\right]\,\sum\nolimits_{n} {\left( {Re\left[ {\delta^{n} }\right]\left( {1 - Re\left[ {o_{n} } \right]} \right)} \right.}\,Re\left[ {W_{nm} } \right] \\&+ Im\left[ {\delta^{n} }\right]\left( {1 - Im\left[ {o_{n} } \right]} \right)Im\left.{\left[ {o_{n} } \right]Im\left[ {W_{nm} } \right]} \right)\} -i\varepsilon \{ (1 \\& - Im\left. {\left[ {Hm} \right]}\right)Im\left[ {Hm} \right]\sum\nolimits_{n} {\left( {Re\left[{\delta^{n} } \right]} \right.\left( {1 - Re\left[ {o_{n} } \right]}\right)} Re\left[ {o_{n} } \right]Im\left[ {W_{nm} } \right] \\&-Im\left[ {\delta^{n} } \right]\left( {1 - Im\left[ {o_{n} } \right]}\right)Im\left[ {o_{n} } \right]Re\left[ {W_{nm} } \right])\} \\\end{aligned} $$
    (12.10)

    where \( \overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\smile}$}}{W}_{lm} \) is the weight between input neuron l and hidden neuron m, I l are the output values of the input neuron l, θ m is the threshold value of the hidden neuron m.

  6. (6)

    The network is trained by presenting each input pattern in turn at the inputs and propagating forward and backward, followed by the next input pattern. Then this cycle is repeated many times. The training stops when any of these conditions occurs: the maximum number of epochs (repetitions) is reached or the error has an acceptable training error threshold.

After network training, a stable network structure is established. The network then can be used to predict output values from new input values, i.e. to solving coefficient inverse problem of identifying elastic and dissipative properties of solids.

4 Application of CANN to Solve Coefficient Inverse Problem for Identification of Elastic (Young’s Modulus) and Dissipative (Quality Factor) Properties of Solids

In order to identify Young’s modulus and quality factor as it was described above used CANN, in which they were output data. As input data, we used displacement amplitudes (12.3), measured on solid surfaces. By training the CANN with a set of input and output data, the first step of the computation process was calculation of natural resonance frequencies, and following definition of steady oscillations of the solid at these frequencies or in their neighborhood with the mechanical energy dissipation.

Figure 12.1 shows four points 2, 3, 4, 5, where displacement amplitudes were measured. The process of the displacement measurements is simulated using calculation of finite element software ANSYS. Table 12.1 presents the input and output data for the “measurement” at the point 2.

Table 12.1 Data from problem of modal and harmonic analysis using ANSYS

CANN has the following configuration in dependence on number of points and the frequency “measurement”:

  1. (i)

    at one measurement point and one frequency, this neural network consists of two neural input values: \( \varvec{U}_{{2\varvec{r}}} + \varvec{iU}_{{2\varvec{i}}} \), \( \varvec{V}_{{2\varvec{r}}} + \varvec{iV}_{{2\varvec{i}}} \) and one neural output value: E + iQ;

  2. (ii)

    at one measurement point and three frequencies, this neural network consists of six neural input values: \( \varvec{U}_{{1\varvec{r}}} + \varvec{iU}_{{1\varvec{i}}} ,\varvec{ V}_{{1\varvec{r}}} + \varvec{iV}_{{1\varvec{i}}} \), \( \varvec{U}_{{2\varvec{r}}} + \varvec{iU}_{{2\varvec{i}}} ,\varvec{ V}_{{2\varvec{r}}} + \varvec{iV}_{{2\varvec{i}}} \), \( \varvec{U}_{{3\varvec{r}}} + \varvec{iU}_{{3\varvec{i}}} ,\varvec{ V}_{{3\varvec{r}}} + \varvec{iV}_{{3\varvec{i}}} \) and one neural output value: E + iQ;

  3. (iii)

    at three measurement points and one frequency, this neural network consists of six neural input values: \( \varvec{U}_{{1\varvec{r}}} + \varvec{iU}_{{1\varvec{i}}} ,\varvec{ V}_{{1\varvec{r}}} + \varvec{iV}_{{1\varvec{i}}} \), \( \varvec{U}_{{2\varvec{r}}} + \varvec{iU}_{{2\varvec{i}}} ,\varvec{ V}_{{2\varvec{r}}} + \varvec{iV}_{{2\varvec{i}}} \), \( \varvec{U}_{{3\varvec{r}}} + \varvec{iU}_{{3\varvec{i}}} ,\varvec{ V}_{{3\varvec{r}}} + \varvec{iV}_{{3\varvec{i}}} \) and one neural output value: E + iQ.

The results of the computer experiments, conducted by using CANN with 2000 patterns, in which, 1800 patterns were used for training and 200 for testing, are show in Tables 12.2, 12.3, 12.4 and 12.5.

Table 12.2 The results of training and testing for CANN with two and six input values at the different points (with the same number of epochs)
Table 12.3 The results of training and testing for CANN with two input values at the point 4 (with different epochs)
Table 12.4 The results of training and testing for CANN with two input values at the point 5 (with different numbers of hidden layers)
Table 12.5 The results of training and testing for CANN with six input values at the points 3, 4, 5 (with different numbers of hidden layers)

Time of training for CANN with different epochs (Table 12.3) is shown in Fig. 12.3. Time of training and testing for CANN with different number of neurons in hidden layer is shown in Fig. 12.4. The relationship between time of training and testing for CANN with different architecture is shown in Fig. 12.5 (data from Table 12.5).

Fig. 12.3
figure 3

Relationship between time of training and testing for CANN with different epochs

Fig. 12.4
figure 4

Relationship between time of training and testing for CANN with different architecture

Fig. 12.5
figure 5

Relationship between time of training and testing in CANN with different architecture

Figures  12.6, 12.7, 12.8 and 12.9 show the results of testing with 200 patterns in CANN (with architecture 2-4-1, see Table 12.2, the row 3). The circles present the data of test E, triangles and squares describe the forecast data E c , obtained by using CANN. The relative errors δ E , δ Q are calculated in the forms: δ E = |E c E|/E, δ Q = |Q c Q|/Q. In this case, corresponding triangles and squares with the identification error no exceeding of 20 %, are shown in Figs. 12.6 and 12.7.

Fig. 12.6
figure 6

Results of testing E at the threshold error of 20 %

Fig. 12.7
figure 7

Results of testing Q at the threshold error of 20 %

Fig. 12.8
figure 8

Results of testing E at the threshold error of 10 %

Fig. 12.9
figure 9

Results of testing Q at the threshold error of 10 %

The similar results for threshold error of 10 % are shown in Figs. 12.8 and 12.9.

Figure 12.10 demonstrates the results of identification in dependence on number of test data in the form of two curves, the solid curves correspond to the required data, and the dashed lines define the forecast data obtained by using CANN. The graph of training error (12.6) with 20,000 epochs is shown in Fig. 12.11.

Fig. 12.10
figure 10

Prediction of results obtained by using CANN

Fig. 12.11
figure 11

Graph of training error

5 Conclusion

In this paper, we developed a method for determination of material elastic and dissipative properties by using data for harmonic oscillations on the resonance frequency based on a combination of the finite element method and CANN. The results of the experiments showed that CANN with one of the following architectures:

  • 2 (input neurons)—4 (hidden neurons)—1 (output neuron)

  • or

  • 6 (input neurons)—6 (hidden neurons)—1 (output neuron)

gives the best result of identification. The time expenses caused by the CANN training were also estimated in this paper. The developed method and computer program could be used to determine the dissipative properties on different frequencies (not only on the first one), as well as in anisotropic elastic solids.