Keywords

1 Introduction

In the framework of inverse crack identification the goal is to predict the unknown crack, defined by its parameters \(P\) [1, 2], based only on accessible structural response measurements \(u\left({P}_{0}\right)\). The most commonly used are the vibrational response and the mechanical responses. However, they face the challenge of the ill-Posedness [3,4,5,6,7,8], as the measured structural response may be insensitive to different cracks, and very different values of \(P\) can correspond to close measurements. Additionally, from the optimization problem perspective, this is a non-differentiable problem.

The crack identity \(P\) is dictated by the considered inverse problem. Thus, the correct numerical simulation of the problem is very important. The second difficulty is related to structural response characteristic in relation to variation to the crack parameters [9]. The existence of local minima is inevitable is these type of problems. Where local optimization algorithms cannot guarantee to attain the global minimum. Plus they require the problem to be differentiable. For these reasons, global optimization algorithms are needed [3, 10, 11].

An efficient algorithm is capable of finding the global minimum, regardless of the starting point in the search field. Global methods simulate a sample of parameters by exploiting only the information provided by the calculation of \(f\left(P\right)\). To get the global minimum of a function using global methods, it would be necessary to go through all the space of the independent variables, which is generally impossible for classical optimization methods, given the size of the space and the number of variables.

In the inverse problem stage, the optimization algorithm is applied. Its goal is to find the closest response to one consequence of the crack to be identified [6]. The structural response is measured at the reference specimen. The optimization algorithm will then search for the crack parameters that correspond to the measured response, By calculating the structural response \(u\left(P\right)\) corresponding to a possible crack \(P\). The fitness function value is the error between this vector and the reference structural response \(u\left({P}_{0}\right)\) caused by the real crack parameters. [12]. Figure 1 depicts the inverse crack identification approach. Where the optimal crack identity is the one that provides a fitness function value less or equal to a limit. This limit varies from one case to another and also on the nature of the crack parameters. It is decided after knowing the fitness function equivalent to the wanted precision. The stopping criterion is considered based on the number of iterations. The number of iterations depends on the optimization algorithm, so the optimal crack identity can be reached if the optimization algorithm is fast enough.

The structural response researchers have suggested several approaches based on ANN [13], Weinstein et al. suggested. One of the early deep learning works in this field of vibration analysis is made by Barai et al. [14] in damage identification Zang et al. [15] suggested an approach based on frequency response functions as input to ANN. Abdeljaber et al. [16] designed a real-time damage identification strategy based on one-dimensional convolutional neural networks. In the studies based on the mechanical response, Stavroulakiset al [17, 18] created a crack identification method based on backpropagation neural networks for structural response generation, using the boundary element method to create the training data. Khaleghi et al. [19] suggested an inner fracture characteristics approach based on artificial neural networks under noisy measurement.

Fig. 1.
figure 1

Inverse crack identification approach

2 Numerical Simulation

In elastostatics, the sum of all forces on the structure is equal to zero, and the displacements are not a function of time [20,21,22]. The equilibrium equations are stated as:

$$\sigma_{ij,j} + b_{i} = 0,{\quad }i,j = 1,2,3$$
(1)

\({\sigma }_{ij}\) is the Cauchy stress tensor. \({b}_{i}\) represents the body-force components. And the strains \({\varepsilon }_{ij}\) are defined by:

$$\varepsilon_{ij} = \frac{1}{2}\left( {u_{i,j} + u_{j,i} } \right),{\quad}i,j = 1,2,3$$
(2)

\({u}_{i,j}\) are the displacement component. Considering that the material is elastic, homogenous and isotropic, Eqs. (1) and (2) are related by Lamé’s equation:

$$\sigma_{ij} = \lambda \varepsilon_{kk} \delta_{ij} + 2\mu \delta_{ij} ,{\quad }i,j,k = 1,2,3$$
(3)

The Kronecker-delta function \({\delta }_{ij}\) is written as follows:

$$\delta_{ij} = \{ 1{\quad }i = j,{\quad }0{\quad}i \ne j \}$$
(4)
$$\lambda = \frac{2\nu \mu }{{\left( {1 - 2\nu } \right)}}$$
(5)
$$\mu = \frac{E}{{2\left( {1 + \nu } \right)}}$$
(6)

where \(\lambda\) Lamé constant, and \(\mu\) is the shear modulus. The strains have to satisfy the following compatibility equations:

$$\frac{{\partial \varepsilon_{ij} }}{{\partial x_{i} \partial x_{k} }} - \frac{\partial }{{\partial x_{i} }}\left( { - \frac{{\partial \varepsilon_{jk} }}{{\partial x_{i} }} + \frac{{\partial \varepsilon_{ik} }}{{\partial x_{j} }} + \frac{{\partial \varepsilon_{ij} }}{{\partial x_{k} }}} \right) = 0,{\quad}i,j,k = 1,2,3{\,}and{\,}i \ne j \ne k$$
(7)

The problem is defined as follows. Let \(\Omega\) denote an open set with boundary \(\Gamma\) subject to the boundary conditions:

$$u = \underline{{u_{i} }} \,{\rm{on}}\,\varGamma_{u} ,$$
(8)
$$t = \underline{{t_{i} }} \,{\rm{on}}\,\varGamma_{t}$$
(9)

Where

$$\varGamma = \varGamma_{u} {\rm{\bigcup }}\varGamma_{t} ,{ }\varGamma_{u} {\rm{\bigcap }}\varGamma_{t} = \emptyset$$
(10)

And

$$\underset{\raise0.3em\hbox{$\smash{\scriptscriptstyle-}$}}{\varOmega } = \Omega {\rm{\bigcup} }\varGamma$$
(11)

The problem is: given \(\underset{\_}{{u}_{i}}\), \(\underset{\_}{{t}_{i}}\) and \({b}_{i}\); find \({u}_{i}(y)\), \(y\in \underset{\_}{\Omega }\). Furthermore, the plane strain is assumed and therefore:

$$\varepsilon_{zz} = \varepsilon_{xz} = \varepsilon_{yz} = 0$$
(12)
Fig. 2.
figure 2

Display of meshing around the crack tips and Von Mises constraint

The simulations in this paper are made using Abaqus. It is a powerful finite element simulation tool used in many fields of mechanical studies. It has special features dedicated to fracture mechanics, which makes this tool very advantageous, such that the crack segment model, the calculation of J integral, and different crack propagation laws.

In this section, we present a simulation of a cracked plate of a rectangular shape with 1 mm thickness, having dimensions of 30 mm height and 10 mm width, it is subjected to traction force from the upper and the lower sides. The Young modulus and Poisson coefficient of the material were respectively \(E=210\,GPa\) and \(\nu =0.3\). The vertical sides have meshed with 80 quadrilaterals and the horizontal side with 30 elements. The meshing of the crack has been divided into two main areas; the area of the crack tip, covering both crack endings, and the area of the middle of the crack. The sweep function is used to mesh the crack tips, with radius of the circle is 0.5 mm. 20 elements along the contour of the circle, and 10 elements along the radius of the circle, resulting in a total of 200 elements at each crack tip as shown in Fig. 2. Along with the the Von Mises constrain.

The majority of crack identification studies are based on boundary data because boundaries are considered the only accessible part of the structure. In the case studied in this paper, the horizontal boundaries (upper and lower boundaries) are subjected to boundary conditions, in real cases; it means interaction with other structures of solicitations, which makes them inaccessible for measurements, leaving the vertical boundaries to be the only source of information.

3 Radial Basis Functions Modelling

Radial Basis Functions (RBF) are very efficient for interpolation between existing data [23, 24]. To approximate a function \(f\left(x\right)\), where \(x\) is an M-dimensional vector, RBF requires the information in a form of a set of \(N\) nodes \({x}_{i}\), for which the values of the function are known. For any new value of \(x\) the interpolation is performed involving just a few nearby nodes. The approach of RBF performs it one continuous function, defined over the whole domain. The approximation is written as the combination of \({g}_{i}\) functions:

$$f\left( x \right) \approx \mathop \sum \nolimits_{i = 1}^{N} \alpha_{i} g_{i} \left( x \right)$$
(13)

This equation is defined once the basis functions \({g}_{i}\) are selected and the coefficients \({\alpha }_{i}\) are known. Where \({\alpha }_{i}\) are the combination coefficients. For the radial basis functions, the Euclidian distance is considered in this paper.

$$g_{i} \left( x \right) = g_{i} \left( {\left\| {x - x_{i} } \right\|} \right), i = 1,2, \ldots ,N$$
(14)

To determine the coefficients \({\alpha }_{i},\) the interpolation needs to be exact in all \(N\) nodes, therefore, the system of \(N\) equations defined by:

$$f\left( {x_{j} } \right) = y_{j} = \mathop \sum \nolimits_{i = 1}^{N} \alpha_{i} g_{i} \left( {x_{j} } \right),{ }j = 1, \ldots ,N$$
(15)

where \({y}_{j}\) are known values of the function in the nodes. Introducing the following matrix notation

$$G = \left[ {g_{1} \left( {x_{1} } \right){ } \ldots { }g_{N} \left( {x_{1} } \right){ } \vdots { } \ddots { } \vdots { }g_{1} \left( {x_{N} } \right){ } \ldots { }g_{N} \left( {x_{N} } \right)} \right];\;\alpha = \left[ {\alpha_{1} ,\alpha_{2} , \ldots ,\alpha_{N} } \right]^{T} ;\;Y = \left[ {y_{1} ,y_{2} , \ldots ,y_{N} } \right]^{T}$$
(16)

The system can be written as:

$$\alpha \cdot G = Y$$
(17)

Matrix Eq. (15) is solved for the interpolation coefficients \({\alpha }_{i}\). The interpolation coefficients are computed once and for all, and involve the known values at the nodes. Once this is achieved, we can approximate the function at any given point [23]. Considering that the results of Eq. (13) are exact in the nodes, and gives interpolated results for any new value of \(x\). The RBF represents one approximation valid for the whole domain in which the original data were situated.

The results is a model that is able of reproducing the original data field, plus interpolated values for any new set of parameters within the initial variables domain. The results of extrapolation outside this domain may be of poor accuracy [23]. We use the RBF model to retrieve the values of the unknown geometric crack parameters in this study. And employed to provide the structural response, representing the direct model.

4 Deep Artificial Neural Network

Deep ANN has been used widely in a wide range of engineering problems [25, 26]. Theoretically, it has the power to connect the problem’s input and target features in nonlinear and complex spaces, given enough neurons in the hidden layers. With the wide availability of ANN tools and open source codes, engineers only have to solve the issue of the number of layers, and the number of neurons in each layer [27].

The information is stored in the node’s weights. Thus for the network to approximate the system’s output, these weights are optimized in a process called training. Figure 3. Illustrates the Deep ANN network. Training of collected data is used to model the characteristics of the system. During the training, the process uses a smaller set of data called the validation dataset is used test the quality of the network predictions, and change the parameters to achieve high precision [28].

Fig. 3.
figure 3

Von Mises constraint ANN structure to determine damage level

The Deep ANN is characterized by multiple hidden layers, besides the input layer and the output layer. The number of hidden layers is denoted in the figure by \(D\), the number of input and output neurons are denoted by \(n\) and \(m\) respectively, and the number of neurons in each hidden layer is denoted by \(l\). Each neuron is connected to all the neurons in the next layer, and the strength of the connection is weighted though \({w}_{ij}^{k}\). Each neuron is assigned a bias value β. In the training phases, the optimization algorithm will find the optimal weights and biases combinations that correspond to the lowest prediction error, according to the testing data set.

5 YUKI Algorithm

YUKI algorithm is a newly suggested method [20], with an innovative search space reduction technique. It uses simple steps to focus on interesting search areas, and dynamically changes the size and position of the search focus. The first idea of YUKI algorithm is to create a local search area smaller than the search domain, it is positioned around the absolute best solution at the current iteration \({{\varvec{X}}}_{{\varvec{b}}{\varvec{e}}{\varvec{s}}{\varvec{t}}}\), and its size is calculated as the distance between this point and the \({{\varvec{X}}}_{{\varvec{M}}{\varvec{e}}{\varvec{a}}{\varvec{n}}{\varvec{B}}{\varvec{e}}{\varvec{s}}{\varvec{t}}}\). Where MeanBest is the centre of the point cloud representing the historical best founts found by each solution. \({\varvec{L}}{\varvec{T}}\) and \({\varvec{L}}{\varvec{B}}\) are the top and bottom boundaries of the local search area respectively. They are calculated as follows:

$$D=\left|{X}_{best}-{X}_{MeanBest}\right|$$
(1)
$$LT={X}_{best}+D$$
(2)
$$LB={X}_{best}-D$$
(3)

The second idea of this algorithm is to dedicate two parts of the population to exploration and exploitation simultaneously, by splitting the population. One part is assigned to exploration outside the local search area, and the other is assigned to focus on searching inside the local search area. The size of the population dedicated to each part is determined by the EXP parameter, which should be between 0 and 1. 0.6 for example, means 60% of the population will be dedicated to exploring outside the focus area, and 40% will focus the search inside the local search area.

Fig. 4.
figure 4

Illustration of the focus zone exploration strategy

This is done by first, generating a random distribution of points inside the local search area, then assigning the dedicated search populations selected for exploration outside the local search area. Exploration spread is directed to look toward the historical best points as follows:

$$E={X}_{loc}-{X}_{best}$$
(5)

where \({{\varvec{X}}}_{{\varvec{l}}{\varvec{o}}{\varvec{c}}}\) denotes a position of the selected point to be directed toward exploration, and \({X}_{best}\) is the historical best point found by this individual particle. The next position will then be calculated as follows:

$${X}_{new}={X}_{loc}+ E$$
(6)

For the solutions dedicated to the exploitation inside the local search zone, we first calculate the distance from the centre Eq. (7), then calculate the next position as in Eq. (8). Where rand here is the random value between 0 and 1. And for each solution. Figure 4 illustrates the YUKI algorithm ideas.

$$F = X_{loc} - X_{best}$$
(7)
$$X_{new} = X_{loc} - rand \times { }F$$
(8)

These simple steps lead to the complex behaviour of the local search area, as it can dynamically reduce its size if new solutions are found inside, or increase the size significantly if better solutions are found outside. The location of this focus area also changes automatically throughout the search. With the progress of iterations, the local search area tends to reduce in size leading to focusing on the optimal position at the end of the search.

6 ANN and RBF for Direct Problem Representation

The boundary displacement responses are collected from 67 simulations. The boundary conditions are constant but crack sizes and positions vary. The size of the crack varies between 0 and 5 mm, and its position is located between –10 and 10 on the Y-axis and –2 and 2 on the X-axis (Fig. 5). Each node in the vertical boundaries is considered a sensor point, thus the total number of 162 data points, 81 from the left side and 81 from the right vertical side boundary.

Fig. 5.
figure 5

Illustration of the simulated problem with crack parameters.

Fig. 6.
figure 6

Displays the training results for the deep ANN network.

These results are used to build surrogate models using RBF and ANN networks. Several ANN designs have been investigated. After statistical analysis, the backpropagation architecture with these layers (60:150:250:200:180) has been selected for further study. Figure 6 displays the training results for this deep ANN network, using conjugate gradient backpropagation with Fletcher-Reeves updates.

Fig. 7.
figure 7

Boundary displacement prediction error, using deep ANN and RBF networks.

Figure 7 Shows the boundary displacement prediction error using deep ANN and RBF. Where the considered example is a crack at the lower left side with center coordinates (–1.75, –9.5) and a size of 3.2 mm. The left side graph compares the predictions to of the real displacement field. And on the right side graph the prediction errors of deep ANN and RBF. These figures show an advantage for Deep ANN in terms of overall accuracy, it also indicates that although RBF has a smoother displacement field, more appropriate in this case than the results found by deep ANN, RBF faces a challenge in predicting the displacement at that boundary node in the proximity of the crack. These results are found in other examples and are indicative of the overall predictions.

7 Crack Identification Results

The crack identification is formulated as an optimization problem, as suggested in Fig. 1, the goal is to find the crack corresponding to the boundary displacement response most close to the measurement, the closeness is calculated as described in Eq. (9). We considered 4 segments of the displacement field, representing the four quarters of the vertical boundaries.

Fig. 8.
figure 8

Crack size prediction based on boundary displacement using Deep ANN and RBF.

Due to the unsymmetric nature of the problem, we assume that this approach help identify the region in which the crack exists more efficiently than if we consider the boundary displacement field as a whole. Three cost functions were considered, expressed in Eqs. (10), (11) and (12).

$$f^{i} \left( P \right) = \frac{{\left\| {u^{i} \left( {p_{0} } \right) - u^{i} \left( p \right)^{2} } \right\|}}{{\left\| {u^{i} \left( {p_{0} } \right)^{2} } \right\|}}\;{\text{with}}\,i\, = 1,2,3,4.$$
(9)
$${F}_{1}\left({P}_{opt}\right)=\mathrm{min}({f}^{1}\left(P\right),{f}^{2}\left(P\right),{f}^{3}\left(P\right),{f}^{4}\left(P\right))$$
(10)
$${F}_{2}\left({P}_{opt}\right)=\mathrm{Mean}({f}^{1}\left(P\right),{f}^{2}\left(P\right),{f}^{3}\left(P\right),{f}^{4}\left(P\right))$$
(11)
$${F}_{2}\left({P}_{opt}\right)={F}_{1}+{F}_{2}$$
(12)

The first function takes into consideration only the boundary quarter corresponding to the minimum fitness value. This is designed to isolate the noisy effect created by the rest of the data points. The second function is similar to the case where we consider the whole boundary displacements, it has the advantage of having better knowledge of the effect of each particular crack on the boundary displacement field.

The last function takes into consideration the advantages of both functions, and it is assumed in this case, that the fitness of wrong crack identities will be amplified due to having the error considered twice. Tables 1, 2 and 3 show the identification errors in terms of crack size and position, the error is calculated as the Euclidian distance between the real crack centre and the centre of the predicted crack. The first crack is C1 with centre coordinates (–1.7, –9.5) and a size of 3.2 mm, the C2 centre is at (1.6, 3.8) and a size of 3.1 mm. C3 is positioned at (–0.15, 0.15) and has a size of 4.6 mm.

Table 1. Identification results based on the first fitness function (Eq. 10).
Table 2. Identification results based on the second fitness function (Eq. 11).
Table 3. Identification results based on the third fitness function (Eq. 12).

The crack identification error varies between 0.22 mm and 1.6 mm in terms of position, and between 0.05 and 1.07 mm in terms of crack size. The results suggest that, overall, the third fitness function corresponds to the highest identification accuracy for both crack size and location. Figure 8 Shows this advantage in crack size prediction. It also shows that Deep ANN is favourable in most cases compared to RBF modelling. This is due to the higher boundary displacement accuracy. However, the fitness evaluation using Deep ANN is very costly compared to RBF. Figure 9 depicts the difference between Deep ANN and RBF in the computational cost of a single identification run using the YUKI algorithm, with a population of 20 and the maximum number of iterations is 100.

Fig. 9.
figure 9

Computation cost using Deep ANN and RBF.

8 Conclusion

In this paper, damage identification in steel plates is presented, using the inverse problem approach. The approach uses the boundary displacement response measured in the presence of an inner crack with different sizes and positions. The finite element method is used to simulate these boundary displacement responses. The inverse problem requires an iterative search with thousands of response evaluations, crack identification researchers investigate alternative solutions to FEM simulation, due to the very expensive computational cost. As well as suggest better-performing search algorithms that require a lower number of iterations.

The performances of Deep ANN and RBF networks for the inverse problem evaluation have been investigated in this work. Using the newly suggested YUKI algorithm. We found that Deep ANN is favourably placed to simulate the boundary displacement response, as it has good stability to unknown crack parameters compared to the RBF method, providing a more consistent prediction error. However, the computation time for RBF is significantly better, which may justify the use of this method.

Boundaries most close to damage are subjected to higher displacement than farther away boundaries. The use of the whole boundary displacement sensors is not favourable for the objective function, because that does not consider the above-mentioned fact, we investigated the segmentation of the boundaries and use cost functions that take into consideration the regional effect of the cracks, and found that the best results are found when considering the average error in all segment plus error in the segment corresponding to the minimum error. This approach will amplify the error in the wrong output, which helps reduce the ill-Posedness of this problem.