Keywords

3.1 Introduction

Finite element methods (FEMs) are the common tools for engineering analysis [1,2,3]. These approaches are adopted for different physical disciplines, including structural dynamics, heat transfer, fluid flow, and electromagnetic potential. In structural engineering, FEMs are sufficient to predict the response of simple structural system. In contrast, the reliability of the FEM results might be decreased when dealing with complex structural systems. Therefore, the finite element model updating (FEMU) techniques are employed to minimize the dissimilarities between the FEM solutions and the real structural data [4,5,6,7]. The FEMU methods are classified in two main classes: (i) direct updating approaches and (ii) indirect updating approaches. Direct updating methods are computationally efficient since they only use one updating step to achieve the structural updating goal. In these methods, the system properties (i.e., mass and stiffness matrices) are directly equated to the structural measured mode shapes. However, the dimensionality of the acquired solutions by the FE analysis is normally larger than the measured data. Thus, direct updating methods may produce unrealistic physical components. On the other hand, the indirect updating methods, also refer to iterative techniques, are guaranteed to produce physically realistic updating results and to increase the correlations between FE solutions and the actual experimental data.

The Bayesian framework is well-known for quantifying the uncertainties that are associated with the system parameters. In Bayesian theory, the uncertainties of the modeled parameters are defined through the posterior probability density function (PDF). The function defines the updating parameters as random variables. Unfortunately, the analytical solutions for this function cannot be attained. That is due to the complexity and the high-dimensionally of this function. Alternatively, sampling methods are used to approximate solutions for complex distribution. Among these methods, the Markov chain Monte Carlo (MCMC) [8, 9], represented by the Metropolis-Hastings (M-H) algorithm [10, 11], is the most known approach to draw samples from complex probability distributions. The generated samples (proposals or solutions) are then used to approximate solutions for the posterior PDF, thereby updating the required FEM. However, MCMC methods are limited when several updating parameters are involved, as well as when the structure under consideration is sufficiently complex. For that reason, further updating methods are advised to merge with the MCMC procedure to improve the ability of drawing new proposals.

The evolutionary algorithms (EAs) are well recognized natural-inspired methods used for optimization problems [12,13,14]. Generally, the evolution-based search algorithm utilizes the availability of the current generation, known as solution or population, to produce a potentially more accurate solution. The interacting population at each step (iteration) is modified through specific operators to produce the next generation. Each individual solution is evaluated by a fitness function. This function is defined as the objective function of the optimization problem. The estimation of the function reflects the solution which the individual provides. EAs are promising techniques for the optimization scope. The evolutionary paradigm is implemented through different classes and for wide range of application.

In this paper, a shuffled complex evolution Markov chain algorithm is used to update a 2-D frame structure with measured data. The algorithm uses a population of solution and evolves through shuffling and partitioning into complexes. Next, the algorithm applies the metropolis acceptance rule to accept or reject the proposed solutions. As a result, this combination of the two procedures generates more accurate updating parameters. The following section explains the Bayesian inference for FEMU problem. Section 3.3 details the implementation of shuffled complex evolution Markov chain. In addition, Sect. 3.3 shows the implementation of the algorithm to sample from the posterior PDF of a Bayesian approach. Section 3.4 shows the application of the algorithm to update a 2-D frame structure with real data. Finally, the paper is concluded in Sect. 3.5.

3.2 Bayesian Inference

The Bayesian approach defines the uncertain parameters as random vector. Each parameter is determined as a stochastic variable as given by Bays’ rule [3, 15,16,17]:

$$ P\left(\left.\boldsymbol{\theta} \right|\mathcal{D},\mathcal{M}\right)\propto P\left(\left.\mathcal{D}\right|\boldsymbol{\theta}, \mathcal{M}\right)\ P\left(\left.\boldsymbol{\theta} \right|\mathcal{M}\right) $$
(3.1)

θ is the vector of updating parameters \( \boldsymbol{\theta} \in \boldsymbol{\Theta} \subset {\mathcal{R}}^d \). \( \mathcal{D} \) is the experimental data of the structural system, represented by the natural frequencies \( {f}_i^m \) and mode shapes \( {\boldsymbol{\phi}}_i^m \). \( \mathcal{M} \) represents the model class of the system, where each model class is defined by a different set of updating parameters. \( P\left(\left.\boldsymbol{\theta} \right|\mathcal{M}\right) \) refers to the prior PDF that offers the previous knowledge of the uncertain parameters θ given \( \mathcal{M} \) and with the absence of \( \mathcal{D} \). \( P\left(\left.\mathcal{D}\right|\boldsymbol{\theta}, \mathcal{M}\right) \) is the likelihood function that describes the difference between the measured data and the analytical solutions. \( P\left(\left.\boldsymbol{\theta} \right|\mathcal{D},\mathcal{M}\right) \) is the posterior PDF of the updating parameters given a model class \( \mathcal{M} \) and the measured data \( \mathcal{D} \). Since only one model class is considered by this paper, \( \mathcal{M} \) is removed for simplicity.

The likelihood function is given by:

$$ P\left(\left.\mathcal{D}\right|\boldsymbol{\theta} \right)=\frac{1}{{\left(\frac{2\pi }{\beta_c}\right)}^{N_m/2}{\prod}_{i=1}^{N_m}{f}_i^m}\exp \left(-\frac{\beta_c}{2}\ {\sum}_i^{N_m}{\left(\frac{f_i^m-{f}_i}{f_i^m}\right)}^2\right) $$
(3.2)

N m is the number of measured modes, and β c is an arbitrary constant. \( {f}_i^m \) and f i are the ith measured and analytical natural frequencies.

The prior P(θ), that provides the initial knowledge of the updating parameter θ, is defined by a Gaussian probability distribution:

$$ P\left(\boldsymbol{\theta} \right)=\frac{1}{{\left(2\pi \right)}^{Q/2}{\prod}_{i=1}^Q\frac{1}{\sqrt{\alpha_i}}}\exp \left(-{\sum}_i^Q\frac{\alpha_i}{2}{\left\Vert {\theta}^i-\left.{\theta}_0^i\right\Vert \right.}^2\right)=\frac{1}{{\left(2\pi \right)}^{Q/2}{\prod}_{i=1}^Q\frac{1}{\sqrt{\alpha_i}}}\exp \left(-\frac{1}{2}{\left(\boldsymbol{\theta} -{\boldsymbol{\theta}}_0\right)}^T\ {\Sigma}^{-1}\left(\boldsymbol{\theta} -{\boldsymbol{\theta}}_0\right)\right) $$
(3.3)

where Q is the number of the uncertain parameters, θ 0 represents the mean value of the uncertain parameters, α i is the coefficient of uncertain parameters, i = 1, …, Q, and the Euclidean norm of ∗ is noted by ‖∗‖.

The posterior PDF \( P\left(\left.\boldsymbol{\theta} \right|\mathcal{D}\right) \) of the updating parameters θ given the measured data \( \mathcal{D} \) can be now described by substituting Eqs. (3.2) and (3.3) into Eq. (3.1):

$$ P\left(\left.\boldsymbol{\theta} \right|\mathcal{D}\right)\propto \frac{1}{Z_s\left(\alpha, {\beta}_c\right)}\exp \left(-\frac{\beta_c}{2}\ {\sum}_i^{N_m}{\left(\frac{f_i^m-{f}_i}{f_i^m}\right)}^2-{\sum}_i^Q\frac{\alpha_i}{2}{\left\Vert {\theta}^i-\left.{\theta}_0^i\right\Vert \right.}^2\right) $$
(3.4)

where,

$$ {Z}_s\left(\alpha, {\beta}_c\right)={\left(\frac{2\pi }{\beta_c}\right)}^{N_m/2}{\prod}_{i=1}^{N_m}{f}_i^m{\left(2\pi \right)}^{Q/2}{\prod}_{i=1}^Q\frac{1}{\sqrt{\alpha_i}} $$
(3.5)

Equation (3.4) represents the final form of the posterior PDF. The same function is employed as the objective function for evolutionary procedure. The dimensionality of the unknown parameters along with the complexity of the uncertain search bounds may affect the accuracy of the solutions. However, in the next section, the shuffled complex evolution Markov chain method is explained and detailed to solve the Bayesian FEMU problem.

3.3 Shuffled Complex Evolution Markov Chain Algorithm

The shuffled complex evolution Markov chain (SCE-MC) [18, 19] is a population-based algorithm for global optimization problems. It aims to infer the target function by combining the advantages of the Metropolis-Hastings sampler with the shuffled complex evolution (SCE) approach [20, 21]. The algorithm applies the controlled random search with competitive evolution and complex shuffling to continuously update the proposal distribution. In addition, the SCE-MC’s evolution procedure evolves the proposed samples to the posterior PDF through sequence of iterations. As a result, the SCE-MC method has the ability to improve the best parameter set found in the search space every iteration. Thus, the population (Markov chains) is continuously evolved toward the optimal region in the search. The main advantage of this algorithm, in comparison to the standard SCE method, is to avoid trapping into a region with lower posterior density, which prevents most of the population to converge toward a single mode or to collapse into relatively small region of a single best parameter set (local minimum) (Fig. 3.1).

Fig. 3.1
figure 1

Flow chart of the SCE-MC algorithm

In addition, the SCE-MC algorithm uses a large number of population, N, which offers the ability for wide exploration of the parameters space, thereby raising the chance to find the global optimum of the parameter set in few iterations. Usually, a population size N = 10d is sufficient to approximate a complex target distribution. The parallel sequence provides a mechanism for a simultaneous different region exploration. Thus, the number of j (number of parallel sequencing) enables the algorithm to seek several regions of attraction and allows to apply the heuristic test to check the convergences. A default set j = d is enough for wide exploration. The user can define a greater number of sequences as the complexity of the posterior function increased. Partition into complexes brings the possibility for each individual (Markov chain) sequence to collect more knowledge about the search space during the evolution step. The shuffling process for these complexes assists the surviving for the sequence through a global sharing of the information which is already collected by each parallel sequence. The implementation of this process produces a robust sampling procedure that uses the MCMC principles for global optimum solutions.

The SCE-MC algorithm uses the Sequence Evolution Metropolis (SEM) procedure to produce new candidates (samples) in each parallel sequence \( {\theta}_z^S \). This can be done by creating draws from an adaptive proposal distribution through the use of the information induced in m chains of ∁z. This paper applies the default SEM procedure as explained in [18]. Another important feature of the SCE-MC algorithm is the convergence to the stationery posterior distribution. The algorithm applies the Gelman-Rubin rule [18] to estimate the convergence status. This rule is to determine the scale reduction \( \sqrt{SR} \) between the chains (sequences) and the variances as given by:

$$ \sqrt{SR}=\sqrt{\frac{t-1}{t}+\frac{j+1}{j.t}\frac{V}{U}} $$
(3.6)

where t is the number within each sequence, V is the variance between the j sequence means, and U is the average of the j within sequence variance for the parameters under consideration, respectively. Note that the product of j and t is identical to the total number of derived samples (number of algorithmic iterations T). The convergence of each parameter occurs when \( \sqrt{SR} \) is close to 1. However, since this score is difficult to achieve, Gelman-Rubin suggested using value less than 1.2 to declare convergence to a stationary distribution. On the other hand, the algorithm can work through pre-defined number of iterations, which can be used to compare two updating procedures at the same number of iterations.

The SCE-MC algorithm for Bayesian FEMU is summarized as follows:

  1. 1.

    Create population: generate N number of chains {θ 1, θ 2, …, θ N} randomly from the prior distribution P(θ i).

  2. 2.

    Compute posterior: calculate the posterior density for each generated chain {P(θ 1| D), P(θ 2| D), …, P(θ N| D)}.

  3. 3.

    Descent ranking: sort N chains in order of decreasing the posterior density and store them in matrix \( \mathcal{R}\ \left[1:N,1:d+1\right] \), where d is the number of the updating parameters and the first row of \( \mathcal{R} \) represents the chain with highest posterior density.

  4. 4.

    Initialize parallel sequence: start the parallel sequencing of the chains, \( {\boldsymbol{\theta}}_1^S \), \( {\boldsymbol{\theta}}_2^S \), …, \( {\boldsymbol{\theta}}_j^S \), such that \( {\boldsymbol{\theta}}_z^S \) is \( \mathcal{R}\ \left[z,1:d+1\right] \), where z = 1, 2, …, j.

  5. 5.

    Complex partitioning: partition the N chains of the matrix \( \mathcal{R} \) into z complexes ∁1, ∁2, …, ∁z, where each complex ∁ contains k chains, such that the first complex includes j(i − 1) + 1 ranked chains, the second complex includes every j(i − 1) + 2 ranked chains of the matrix \( \mathcal{R} \), and so on, where i = 1, 2, …, k.

  6. 6.

    Sequence evolving: evolve each of the parallel sequences according to the Sequence Evolution Metropolis procedure.

  7. 7.

    Shuffle complexes: unpack all complexes ∁ back into \( \mathcal{R} \), then rank the chains in order of decreasing posterior density and reshuffle the N chains into j complexes according to the procedure specified in step 5.

  8. 8.

    Termination: check the Gelman-Rubin convergence criteria or use maximum number of iterations. If any of these conditions is satisfied, stop. Otherwise return to step 6.

3.4 Application: Updating the Young’s Modulus

A 2-D steel frame structure, as shown by Fig. 3.2, is used to demonstrate the performance of the SCE-MC algorithm for Bayesian FE model updating. The structure is fixed to the ground from its bottom side [22]. The overall height of the frame is 1.5 m and the span of the beam is 0.5 m. Each story of the structure has the same height of 0.5 m. The structure is assembled with five beams; two vertical beams, making up the right and the left side of the structure, are noted as Beam-1and Beam-2, respectively. The other three are horizontal beams noted as Beam-3 at the bottom side, Beam-4 in the middle, and Beam-5 at the top. All the beams have the same cross-sectional dimensions: their width is 75 mm, and their thickness is 5 mm. The material density of the frame is 7850 kg/m3, and the elastic modulus is 2.0 × 1011N/m2.

Fig. 3.2
figure 2

the 2-D frame structure

An instrumental hammer with a rubber tip was used to excite the frame, where the frequency range of interest is about 0–100 Hz. Aaccelerometers with magnet bases are mounted on the vertical and horizontal beams to measure the response. The first five natural frequencies and mode shapes in the range of 0 – 100 Hz were extracted. The measured natural frequencies are ω m = {4.23, 14.03, 25.45, 44.81, 58.12} Hz. the FEMU application was implemented through the MATLAB environment [23], and the FE model was developed using the structural dynamic toolbox [24]. A 2-D geometrical beam element is used to simulate the frame model. The obtained analytical frequencies with the mode shapes are illustrated in Fig. 3.3.

Fig. 3.3
figure 3

The analytical frequencies and the mode shapes of the 2-D frame structure

In this application, the modulus of elasticity of the five structural beams (Beam-1 to Beam-5) is selected as the updating parameters. The updating vector is θ = {E m1, E m2, E m3, E m4, E m5}, where E m1 represents the elastic modulus of Beam-1 and so on. Then, θ = {2 × 1011, 2 × 1011, 2 × 1011, 2 × 1011, 2 × 1011}. The updating parameters are bounded by θ max and θ min to keep the updated values realistic, where θ max = {2.4 × 1011, 2.4 × 1011, 2.4 × 1011, 2.4 × 1011, 2.4 × 1011} and θ min = {1.6 × 1011, 1.6 × 1011, 1.6 × 1011, 1.6 × 1011, 1.6 × 1011}. The variance vector is σ 2 = {5 × 1011, 5 × 1011, 5 × 1011,  5 × 1011, 5 × 1011}, and the total number of iterations is t s = 500. Meanwhile, the number of complexes z in the SCE-MC algorithm is set to 5 and the population size N = 50. On the other hand, the M-H method is set as its default settings with increasing the size of its move.

In this application, the SCE-MC algorithm and the canonical M-H method are applied to update the same structure, this aims to compare the two algorithms and to highlight the improvements gained when combining the M-H criterion with the evolutionary mechanism. Both algorithms were tested to update the frame structure for several independent runs. Each updating trial was set to the same algorithmic settings and parameters. Table 3.1 presents the updated parameters by the SCE-MC algorithm and the M-H method. The initial values of the elastic modulus are given in the table for the five updating elements. The final updated values as shown in the table are different to the initial states. This shows that both algorithms produced realistic updated parameters that keep physical meaning for the updating feature. The table also includes the coefficient of variation (cov) of the updated parameters from each algorithm. This coefficient measures the level of precision within the samples produced by the updating methods. The cov is defined as the ratio of the standard deviation Α to the mean value 𝜃𝑖 (the value of the final updated parameter), i.e., cov\( =\frac{\sigma_i}{\theta_i} \).

Table 3.1 The updated elastic modulus of the 2-D frame structure using the M-H and the SCE-MC

Generally, the cov values obtained by the SCE-Mc algorithms are better than the M-H method. The sampling precision that happens in the SCE-MC procedure is due to the use of large number of population. Thus, the selected solutions reflect the best results among the other solutions. Contrary, the M-H applies a single move only to propose the new solutions state. However, the cov of the still reflect a quiet acceptable range.

Figure 3.4 illustrates the boxplots of the generated samples by SCE-MC algorithm. On each box, the central dark blue line represents the median, and the light blue box spans indicate the 25th and 75th percentile (the range that bounds 25% to 75% of samples). The whiskers (dark blue horizontal lines) extend to the most extreme samples position (from 1% to 99% of samples are placed within this range), and red dot symbol denotes the outliers. This figure indicates the sampling performance of the SCE-MC algorithm. The convergence toward the high region of solutions can be seen from the figure as 25~75% of the samples are concentrated around the mean (a small dark blue square at the central). The outliers are expected to appear, since the algorithm uses a population size of 50 Markov chains, which creates a challenge to converge all set of chains simultaneously.

Fig. 3.4
figure 4

Boxplot of the random samples by the SCE-MC algorithm

The correlations between the updated parameters are shown in Fig. 3.5. A small rate correlation demonstrates that both updating parameters are weakly correlated (<0.3), while a larger rate means a highly correlation (>0.7). The positive or negative correlation indicates that the variables relatively affect each other. The SCE-MC algorithm has successfully sampled from the posterior PDF since no highly absolute correlation is observed.

Fig. 3.5
figure 5

The correlation between the updating parameters obtained by SCE-MC algorithm

The updated natural frequencies are given in Table 3.2. The results indicate that the SCE-MC updating method has updated the natural frequencies for the five modes with outcomes that are better than the initial FE model. The percentage absolute mode error \( \frac{\left|{f}_i^m-{f}_i\right|}{f_i^m} \) is included in the table to compare the differences between the measured and the updated frequencies. The table also presents the total average error (TAE), which gives a general overview of the updated modes, and is determined due \( TAE=\frac{1}{N_{\omega }}{\sum}_{i=1}^{N_{\omega }}\frac{\left|{f}_i^m-{f}_i\right|}{f_i^m} \), where N ω is the number of the updated frequencies. The SCE-MC has successfully minimized the TAE from its initial state to 2.67%. On the other hand, the M-H algorithms has completed the updating process with 3.8% for the TAE. It can be shown here the SCE-MC has the advanced ability to update the frame structure and to produce efficient updating solutions.

Figure 3.6 shows the convergence of the TAE for the SCE-MC and the M-H algorithms when updating the natural frequencies of the 2-D frame structure. The SCE-MC algorithm utilizes 50 Markov chains and 5 evolutionary complexes to implement the updating process. Thereby, the algorithm tests and shuffles a number of 50 solutions at each iteration. This mechanism provides the algorithm with the ability to reach a quick convergence. It shows in the plot that the algorithm started with TAE which is higher than the TAE of initial FE model. This indicates the effectiveness of the evolving process which is implemented by the algorithm to explore a wider region of the search space and iteratively select the best set of solutions. The convergence of the SCE-MC method, as seen in the figure, is achieved before the first 100 iterations.

Table 3.2 The updated natural frequencies using M-H and the SCE-MC algorithms
Fig. 3.6
figure 6

Convergence of the total average error of the 2-D frame structure using the SCE-MC algorithms

For the purpose of comparing the results of both updating methods, the M-H and the SEC-MC methods, the M-H algorithm is set to have a relatively large jump at iteration, which facilitates the M-H procedure to have its maximum performance and quickest convergence within 500 iterations only for the FEMU application. This is used to highlight the limitations for each updating method when used to solve the same updating conditions.

3.5 Conclusion

This paper has introduced the shuffled complex evolution Markov chain algorithm for Bayesian updating problem. The presented algorithm merges the advantages of the Metropolis-Hastings method, controlled stochastic search, competitive evolution, and complex shuffling to evolve population of samples to approximate the posterior PDF of the Bayesian model. The algorithm has been tested to update a FE model of a 2-D frame structure. The Young’s modulus is chosen to be updated. Five updating elements are considered for the updating problem. In addition, the posterior function includes five natural frequencies which are collected from modal experiment. It is found that the SCE-MC algorithm requires large number of population to propose desired samples and that is correlated with the complexity of the posterior function. The algorithm is also affected by the number of the complexes used to partition and sequencing the population. Increasing the number of the complexes offers more opportunities to rank and sequence the solutions, therefore giving the chance to converge to a higher region of solutions. However, this may raise the computational demand of the problem.

The updated parameters have reflected realistic updating values. The obtained set of solutions has minimized the total average error of the updated natural frequencies. The obtained results were compared with the outcomes of the M-H algorithm and found that the performance of the SCE-MC method is more efficient. It’s advised for future work to investigate the SCE-MC algorithm with another evolutionary Markov chain method.