Abstract
Knowing the thermo-mechanical history of a part during its processing is essential to master the final properties of the product. During forming processes, several parameters can affect it. The development of a surrogate model makes it possible to access history in real time without having to resort to a numerical simulation. We restrict ourselves in this study to the cooling phase of the casting process. The thermal problem has been formulated taking into account the metal as well as the mould. Physical constants such as latent heat, conductivities and heat transfer coefficients has been kept constant. The problem has been parametrized by the coolant temperatures in five different cooling channels. To establish the offline model, multiple simulations are performed based on well-chosen combinations of parameters. The space-time solution of the thermal problem has been solved parametrically. In this work we propose a strategy based on the solution decomposition in space, time, and parameter modes. By applying a machine learning strategy, one should be able to produce modes of the parametric space for new sets of parameters. The machine learning strategy uses either random forest or polynomial fitting regressors. The reconstruction of the thermal solution can then be done using those modes obtained from the parametric space, with the same spatial and temporal basis previously established. This rationale is further extended to establish a model for the ignored part of the physics, in order to describe experimental measures. We present a strategy that makes it possible to calculate this ignorance using the same spatio-temporal basis obtained during the implementation of the numerical model, enabling the efficient construction of processing hybrid twins.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Introduction
Metal Casting is one of the oldest materials forming technique which is widely employed in industrial environment. It enables manufacturing complex shaped parts with high productivity and less raw consumption [1]. Gravity casting is the simplest form of casting that consists of pouring molten alloy into a mould cavity with no force other than gravity, where it cools and solidifies. The mould can be made of sand, metal or some other materials [2]. The permanent mould casting is a process that uses a metal mould namely tool steel [3], iron and bronze [4]. The most stringent requirement on permanent moulds is their cooling ability. They are characterized by high thermal conductivity which allows to increase the rate of heat transfer and to reduce the solidification time. As a consequence, the produced cast parts present better dimensional tolerances, superior surfaces finishing, and higher mechanical properties [3, 4]. In industry, the most up-to-date application of permanent mould casting is the aluminium alloys due to their excellent properties such as excellent cast ability [5], high electrical and thermal conductivity [6,7,8], low density, low weight and high strength to weight ratio [8, 9].
To ensure high quality casting products, the casting stages need to be well controlled starting by mould preparation, alloy melting, pouring, and finally solidification process. Inaccurate supervision at these stages leads to casting defects [10]. The cooling stage has a significant effect on the microstructure of the cast parts, which means on their mechanical properties. It is necessary to understand the heat transfer process inside the mould to ensure the required mechanical properties in the casting. The heat released during the solidification is transferred within the mould by conduction. Once the heat reaches the mould walls, it is transferred to the air essentially by natural convection [11]. It is well known that for aluminium alloys, the cooling rate directly affects the microstructure morphology and the size of the grains: Raising up the cooling rate during the casting can significantly refine the microstructure and thereby improve the mechanical properties of produced parts [12,13,14]. The material and geometry of permanent metal mould contribute on the heat transfer process, that is, on the casting cooling [11]. In permanent mould casting, it is highly recommended to have homogeneous distribution of temperature in the mould. Non-uniform cooling causes defects in the cast parts such as low residual stress, hot spots and distortions in the form [15]. These casting defects could be reduced by using “cooling channels" moulds. They date back to 1990 and were initially suggested for injection moulding [16, 17] and then were extended to others fields such as extrusion [18], hot sheet metal forming [19], forging [20], and die casting [15, 21, 22]. Karakoc et al. showed, in references [22] and [15], that the porosity in the cast parts was reduced by \(43\%\) and the average particle size of the cast parts was \(13.5\%\) smaller than those parts obtained with standard moulds. Both of these studies were carried out through experimental methods and numerical simulations. Norwood et al. were also used the simulation tools to optimize the design of cooling channels to ensure a high product quality and minimize production costs [21].
Today, numerical simulations are widely used in casting optimization process. However, for an optimized casting configuration, the simulations analyses were generally based on a particular set of parameters. In addition, the requirement of very accurate and reliability data increases significantly the calculation time of numerical simulations that means the computing coasts. Thus, the numerical simulations in casting process is still interesting through the use of artificial intelligence. It is hence possible nowdays in metal casting processes to applied powerful tools and models developed with reasonable number of simulations that allows predicting parts defects and controlling complex processes [23,24,25]. For example, Jiang et al. used back propagation neural network models to establish a relationship between the continuous casting parameters and the cooling rate which was based on secondary dendrite arm spacing compute [26]. They showed that this model has a higher accuracy in the optimization of the continuous casting technology. Others researchers used also artificial neural networks and they were more focused on the cooling-solidification process and the heat transfer coefficient as well [27, 28]. Susac et al. applied artificial neural network to predict the thermal field of permanent mould based on the thermal history of the aluminium cast parts [28]. Vasileiou et al. proposed a genetic optimization algorithm aided with numerical simulations to determine the heat transfer values in casting [29]. However, for every new casting change in material and/or in shape, this approach should repeat again. Researchers tried to developpe interesting approaches for thermal field evolution in the cast and in the mould as well. Despite these efforts, most of the proposed approach are limited to the cast part design, casting process parameters, and also to the number of input parameters. The present work proposes a new approach combining physics-based reduced order models, enabling parametric studies, and data-driven model enrichment in the so-called hybrid modelling framework, enabling the highest accuracy with respect to the experimental measurements, while proceeding under the stringent real-time constraint.
Empowering engineering from the use of surrogates
Efficient design and system control are needed for quick evaluations of the system response for any choice of the parameters involved in the associated model. Usual numerical simulation techniques remain unable to provide results under the stringent real-time constraints imposed by control.
Parametric models, also called surrogates, metamodels or response surfaces, make it possible to attain such feedback rates. Then, on top of these surrogates, simulation, optimization, uncertainty propagation or simulation-based control become attainable even under the stringent real-time constraint. Thus, the challenge of developing efficient simulations is translated into the one of an efficient construction of such surrogates, that is far from being a trivial task.
In fact, if one assumes a multivalued input \(\textbf{X}\) and an associated multivalued output \(\textbf{Y}\), the surrogate is no more than the mapping \(\textbf{Y}=\textbf{F}(\textbf{X})\), where \(\textbf{F}(\textbf{X})\) constitutes the searched model, that in general consists of a linear or a nonlinear regression.
Constructing a regression is not difficult, conceptually speaking. However, the amount of data needed for this purpose strongly depends on the model complexity.
Since complexity will depend on the dimension of the data (number of features involved in \(\textbf{X}\)) and variables to model (size of \(\textbf{Y}\)), one is tempted to proceed reducing the data dimensionality prior to create the regression.
Data dimensionality reduction can be performed by using a linear reduction—for instance by employing principal component analysis, PCA—or a nonlinear one, making use of manifold learning techniques, for instance, or in a more transparent manner, by training autoencoders able to map the data into a reduced latent space.
Usually in the case of engineering, and more particularly in casting process simulation, where the temperature field is expected to depend on few process parameters (like in this paper the temperature of the fluid flowing into the so-called cooling channels disposed in the mould) we look to infer 3D fields from few features. Thus we firstly need to reduce computer memory storage space and enable real-time predictions for temperature field. Then we can move to creating a regression (linear or non-linear) between the features and the reduced description of the temperature field.
In turn, this regression can be linear (even when non-linear approximation functions are involved) or non-linear. Polynomial linear regressions are very usual, and they were adapted to address multidimensional problems by making use of separated representations [30, 31].
Regularization allows us to address rich approximations while keeping the amount of data to a minimum [32]. These situations result, in general, in underdetermined linear systems, that need for appropriate regularizations to avoid overfitting. Elastic Net regularizations combining the Ridge L2-regularization, that prevent overfitting, and the Lasso L1-regularization, that promotes sparsity, are widely adopted [33].
When the amount of data is large enough and it is expected to be distributed on a nonlinear manifold, artificial neural networks, ANN, [34] become an appealing choice.
Filling the gap between knowledge and observations: the hybrid twin
A particular situation occurs when physics is solved very efficiently by employing surrogates, whose construction has just been addressed, but a significant gap between the predictions and the observations is noticed. Such a gap reflects the limitations of the considered model, that can be inaccurate or incomplete with respect to the addressed physics. In this situation, two direct alternatives exist: (i) refine the physics-based model to improve the prediction performance; or (ii) correct (or enrich) the physics-based model by adding a data-driven model of the observed gap—something that we refer to as modelling the ignorance (i.e., the gap between measures and simulations). This second route is at the origin of the so-called hybrid-twin concept, addressed in our recent works [35,36,37,38,39].
The main advantage of this augmented framework is twofold. First it offers the possibility of explaining the (usually) most important part of the resulting hybrid (or augmented) model: the one concerning the physics-based contribution. Second, with a deviation less nonlinear that in the case of the observed reality (the physics-based model contains an important part of such nonlinearity), less data becomes sufficient for constructing the data-driven model.
Methods
This section revisits usual surrogate constructors that make use of separated representations, and proposes an appealing alternative that overcomes these. Our objective in this study is to elaborate a parametric solution with a representation compatible with the use of machine learning techniques, so as to enable the prediction of new scenarios associated with arbitrary parameter choices.
A space-time and parameters separated representation
We consider a field T defined in the physical domain, \(\textbf{x} \in \varOmega _x \subset \mathbb {R}^D, \ D=2,3\). This field evolves in time \(t \in [0, +\infty ) \). Our problem depends on a set of parameters \(\textbf{p} = p_1, p_2, \ldots , p_n, \textbf{p} \in \varOmega _p \subset \mathbb {R}^n\).
It is assumed that a design of experiment makes it possible to obtain the evolution of the field T, in space and time, for several combinations of parameters \(\textbf{p}\). Our solution is then written in a general form \(T(\textbf{x}, t, \textbf{p})\).
The representation of this solution, specifically according to the parametric dimension, is discrete. Artificial intelligence plays the role of interpolating (or extrapolating) from the set of parameters already considered in the training stage.
In those approaches we have developed so far, we used a non-intrusive dimensionality reduction that expresses the solution from a finite sum of products of functions. For this purpose, we rely on the singular value decomposition. In order to apply this singular value decomposition approach, we need to operate on a discrete representation of the field T. In its classical form, the singular value decomposition decomposes the field into sums of tensor products of two discretized functions. A simple choice is to consider space and time on the one hand, and parameter space on the other. The reader can refer to [40] to see an example of the application of this approach.
The continuous form reads:
This form corresponds to a discrete form which could be written with the index notation as
where the subscripts i and j, refer here to the degrees of freedom in space, time and parameter dimensions, respectively.
The previous form can be rewritten in the tensor form
The determination of this form can be made in a direct way, by using a classical calculation based on the eigenvalue decomposition. However, in what follows, we use an iteration procedure, easily generalizable later to more dimensions, the so-called high-order singular value decomposition.
To find the series \((\textbf{F}^1, \textbf{H}^1), (\textbf{F}^2, \textbf{H}^2), \ldots \) we assume that the solution at iteration \(k-1\) is known and given by
where \(\tilde{\mathbb {T}}\) represents the field discrete approximation.
The difference between the initial discrete field and the approximated one is noted by \(\mathbb {T}' = \mathbb {T} - \tilde{\mathbb {T}}\), which represents the approximation residual. The associated iteration algorithm solves:
until the convergence (fixed point) is reached.
The enrichment stops when the norm of the residual \(\mathbb {T}'\) becomes lower than a tolerance criterion, fixed by the user. Here we assume that the enrichment process stops after K couples have been computed.
It follows that vectors \(\textbf{H}^k\) contains the parameter weights at each considered choice of parameter \(\textbf{p}\), \(\textbf{p}^j\). Thus, the component \(\textbf{H}^k_j\), \(k=1, \ldots , K\), is related to \(\textbf{p}^j=(p^j_1, \ldots ,p^j_n)\).
Thus, one is tempted to train, from the available data couples \((\textbf{p}^j,\textbf{H}^k_j)\), an AI-based regression to evaluate the scalar \(H^k\), \(\forall k\), for any other value of \(\textbf{p}_{\text {new}}\), noted by \(H^k(\textbf{p}_{\text {new}})\), from which the reconstructed espace-time solution \(\textbf{T}_{xt}(\textbf{p}_{\text {new}})\) reads
Separating space and time
The approach that we have just presented fails to address problems with too many degrees of freedom in space and time. It is therefore useful to separate the temporal dimension from the spatial one (see [41] for an example).
The simplest option consists of performing a singular value decomposition in space and time for each solution associated to the parameters choice \(\textbf{p}^h=(p_1^h, p_2^h, \ldots )\), \(h=1, \ldots ,H\). This SVD allows us to write
whose discrete form reads
and, in tensor form
This expression does not allow us to build a response surface on the parametric space. To this end, we must express our different functions in a common approximation basis. To avoid redundancies between the different functions \(^h\textbf{F}^k\) and \(^h \textbf{G}^k\), obtained during the performed simulations for different parameter choices, and in order to guarantee the orthogonality of the basis, a proper orthogonal decomposition is performed.
Let \(\mathbb {Q}\) be the matrix composed by the functions \(^h\textbf{F}^k\) for different parameters choices \(\textbf{p}^h\)
The resulting orthonormal eigenvectors are noted as \(\textbf{B}_1, \textbf{B}_2, \ldots \).
By selecting the r eigenvectors associated with the r highest decomposition eigenvalues, the space approximation basis reads
To express the basis obtained for a set of parameters h into the global basis Eq. 12, we define the coordinates matrix \(^h\!\beta \) enabling the expression of \(^h\mathbb {F}=[^h\textbf{F}^1, \ldots , ^h\textbf{F}^K]\) into the common basis Eq. 12, according to
whose solution results from
The same rationale applies on the time vectors, leading to
Thus, finally, the approximation reads
or, by defining the new matrix \( ^h\!\alpha = ^h\!\beta (^h\!\gamma )^T \), it results
Artificial intelligence intervenes here to obtain each component of the matrix \(\alpha \), \(\alpha _{ij}(p_1, p_2, \ldots ,p_n)\) from the existing knowledge: \(^h\alpha _{ij}(p_1^h, p_2^h, \ldots , p_n^h)\), \(h=1, \ldots ,H\).
The major drawback of such an approach is that the \(\alpha \) coordinate matrix is not diagonal. This leads to a very large number of \(\alpha _{ij}\) values involved in the training process. In addition, the numerous projections into the common truncated POD approximation basis introduce an additional error.
The proposed approach: a space, time and parameter separated representation
In this section we propose an approach that combines the advantages of the two procedures just described.
This approach relies on a high-order singular value decomposition involving three functions:
whose discrete form reads
Following the rationale previously introduced, the approximation is obtained by successive enrichments (until obtaining the desired accuracy at \(k=K\)) and at each enrichment step k iterating until convergence, that is, until attaining the fixed point, according to
Using the same rationale previously described, from the couples \((\textbf{p}^h,\textbf{H}_h^k)\), \(k=1,\ldots ,K\), a regression is constructed to infer the scalars \(H^k(\textbf{p}_{\text {new}})\), \(\forall k\), related to the parameters choice \(\textbf{p}_{\text {new}}\).
The reconstructed solution reads
The main steps of this methodology are summarized in Fig. 1
Case study
The problem here consists in the metal cooling that fills a mould during the casting process. The mould cavity is created using tool steel and endowed with cooling channels. The metal used to fill the cavity is an aluminium-silicon alloy.
We denote by \(\varOmega _1\) the domain filled by the metal and by \(\varOmega _2\) the mould, being \(\varGamma \) the interface between the metal and the mould. \(\varGamma _2\) represents the interface between the mould and the surrounding environment occupied by the air.
In the mould there are five cooling channels where the cooling liquid circulates. The interfaces between the mould and the cooling channels are denotes \(C_i, i=1,\ldots ,5\).
The thermal properties including the metal (with subscript 1) and the mould (with subscript 2) are given below (all quantities are expressed in the international units system):
-
The convection coefficient on \(\varGamma \) is denoted \(h_{12} = 500\) for the external boundary and 300 for the internal one.
-
The convection coefficient on \(\varGamma _2\) is denoted \(h_{\text {air}} = 20\).
-
The convection coefficient between the mould and the cooling liquid on \(C_i\), \(\forall i\), is denoted \(h_{c} = 10^4\).
-
The product of the density by the heat capacity for the part is \(\rho _1 C_{p_1} = 5.4 \ 10^6\).
-
The product of the density by the heat capacity for the mould is \(\rho _2 C_{p_2} = 1.5 \ 10^6\).
-
The conductivity of the metal is \(\lambda _1 = 70\).
-
The conductivity of the mould is \(\lambda _2 = 40\).
-
The air temperature outside the mould is \(T_{\text {air}}=20\).
The system of equations to solve during the time interval \([0, t_{max} = 300]\) is given by
for \((\textbf{x},t) \in \varOmega _1 \backslash \varGamma \times (0, t_{\text {max}}]\),
for \((\textbf{x},t) \in \varOmega _2 \backslash (\varGamma _2 \cup C_1 \ldots \cup C_5 ) \times (0, t_{\text {max}}]\).
These equations are subjected to the boundary conditions
where \(p_i\) refers to the temperature of the fluid circulating inside the channels and the superscripts \(\varGamma ^+\) and \(\varGamma ^-\) the two sides of the interface.
The variational formulation for the problem on \(\varOmega _1\) with a test field \( \varPsi ^*\) writes
This can be rewritten as
By skipping the details of the integration using the Galerkin approach with peace-wise linear functions the discrete system writes after simplification of the test field
In Eqs. 32 and 33 we have kept the same order of the different contributions so that the reader can make the correspondence between the different terms.
A similar approach for the domain \(\varOmega _2\) gives the following system
where the new terms \(\mathbb {E}_2\) and \(\textbf{J}_2\) account for the contributions of the convective heat transfer with air and with coolant.
The coupled system to be solved writes finally
In order to take into account the phase change latent heat for the metal, we use effective value of \((\rho _1 C_{p_1})_{\text {eff}}\):
The introduction of this relation to model latent heat effects comes from [42] and [43]. The idea consists to replace the constant value of \(\rho _1 C_{p_1}\) by an effective value that is augmented by a new curve in the form of a smoothed Dirac function. The area under this curve represents the latent heat and controlled by the parameter A. \(\delta \) is the phase change temperature range. It characterizes the global width of the curve. It is homogeneous to a temperature. \(T_\varphi \) is the temperature around which the phase change occur.
The numerical values considered in our study are \(A=3.3.10^8\), \(\delta = 1.1\), \(T_\varphi = 380\).
The simulation is done with an implicit approach in time and with a time step equal to 1 second in a time interval of 300 seconds.
The five variable parameters in this study are the temperatures of the fluid circulating in the five cooling channels. They will be denoted by \(\textbf{p} = p_1, \ldots , p_5\).
The domain of this study is presented in Fig. 2. The casted part has a width equal to 0.1 and a height equal to 0.06. The external dimensions of the mould are \(0.16 \times 0.12\). The computational mesh is represented in Fig. 3.
Figure 4 shows the thermal field on the mould and metal assembly, when our five parameters are uniformly set to 20. However, to better identify the distribution of temperature in each region, an exploded representation is given in Fig. 5. The initial conditions are such that temperature is equal to 500 degrees Celsius for the metal and 100 degrees Celsius for the mould. The illustrations of figures are after 300s cooling time.
Another illustration is shown in Figs. 6 and 7 where we deliberately unbalanced different temperatures in the cooling channels to see the consequence on the thermal distribution, in both, the part and in the mould.
Parametric surrogate
A design of experiments was generated with 200 simulations. Each of the simulations start from the initial temperature field described above, and the temperature evolution is calculated during 300 seconds.
From these 200 simulations, 150 were used in the model training, 30 were used for testing, and the remaining 20 will be used for validation purposes as discussed later.
These 200 configurations, consisting of different parameters choices \(\textbf{p}^h\), were generated using the Latin hypercube sampling. The interval in which the different parameters take their values is [0, 100].
Even if, during the simulations, we are interested in the thermal field in the global domain, part and mould, during the machine learning phase, we will focus only on the domain defined by the cavity because indeed our interest focus in controlling the evolution of the temperature in the part, which can affect its properties in service, from the level of residual stresses.
The temperature field in the domain defined by the cavity was then stored for the 300 iterations and for the different parameters, in a three-dimensional matrix description \(\mathbb T\). The application of the singular value decomposition on this matrix leads to different modes in space, in time, and in the parameters space.
Figure 8 shows the first four modes of the decomposition. The left column depicts the modes in space. The central column presents the time modes. Finally, the right column presents the parametric modes. In the figures of the right column, the order of the points is completely arbitrary. In order to simplify the visual representation, we represented only \(10\%\) of the points in the design of experiments, that is 18 over the 180 (training and test sets). On the x-axis each point represents a parameter data-point \(\textbf{p}^h\) (the five temperatures of the cooling fluid circulating in the five channels) and on the y-axis the associated value of function \(\textbf{H}_h^k\).
The relative norm of the residue represented in Fig. 9 proves that the first mode is the most relevant, and that 40 allows reducing the error by three orders of magnitude.
Two training set one based on random forest approach and another based on polynomial interpolation were performed on all the H-functions, the training set consisting of the couples \((\textbf{p}^h,\textbf{H}_h^k)\), \(\forall h, \ \forall k\).
The graphs in Figs. 10 and 11 represent the performance of the predictions for the first 12 functions \(H^k, k=1,\ldots ,12\). For each of the functions we represent the inferred value versus the value given by the simulation (considered as a reference value) and we indicate at the top of each image, the performance of the training.
These performances are quantified from the root mean square error (RMSE) and the R2 coefficient. The first line deals with he set used for training and the second line is for the set used for the test. By comparing both approaches, it turns out that in this case the polynomial approach performs better. An approach based on neural networks (not presented here) provides results that are very close to those obtained from the polynomial regression.
In order to quantify the performances of our method on the sets of parameters used for the validation, we will directly compare the thermal fields with the reference simulations. Indeed, these simulations used for the validation did not intervene in the singular value decomposition. We will use the modal basis extracted from the SVD built on the training simulations combined with the estimation of the H-functions based on AI-based regressions.
The comparison made directly on the thermal field on all the 20 simulations used for the validation, gives deviations which do not exceed 0.4 degree in the temperature values. Figures 12 and 13 concern two arbitrary combinations of parameters, and depict the temperature field at 100 and 300 seconds. The thermal fields presented here are the ones obtained by reconstruction from the use of the surrogate. The bottom figures represent the error with respect to the reference solution. These errors remain relatively small and are acceptable for a prediction of the thermo-mechanical properties induced by the thermal field.
Construction of the hybrid twin
Our objective in this part is to set up a hybrid twin of the casting process. This twin shall be able to learn the difference between numerical simulations and experimental observations. As we have not yet developed experiments for the case presented above, we will generate the experimental data synthetically.
We use the numerical model previously developed as the basis, while increasing the conductivities by \(10\%\) and reducing the convection coefficients by \(10\%\). From now on, we note by experimental results the numerical data generated under these conditions.
The experimental observation is normally limited to a set of thermocouples. In our case this set is presented in Fig. 14. The indexes of the eight nodes where thermocouples are placed are noted by \(i_1, i_2, \ldots , i_8\).
For a set of parameter \(\textbf{p}^h\) we denote by \(^{h}\mathbb {T}_{i'j}^{\text {Exp}}, i'=i_1, \ldots , i_8, j=1,\ldots ,300\) the matrix containing the experimental temperature evolution at the eight thermocouples for the 300 simulation time steps. We also denote by \(^{h}\mathbb {T}_{i'j}^{\text {Num}}\) the matrix containing the simulated temperature evolution at the same nodes where the thermocouples are located.
Our aim is to establish a correction model based on the tensor decomposition of the numerical simulation \(^h\mathbb {T}_{ij}^{\text {Num}} = \displaystyle \sum _{k} \textbf{F}^k_i H_{h}^k \textbf{G}^k_j \).
Let us denote the difference between experiments and simulation (or model’s ignorance), at each thermocouple location, by
Ignorance model learnt through a minimization procedure
In order to express this difference in the same space-time basis \((\textbf{F}^k_i, \textbf{G}^k_i)\), the following minimization problem should be solved:
It is important to mention at this point that in this minimization we constrain the difference (ignorance) to be written using the functions defined in the space-time description of the numerical simulation. This can sometimes be slightly restrictive. Later we will propose a less restrictive approach later.
In the present case order to alleviate the minimization procedure, we limit the time period in which the minimization applies to the interval \(j'=200, \ldots , 300\).
In Fig. 15 the time evolution of the temperature at the eight thermocouples locations are illustrated for the choice of the parameters indicated in the figure. In Fig. 16 the deviation (ignorance) between the numerical predictions and the experimental observation is shown. In this figure the dashed lines represent the results of the reconstructed model by using the optimisation procedure described above.
The numerical prediction of our model is provided in Fig. 17. The reconstructed ignorance is illustrated in Fig. 18(left). All these figures are produced with the final time (\(t=300\)) and with the set of parameters indicated in Fig. 15.
The predictions obtained by using the numerical model enriched with the one of the ignorance is represented in Fig. 19(left). Figure 20 shows the experimental temperature at the thermocouples location. Finally, Fig. 21(left) gives the global error of the hybrid twin model, where an impressive error reduction can be identified.
Ignorance model learnt from a projection procedure
We propose here a more general procedure that alleviates some of the constraints of the previous procedure. Here we will use slightly different notations. The Eq. 19 is rewritten using the Khatri-Rao product (\(\odot \)) generalized for three matrices.
By defining the following matrices
Equation 19 can be rewritten as
The simulation matrix \(\mathbb {T}\) has dimension \((N \times t \times d)\) and the size of \(\mathbb {F}\) is \((N \times K)\) where N is the number of nodes involved in the cavity mesh, t the number of time steps, d the DoE size and K is the number of modes.
Concerning the ignorance matrix, with \(n=8\) thermocouples, the matrix size becomes \((n \times t \times d )\). This ignorance matrix reads
where
This matrix has been obtained from a new iterative SVD (involving \(K'\) modes) completely independent of the one that served to decompose the simulation solution.
The main idea consists in expressing this new decomposition using the space-time functions of the numerical decomposition. Let us denote by \(\mathbb {F}'_{(n\times K)}\) the selection of the n rows of the matrix \(\mathbb {F}\). The coordinates of the matrix \(\bar{\mathbb {F}}\) into the basis \(\mathbb {F}'\) define matrix \(\textbf{a}_{(K \times K^\prime )}\)
that defining an undetermined problem, its solution must be regularized. In order to preserve the sparsity this system is solved subjected to the L1-norm minimisation. Thus the obtained solution \(\textbf{a}\) selects naturally the more adequate functions of the numerical basis to express the ignorance.
Concerning the time basis, the coordinates of the matrix \(\bar{\mathbb {G}}\) into the basis \(\mathbb {G}\) results in matrix \(\textbf{b}_{(K \times K')}\)
that being usually overdetermined, a classical minimization procedure performs well (but a L1 norm could be applied if the system becomes undetermined)
It is now possible the write the ignorance defined in Eq. 40 by using the space-time basis that comes from numerical simulation
that can be then extended to the whole space domain by simply replacing \(\mathbb {F}'\) by \(\mathbb {F}\)
In order to compare the performance of this projection based approach in relation to the minimization based approach described in the previous section, the new proposed procedure is applied to the case-study previously addressed.
The reconstructed ignorance is illustrated in Fig. 18(right). The superposition of the ignorance with the numerical model is depicted in Fig. 19(right). Finally Fig. 21(right) gives the global error of the hybrid twin model, proving its exceptional performance.
In the particular case of our so-called experimental solution that has been obtained numerically, temperature filed could be known everywhere in the computational domain, as illustrated in Fig. 22. Thus, the global error of the hybrid twin model can be obtained for both, minimization and projection procedures as depicted in Fig. 23.
The performance of the projection method are better in terms of error values but also in terms of error distribution over the domain.
Remark
In order to specify an order of magnitude on the resolution and storage cost we give a small illustration on the studied case. In our study we have a problem which contains N degrees of freedom in space (about 1500), t time steps (about 300) and d combinations of parameters (about 200). The decomposition of the tensor \(\mathbb {T}\) as written in Eq. 19 takes around 5 CPU-seconds for each fixed point iteration involving about 100 alternating resolution of Eqs. 20, 21 and 22. If 50 enrichments are performed the entire decomposition takes about 250 CPU-seconds. In terms of memory storage we are always dealing with a tensor size containing \(9 \cdot 10^7\) real values that represents 0.7 Gigabytes assuming a double precision of float number representation. In such situation if we imagine that one would use a more refined mesh which involves twice more degrees of freedom in the physical space representation (2N) thus the total used memory is multiplied by 2. It is the same for the CPU cost of the resolution of Eqs. 20, 21, 22. In fact the CPU evolution here is linear and not quadratic because the latter system does not contain inversion but just a set of matrix product operations. However for the hybrid model as we have very little experimental information (\(n=8\) instead of N) the costs of calculation and storage are much reduced.
Conclusion
The casting twin addressed in the present paper was developed on a combination of a singular value decomposition strategy with machine learning-based regressions. This approach has been extended to establish a model of ignorance when experimental data is available. To our knowledge, the combination of singular value decomposition with machine learning-based regressions has very rarely been applied to processes in general and we have not found any work in the literature concerning the specific casting process. In most studies using artificial intelligence for processes, inputs and outputs are related to more macroscopic quantities. This new proposed methodology was applied to a casting part where the different temperatures of the fluid circulating in the cooling channels were considered as variable parameters. The errors of the digital twin, as well as the hybrid twin, were evaluated at different instants of the cooling process and compared to a reference solution.
The error and performance of the parametric surrogates and the hybrid twin were convincing, proving the potential of the proposed approach. Less than one degree Celsius was noted for model accuracy. This remains largely within the tolerance interval in the temperature prediction of such a process.
The machine learning part convincingly showed the ability of the artificial intelligence models used to determine a response surface with completely satisfactory metrics. Both regressions tested (Random Forest and Polynomial) gave rise to RMSE errors less than 0.1 for training and testing sets associated to a determination coefficients generally higher than 0.8. The application framework of the strategy put in place within the framework of this work can be extended to any transient problem without being limited to shaping processes. This can be in particular the case of velocity field evolution in a transient flow, or for example the evolution of chemical concentration in a transient non-homogeneous problem.
References
Lehmhus D (2022) Advances in metal casting technology: a review of state of the art, challenges and trends; part i: changing markets, changing products. Metals 12(11)
Campbell J (2015) Complete casting handbook. Butterworth-Heinemann
Kridli GT, Friedman PA, Boileau JM (2021) Chapter 7 - manufacturing processes for light alloys. In: Mallick PK (ed) Introduction to aerospace materials. Woodhead Publishing, pp 267–320
Mouritz AP (2012) Introduction to aerospace materials. In: Mouritz AP (ed) Introduction to aerospace materials. Woodhead Publishing, pp 1–14
Fang HC, Chao H, Chen KH (2014) Effect of zr, er and cr additions on microstructures and properties of al-zn-mg-cu alloys. Mater Sci Eng A 610:10–16
Miladinovic S, Stojanovic B, Gajevic S, Vencl A (2023) Hypereutectic aluminum alloys and composites: a review. Silicon 15:2507–2527
Rams J, Torres B (2022) Casting aluminum alloys. In: Caballero FG (ed) Encyclopedia of materials: metals and alloys. Elsevier, pp 123–131
Bayraktar S, Hekimoglu AP (2021) Chapter 19 - current technologies for aluminum castings and their machinability. In: Davim JP, Gupta K (eds) Advanced welding and deforming. Elsevier, pp 585–614
Mouritz AP (2012) Production and casting of aerospace metals. In: Mouritz AP (ed) Introduction to aerospace materials. Woodhead Publishing, pp 128–153
Tiwari SK, Singh RK, Srivastava SC (2016) Optimisation of green sand casting process parameters for enhancing quality of mild steel castings. Int J Product Qual Manag 17:127–141
Ahmadein M, Ammar HE, Naser AA (2022) Modeling of cooling and heat conduction in permanent mold casting process. Alex Eng J 61:1757–1768
Gunduz M, Kaya H, Cadirli E, Ozmen A (2004) Interflake spacings and undercoolings in al-si irregular eutectic alloy. Mater Sci Eng A 369:215–229
Gras Ch, Meredith M, Hunt JD (2005) Microdefects formation during the twin-roll casting of al-mg-mn aluminium alloys. J Mater Process Technol 167:62–72
Wang L, Sun Y, Bo L, Zuo M, Zhao D (2019) Effects of melt cooling rate on the microstructure and mechanical properties of al-cu alloy. Mater Res Express 6:116507
Kurtulus K, Bolatturk A, Coskun A, Gürel B (2021) An experimental investigation of the cooling and heating performance of a gravity die casting mold with conformal cooling channels. Appl Therm Eng 194:117105
Sachs E, Wylonis E, Allen S, Cima M, Guo H (2000) Production of injection molding tooling with conformal cooling channels using the three dimensional printing process. Polym Eng Sci 40:1232–1247
Feng S, Kamat AM, Pei Y (2021) Design and fabrication of conformal cooling channels in molds: review and progress updates. Int J Heat Mass Transf 171:121082
Bertoli C, Stoll P, Philipp, Hora P (2019) Thermo-mechanical analysis of additively manufactured hybrid extrusion dies with conformal cooling channels. In: COMPLAS XV: proceedings of the XV international conference on computational plasticity: fundamentals and applications. CIMNE, pp 519–528
Muller B, Gebauer M, Polster S, Neugebauer R, Malek R, Kotzian M, Hund R (2013) Ressource-efficient hot sheet metal forming by innovative die cooling with laser beam melted tooling components. In: High value manufacturing: advanced research in virtual and rapid prototyping: proceedings of the 6th international conference on advanced research in virtual and rapid prototyping, pp 321–326
Behrens BA, Bouguecha A, Vucetic M, Bonhage M, Malik IY (2016) Numerical investigation for the design of a hot forging die with integrated cooling channels. Proc Technol 26:51–58
Norwood AJ, Dickens PM, Soar RC, Harris R, Gibbons G, Hansell R (2004) Analysis of cooling channels performance. Int J Comput Integr Manuf 17:669–678
Karakoc C, Dizdar KC, Dispinar D (2022) Investigation of effect of conformal cooling inserts in high-pressure die casting of alsi9cu3. Int J Adv Manuf Technol 121:7311–7323
Yang Xw ZHU, JC, Nong ZS, Dong HE, LAI ZH, Liu Y, Liu FW, (2013) Prediction of mechanical properties of a357 alloy using artificial neural network. Trans Nonferrous Met Soc China 23:788–795
Suleiman LTI, Bala KC, Lawal SA, Abdulllahi AA, Godfrey M (2020) Applications of artificial intelligence techniques in metal casting-a review, pp 97–102
Cemernek D, Cemernek S, Gursch H, Pandeshwar A, Leitner T, Berger M, Klösch G, Kern R (2022) Machine learning in continuous casting of steel: a state-of-the-art survey. J Intell Manuf 33:1561–1579
Jiang LH, Wang AG, Tian NY, Zhang WC, Fan QL (2011) Bp neural network of continuous casting technological parameters and secondary dendrite arm spacing of spring steel. J Iron Steel Res Int 18:25–29
Bouhouche S, Lahreche M, Bast J (2008) Control of heat transfer in continuous casting process using neural networks. Acta Automatica Sinica 34:701–706
Susac F, Tăbăcaru V, Baroiu N, Viorel P (2018) Prediction of thermal field dynamics of mould in casting using artificial neural networks. In: MATEC Web Conferences, vol 178. EDP Sciences, p 06012
Vasileiou AN (2015) Determination of local heat transfer coefficients in precision castings by genetic optimisation aided by numerical simulation. J Mech Eng Sci 229:735–750
Borzacchiello D, Aguado JV, Chinesta F (2019) Non-intrusive sparse subspace learning for parametrized problems. Archives of computational methods in engineering 26:303–326
Ibanez R, Abisset-Chavanne E, Ammar A, Gonzalez D, Cueto E, Huerta A (2018) Duval JL, Chinesta F (2018) A multi-dimensional data-driven sparse identification technique: the sparse proper generalized decomposition. Complexity, 5608286
Brunton S, Proctor JL, Kutz N (2016) Discovering governing equations from data by sparse identification of nonlinear dynamical systems. PNAS 113(15):3932–3937
Sancarlos A, Champaney V, Cueto E, Chinesta F (2023) Regularized regressions for parametric models based on separated representations. Adv Model and Simul in Eng Sci 10:4
Goodfellow I, Bengioand Y, Courville A (2016) Deep learning. MIT Press, Cambridge
Chinesta F, Cueto E, Abisset-Chavanne E, Duval JL, El Khaldi F (2020) Virtual, digital and hybrid twins: a new paradigm in data-based engineering and engineered data. Archives of Computational Methods in Engineering 27:105–134
Sancarlos A, Cameron M, Abel A, Cueto E, Duval JL, Chinesta F (2021) From rom of electrochemistry to ai-based battery digital and hybrid twin. Archives of Computational Methods in Engineering 28:979–1015
Moya B, Badias A, Alfaro I, Chinesta F, Cueto E (2022) Digital twins that learn and correct themselves. Int J Numer Methods Eng 123(13):3034–3044
Argerich C, Carazo A, Sainges O, Petiot E, Barasinski A, Piana M, Ratier L, Chinesta F (2020) Empowering design based on hybrid twin: application to acoustic resonators. Designs 4:44
Sancarlos A, Le Peuvedic JM, Groulier J, Duval JL, Cueto E, Chinesta F (2021) Learning stable reduced-order models for hybrid twins. Data Centric Engineering 2:e10
Nouri M, Artozoul J, Caillaud A, Ammar A, Chinesta F, Köser O (2022) Shrinkage porosity prediction empowered by physics-based and data-driven hybrid models. Int J Mater Form 15(3)
Sancarlos A, Champaney V, Cueto E, Chinesta F (2023) Regularized regressions for parametric models based on separated representations. Advanced Modeling and Simulation in Engineering Sciences 10(1)
Samarskii AA, Vabishchevich PN, Iliev OP, Churbanov AG (1993) Numerical simulation of convection/diffusion phase change problems-a review. Int J Heat Mass Transfer 36(17):4095–4106
Heim D, Clarke JA (2004) Numerical modelling and thermal simulation of pcm-gypsum composites with esp-r. Energy and Buildings 36(8):795–805
Acknowledgements
This material is based upon work supported in part by the Army Research Laboratory and the Army Research Office under contract/grant number W911NF2210271. This work has also been partially funded by the Spanish Ministry of Science and Innovation, AEI /10.13039/501100011033, through Grant number PID2020-113463RB-C31 and the Regional Government of Aragon and the European Social Fund, group T24-20R. The support of ESI Group through the Chairs at ENSAM and Universidad de Zaragoza is also gratefully acknowledged.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Ammar, A., Ben Saada, M., Cueto, E. et al. Casting hybrid twin: physics-based reduced order models enriched with data-driven models enabling the highest accuracy in real-time. Int J Mater Form 17, 16 (2024). https://doi.org/10.1007/s12289-024-01812-4
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s12289-024-01812-4