Abstract
We present a new grid movement strategy, tested for a generic fluid-structure interaction (FSI) test case. The test case describes a flat plate with a prescribed rotational movement in a turbulent channel flow. The transient turbulent flow field is calculated with a low-Re RANS model. To account for the deforming fluid domain two different grid movement methods are compared. Using transfinite interpolation with a grid-point distribution fitted to the stationary starting conditions as grid moving method leads to errors for the drag-coefficient. By employing a normalized wall distance adaptive method it is possible to fulfill the near-wall resolution requirements within every time step and, to thereby, get more accurate results. The parallelization is achieved by domain decomposition and is evaluated using a strong scaling experiment.
Access provided by Autonomous University of Puebla. Download chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
1 Introduction
For many engineering problems the interaction between a moving structure and a turbulent flow is of importance. Frequently, numerical models are used to predict the behavior of such systems. The overall solution fidelity strongly depends on the quality of the turbulent flow model used to calculate the forces acting on the structure.
We employ the finite volume method with an URANS (Unsteady Reynolds Averaged Navier Stokes) modeling approach. Here, the underlying spatial discretization is decisive for efficiency and accuracy. Furthermore, the grid resolution has to fulfill validity requirements implied by the turbulence model in the near wall regions. These are formulated in terms of a velocity normalized wall distance y + for the cells adjacent to the wall.
The need to change the grid that represents the deforming fluid domain induces a conflict of aims between geometrical cell quality and resolution requirements. Standard techniques for calculating the movement of grid points such as spring analogy methods or interpolation methods (see, for example [1–3]), are designed to keep the initial distribution constant relatively to the structure. These do not address the conflict of aims explicitly and may produce wrong flow solutions, if the resolution requirements change.
In this contribution we want show, that it is possible to meet the near wall resolution requirements of a RANS model in a generic test case by relocating grid-points according to the flow situation. The idea is, regarding that it is necessary to solve a grid movement problem for the deforming domain anyway, to incorporate solution information for a more suited fluid grid. We use the Target-Matrix-Paradigm (TMP) introduced in [4] to formulate the grid movement problem as an optimization problem. It offers the possibility to express different aspects of cell quality within a uniform principle.
First, we present the governing equations that are relevant for computing the turbulent flow fields and understanding why near wall resolution is of interest, then we describe the numerical methods used. Details about the parallelization of the grid movement strategy are given. After that, the test-case is described and evaluated for a reference grid movement method using transfinite interpolation and the y + adaptive movement method. The test-case is also used to evaluate the parallel efficiency of the presented algorithm.
2 Governing Equations
As we consider prescribed structure motion, only the fluid continuum has to be described. For an incompressible Newtonian fluid with constant properties, the following Reynolds-averaged equations in Einstein notation describe the turbulent mean flow field:
where u i represents the mean velocity, p is the mean static pressure, ρ is the density, μ is the dynamic viscosity, t is the time, and x i are spatial coordinates. We close the system of equations with the Boussinesq eddy viscosity assumption
and two additional transport equations for the turbulent kinetic energy k and the dissipation rate ε to model the turbulent viscosity μ t (Chien [5]):
with
\(C_{\epsilon _{1}} = 1.35\), \(C_{\epsilon _{2}} = 1.8\), C μ = 0. 09, \(\sigma _{k} = 1\) and \(\sigma _{\epsilon } = 1.3\) are model constants. The source terms (8), (9) as well as the damping functions (10), (11) introduce a dependency on the wall distance y +, which is necessary to make the equation for k and ε valid down to the viscous sublayer.
3 Numerical Methods
Here we explain the main numerical methods involved when simulating the interaction of fluid and structure. As we only consider prescribed structure motion, methods to describe structure deformation are not outlined here.
3.1 Flow Solver
To solve the above system of partial differential equations, we use a finite volume scheme, implemented in our inhouse code FASTEST [6]. It is a boundary-fitted block-structured multigrid solver using hexahedral cells. The velocity-pressure coupling is achieved via the SIMPLE algorithm. For the time discretization, an implicit second order accurate backward differencing scheme is used. The convective fluxes are discretized with the upwind-scheme, whereas the central differencing scheme is used for the diffusive fluxes. The strongly implicit procedure is used to solve the systems of linear equations. The code is parallelized using domain decomposition methods. Each block of the grid is extended with a surface layer of halo cells to implement inter block communication. If blocks belong to different domains, MPI is used for the communication.
3.2 Grid Movement Strategy
For turbulent flows, the spatial discretization can be the source of two different kinds of errors. The obvious one is the discretization error which results from the fact that the continuous solution variables can not be perfectly represented with a finite number of unknowns. The other one is related to turbulence modeling, where the solved equations are only approximately describing the physical behavior. A special characteristic of many turbulence models is that the discretization is part of the model. In RANS models, the grid dependency is introduced with the near wall treatment approach, regardless of whether a wall function or a special near wall formulation is used. For this work it can be seen that in a finite-volume method equations (4) and (5) are associated to cell midpoints after discretization. In large eddy simulations, the filter width and the subgrid scale model are mesh width dependent. Hybrid models compare a turbulent length scale to the local mesh width for modeling turbulent viscosity. For the simulation of FSI problems, where the flow and the geometry are changing, the numerical method has to use a grid movement strategy, which is capable of limiting discretizaion and turbulence model errors. A new approach accomplishing this is explained first in this subsection. Afterwards, a method based on transfinite interpolation serving as reference in Sect. 4 is outlined.
3.2.1 Optimization Approach
3.2.1.1 Working Principle
To account for the deforming fluid domain, we use an optimization approach. In every new time step, the movement of the boundary grid points X b is given by a prescribed motion. We want to find displacements for n inner grid points X i such that an objective function f becomes minimal:
We use the target-matrix-paradigm introduced in [4] to formulate a relation q(X i ) between grid points and the quality of the cells they belong to. Here, quality is defined in terms of deviation from an optimal cell. For detailed information see [4]. In the near wall region, we define the optimal cells as rectangular and scaled in wall normal direction according to the y+ requirements. In the rest of the field, optimal cells are chosen to be just rectangular with an arbitrary size.
The corresponding objective function for m cells can be defined as:
In order to solve the optimization problem (13), we use the method of steepest descent.
To illustrate the working principle, Fig. 1a shows a simple square grid for which a displacement of it’s grid points for a given boundary deformation has to be found. After moving the boundary points to fit the new domain in Fig. 1b, the optimization problem can be formulated. Defining the optimal cell geometry to be square, one can obtain the valid discretization shown in Fig. 1d. Specifying an additional target size for every cell—as for example a linear distribution of cell heights along the vertical axis in Fig. 1c—can be used to adapt the grid to changing flow conditions. Figure 1e shows the resulting grid.
3.2.1.2 Parallelization
Simulating turbulent flow fields for technical applications can be computationally intensive. In this case, required computing resources can only be provided by parallel hardware. Therefore, the parallel software performance must scale with the number of compute units. According to Amdahl’s law [7], the speed-up is limited to the inverse of the serial portion of the software. So even if grid movement is only a small portion of the overall solution process it has to be suitably parallelized.
For an efficient parallelization, we use the same domain decomposition as for the flow solver. This avoids data transfer as the grid movement can be calculated on the processor on which it is needed. Also the grid movement code can directly work on the same grid representation in memory. Defining the optimization problem (13) for each subdomain, boundary points would be fixed as only inner points are optimized. Overlapping one vertex per dimension more than the subdomain boundary from one side, as depicted in Fig. 2, allows subdomain interfaces to be part of the solution. The overall solution is found by iteratively doing steepest descent iterations on each subdomain and updating the overlapping vertices at the boundaries. Figure 3 illustrates the exchange pattern. Convergence is measured by the relative change of the L 2 norm over all displacements between two consecutive iterations.
3.2.2 Blockwise Transfinite Interpolation
In this method the grid is regenerated in every deformation step. See [8] for detailed information. As we have a block structure it is sufficient to recreate the topological elements of the grid in hierarchical order. First, block edges are calculated with cubic splines between the given, new endpoints. By requesting constant angles at the endpoints with respect to the initial grid the splines are uniquely defined and orthogonality characteristics are preserved. The distribution of grid points along the length of the edges is also taken from the initial grid. Figure 4 illustrates this step. Transfinite interpolation [9] is then used to calculate new faces as illustrated in Fig. 5. Analogously, inner block points are calculated from the new faces.
4 Results and Discussion
4.1 Description of the Test Case
As test case, we use a two-dimensional flow in a channel with a length of 2 m and a height of 0. 45 m. In it’s center, a flat plate with a length of 0. 12 m, a height of 0. 006 m and rounded edges, which inclines with a constant angular velocity of 175 s−1, serves as obstacle. At the inflow, a parabolic profile is chosen to avoid high gradients triggering the refinement mechanism at the edges. The Reynolds number based on the channel height and center inflow velocity \(u_{max} = 4\,\text{ m/s}\) is Re = 2. 21 ⋅ 105. Figure 6 illustrates the setup after 5 ms simulation time. The simulations are started from the stationary non-inclined solution.
The starting grid shown in Fig. 7 is an O-Grid type mesh consisting of 274,880 control volumes. The treatment of turbulence in wall proximity, as introduced in Sect. 2, is based on the first control volume being located in the viscous sublayer. Thus y + should not exceed 5 and the initial grid has been chosen accordingly as shown in Fig. 8d.
4.2 Evaluation and Discussion
In order to demonstrate that the near wall resolution can be controlled, Fig. 8 contains plots of y + on the plate’s top wall for both methods at different simulation times. The reference method does not change the near wall resolution. So in consequence of the accelerated flow towards the trailing edge, the y + values surpass the model limit of 5. For the y + adaptive method, we can not observe a raising tendency. A visual impression of the resulting grids after 25 ms is shown in Fig. 9.
It has to be mentioned that the exceeding y + values could be avoided by choosing a finer initial mesh. But such a trial and error strategy might not be feasible for more complex FSI scenarios.
In order to determine the impact of under-resolved boundary layers on the quality of results, Fig. 10 shows both drag coefficients of the plate over time. Clearly, the solution obtained with transfinite interpolation as grid movement method produces lower drag forces as the velocity gradients on the wall are not represented correctly.
4.3 Parallel Performance
All performance measurements were carried out on the Lichtenberg high performance computer of TU Darmstadt with two Intel Xeon E5-2670 processors per node, each with 8 cores running at 2. 6 GHz base and 3. 3 GHz maximum turbo clock frequency per dual socket mainboard forming a network node. For inter node communication, FDR-10 InfiniBand is used. Four nodes were exclusively used to measure the strong scaling behavior. The MPI tasks are mapped to the cores in a sequential fashion, meaning that a minimum number of NUMA-levels is used. Figure 11 shows runtime results, taken for one time step on different numbers of cores. In order to avoid random effects, the measurements are repeated as many times as cores are used and then averaged. Solving the turbulent flow field takes the dominant portion of runtime for all numbers of cores. By looking at the corresponding speedup results in Fig. 12, it can be observed that the grid movement solver has a similar scaling behavior as the flow solver. The difference indicates higher losses in algorithmic efficiency for executing the steepest descent method on the decomposed domain. The intra-processor scaling up to 8 cores is afflicted by the turbo mode and shared memory bandwidth. The computational resources do not scale linearly in that region. Thus, in Fig. 13 the overall parallel efficiency is calculated based on the speedup towards the 8 core runtime.
5 Conclusion
We have presented a turbulent test case with prescribed structural motion, for which we applied a y + adaptive and a non-solution-adaptive grid movement strategy. The conducted numerical experiments showed, that it is important to control the grid resolution for simulations with changing flow conditions as occurring for FSI scenarios. The grid moving strategy outlined in this work in principle is not restricted to just limiting y +. Other solution dependent criteria are possible, as for example a posteriori error estimates aiming to reduce discretization errors.
We have also shown, that the solution adaptive grid movement can be calculated in an acceptable portion of total run time for serial as well as for parallel computation. The runtime results are of course test case dependent. Here the ratio of maximum displacement to a characteristic length of the computational domain is an important influence factor and has to be investigated in future work.
References
Degand, C., Farhat, C.: A three-dimensional torsional spring analogy method for unstructured dynamic meshes. Comput. Struct. 80(3–4), 305–316 (2002)
de Boer, A., van der Schoot, M., Bijl, M.: Mesh deformation based on radial basis function interpolation. Comput. Struct. 85(11–14), 784–795 (2007)
Spekreijse, S.P., Prananta, B.B., Kok, J.C.: A simple, robust and fast algorithm to compute deformation of multi-block structured grid. Technical Report - National Aerospace Laboratroy NLR, NLR-TP-2002-105 (2002)
Knupp, P.: Introducing the target-matrix paradigm for mesh optimization via node-movement. Eng. Comput. 28(4), 419–429 (2012)
Chien, K.-Y.: Predictions of channel and boundary-layer flows with a low-Reynolds number turbulence model. AIAA J. 20(1), 33–38 (1982)
Department of Numerical Methods in Mechanical Engineering: FASTEST - User Manual. Technische Universität Darmstadt, Darmstadt (2005)
Amdahl, M.: Validity of the single processor approach to achieving large scale computing capabilities. In: Proceedings of the 18–20 April 1967, Spring Joint Computer Conference, pp. 483–485. ACM, New York (1967)
Pironkov, P.: Numerical simulation of thermal fluid-structure interaction. Ph.D. thesis, Technische Universität Darmstadt (2010)
Smith, R.E.: Transfinite interpolation (TFI) generation systems. In: Weatherill, N.P., Thompson, J.F., Soni, B.K. (eds.) Handbook of Grid Generation. CRC Press, Boca Raton (1999)
Acknowledgement
This work is supported by the ‘Excellence Initiative’ of the German Federal and State Governments and the Graduate School of Computational Engineering at Technische Universität Darmstadt.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Kneißl, S., Sternel, D.C., Schäfer, M. (2015). Parallel Algorithm for Solution-Adaptive Grid Movement in the Context of Fluid Structure Interaction. In: Mehl, M., Bischoff, M., Schäfer, M. (eds) Recent Trends in Computational Engineering - CE2014. Lecture Notes in Computational Science and Engineering, vol 105. Springer, Cham. https://doi.org/10.1007/978-3-319-22997-3_5
Download citation
DOI: https://doi.org/10.1007/978-3-319-22997-3_5
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-22996-6
Online ISBN: 978-3-319-22997-3
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)