Keywords

1 Introduction

The increasing energy consumption in data centers (DCs) and its associated cost and environmental impacts emphasizes the importance of alternative energy conservation strategies, while airflow management is regarded as one of the effective approaches to improve thermal environment and cooling efficiency in DCs [1]. According to Lu et al. [2], airflow management is defined to be the approach to keep the operating parameters (IT equipment) within the recommended ranges with minimum energy consumption. Many researchers have studied airflow management to optimize thermal environment and improve cooling efficiency in DCs.

Schmidt and Lyengar [3] numerically analyzed the suitability of a raised floor air supply and overhead air supply design in very high-density data centers. The results showed that under-floor air supply design was appropriate for higher chilled-air supply cases to produce cooler rack intake air, while overhead air supply design was suitable for lower chilled-air supply cases to yield a cooled rack intake temperature. Yuan et al. [4] investigated and improved the airflow performance of airflow distribution system in DC by applying under-floor flexible baffles (UFBs). The results show that the use of UFBs can efficiently reduce rack hotspot and make the exhaust air temperature distribution more uniform. Furthermore, Zhang et al. [5] investigated the optimization of airflow organization in a small-scale data room with cold aisle closure. The results show that not only the cold aisle closure can eliminate by-pass air, but also it can optimize thermal environment. In addition, Ni et al. [6] numerically analyzed the thermal environment and airflow organization in a small DC and investigated the optimum airflow distribution conditions. The results show that, under the optimum airflow condition, the thermal environment and cooling efficiency were successfully improved and the energy consumption decreased. Yuan et al. [1] experimentally and numerically investigated the impacts of flexible baffles attached on the front door of racks on the optimization of thermal environment. They concluded that the use of flexible baffles on the front door of racks could effectively improve the thermal environment and weaken rack hotspots.

All these researches all used airflow management to contribute to the optimization of thermal environment and achievement of energy saving in DCs. The novelty of this paper is to optimize airflow by setting baffles on the lower-side terminal of servers. The aim of this research is to compare the effects of different baffle angle on the airflow distribution and energy saving.

2 Methods

2.1 Data Center Description

The studied DC is located in Nanjing, China, and it works 24/24, 365 days/year providing network service for a campus. Figure 1 shows the 3-D view of the studied DC. The dimension of the DC is 11 m (L) × 8 m (W) × 4 m (H) with under-floor plenum height of 0.45 m. The air-supplying mode of the computer room air-conditioners (CRACs) is under-floor air supply and direct air return, while the aisle configuration mode is closed aisle containment and open hot aisle. The dimension of CAC is 5.4 m (L) × 1.8 m (W) × 2.2 m (H). There are two rows of 18 racks in total, which are evenly placed face to face on both sides of the CAC. The geometric dimensions of the rack are the same, whose size is 0.6 m (L) × 1.2 m (W) × 2.2 m (H). The supply air temperature of CRACs is 22 °C, while the supply air speed is 5.33 m/s.

Fig. 1
figure 1

3-D view of the studied data center

2.2 Numerical Description

In this study, Airpak 3.0 package is used to establish the DC model based on an actual operating DC located in a campus, in Nanjing, China. All the geometry sizes and facility parameters are set corresponded to the actual DC. The specification and parameters are summarized in Table 1. As shown in Table 2, there are seven cases in total in this study. The baffle angle is defined as the angle between baffle and horizontal plane. In case 2, the baffles are parallel to the horizontal plane; thus, the baffle angle is 0°. For each case, the simulation converged in about 2 h with an Intel Core i7 processor.

Table 1 Specification and parameters setting of the studied DC in the numerical part
Table 2 Different cases

Concerning the scale of DC, racks, servers, and other conditions, standard k–ε model is applied in this paper. This model has been widely recognized by many researchers and has gotten good simulation results [7,8,9]. There are four assumptions proposed to simplify the matter: (i) ignorance of air leakage, (ii) no radiation effects, (iii) air is incompressible fluid, and (iv) the flow was considered as steady turbulent flow. The convergence criteria for flow and continuity are 10–3, while that for energy is 10–6. The mesh type was set as unstructured hexahedral, while the meshes around the servers and vents were generated by pre-object meshing parameters. An enhanced wall treatment is used to solve the near-wall region.

A second-order upward scheme was used to discretize the convective term, while a semi-implicit method for a pressure-linked equation algorithm is used to achieve the pressure–velocity coupling. The data center was considered a confined space envelope with a defined solid-wall boundary. CRACs supplied cooling air through the perforated tiles. Thus, the velocity inlets should be the perforated tiles. The raised floor, the walls, and the ceiling were regarded as adiabatic boundaries, while the servers were comprised of five partitions and four uniform heat sources. In addition, the back doors of rack were regarded to be the velocity outlets. The mesh number for 7 cases was all beyond 1,800,000. Airpak packages could automatically check the mesh quality, while more than 95% of the meshes had a quality of 1; thus, the mesh quality was satisfactory.

Equations (1)–(4) express the governing equations for incompressible fluid. The equations for continuity, momentum, and energy conservation are shown as follows.

Continuity equation:

$$\nabla \cdot \vec{u} = 0$$
(1)

Momentum equation:

$$\frac{\partial }{{\partial {\text{t}}}}\left( {\vec{u}} \right) + \vec{u} \cdot \nabla \vec{u} = - \nabla p + \nabla \cdot \left( {\overline{\overline{T}} } \right) + \vec{g}$$
(2)

where p is the static pressure, T is the stress tensor, and \(\vec{g}\) is the gravitational body force.

The stress tensor T is given by

$$\overline{\overline{T}} = v_{\text{eff}} \cdot \nabla u$$
(3)

Energy conservation equation:

$$\frac{\partial }{{\partial {\text{t}}}}\left( h \right) + \nabla \cdot \left( {h\vec{u}} \right) = \nabla \cdot \left[ {k_{\text{eff}} \nabla T} \right] + S$$
(4)

where \(v_{\text{eff}}\) and \(k_{\text{eff}}\) are the effective fluid viscosity and thermal conductivity, respectively; \(\vec{g}\) is the gravitational acceleration vector, and S represents the volumetric heat generation.

3 Results and Discussion

Figure 2 shows the exhaust air temperature distribution of rack B4 in seven different cases. It can be seen that there was an obvious heat accumulation in the lower part of the rack. When the baffles parallel to servers were set at the lower-side terminal of servers, the phenomenon of heat accumulation became worse. However, when the baffle angle increased from 0 to 45°, the heat accumulation was slowly alleviated. However, when the angle continuously increased from 45 to 75°, thermal environment is deteriorated again. Thus, the optimum thermal distribution appeared when 45°, and 8-cm baffles were applied. Under this circumstance, temperature distribution at the upper and lower parts of the server was more uniform, while the heat accumulation was successfully eliminated.

Fig. 2
figure 2

Temperature cloud map of rack B4 in seven different cases

Figure 3 shows the temperature of rack hotspot at the rack B4 in seven different cases. According to Yuan et al. [1], the rack hotspot was defined to be the point where the maximum exhaust temperature of each rack was located. As shown in the figure, the temperature of rack hotspot stood at 32.6 °C in case 1 and increased dramatically to a peak of 34.5 °C in case 2 with 8-cm and 0° baffles, where there was an exponential decline to an all-time low of 30.7 °C in case 5 with 8-cm and 45° baffles. However, when the baffle angle continued to grow from 45 to 75°, the temperature of rack hotspot in the rack B4 recovered to 31.5 °C in case 7. Thus, case 5 with 8-cm and 45° baffles performed the best in terms of rack hotspot temperature. The temperature drop of rack hotspot reached 1.9 °C, which reduces the risk of the damages to servers.

Fig. 3
figure 3

Temperature of rack hotspot in each case

Figure 4 shows the temperatures in six different heights of the rack B4 in seven cases. The temperatures of six different heights in case 2 and case 3 were all bigger than those in case 1. In case 4, and 6–7, more than one temperature was bigger than that in case 1. Thus, all these cases could not meet the requirement of temperature drop. However, only in case 5, all the temperatures in six different heights were lower than that in case 1. Therefore, the application of 8-cm and 45° baffles in case 5 could realize temperature drop of all six heights in the rack B4 at the same time.

Fig. 4
figure 4

Temperatures in six different heights of the rack B4 in seven cases

4 Conclusions

The studied data center is an actual running one located in Nanjing, China. The numerical model is established totally based on the geometric sizes and operating parameters of the actual data center. The conclusions of this study are as follows:

  • The application of baffles set at the lower-side terminal of servers can optimize the airflow distribution to varying degrees, when the baffle angle is bigger than 0°.

  • The optimum thermal environment can be achieved when 8-cm and 45° baffles are applied, while the temperature distribution at the upper and lower parts of the server is more uniform and the heat accumulation was successfully eliminated.

  • Under the optimum condition, all the temperatures of six different heights of rack drops, while the temperature drop of rack hotspot reached 1.9 °C reducing the risk of the damages to servers.