• No results found

Validated thermal air management simulations ofdata centers using remote graphics processing units

N/A
N/A
Protected

Academic year: 2021

Share "Validated thermal air management simulations ofdata centers using remote graphics processing units"

Copied!
6
0
0

Loading.... (view fulltext now)

Full text

(1)

Validated thermal air management simulations of

data centers using remote graphics processing units.

Johannes Sj¨olund

∗†

, Mattias Vesterlund

, Nicolas Delbosc

, Amirul Khan

§

, Jon Summers

∗§ ∗RISE SICS North, Lule˚a, Sweden

Lule˚a University of Technology, Lule˚a, SwedenDassault Syst`emes Madrid, Madrid, Spain §University of Leeds, Leeds, United Kingdom

Email: {johannes.sjolund,mattias.vesterlund,jon.summers}@ri.se, {a.khan,j.l.summers}@leeds.ac.uk

Abstract—Simulation tools for thermal management of data centers help to improve layout of new builds or analyse thermal problems in existing data centers. The development of LBM on remote GPUs as an approach for such simulations is dis-cussed making use of VirtualGL and prioritised multi-threaded implementations of an existing LBM code. The simulation is configured to model an existing and highly monitored test data center. Steady-state root mean square averages of measured and simulated temperatures are compared showing good agreement. The full capability of this simulation approach is demonstrated when comparing rack temperatures against a time varying workload, which employs time-dependent boundary conditions.

Index Terms—data centers, thermal management, CFD, GPU, lattice Boltzmann methods

I. INTRODUCTION

The electrical consumption of Information Communication Technology (ICT), data networks, personal devices and data centers, showed an annual growth of 7% during 2007 to 2012 and estimated to double every 10 years. The full ICT electrical energy footprint was estimated to be 900TWh in 2012, corresponding to 4.6% of the total global electricity consumption, which had increased from the 3.9% in 2007 [1]. In a study where 100 data centers were examined, the metric called power usage effectiveness (PUE) was examined and values were found to be from 1.33 to 3.85, with an average of 2.13 and where closer to 1 is most favourable. Nearly half had a PUE above 2, signifying an energy overhead on the ICT of more than 100%. Furthermore, air conditioning systems were often found ineffective using between 21 to 61% of the total facility energy, averaging at 38% [2].

In recent years there has been notable changes in the data center operating environmental conditions, leading to higher operating temperatures to save cooling costs and the adoption of economization. There is a wider acceptable operating tem-perature range that can reduce internal server fan speeds and result in raised exhaust temperatures [3]. However, the facility fan power can be at a higher level than experienced when operating according to best practices, since there is a require-ment for greater fan work as part of the facility equiprequire-ment when higher temperature differences are to be generated [4]. To ensure effectiveness of large data centers, tools that enable sensible decisions about the cooling arrangement are needed

that accounts for cooling/heating sources and external thermal factors [5].

Thermal management of data centers includes the crucial matter of preventing hot air exiting the rear of the racks and recirculating to be ingested by front inlets or even the over provision of cold air that returns to the cooling units without passing through the racks. These issues often depend on supplying sufficient cold air to the front of the racks. Thermal management can also be improved by the addition of curtains, partitions, drop ceiling and ducting that targets cool air to or hot air from the racks [6]. When the airflow demand of the racks is targeted or contained and matches that supplied by the cooling units, then proper cooling is generally assured with minimised air short circuits and mixing of cold and hot air streams [7].

The preferred approach to cooling of data centers includes free cooling and aisle containment, since these technologies can yield greater elevated temperatures and improved effi-ciency gains over legacy data centers [2]. With pressurised aisle containment, it is possible for racks to experience up to 20% leakage, where the supplied air bypasses the servers when there is a large pressure difference between the hot and cold aisles. Improving the rack design and blocking leakage paths can lead to nearly 9% performance improvements. By optimizing the pressure difference between the hot and cold aisles, it is also possible to achieve a 16% reduction in total energy consumption [8].

Developing analytical optimization models to represent the dynamics and physical characteristics of the different sub-systems in a data center can reduce electricity costs by 3% and the ventilation and air conditioning energy by 8% [9]. One often studied subsystem in a data center is the dynamics of the airflow, where air temperatures and velocities are of interest to determine the most appropriate layout of the data center for new sites or performing root cause analyses in terms of poor thermal management of existing data centers. This is achieved by creating numerical models of the data center where simulation software can be employed as a predictive tool, making use of Computational Fluid Dynamics (CFD) methods.

(2)

II. DATACENTERCOMPUTATIONALFLUIDDYNAMICS

The application of CFD enables the creation of a complete computer model of the data center, with raised floor, cooling units, perforated tiles and server racks. Such simulations provide detailed distributions of air velocity, pressure and temperature in the data hall. It is then possible to determine airflow and thermal management issues, such as hotspots [7], which can be prevented without overcooling the entire data center and in addition, trends in the rack inlet temperature distributions can be captured. Moreover, transient simulations can be used to analyse different failure scenarios, such as determining the time until overheating of the IT equipment when either the chilled water pump or Computer Room Air Handling (CRAH) fans cease to work [10].

CFD can be applied to exergy destruction analyses to identify zones with inefficiencies and energy losses that results from inadequate deliverly of server air cooling [11]. In con-trast, CFD can also be used for detailed analyses of different floor tile designs and their influence on thermal performances of aisles from features, such as vane angle, number of grills and their percentage openness [12]. The CFD simulations must be validated against reality to provide trustworthyness of the model, which includes an appropriate choice of turbulence model, constraints and boundary conditions that can yield simulation results closer to reality [13].

Transient and real-time data center simulations are chal-lenging, since the dynamic environment thermal profile must be generated fast enough to capture server thermal generation interference [14]. By using real-time visualization of an entire data center, the thermal transients of individual servers can be assessed, which helps to determine sources of potential hotspots before they occur i.e. servers exceeding their specified temperature threshold [15].

The lattice Boltzmann method (LBM) is a recently devel-oped method for modeling fluid mechanics problems [16]. The method has been applied to many physical situations including airflows in data centers [17], where a successful comparison of the simulated results against CFD simulations based on the finite volume and finite element methods is demonstrated. Comparing different computational approaches for data center airflow simulations, the LBM demonstrates clear advantages in computational performance and applicability for transient and real-time flow simulations if executed on GPUs [18]. The purpose of this paper is two-fold:

• to demonstrate the real-time simulation and visualisation capability of the LBM on GPUs that are remotely located inside a data center without any performance degradation, and

• to validate the LBM simulation air velocity and temper-ature field results against real measured results from a highly monitored test data center.

The contribution of this work is therefore a strategy for execut-ing an existexecut-ing optimised GPU based LBM code developed by Delbosc [19] on remote GPUs and to validate for the first time

the numerical results of data center air flows against detailed experiments.

III. THELATTICEBOLTZMANNMETHOD

The LBM is originally based on the concept of a cellular automaton, which is a discrete computational model based on a regular grid of lattice sites. The grid can have any finite number of dimensions and each site has a finite number of states and at time t = 0 the state of each site is initialized to some predetermined value. As time progresses in discrete steps such that t = t + ∆t the state of each site is changed according to fixed rules that involve the state of the lattice site and that of its neighbours on the grid.

Applying the LBM as a CFD modeling tool, the states of the lattice sites are given by continuous distribution functions, and the rule of the automaton is based on the discrete lattice Boltzmann equation, which is a special discretisation of the Boltzmann equation [20].

A. The Boltzmann Equation

The Boltzmann equation is derived from kinetic theory, where gases are composed of particles following the laws of classical mechanics and particle interactions are described sta-tistically. Groups of particles are affected by adjacent groups and particles streaming between each other in a billiard ball-like fashion. A single particle distribution function f(1) is

sufficient to describe all properties of non-dilute gases [21], with

f(1)(~x, ~p, t) (1)

which gives the probability of finding a particle at location given by the displacement~x with momentum ~p at the time t. When an external force ~F acts on the particles, their future positions and momentum are described by

f(1)(~x + d~x, ~p + d~p, t + dt) (2) as long as no collisions occur between particles. This is called the particle streaming motion and captures fluid advection, which is the transport of fluid properties by the bulk motion. Accounting for particle collisions, the evolution of this dis-tribution function by particle interactions over time can be described by the Boltzmann Equation

 ~u · ∂ ∂~x + ~F · ∂ ∂~p+ ∂ ∂t  f(1)(~x, ~p, t) = Γ(+)− Γ(−). (3) The left hand side describes the streaming motion introduced by the external force ~F during the time interval dt. The right hand side contains the so called collision operators, which act on the fluid velocity~u. Γ(−)represents the number of particles

starting at(~x, ~p) and not arriving at (~x + d~x, ~p + d~p) due to particle collisions. Conversely,Γ(+)is the number of particles

not starting at at(~x, ~p) but ending up at (~x + d~x, ~p + d~p) [21]. The right hand side of the Boltzmann equation captures fluid diffusion.

The combined streaming and collision motion on the par-ticle level as represented by the Boltzmann equation yields what is called advection-diffusion transport processes that are central to modeling fluid flows.

(3)

B. Discrete Lattice Boltzmann

The goal of LBM programs is to provide a numerical solu-tion to the Boltzmann Equasolu-tion, using a discrete approximasolu-tion called the Discrete Lattice Boltzmann Equation which can be written as [19]

fi(~x + ~ei∆t, t + ∆t) = fi(~x, t) + Γ(fi(~x, t)). (4)

Therefore the basis of the LBM algorithm is the discretization of space and time into a lattice of suitable number of dimen-sions. The unique solution to the equation varies depending on these properties and the initial and boundary conditions of the fluid domain.

For a specific problem domain, the conversion of length in m and velocity in ms−1 defines the lattice time step,∆t. By

defining a dimensional physical length,Lphysand dimensional

velocity,Vphys ∆t = CL CU whereCL= Lphys Llbm andCU = Vphys Vlbm . (5) Therefore the number of lattice sites, Llbm, in any direction

and the choice ofVlbm, less than 0.2 for stability, dictates the

time step. A grid spacing∆x on a 3D lattice can be expressed fromLphysand the number of lattice sites in the whole domain

N = Nx· Ny· Nz with

∆x = L√3phys

N (6)

Time in the simulation domain is not continuous as in nature, but is measured in constant time steps ∆t. A time step is completed after all sites in the lattice have been updated from the previous time. This means that for each update, a constant time period of ∆t seconds in simulated time has passed. If ∆t is equal to, or greater than the time it took to compute the lattice update, the simulation is considered real-time or faster. Each direction vector ~ei on the 3D D3Q19 lattice shown

in Figure 1 is called a lattice velocity vector and is scaled such that during a time step, ∆t, a particle can move exactly from one site to an adjacent one. When velocity is measured in lattice units per time step (∆x ∆t−1) the magnitude of the ithlattice velocity is given as k~e

ik.

C. The LBM Algorithm

The collision operatorΓ in the Boltzmann Equation (3) can be implemented in multiple ways, the simplest employs the Bhatnagar-Gross-Krook (BGK) approximation [21], which is modeled by fi(~x + ~ei∆t, t + ∆t) = fi(~x, t) − ∆t(fi(~x, t) − fieq(~x, t) τ + ∆twi~ei· ~F c2 . (7)

As with the Boltzmann Equation (3), Equation (7) contains a streaming part, represented byfi(~x +~ei∆t, t + ∆t) = fi(~x, t),

a collision term ∆t(fi(~x,t)−fieq(~x,t))

τ , where ~x is the particle

displacement and forcing term ∆twi~ei· ~F

c2 . The functionf

eq i is

called the equilibrium distribution function given as [21]

fieq(~x, t) = wiρ(~x, t)  1 +~ei· ~u c2 s +(~ei· ~u) 2 2c4 s − ~u 2 2c2 s  (8)

where cs is the speed of sound and the vector product is

defined as the inner product. The macroscopic fluid density and velocity are given by

ρ(~x, t) =X i fi(~x, t) and ~u(~x, t) = 1 ρ X i ~eifi(~x, t) (9)

where the lattice weights are given in Table I. In equation (7)

~e14 ~ e15 ~e5 ~ e18 ~e11 ~ e7 ~ e1 ~e3 ~ e0 ~ e10 ~ e2 ~ e8 ~e4 ~e12 ~ e16 ~ e6 ~ e17 ~ e13 ~ e9 z y x

Fig. 1. D3Q19 lattice site with its lattice velocities ~ei.

TABLE I

D3Q19LATTICE VELOCITIES~eiAND WEIGHTSwi.

~e0 = ( 0, 0, 0) w0 = 1/3 ~e1 = ( 1, 0, 0) w1 = 1/18 ~e2 = (-1, 0, 0) w2 = 1/18 ~e3 = ( 0, 1, 0) w3 = 1/18 ~e4 = ( 0,-1, 0) w4 = 1/18 ~e5 = ( 0, 0, 1) w5 = 1/18 ~e6 = ( 0, 0,-1) w6 = 1/18 ~e7 = ( 1, 1, 0) w7 = 1/36 ~e8 = (-1,-1, 0) w8 = 1/36 ~e9 = ( 1,-1, 0) w9 = 1/36 ~e10 = (-1, 1, 0) w10 = 1/36 ~e11 = ( 1, 0, 1) w11 = 1/36 ~e12 = (-1, 0,-1) w12 = 1/36 ~e13 = ( 1, 0,-1) w13 = 1/36 ~e14 = (-1, 0, 1) w14 = 1/36 ~e15 = ( 0, 1, 1) w15 = 1/36 ~e16 = ( 0,-1,-1) w16 = 1/36 ~e17 = ( 0, 1,-1) w17 = 1/36 ~e18 = ( 0,-1, 1) w18 = 1/36

the distribution functionfiis relaxed towards the equilibrium

functionfieq at a collision frequency of 1/τ . The constant τ

is fixed to correspond to the correct kinematic viscosity,ν, of air given by [21] ν = 1 3( τ ∆t− 1 2) ∆x2 ∆t . (10)

The inclusion of the energy equation via an additional particle distribution function, buoyancy effects as a body force to the momentum equation, a turbulence model and appropriate boundary conditions are described in detail in [17]–[19].

IV. THELBM GPUCODE IMPLEMENTATION

The original codebase was written using a combination of C++, CUDA1 and LUA2. The code depends on shared

dy-namic visualisation libraries, including the OpenGL toolkit, X

1https://developer.nvidia.com/cuda-toolkit 2https://www.lua.org

(4)

Window System (X11), OpenGL Extension Wrangler Library (GLEW), and the Joint Photographic Experts Group (JPEG) image codecs.

A. Remote OpenGL Visualization through VirtualGL

While the CUDA framework used by the LBM code did not need any graphics rendering capability to perform the CFD simulation, the user interactive part of the program, such as keyboard and mouse event handling and graphical image rendering required an X11 server with hardware accelerated OpenGL functionality. In this configuration, the application is allowed to directly access the GPU hardware through the Direct rendering Interface (DRI), allowing for hardware accelerated graphics direct rendering to take place.

When an OpenGL application running on a remote headless server is accessed through a remote access system such as a Virtual Network Computer (VNC) or X11-forwarding through Secure Shell (SSH), LibGL creates GLX protocol messages and sends them to the local client X11 server via a network socket. The local client then passes the messages to the local 3D rendering system for displaying on a monitor3. There

are two main problems with this approach. First, in the case where the LBM code is executed through X11-forwarding and rendered on a local client, some openGL extensions require that the application has direct access to the GPU causing restrictions over a network. Second, 3D graphics data such as textures and large geometries can occupy several megabytes of space. Since an interactive 3D application requires tens of frame updates per second to be free of lag, indirect rendering requires an extremely high bandwidth and latency. A

cross-Application Xlib libGL GLX 3D Driver OpenGL

Remote Server Local Client Client 0 VirtualGL 3D X Server Rendered image Rendered image VNC Server Rendered image X11 cmds X11 events VNC Viewer stream Client 1 Client N ... Keyboard/mouse events Network Monitor

Fig. 2. In-Process GLX Forking with an X11 proxy over a network, in the form of a VNC server and client.

platform VNC client solution to the problem of OpenGL rendering on a remote server can be found in the VirtualGL4

software toolkit. This is called in-process GLX forking and involves interposing application GLX calls and redirecting them to the server GPUs. The rendered images can be sent to an X11-proxy such as a VNC server. The local client can then connect to the VNC server for visualisation of the simulations running on a remote server.

3https://dri.freedesktop.org/wiki/libGL 4https://www.virtualgl.org/

B. Multithreading

The original LBM code by Delbosc [19] was written as a single-threaded application and all code was executed in a single loop. This approach limited the performance of the LBM application in the graphical OpenGL visualization part, which is not an issue when the GPU is local. Even though VirtualGL allows excellent performance for remote OpenGL rendering through VNC, it introduces an overhead when interposing GLX calls and transporting the rendered image to the VNC X11-proxy. When visualising on a local GPU, the rendering time of OpenGL visualization is negligible, but several CUDA kernel simulations steps can be performed during the VirutalGL overhead.

Adding multi-threaded support to the LBM code allows the CUDA kernel to execute as often as possible, while also allowing the user to modify the execution parameters, such as simulation boundary conditions. A single CPU thread runs in an infinite loop, always trying to execute the kernel again as soon as the previous execution has completed. Through the common kernel interface, other threads are able to signal suspend and resume of kernel executions as well as reading simulation data and setting simulation boundary conditions. Thread access to the kernel is configured and protected by Mutual Exclusion (mutex) locking to ensure no race conditions occur. The order in which mutex locking grants access to shared memory is based on different thread privileges. Adding a common communication interface enables mutex locking for thread synchronization.

The overhead from VirtualGL rendering is eliminated with respect to the amount of CUDA kernel executions performed during a certain time period. The low priority simulation kernel execution thread runs on a dedicated GPU stream as often as possible. The GPU thread responsible for render-ing the OpenGL visualization copies simulation output from this thread and streams when a new visualization frame is required based on a set frame rate. Copying is asynchronous using device-to-device copy (between GPU memory banks) to a memory buffer, resulting in rendering being performed independently of simulation kernel execution. There remains a small overhead from performing the visualization and dis-abling the visualization by minimizing or hiding the drawing window improves the simulation performance. The technical detail of the multi-threaded implementation can be found in the masters thesis of Sj¨olund [22].

V. DATACENTERCFD MODEL

The racks in the test data center are configured with a central hot aisle as depicted in Figure 3, which includes a schematic of the thermal airflows. The LBM in its simplest form makes use of regular grids with Euclidian coordinates and therefore modeling sloped surfaces requires a high resolution lattice, therefore curved features have a simplified geometry, whilst maintaining the same areas of inlet and exhaust airflows.

The floor, walls and ceiling of the room are modeled using the half-way bounce-back scheme, [19], which defines zero air velocity along these boundaries. Likewise, the temperature

(5)

7000 6484 3200 From Heat Exchanger To Heat Exchanger Water with 30% Ethylene Glycol

Fig. 3. Schematic heat flow in the test data center at RISE.

distribution functions were also implemented with bounce-back to maintain no heat transfer at these surfaces. The boundary conditions for the CRAC and rack inlets and outlets requires specialized definitions.

A. CRACs and Server Racks

Conditioned air inlets from the CRACs blow cold air at a constant temperature, Tsupply, and flow rate,Qsupply. The

return CRAC inlets take a constant static pressurepreturn by

setting the velocity and temperature gradients to zero, [19]. Each rack in the test data center contains between 16-30 servers and power consumption is monitored on a per-rack basis. The air inlet at the front of the per-racks is modeled using a zero-gradient boundary condition with constant flow rate, Qin. Each server contains case mounted fans which

provide a temperature dependent server flow rate,Qserver. The

temperature on the back of the racksToutis thus dependent on

the inlet temperature, Tin, plus a temperature increase, ∆T ,

which is affected by the server workload, which in turn relates directly to its power consumption P and fan flow rate Qout.

This is given by [17]

Tout= Tin+ ∆T, ∆T =

P · ν Qout· k · P r

, (11)

where the constants ν = 1.568 · 10−5 m2/s is the kinematic

viscosity,k = 2.624·10−5kW/m K is the thermal conductivity

andP r = 0.707 is the Prandtl number of air at 30◦C.

B. Simulation Input Data

The data center LBM model has unknown input parameters, namely volumeteric flow rates and air exhaust temperatures of the CRAC units, Qsupply,i andTsupply,i withi = 1 . . . 4,

rack flow rates, Qout,j, and temperature increases, ∆Tj,

j = 1 . . . 10. Temperature increases depend on rack power consumption Pj. Sensor data was recorded every minute on

23/01/2018 from 09:00 for 36 hours from the test data center to provide these time varying input parameters.

The CRAC air flow rate, Qsupply, was measured using a

mass flow sensor and the air supply temperature Tsupply was

simply measured by a thermometer. The server rack power consumption, Pj, was calculated as the sum of the three

monitored phases to provide the∆Tj for each rack. However,

there is no direct measurement of the air flow rate through the racks, apart from the monitored rotations per minute (RPM) of the fans. Flow rates are approximated using specifications of the rack fans and the fan affinity laws, where fan power is proportional to the cube of the shaft speed. This means the ratio between maximum input powerPmaxand an operational

point Pop is equal to the cube of the ratio of maximum fan

speed fmax to operational speed fop. Around this operating

point, the power ratio can be assumed to be proportional to the volumetric flow rates Qmax and Qop. Since each server

has nf ans= 6 integrated fans and experimental data logs of

average fan speeds, the boundary conditions are available for server air flow rates by using an average fan speed frack in

RPM of all rack servers. The total flow rate for a rack is

Qout= frack· nf ans· nservers·

Qop

fop

(12)

= frack· nf ans· nservers· Qmax

(fop)2

(fmax)3

. (13)

C. Simulation Output Data

Both front and back of the racks in the test data center are fitted with temperature sensor strips that provides three temperature measurements at different heights above the floor. The CRAC units contain integrated sensors that record tem-peratures of the intake and exhaust air.

The lattice sites in the simulation correspond to the positions of the three rack based temperature sensors and are sampled during simulation runtime. The CRAC unit intake and exhaust temperatures are averaged from the lattice sites adjacent to sites containing the boundaries.

Temperature readings from both experimental and simula-tion data were averaged over one minute, after which they were recorded in CVS format files for the four CRACs and ten racks.

VI. VALIDATION ANDRESULTS

Temperature measurements of the back and front of the racks was performed using temperature sensor MCP9808, which according to specification has a ±0.5◦C accuracy. The CRACs have integrated sensors for temperature and mass flow rate with an unknown sensor accuracy.

Different lattice resolutions were tested and 36lu per metre, or 2.7 cm per lu, resulted in ≈ 7 × 106 lattice sites and

reasonably accurate flow rates with tractable computational times. The simulated average root mean square (RMS) tem-perature difference of the CRAC inlets were found to be approximately 1◦ C off from experiments. The steady-state temperature differences between measured and simulated for the rack inlets (front) and exhausts (rear) are shown in Table II, the average RMS difference between simulation and measured at the lowest point in the racks was accurate to within ±1◦C,

the middle within ±2◦C and at the top ±4C.

Since the LBM simulations are time-dependent it is possible to look at varying the power consumption of the racks. This is achieved in the test data by changing the server

(6)

TABLE II

RMSOF THE DIFFERENCE BETWEEN SIMULATED AND MEASURED TEMPERATURES IN◦CAT DIFFERENT RACK POSITIONS. Rack FrontBot FrontMid FrontTop BackBot BackMid BackTop 1 0.275 2.67 3.28 2.28 2.59 7.07 2 0.838 1.8 4.05 3.32 2.64 3.29 3 1.04 1.34 3.82 5.82 3.12 2.44 4 0.617 2 5.28 2.21 1.4 1.07 5 0.78 2.22 3.03 0.859 2.46 1.89 6 1.11 1.25 1.63 1.43 1.06 5.11 7 0.226 1.31 1.17 2.59 1.99 5.72 8 1.11 1.19 4.48 1.4 2.53 0.653 9 0.656 1.63 4.71 0.859 1.57 0.707 10 0.665 3.04 4.46 2.36 3.93 1.93

workloads, which is a feature of the test data center. With a time varying power consumption profile as an time varying boundary condition it is then possible to obtain a time series of the rack temperatures. Figure 4 shows the middle front and rear temperature as a function of time against the varying power consumption profile of rack 6 in the data center.

3,500 4,000 P o w er (W) 20 25 30 35 40 45 50 F ron t Mid ( ◦C) 200 400 600 800 1,000 1,200 1,400 1,600 1,800 2,000 2,200 20 25 30 35 40 45 50 Time (min) Bac k Mid ( ◦C) LBM Exp.

Fig. 4. Comparison of simulated and experimental temperatures for rack 6.

VII. CONCLUSIONS

The computational and visualisation advantages of imple-menting an LBM code with OpenGL functionality using Vir-tualGL offers the possibility to execute thermal management interactive simulations of data centers on remote GPUs. An existing single-threaded LBM on GPU code has been suc-cessfully augmented to a multi-threaded version for increased throughput when VirtualGL for remote GPU execution is used. The multi-threaded version is then configured to acquire time-varying boundary conditions to validate the simulations against a highly monitored test data center. The validation results demonstrate a good comparison, however there is diver-gence in the measured and simulated temperatures at increased distances from the data center floor. It is speculated that the buoyancy and turbulence modeling do not fully represent reality in addition to the fact that servers are not individually modeled, but lumped together at the rack level.

REFERENCES

[1] W. Van Heddeghem, S. Lambert, B. Lannoo, D. Colle, M. Pickavet, and P. Demeester, “Trends in worldwide ICT electricity consumption from 2007 to 2012,” Comput. Commun., vol. 50, pp. 6476, Sep. 2014. [2] J. Ni and X. Bai, “A review of air conditioning energy performance in

data centers,” Renew. Sustain. Energy Rev., vol. 67, pp. 625640, Jan. 2017.

[3] T. C. ASHRAE, “Data Center Networking Equipment - Issues and Best Practices,” Whitepaper Prep. by ASHRAE Tech. Comm., 2015. [4] T. C. ASHRAE, “9.9 (2011) Thermal guidelines for data

process-ing environmentsexpanded data center classes and usage guidance,” Whitepaper Prep. by ASHRAE Tech. Comm., vol. 9, 2011.

[5] C. Conficoni, A. Bartolini, A. Tilli, C. Cavazzoni, and L. Benini, “HPC Cooling: A Flexible Modeling Tool for Effective Design and Management,” IEEE Trans. Sustain. Comput., pp. 11, 2018.

[6] J. Cho, T. Lim, and B.S. Kim, “Cooling systems for IT environment heat removal in (internet) data centers,” Journal of Asian Architecture and Building Engineering, vol. 7, no. 2, pp.387-394, 2008.

[7] S. V. Patankar, “Airflow and Cooling in a Data Center,” J. Heat Transfer, vol. 132, no. 7, pp. 117, 2010.

[8] M. Tatchell-Evans, N. Kapur, J. Summers, H. Thompson, and D. Oldham, “An experimental and theoretical investigation of the extent of bypass air within data centres employing aisle containment, and its impact on power consumption,” Appl. Energy, vol. 186, pp. 457469, Jan. 2017.

[9] L. Cupelli, T. Schutz, P. Jahangiri, M. Fuchs, A. Monti, and D. Muller, “Data Center Control Strategy for Participation in Demand Response Programs,” IEEE Trans. Ind. Informatics, 2018.

[10] J. Athavale, Y. Joshi, and M. Yoda, “Experimentally Validated Compu-tational Fluid Dynamics Model for Data Center With Active Tiles,” J. Electron. Packag., vol. 140, no. 1, pp. 110, Mar. 2018.

[11] L. Silva-Llanca, A. Ortega, K. Fouladi, M. del Valle, and V. Sundar-alingam, “Determining wasted energy in the airside of a perimeter-cooled data center via direct computation of the Exergy Destruction,” Appl. Energy, vol. 213, pp. 235246, Mar. 2018.

[12] S. Khalili, M. I. Tradat, K. Nemati, M. Seymour, and B. Sammakia, “Impact of Tile Design on the Thermal Performance of Open and Enclosed Aisles,” J. Electron. Packag., vol. 140, no. 1, pp. 112, Mar. 2018.

[13] E. Wibron, A.-L. Ljung, and S. Lundstr¨om, “Computational Fluid Dynamics Modeling and Validating Experiments of Airflow in a Data Center,” Energies, vol. 11, no. 3, pp. 644, 2018.

[14] M. A. Oxley, E. Jonardi, S. Pasricha, A. A. Maciejewski, H. J. Siegel, P. J. Burns, and G. A. Koenig, “Rate-based thermal, power, and co-location aware resource management for heterogeneous data centers,” J. Parallel Distrib. Comput., vol. 112, pp. 126139, Feb. 2018.

[15] R. Ullah, N. Ahmad, S. U. R. Malik, S. Akbar, and A. Anjum, “Simulator for modeling, analysis, and visualizations of thermal status in data centers,” Sustain. Comput. Informatics Syst., Jan. 2018. [16] S. Chen, and G.D. Doolen, “Lattice Boltzmann method for fluid flows,”

Annual review of fluid mechanics, vol. 30, no. 1, pp. 329-364, 1998. [17] G. N. de Boer, A. Johns, N. Delbosc, D. Burdett, M. Tatchell-Evans, J.

Summers, and R. Baudot, “Three computational methods for analysing thermal airflow distributions in the cooling of data centres,” Int. J. Numer. Methods Heat Fluid Flow, vol. 28, no. 2, pp. 271288, Feb. 2018. [18] N. Delbosc, J.L. Summers, A.I. Khan, N. Kapur, and C.J. Noakes, “Optimized implementation of the Lattice Boltzmann Method on a graphics processing unit towards real-time fluid simulation,” Computers and Mathematics with Applications, vol. 67 no. 2, pp. 462-475, 2014. [19] N. Delbosc, “Real-Time Simulation of Indoor Air Flow using the Lattice

Boltzmann Method on Graphics Processing Unit” (Doctoral dissertation, University of Leeds), 2015.

[20] X. He, and L.S. Luo, “Theory of the lattice Boltzmann method: From the Boltzmann equation to the lattice Boltzmann equation,” Physical Review E, vol. 56, no. 6, p.6811, 1997.

[21] M.C. Sukop, and D.T. Thorne, “Lattice Boltzmann Modeling - An Intro-duction for Geoscientists and Engineers,”, Springer Berlin Heidelberg, 2006.

[22] J. Sj¨olund, “Real-time Thermal Flow Predictions for Data Centers,” Master Thesis, Lule˚a University of Technology University, 2018.

References

Related documents

46 Konkreta exempel skulle kunna vara främjandeinsatser för affärsänglar/affärsängelnätverk, skapa arenor där aktörer från utbuds- och efterfrågesidan kan mötas eller

För att uppskatta den totala effekten av reformerna måste dock hänsyn tas till såväl samt- liga priseffekter som sammansättningseffekter, till följd av ökad försäljningsandel

Coad (2007) presenterar resultat som indikerar att små företag inom tillverkningsindustrin i Frankrike generellt kännetecknas av att tillväxten är negativt korrelerad över

The increasing availability of data and attention to services has increased the understanding of the contribution of services to innovation and productivity in

Generella styrmedel kan ha varit mindre verksamma än man har trott De generella styrmedlen, till skillnad från de specifika styrmedlen, har kommit att användas i större

Parallellmarknader innebär dock inte en drivkraft för en grön omställning Ökad andel direktförsäljning räddar många lokala producenter och kan tyckas utgöra en drivkraft

Närmare 90 procent av de statliga medlen (intäkter och utgifter) för näringslivets klimatomställning går till generella styrmedel, det vill säga styrmedel som påverkar

Den förbättrade tillgängligheten berör framför allt boende i områden med en mycket hög eller hög tillgänglighet till tätorter, men även antalet personer med längre än