Data center environmental control

From Wikipedia, the free encyclopedia
Jump to: navigation, search

Data center environmental control is a constructive generic framework for maintaining temperature, humidity, and other physical qualities of air within a specific range in order to allow the equipment housed in a data center to perform optimally throughout its lifespan.

Air flow[edit]

Air flow management addresses the need to improve data center computer cooling efficiency by preventing the recirculation of hot air exhausted from IT equipment and reducing bypass airflow. There are several methods of separating hot and cold airstreams, such as hot/cold aisle containment and in-row cooling units.[1]

Overheating of data center equipment can result in reduced server performance or equipment damage due to hot exhaust air finding its way into an air inlet. Atmospheric stratification can require setting cooling equipment temperatures lower than recommended. Mixing the cooled air and exhausted air increases refrigeration costs.


IT vendors recommend maintaining temperature of 70–75 °F (21–24 °C). The American Society of Heating, Refrigerating and Air-Conditioning Engineers (ASHRAE) states that the recommended temperatures is between 68–77 °F (20–25 °C), with an allowable range spanning 59–90 °F (15–32 °C). Research has shown, however, that the practice of keeping data centers at or below 70 °F (21 °C) may be wasting money and energy.[2] Overcooling equipment, in environments with a high relative humidity, can expose equipment to a high amount of moisture that facilitates the growth of salt deposits on conductive filaments in the circuitry.[3]

Although a high cooling temperature is more efficient, overcooling of smaller server rooms can provide reserve thermal protection in the event of cooling system failure, or in outage situations where the computers are provided emergency backup power but cooling systems are not. In such situations, the equipment can continue to run and tolerate several degrees of temperature rise before system damage from overheating occurs. Environmental reserve cooling capacity may not be feasible in dense equipment environments such as datacenters, but those locations are more likely to have backup cooling and emergency power for cooling.

Summary of vendor datacenter temperature recommendations
Vendor Low (C°/F°) High (C°/F°) Optimal
Avtech[4] 20/68 24/75 -
ASHRAE[5] 18/64.4 27/80.6 -
Cisco[6] 18/64.4 27/80.6 -
Google[7] - - 26.7/80
Dell[8] 24+/Upper 70 F° 26+/Lower 80 F°
HP[9] 18/64.4 27/80.6
IBM[10] 18/64.4 27/80.6
Oracle[11] 21/70 23/74 22/72

Rack hygiene[edit]

Blanking plates and other fittings around the edge, top, floor, or the rack direct air intake so that only air from the cold aisle reaches equipment intakes and prevent leakage of exhaust air into the intake area. Fans on the top or rear doors of the cabinet ensure a negative pressure for exhaust air coming out of equipment. Effective airflow management prevents hot spots, which are especially common in the top spaces of a rack, and allows the temperature of cold aisles to be raised.

Aisle containment[edit]

Containment of hot/cold aisles and ducting hot air from cabinets are intended to prevent cool/exhaust air mixing within server rooms. Generally rows of cabinets face each other so that cool air can reach the equipment air intakes at the set temperature point for the room. A more recent addition to the consideration of above floor containment is below floor air flow control. A range of underfloor panels can be fitted within the raised floor plenum to create efficient cold air pathways direct to the raised floor vented tiles.

Containment is generally implemented by physical separation of the hot and cold aisles, using blanking panels, PVC curtains or hard panel boards. Containment strategies could differ based on various factors including server tolerance, ambient temperature requirements and leakage from data centers.[12][dead link]

Access floor[edit]

The type of access floor a facility uses can also greatly affect its air flow capabilities and limitations. Specifically, a raised access floor allow for under-floor cable management as well as air flow. There are two main classifications of such access floors: low-profile floors (under 6 feet in height) as well as standard/traditional access floors (over 6 feet).[13]

See also[edit]


  1. ^ Fox, Mike (2012-02-15). "Stulz announced it has begun manufacturing In Row server cooling units under the name "CyberRow".". DataCenterFix. Retrieved 27 February 2012. 
  2. ^ Cosmano, Joe (2009), Choosing a Data Center (PDF), Disaster Recovery Journal, retrieved 2012-07-21 
  3. ^ Garrett, David (2004), Heat Of The Moment, Processor, retrieved 2012-07-21 
  4. ^ "AVTECH - Recommended Data Center Temperature & Humidity". Retrieved 2015-08-01. 
  5. ^ "2008 ASHRAE Environmental Guidelines for Datacom Eq uipment" (PDF). 
  6. ^ "Data Center Power and Cooling white paper" (PDF). 
  7. ^ "Efficiency: How we do it – Data Centers – Google". Retrieved 2015-08-01. 
  8. ^ "Data Center Operating Temperature: The Sweet Spot" (PDF). 
  9. ^ "Applying 2011 ASHRAE data center guidelines to HP ProLiant-based facilities" (PDF). 
  10. ^ "Environmental design criteria". Retrieved 2015-08-02. 
  11. ^ "C H A P T E R 2 - Environmental Requirements". Retrieved 2015-08-02. 
  12. ^ Data center containment - Hot aisle or cold aisle?
  13. ^ "What Type of Access Floor Should I Choose?". Netfloor USA. Retrieved 4 October 2015. 

Further reading[edit]