Archnes: Jeremy
Archnes: Jeremy
By MASSACHUSETTS INSTJITTE
By OF TECHNOLOGY
Signature of Author
( S l May 7, 2010
Department of Engineering Systems, MIT Sloan School of Management
Certified by
Leon Glicksman, Thesis Supervisor
Professor of Building Technology & Mechanical Engineering
Certified by
Accepted by _________
t bDebbie Berechman
Executiv ector of MBA Progra IT Sloan School of Management
Accepted by
Nancy Leveson
Chair, Engineering Systems Division Education Committee
Professor of Aeronautics and Astronautics, Engineering Systems at MIT
Developing a Low-Cost, Systematic Approach to Increase an
Existing Data Center's Energy Efficiency
By
Jeremy M. Stewart
Submitted to the MIT Sloan School of Management and the Engineering Systems Division on
May 7, 2010 in partial fulfillment of the requirements for the degrees of
Abstract
Data centers consume approximately 1.5% of total US electricity and 0.8% of the total world
electricity, and this percentage will increase with the integration of technology into daily lives. In
typical data centers, valued added IT equipment such as memory, servers, and networking account
for less than one half of the electricity consumed, while support equipment consumes the remaining
electricity.
The purpose of this thesis is to present the results of developing, testing, and implementing a low-
cost, systematic approach for increasing the energy efficiency of data centers. The pilot process was
developed using industry best practices, and was piloted at a Raytheon site in Garland, TX. Because
the approach is low-cost, there is an emphasis on increasing the energy efficiency of data centers'
heat removal and lighting equipment.
The result of implementing the low-cost systematic approach, consisting of both technical and
behavior modifications, was a 23% reduction in electricity consumption, leading to annual savings of
over $53,000. The improvement to the heat removal equipment's energy efficiency was 54%. In
addition to presenting the results of the pilot, recommendations for replicating the pilot's success are
provided. Two major managerial techniques are described - creating an aligned incentive structure in
both Facilities and IT departments during the data center design phase, and empowering employees
to make improvements during the use phase. Finally, a recommended roll out plan, which included a
structure for Data Center Energy Efficiency Rapid Results Teams, is provided.
Thesis Advisors:
Leon Glicksman, Professor of Building Technology & Mechanical Engineering
Sarah Slaughter, Senior Lecturer, MIT Sloan School of Management
Acknowledgements
First, I would like to acknowledge the Leaders for Global Operations (LGO) program at MIT, my
peers in the LGO class of 2010, and my advisors Sarah Slaughter and Leon Glicksman for their
support of this work.
Second, I would like to thank the amazing team I worked with at Raytheon Intelligence &
Information Systems - Roy Villanueva and Joe Hall - for their support during and after the
internship that produced this thesis.
Third, I would like to thank my friends, colleagues and mentors at Boeing. I appreciate your advice,
faith and sponsorship during the two year LGO program. I look forward to the making an impact
and addressing the challenges that lie ahead.
Finally, and most importantly, I would like to thank my wife, my parents, and my in-laws for their
love and support. I realize that the LGO program is one step in a lifelong journey, and I am glad
that I have y'all by my side.
TABLE OF CONTENTS
1 INTRODUCTION...... ................................................ 9
2 BACKGROUN D ................................................................................................................ 12
5.3 IMPLEMENTED IMPROVEMENTS AND THEIR AFFECTS ON THE PILOT DATA CENTER ............ 66
6.2.3 ALLOW FEEDBACK AND INCREASED KNOWLEDGE TO BE SHARED AMONGST THE RRTS .......... 84
R BIBLIOGRAPHV ..................................................
ER'OR!I BnOOKMARKW MC~r TN DPTXTfl
Since the high-cost for electricity leads to a high overhead rate that lowers their likelihood of IIS
winning cost-based defense contracts, there is an intrinsic reason for the IIS IT organization to strive
to increase data centers' energy efficiency. Furthermore, because most of the electricity consumed by
IIS is derived from fossil fuels, their energy consumption leads to additional GHG in the
environment. This provides the Facilities organization incentive to help increase the energy efficiency
of data centers. Since there are two organizations - IT and Facilities - who are incentivized to
increase the energy efficiency of data centers, behavior change techniques were developed in such a
way to apply to both organizations. Additionally, organization alignment, interaction, and incentives
were considered levers to use to increase the teamwork necessary for large organizations to achieve
energy efficiency reductions.
3 (EPA 2010)
1.3 Thesis Overview
The document is organized as described below:
Chapter 1 outlines the general motivation for the thesis and provides an overview of the thesis
contents.
Chapter 0 provides the company background and a brief discussion of data centers and the
Chapter 3 presents the hypothesis for the study undertaken.
Chapter 4 describes existing knowledge on how to improve the energy efficiency of data centers,
and outlines the developed pilot process that was used to test the hypothesis
Chapter 5 details the pilot data center and provides the before and after electricity consumption
data.
Chapter 6 provides recommendations for improving the energy efficiency of data centers through
engineering design and managerial technique, and through extensive roll-out of the piloted process
with Rapid Results Teams.
Chapter 7 presents an overview of the business case, including a summary of key findings.
2 Background
2.1 Company Background - Raytheon
The Raytheon Company, established in 1922, is a technology leader specializing in defense, homeland
security, and other government markets throughout the world. With a history spanning more than
80 years, Raytheon provides electronics, mission systems integration, and other capabilities in the
areas of sensing, effects, communications and intelligence systems, as well as a broad range of
mission support services. Raytheon has around 75,000 employees worldwide and generated $25
billion in 2009 sales. 4 Raytheon has the following six business units: 5
Integrated Defense Systems (21% of sales and 29% of operating profits in 2008) is a leading
provider of integrated joint battle space (e.g., space, air, surface, and subsurface) and homeland
security solutions. Customers include the U.S. Missile Defense Agency (MDA), the U.S. Armed
Forces, the Dept. of Homeland Security, as well as key international customers. Main product lines
include sea power capability systems, focusing on the DDG-1000, the Navy's next-generation naval
destroyer; national & theater security programs, including the X-band radars and missile defense
systems; Patriot programs, principally the Patriot Air& Missile Defense System; global operations;
and civil security and response programs.
Intelligence & Information Systems (13% of sales and 9% of profits) provides systems,
subsystems, and software engineering services for national and tactical intelligence systems, as well as
for homeland security and information technology (IT) solutions. Areas of concentration include
signals and image processing, geospatial intelligence, air and space borne command & control,
weather and environmental management, information technology, information assurance, and
homeland security.
Missile Systems (21% of sales and 19% of profits) makes and supports a broad range of leading-
edge missile systems for the armed forces of the U.S. and other countries. Business areas include
naval weapon systems, which provides defensive missiles and guided projectiles to the navies of over
30 countries; air warfare systems, with products focused on air and ground-based targets, including
the Tomahawk cruise missile; land combat, which includes the Javelin anti-tank missile; and other
programs.
Space & Airborne Systems (17% of sales and 19% of profits) makes integrated systems and
solutions for advanced missions, including traditional and non-traditional intelligence, surveillance
and reconnaissance, precision engagement, unmanned aerial operations, Special Forces operations,
and space. SAS provides electro-optic/infrared sensors, airborne radars for surveillance and fire
control applications, lasers, precision guidance systems, electronic warfare systems and space-
qualified systems for civilian and military applications.
Technical Services (10% sales and 6% profits) provides technical, scientific, and professional
services for defense, federal government and commercial customers worldwide, specializing in
mission support, counter-proliferation and counter-terrorism, range operations, product support,
homeland security solutions, and customized engineering services.
There are 9,200 employees at various IIS sites, which are shown in Figure 2. The headquarters of
IS, which is located in Garland, TX, consumes about 42% of the total IIS energy consumption, as
shown in Figure 3. Eighty percent of IIS employees have received security clearances from the US
government. The secret nature of the IIS business is important in that most areas of the company
are considered a Sensitive Compartmented Information Facility (SCIF); therefore, information and
data is not easily allowed to be taken in and out of the data centers. In addition, data centers that are
considered SCIFs must be essentially self-sustaining. As the Physical Security Standards for Special
Access Program Facilities states 7, "walls, floor and ceiling will be permanently constructed and
attached to each other. To provide visual evidence of attempted entry, all construction, to include
above the false ceiling and below a raised floor, must be done in such a manner as to provide visual
evidence of unauthorized penetration."
lrora uoaiTUnthicum
Springfield
Figure 2: Overview of Major IIS Site Figure 3: 2008 IIS Energy Consumption by
Locations 8 Major Site
LIV
1961 -1972
1972-1995 E-Systenorporated
1995
-Preset R oytadn
Figure 4: Photo of the Garland site in the 1950s10 Figure 5: Timeline of the Garland Site"
7%
2%
As shown in Figure 8, the constant plug load makes up the majority of the energy consumed at
Garland. To be the leading provider of intelligence and information solutions, Raytheon Intelligence
and Information Systems (IIS) must have the latest computer equipment, particularly in their data
centers. Therefore, the constant plug load in the data centers is comprised of IT equipment, cooling
equipment used to remove the heat produced by the IT equipment, lighting, and electricity loss that
results from transformers. In addition, emergency lighting and required evacuation systems are
responsible for constant electricity consumption, albeit a minor amount. Finally, it is likely that some
employees leave equipment - both intentionally due to program requirements and unintentionally
due to long standing behavior patterns - turned on 24 hours, and this equipment is included in the
constant load. The total constant plug load for Garland is about 5,200 kW per hour, and is about
89% of the total electricity consumed annually at the Garland site.
A varying amount of electricity, labeled "February Daily Load" and "June Daily Load" in Figure 8, is
consumed each weekday as IIS employees at the Garland site use computers, printers, and other
equipment. As shown in Figure 8, this daily load, which is considered the variable plug load, is
consistently about 1000 kW per hour every month of the year in 2009, and is about 7% of the total
electricity consumed at the Garland site.
2w 4
0 40
r 1
72E231 t 2 3 4 02 8
72 12 17
2011
-
-
]
Z- 1.
z.km.
lmate Zones
l1. than2W00
Zn 2. ls tMan
Zn 3 ileS-
2,
than
20
OD gne
m
tn.
7,000HDD.
5500-000HDD.
D and
CD Mn4,000-5,499
HDD.
.t. .,000 cDOand &an4,000
zn. s2,000 cDnor.nd I.n..,0
The final consumer of electricity at the Garland site is the Heating, Ventilating, and Air Conditioning
(HVAC) systems. As shown in Figure 8, the amount of electricity consumed in February is lower
every hour than the electricity consumed in June. The minimum amount of electricity consumed in
June is higher than the minimum electricity consumed in February due to the HVAC required to cool
the buildings at all times, even weekends and nights. The increased weekend cooling requirements in
June (as well as any non-winter month) are due to the extremely hot weather in Garland, as shown in
Figure 9. The electricity required to achieve this minimum cooling varies month to month, and is
illustrated as "Minimum to cool in June" in Figure 8. In addition to a minimum amount of cooling,
HVAC is required to keep occupants comfortable while they are working at the Garland site. To
determine how much electricity was required to achieve this "comfort cooling", the weekend
electricity during June was observed, as shown in Figure 8. The total HVAC, which is the
summation of the "minimum cooling" and "HVAC cooling" each month, averages about 555 kW
per hour, and is about 4% of the total electricity consumed at the Garland site.
100
1400 7000 100
11200
11000 70
20
E 60-
600 50D
60
03
400 4 aaN00
a
0 a20
Z 200~ J 10
D'a -00
00
M1Tj Nabi Gas mTte Ekec"~~ -Avg TenW*sWe(F)
Figure 10: Garland Natural Gas Consumption, Figure 11: Garland Natural Gas and
2005- 2008 Electricity Consumption, 2005- 2008
80,000
70,000 -
12%Decrease
50,000
40,000 --
30,000 -
W20;000 - -
10,000
0.
Figure 12: Example of energy consumption Figure 13: Breakdown of Garland energy
over a one year period at the Garland IIS consumption, 2005-2009
site
From the analysis of the energy consumed at Garland, it is obvious that the constant load deserves
special attention. Since the majority of the constant load is related to data centers, a thorough
process for improving the operating energy efficiency was developed and piloted using industry best
practices.
2.2 Data Center Overview
2.2.1 Introduction
Even though most consumers do not know what a data center is, anyone who accesses the Internet
uses data centers. As shown in Figure 14, a data center is a room that contains "Information
Technology Equipment" - data processing electronic equipment such as servers, network equipment,
and storage devices (memory). This electronic equipment is arranged in racks, as shown in Figure 15.
Practically all companies in all sectors of the economy have at least one data center and many
companies have multiple data centers to ensure required computing can occur. For example, if a
company's employees have access to the internet, have a company email address, or store files on a
company file server, than the company likely has data centers that support their computing
infrastructure.
Because data centers are used to house computing equipment, they typically have no windows and
have a minimum amount of incoming fresh air. Data centers range in size from small rooms (server
closets) within a conventional building to large buildings (enterprise class data centers) dedicated to
housing servers, storage devices, and network equipment.17 Large data centers are becoming
increasingly common as smaller data centers consolidate.18 In addition, in most cases data centers are
used for many years, resulting in a mixture of state-of-art and obsolete computing equipment.
2.2..1 EectrcityRequred
o DeiverPowe
The irs elctrcityreqireentfor daa cnte is he owe deiver sytemtha suplie elctrcit
2.2.2.1 ElectricityRequid Towelie oe
Network
Rooms 2008
benefial209
2enef(Sal209
through transformers and switchgears. The electricity is then sent to an uninterruptible power supply
(UPS), which is essentially a set of batteries that ensure that electricity is always available to the data
center. This uninterruptible supply of electricity is important because most data centers operate 24
hours a day, 7 days a week. However, grid disruptions can stop the supply of electricity from the
utility, making the UPS a key component of the power delivery system. As shown in Figure 18,
electricity is transformed from the UPS to the power distribution unit (PDU), which then sends the
electricity to the electrical panels that supply electricity to the data center IT equipment.
100%
277/480V (90%-
AC UPS
277/48oV PDU 170%
LO
AC
T T IT I
UP$tris PD Wrng Swthga
Figure 18: Diagram of Electricity Flow Figure 19: Typical Electricity Consumption
through the Power Delivery System 22 ,23 of a Data Center24
As shown in Figure 18, in the UPS the electricity is converted from AC to DC to charge the
batteries. Power from the batteries is then reconverted from DC to AC before leaving the UPS.
Electricity consumed in this power delivery chain accounts for a substantial portion of overall
building load. 2s Inherent losses are present for the power delivery system, as shown in Figure 19, and
increase if the components used are oversized for the data center they are serving.
22 (Rasumssen 2006)
23 (US Environmental Protection Agency 2007)
C 25
-20
Zi E
0a 1
r 15
m 10
55
Uj 0
2000 2001 2002 2003 2004 2005 2006
U Server Electricity EStorage Device Electricity *Networking Equipment Electricity
Figure 20: Estimated Electricity Consumed by the IT Equipment in US Data Centers from
2000-200626
To put the heat load that must be removed from data centers in perspective, a fully populated rack of
blade servers consumes 30 kW of electricity (720 kW per day)28, which is the equivalent of 43 average
UPS
13%
IT Loads
71%
Lighting
10%
Power Dist.
4%
Personnel
2%
Figure 21: Relative contributions to the total thermal output of a typical data center30
There are five basic methods to collect and transport unwanted heat from the IT environment to the
outdoor environment, and each method uses the refrigeration cycle to transport orpump heat from
the data center: 31
" Air cooled systems (2-piece)
* Air cooled self-contained systems (1-piece)
* Glycol cooled systems
* Water cooled systems
* Chilled water systems
The five methods are primarily differentiated in the way they physically reside in the IT environment
and in the medium they use to collect and transport heat to the outside atmosphere. Obviously, each
method has advantages and disadvantages. The decision on which cooling system to choose should
be based on the uptime requirements, power density, geographic location and physical size of the IT
Most large data centers use chilled water systems to remove the heat from data centers, as shown in
Figure 22. Heat generated in the data center is transported to the top of the computer room air
conditioner (CRAC) by the air circulating in the data center. The condenser coil then uses chilled
water to complete the refrigeration process, which is shown in Figure 23, on the heated air.
Heat fram
Figure 22: Diagram of a Chilled Water Figure 23: Heat Energy Removal Via the
System used to Remove Heat from a Data Refrigeration Cycle34
Center33
The major consumers of electricity of the heat removal system are the pumps used to transport the
heat-carrying medium from the data center to the outside atmosphere, the computer room air
conditioners (CRAC), and the ongoing upkeep/production of the heat-carrying medium. The
amount of electricity these consumers use is different for each system. For example, the chilled
water system uses electricity to pump the water from the cooling tower to the water chiller, to pump
the chilled water from the water chiller to the CRAC, and to pump the warm water from the CRAC
to the outside atmosphere. In addition to pumping water, the CRAC uses electricity to reheat air if
the humidity of incoming air from the data center is too low. Finally, electricity is used to reduce the
temperature of the water both in the cooling tower and in the water chiller.
100%
80%
~20%
0%
ASCE
(2001} ASCE
(2002) ACEEE162 GreenGrid Engineering
AHUs Systers
Fans,CRACs,
ChillerPlantandPumps
Figure 24: Typical Electricity Consumption of by the heat removal system of a Data
73
Center3s5 8
2.2.3 Impact of Data Centers
Most data centers are more energy intensive than other buildings. This is due to the high power
requirements of the IT equipment and the power and cooling infrastructure needed to support this
equipment. In fact, data centers can be more than 40 times as energy intensive as conventional office
buildings. 39 As an aggregate in 2005, data centers accounted for 1.5% of the total US electricity
consumption and 0.8% of the total world electricity consumption.40 To put this in perspective, in
2007 the carbon dioxide emitted because of data center energy consumption was more than the
carbon dioxide emitted by both Argentina and the Netherlands. 41 This percentage of total electricity
consumed - and therefore the impact of data centers - is expected to increase over the next decade
for various reasons, such as:
4 (Koomey 2007)
41 (Kaplan, Forrest and Kindler 2008)
MEN
* Industries (banking, medical, finance, etc.) moving from paper records to electronic records.
* Increased use of electronic equipment, such as global positioning system (GPS) navigation
and radio-frequency identification (RFID) tracking in everyday activities.
* Increased access to information technology in developing countries such as India, China, etc.
* Increased storage of records (both original and duplicates) for websites, databases, emails,
etc.
As the move to a digital way of life continues, the electricity consumed by data centers will increase
as shown in Figure 25.
140
4- Historical nergy use Future nergy Historical trends
S120 use projections 4 scenario
Current efficieny
trends scenario~
~100
emproved operation
Figure 25: Projected Electricity Usage of Data Centers for Different Scenarios 42
As illustrated in Figure 25, the level of efficiency, both in the IT equipment and the data center
operating principles will have an enormous impact on how much electricity data centers consume.
The different assumptions for each efficiency scenarios can be seen in Appendix II - Summary of
Alternative Efficiency Scenario Assumptions.
5).
3. Integrate lessons learned into the initial improvement process (as discussed in Chapter 6).
In order to measure the affect of the improvement process, data was collected prior to, during, and
after the implementation of the improvement process.
Figure 26: Screen shot of HP Power Advisor, which helps customers estimate the load of
future racks in their data centers 46
As shown in Figure 26, the user of the software is able to input what servers, storage devices, and
networking equipment will comprise their racks. In addition, the user is able to estimate utilization
of the rack, so that a realistic estimate of the typical power requirements can be determined. If this
type of process is used to design the entire data center, the energy efficiency will be improved.
In addition, to better estimating the future IT equipment load, the data center can be designed not
only to accommodate current requirements, but also to allow for future growth if required. Indeed,
planning for the present and the future can improve overall system efficiency. Examples of adding
redundant capacity and sizing for true peak loads are as follows:48
e Upsize duct/plenum and piping infrastructure used to supply cooling to the data center.
" Utilize variable-speed motor drives on chillers, pumps for chilled and condenser water, and
cooling tower fans.
* Pursue efficient design techniques such as medium-temperature cooling loops and waterside
free cooling.
* Upsize cooling towers so that chiller performance is improved.
eliminating electricity use from the thermal component, leaving only the electricity
requirements of the auxiliary pump and motor loads.
* High-reliability generation systems can be sized and designed to be the primary power
source while utilizing the grid as a backup. Natural gas used as a fuel can be backed up by
propane.
* Where local utilities allow, surplus power can be sold back into the grid to offset the cost of
the generating equipment. Currently, the controls and utility coordination required to
configure a data center-suitable generating system for sellback can be complex; however, but
efforts are underway in many localities to simplify the process.
100% 2LM
Figure 27: UPS efficiency as a function of Figure 28: Effect of internal UPS loss on
load comparing latest generation UPS to Efficiency 53
historic published data5 2
1222
~I 0 0
0_ _ _ _
In addition to purchasing new servers, storage equipment, and networking equipment to replace
obsolete equipment, data center managers can save a substantial amount of energy simply be
removing old, unused IT equipment from their data center. Oftentimes, new servers are installed
into data centers to replace old servers, but they often do not replace those servers immediately.
Rather, there is a transition time in which both the old and new servers are in use. Once the
transition is complete, the data center manager sometimes forgets to remove the old, unused server.
These servers will almost never be utilized (with only the occasional spikes of utilization when
standard housekeeping tasks - backups, virus scans, etc. - run); however, the machines continue to
consume power and produce heat that must be removed.5 9 Unused servers can be identified by
analyzing network statistics, and should be decommissioned to save energy. 60
58 (Fanara 2009)
s9 (Blackburn 2008)
60 (Blackburn 2008)
61 (Blackburn 2008)
Figure 30: Comparison of Power Consumption when "Performance State" of AMD servers6 2
The second way to save energy using a power management strategy is to simply turn off equipment
when it is not in use. Most IT equipment has the ability to completely shut down during a certain
time of the day (for example, 6pm-8am), and/or for certain periods of the year (for example, the
December holiday shutdown that many businesses have). However, in most cases the option must
be turned on and managed closely by the data center manager.
62 (Blackburn 2008)
63 (US Environmental Protection Agency 2007)
0.102
5
002 0.010
07
43 0 1 0 3 0 5 0 7 0 9 0
Figure 31: Average CPU utilization of more Figure 32: Server power usage and energy
than 5,000 servers during a six-month efficiency at varying utilization levels, from
period. Servers are rarely completely idle idle to peak performance. Even an energy-
and seldom operate near their maximum efficient server still consumes about half its
utilization.64 full power when doing virtually no work.65
Two solutions exist to increase the utilization and thus energy efficiency of IT equipment - physical
consolidation and computer consolidation (virtualization). Several examples of physical
consolidation are shown in Figure 33. Physical consolidation increases both operating efficiency and
heat removal efficiency since the consolidation increases the utilization of the servers in use, and
allows data center managers to decrease the areas of the data center in which heat must be removed.
In addition, some of the extremely underutilized servers may be able either decommissioned or used
for a different application, saving the data center manager in energy usage and required expenditures
for new servers.
Virtualization, as shown in Figure 34, is a type of consolidation that allows organizations to replace
several dedicated servers that operate at a low average processor utilization level with a single "host"
server that provides the same services and operates at a higher average utilization level.66
Virtualization may offer significant energy savings for servers because although many services are
only occasionally busy, the power consumed by the idle hardware is almost as much as the power
required for active operation. 67 When these servers are virtualized, a smaller number of servers can
provide significant energy savings because fewer servers are required to meet the computing
64 (Google 2007)
65 (Google 2007)
66 (US Environmental Protection Agency 2007)
67 (Hirstius, Jarp and Nowak 2008)
requirements, and each of the servers operates at a higher utilization and therefore higher energy
efficiency.
Virtual
Virtual Machine
Machine achine Virtual
- _ | Machine
There are three major barriers that must be overcome to make liquid cooling more prominent in data
centers, as follows:
1) the established mindset that air cooling is the best practice in data centers.
2) the fear that liquid cooling can destroy equipment if failure occurs.
3) the new infrastructure, including a new plumbing system, which must be put into place to
support a liquid cooling strategy.
If these barriers can be overcome or mitigated, the potential savings from direct, liquid cooling are
substantial. Not only can the computer room air conditioners (CRAC) be eliminated, but also in
some environments the water from a chilling tower is cold enough without mechanical chilling to
73
remove the heat from data centers.
71 (Sawyer 2004)
72 (Hwang 2009)
73 (Greenburg, et al. 2007)
water directly from a cooling tower to the CRAC units in order to remove heat from the data center,
as shown in Figure 36.
H"s from
Air EconomizerCuprmn
Hot air isflushed outside, and
outside ar isdrawn in Emini
AkA
3 ~~n tc 11
17 __
41
Figure 36: Diagram of a Chilled Water Figure 37: Diagram of a Data Center relying
System using chilled water directly from the only on heat removal through an Air-Side
cooling tower to remove heat from the Data Economizer 75
Center74
Water-side economizers are best suited for climates that have wet bulb temperatures lower than 55*F
for 3,000 or more hours per year, and can reduce the chilled water plant energy consumption by up
to 75%.76 While water-side economizers are considered low-risk by data center professionals, these
77
same professionals are split on the risk when using air-side economizers.
As shown in Figure 37, air-side economizers simply transport cool outside air into the data center
instead of using chilled water to remove heat from the data center. Heat produced from the IT
equipment is flushed outside. Two characteristics of the outside air raise concern with data center
professionals - humidity and contamination. However, tests have shown that these fears may be
unjustified, and that air-side economizers can reduce the electricity consumption of a data center by
60%.78
74 (Evans, The Different Types of Air Conditioning Equipment for IT Environments 2004)
75 (Atwood and Miner 2008)
76 (Greenburg, et al. 2007)
77 (Greenburg, et al. 2007)
78 (Atwood and Miner 2008)
4.2.3.3 Improve Airflow Management
The goal of air management in a data center is to separate the hot air generated from the IT
equipment from the cold air that cools the IT equipment. Airflow management is one of the most
practical ways to improve the overall energy efficiency of data centers, and very large increases in
energy efficiency can result from robust airflow management. There are several reasons to resolve
79
heat removal problems in the data center, as follows:
e There are practical, feasible, and proven solutions.
* Many fixes can be implemented in existing data centers.
e Large improvements (20% or more) can result from little to no investment.
* Both IT people and facilities people can contribute to fixing the problem.
* Solutions are independent of facility or geographic location.
e They lend themselves to correction through policies that are simple to implement.
The potential solutions to heat removal problems are described below, and are organized from
relatively difficult to easy solutions.
79 (Rasmussen, Avoidable Mistakes that Compromise Cooling Performance in Data Centers and
R Rcowd R RCold R
R
aR R
a a a *ot a mAl.l[ a go a Aile a et
Nel C Male JO
c ixn" C ixn C "xin C
k k k k kk k Ik
Cold Ai' Raised Floor *i' Cold A Raised Floor tId A'
Figure 38: Rack arrangement without hot and Figure 39: Hot/cold aisle rack arrangement
cold aisles (side view) Red indicates hot air, (side view). Red indicates hot air, while blue
while blue indicates cold air. indicates cold air.
Other problems exist in the arrangement shown in Figure 38. For example, because the hot air is not
evacuated into the plenum, it must travel directly over the racks in order to return to the CRAC. As
this hot air travels from rack to CRAC, there are many opportunities for it to take a path of lesser
resistance. Unfortunately, oftentimes these paths of lesser resistance are into another rack. This
situation also arises because the rack is pulling in air in an attempt to cool the IT equipment. Not
only does this create the same situation previously discussed, but it also leads to a situation of the hot
air not returning to the CRAC, which leads to inefficiencies in the operation of the CRAC. If the
incoming air to the CRAC does not meet the CRAC's minimum incoming air temperature setpoint,
the electric strip heaters inside the CRAC will heat up the incoming air until the incoming
temperature requirement is eventually met.
A much better arrangement for a data center is shown in Figure 39, in which the inlets of the rack are
facing each other, and the outlets of racks are coordinated so that mixing of hot and cold air is not
possible. In addition, perforated ceiling tiles are added in order to utilize the ceiling plenum both as a
way to create a path of least resistance for the hot air and also as a path for the hot air to flow back
to the CRAC. This hot/cold aisle arrangement meets the basic goal of segregating the cold and hot
air.8 Additional improvements, such as proper placement of perforated floor and ceiling tiles can be
made to improve the airflow management once the hot/cold aisle arrangement is set-up.
Two mistakes are common when it comes to perforated floor tile placement. The first is placing the
perforated tile anywhere besides the cold aisle. One example of this situation, with the perforated
floor tile, or "floor vent", placed too close to the CRAC unit, is shown in Figure 40. This situation
creates a "short circuit", in which the cold air from the CRAC returns to the CRAC unit without
cooling the IT equipment. As discussed previously, this returning cold air is reheated by the CRAC,
which is a waste of electricity.
Figure 40: Example of poor placement of Figure 41: Diagram showing heat removal
perforated floor tiles, or "floor vents", as it is capacity of perforated floor tiles depending on
referred to here8 3 the tile's airflow 84
The second common mistake made with perforated floor tile is that the incorrect number of floor
tiles is placed within the cold aisle. As shown in Figure 41, the perforated tile's airflow is directly
related to how much heat it can remove from IT racks. Therefore, the perforated tiles should be
places in such a way to remove the heat from the areas in which it is generated. For example, a rack
that generates 1 kW of heat (consumes 1 kW of electricity) should have cold air flowing at 160 CFM
in front of the rack. This airflow could be achieved either from one tile directly in front of the rack,
or from two floor tiles in the cold aisle near the rack.
82 (Rasmussen, Avoidable Mistakes that Compromise Cooling Performance in Data Centers and
7FQ
Figure 42: Rack arrangement showing layout of CRAC units in relation to the hot/cold aisle
(top view). This diagram is for a raised floor data center. 88
This alignment of the aisle with the CRACs ensures that the incoming air into the CRACs will be as
warm as is possible, which will increase the efficiency and therefore the capacity of the CRACs. In
addition, this aisle/CRAC alignment minimizes the opportunity for hot and cold air mixing,
especially if perforated ceiling tiles are placed in the hot aisles. In addition to aisle/CRAC alignment,
CRACs should be placed so the path between the CRAC and the cold aisle it is serving is minimized.
91 (Dunlap 2006)
92 (Rasmussen, Avoidable Mistakes that Compromise Cooling Performance in Data Centers and
One way to prevent the waste of costly cold air is to use air containment, brush-type collars kits, as
shown in Figure 45.97 While this technology is well established, it is not fully deployed because data
center managers oftentimes do not realize the impact that cable cut-outs have on the airflow and
electricity usage. Another way to prevent the waste of costly cold air is to run all wires (networking
and power) in the data center above the racks, which prevents the needs for cable cut-outs in floor
tiles. This solution is easiest to implement when the data center is first designed and built, since
moving the power and networking cables requires all IT equipment to be powered off.
95 (Neudorfer 2008)
96 (The Green Grid 2009)
97 (Neudorfer 2008)
N - -- - - dhftmg
* t *
41
4N*
Figure 46: Example of rack without blanking Figure 47: Missing racks in a row of racks
panels (left) and with blanking panels (right)98 allow hot air to re-circulate
If hot air is allowed to re-circulate and re-enter the IT equipment, an overheating of the equipment
can occur. In addition, when the hot air generated by the IT equipment is re-circulated, it does not
return to the CRAC, which lowers the CRAC's efficiency and cooling capacity. There are two
reasons that blanking panels and blanking racks are not fully deployed, as follows: 99
1) Data center managers believe they serve only aesthetic purposes.
2) They are difficult and time consuming to install (however, new snap-in blanking panels can
help alleviate this problem).
Both of these factors are not technical in nature, and require behavior change and aligned incentives
in order to ensure that blanking panels and blanking racks are used.
99 (Rasmussen, Improving Rack Cooling Performance Using Airflow Management Blanking Panels,
Revision 3 2008)
1o (VanGilder and Schmidt 2005)
Figure 48: Power supplies acting as a raised floor Figure 49: Network cabling acting as a
obstruction 01 raised floor obstruction 02
One of the best ways to minimize raised floor obstructions is through frequent or interval cleaning
and inspection. Another way to minimize raised floor obstructions is to run all wires (networking
and power) in the data center above the racks. As stated earlier, this solution is easiest to implement
when the data center is first designed and built, since moving the power and networking cables
requires all IT equipment to be powered off.
Lighting in data centers often provides an opportunity for electricity savings. Relatively low-cost
timers and motion-activated lighting can reduce the amount of time that lights are on. The impact of
Table 3: Strategies for Reducing Electricity Consum tion in New Data Centers
Data Center Improvement Capital Labor Behavior
Component Description Required' Required Change
Power Supply Rightsizin
Add on-site power
Power Supply generation to reduce 2-
or eliminate UPS
Heat Use in-rack water
Removal cooling
Heat Use Cooling
Removal Economizer
1Red = "Yes", Yellow = "In some Instances", Green = "No"
2(Rasmussen,
Implementing Energy Efficient Data Centers 2006)
3Assumed inefficiency
of UPS was eliminated (Rasmussen, Electrical Efficiency Modeling of
Data Centers 2006)
On the other hand, strategies that could be used to improve the electricity efficiency of existing data
centers are shown in Table 4. Of the improvement opportunities listed, only "improve airflow
management" and "install energy efficient lighting" were piloted. Improving the UPS was not an
option because the data center did not utilize a UPS. However, the efficiency of the transformer
utilized by the pilot data center was calculated for information purposes only. Consolidation,
replacing obsolete equipment, and virtualization were not piloted because each of these methods
requires capital, and because the IT equipment contained in the pilot data center was governed by
existing contracts between Raytheon and the US Government. While Raytheon should pursue
modifying existing contracts and putting emphasis of the IT equipment flexibility, it was out of the
scope of this thesis. Managing the IT power consumption was not included in the pilot because
proprietary agreements prevented the analysis of the power settings on the individual IT equipment
contained in the data center.
Install Energy
Yes 1-3%
Efficient Lighting
1Red = "Yes", Yellow = "In some Instances", Green = "No"
2(Rasmussen,
Implementing Energy Efficient Data Centers 2006)
3 Sum of More Efficient Air Conditioner Architecture, More Efficient Floor Layout, Coordinate Air
Conditioners, Locate Vented Floor Tiles Correctly, Reduce Gaps between racks/floors/walls,
and Install Blanking Panels and Floor Grommets
Of the two remaining improvement opportunities ("improve airflow management" and "install
energy efficient lighting"), a phased process was developed based on the difficulty and cost of
implementing each improvement. As shown in Figure 50, there are short-term and medium/long
term items within the improvement process. Because of time and budget constraints, only short-
term, low-cost improvements were implemented in the pilot data center and the following data was
analyzed to understand the impact of each improvement phase:
1) Airflow in front of racks
2) Air temperature at input and output of representative racks
3) Air temperature in plenum and near ceiling
4) Power load of CRACs
5) Instantaneous chilled water
Short Term. Low Cost Improvements
Table 5: Comparison of typical and actual characteristics of the pilot data center
Typical System Characteristics 04 Actual Characteristics of Pilot Data Comparison
Center
Under-floor or overhead air distribution Under-Floor air distribution, and four Same
systems and a few in-room CRAC units. in-room CRAC units
CRAC units are more likely to be water CRAC units utilize chilled water from Different
cooled and have constant-speed fans central chilled water plant for cooling,
and are thus relatively low efficiency and have constant-speed fans
Operational staff is likely to be minimal, 1-4 employees manage IT equipment Same
which makes it likely that equipment staff, but are not responsible for energy
orientation and airflow management are efficiency
not optimized
Air temperature and humidity are tightly Air temperature and humidity is Same
monitored monitored 24/7 for out of range
conditions
Power and cooling redundancy reduce Two CRACs would probably meet Same
overall system efficiency requirements, four are present
The data center is a test and development area, and some racks leave the room after they have been
tested. Because of this testing environment, the number of racks varies from 45-55. There is an 18-
inch raised floor and ceiling plenum in the room. Finally, as shown in Figure 51, there are four
peripheral computer room air conditioning units (CRACs) in the room, which are connected to
building's chilled-water system.
F .. 1: Dit D C
.............. I I
Ipeazk Imp . Ch. D 9/29/99 - 10/95/9
S I I I I
09:49 24 Hours/div 08:23
Figure 52: Current profile of the IT equipment over a one week period
Max Avg
Min Current Current Current Average Daily
(Arms) (Arms) (Arms) Voltage Average Load (kW) Load (kWh)
Phase A 171 198 184.5 279.2 52 1236
Phase B 147 174 160.5 279.2 45 1075
Phase C 168 190 179 279.2 50 1199
146 3511
______________ 220V _______________
Max Avg
Min Current Current Current Average Daily
(Arms) (Arms) (Arms) Voltage Average Load (kW) Load (kWh)
Phase A 396 418 407 120.15 49 1174
Phase B 352 366 359 120.15 43 1035
Phase C 398 422 410 120.15 49 1182
141 3391
The baseline condition of the pilot data center's cooling system is one of severe underutilization.
Each of the four units is nominally rated at 422,000 BTU/hr or 124 kW of cooling, so there is almost
500 kW of cooling capacity in the data center. Since the IT load is 141 kW (reference Table 6), there
is 3.5 times the amount of cooling capacity needed in the pilot data center. In addition to calculating
the cooling capacity and comparing it to the IT-generated heat load, the load required to run the data
center's CRAC units was monitored with a submeter. As shown in Figure 53, the load, which results
from fan power and electric strip heaters (to reheat/humidify), is very cyclical. The average electrical
load was calculated to be about 73 kW.
r~ -x
__
_ __ _ __ -
A, l ~'t
rw.
Ip~n.I~i~ i'B~i9 - 1i1f%
?5 4 ktxrs'or
Figure 53: Current of the data center CRAC units on a three day (left) and one day (right)
scale. Highlighted area in three day chart is area of one day chart.
This cyclical nature of the electrical load on the CRAC units indicates that the units are cycling
between humidifying and dehumidifying functions. However, because this data center is closed to
the outside atmosphere, there is no need to control the humidity. Since humidity is not a factor, the
CRAC units were overcooling the room, which resulted in a need to reheat the air that cycled
through the CRAC unit, only to remove the heat prior to sending the air into the raised floor. In
addition to measuring the electricity required by the CRACs, the chilled water valves on the units
were monitored, as shown in Table 7. The measurement showed that the chilled water required was
65 tons, which according to the calculations shown in Table 7, required 34.7 kW of electricity to
produce. Therefore, the total electricity required for heat removal equipment is 108 kW.
Table 8: Total baseline electrical load for the pilot data center
Component Electrical Daily Electricity Baseline %
Load (kW) Usage (kW/day)
Power Delivery System (loss 5 120 2
from transformer)
IT Equipment 141 3384 55
CRAC 73 1752 28
Chilled Water 35 840 14
Lighting 2.7 65 1
Total - 256.7 6161 100
In order to understand the electricity measurements and the data center efficiency, the power usage
effectiveness (PUE), as illustrated in Figure 54, was calculated for the baseline condition of the pilot
data center. A mathematically ideal PUE would be 1.0, in which 100% of the power from the utility
is delivered to the IT equipment, with none being used by the infrastructure. Any value over 1.0
represents the "infrastructure tax" to run the data center. 107
kaD 4, P~ I USEFULOUT
Efficiency = TOTAL IN
W Transformer Loss
Lighting 2%
1%
Figure 54: Detail of power consumption in a Figure 55: Baseline electricity usage for the pilot
data center efficiency model 08 data center
The average PUE for industry data centers is 2.5, and data centers with the most efficient equipment
and no over provisioning of capacity have been able to reach a PUE of 1.6.10 9 The pilot data center,
which has an electricity breakdown as shown in Figure 55, has a PUE of 1.82. However, there is a
caveat to this value- there is no UPS in the pilot data center. The UPS in data centers can consume
as much as 25% of the total electricity required for the data center.110 In any case, the PUE provided
a method to compare the effectiveness of the improvement process as it was implemented in the
pilot data center. In addition to electrical load, airflow and air temperatures were measured in order
to determine what affect the improvement process had on these critical aspects of the data center.
,L =2.76 kW
-- - 0
----
----
10 -- uIe
JS equired
43 CFM
I too M a W 71W IN W lo
4721 [%A) 141,M tUSZ) a"3 a )
"A3 PIA 143) 4"A1
1 11 21 31 41 51
Rack
Number
Figure 56: Variance in airflow in the pilot Figure 57: Comparison of actual and required
data center (baseline condition) airflow for pilot data center (baseline condition)
As shown in Figure 57, the baseline airflow (683 CFM) was much higher than required to remove the
heat generated by the IT equipment (141 kW/51 racks, or an average of 2.76kW/rack) in the pilot
data center. In fact, the airflow was 167% of the required airflow. This condition is not surprising,
considering that the four CRACs serving the pilot data center had the capacity to provide 3.5 times
the amount of cooling required according to the IT equipment load.
As shown in Figure 58, the air temperature in front of the racks was between 62 and 70.90F; the
average was 67.0F and the standard deviation was 2.4oF. None of the air temperatures are close to
the 80.6oF maximum that inlet air temperatures can be according to ASHRAE; indeed, most
measurements are closer to the 64.40F minimum specified by ASHRAE, and four data points are
below the minimum.11' This is an indication that the CRAC setpoints is too low, and electricity
could be saved by raising the setpoints while still meeting the ASHRAE requirements.
S 70-
65
E
- 50 - -
45
40
1 4 7 10 13 16 19 22 25 28 31 34 37 40 43 46 49 52 55
RackNumber
Figure 58: Variance in air temperature directly in front of each rack in the pilot data center
(baseline condition).
As shown in Figure 59 and Figure 60, the air temperature in the hot aisles (outlet of the racks) is
generally higher than the air temperature in the cold aisles (inlet of the racks). Also, the air
temperature in the raised floor is 68-69-F in three of the four racks, while the air temperature in the
raised floor in front of Rack 21 was 710F, which could be an indication of blockage in between the
CRAC output and the perforated tile in front of Rack 21. While the ceiling had a plenum, it was not
used to evacuate air when the pilot began, as illustrated by the lower air temperature in the plenum
than in room in some cases. This lack of evacuation of air from the hot aisle likely resulted in a
"short circuit" and air mixing, which could have led to the higher temperatures as the distance from
the ground increased in the cold aisle. This higher temperature near the top of racks could have also
been due to missing blanking panels, which allowed air from the hot aisle to return to the cold air
and re-enter the IT equipment.
...
....
..... ....
........
...
.....
..
..
.....
.....
.....
J
17 T1,
FiF , 5 'Rack.
Rdck
The air temperature in the ceiling plenum, 6 inches below the ceiling (and above racks), and near two
of the CRAC units was measured. As shown in Figure 61, the average air temperature 6 inches from
the ceiling and above the racks varied from 78-84-F, with a standard deviation of various
measurement points exhibiting a standard deviation up to 1.2oF. This relatively large standard
deviation indicates that the hot and cold air was mixing. The temperature in the ceiling plenum
varies from 73-78oF, and individual measurement points exhibited a standard deviation of about
0.30F. Most concerning was the air temperature above CRAC B and C, which was 76oF and 820F,
respectively. This air temperature is lower than the air temperature in other parts of the room, as
shown in Figure 61. This low air temperature, especially near CRAC B, indicates that the hot air
generated by the IT equipment was not returning to the CRAC units, resulting in inefficient use of
energy for both the CRAC units and the chilled water production.
Figure 61: Diagram showing the average air temperature at various locations within the pilot
data center (note only a portion of the data center is shown here). The average air
temperature shown was derived from data taken over a one-week period.
Figure 64: Picture showing randomly Figure 65: Picture showing missing
placed perforated tile blowing pieces of doors from racks (baseline condition)
paper into the air (baseline condition)
5.2.8 Summary of Baseline Condition
After reviewing the baseline data of the pilot data center, the issue, solution, and hypothetical result
were developed, as shown in Table 9.
5.3 Implemented Improvements and their Affects on the Pilot Data Center
In order to understand how the efficiency of a data center could be improved, the short-term, low-
cost improvement process developed from industry best practices, as shown in Figure 66, was
implemented in the pilot data center. A detailed diagram showing the changes made can be found in
Appendix IV - Changes made to the pilot data center.
Turn MhE)m
Ouff frssnwl tsui
For logistical reason, the improvements were implemented in two phases; however, all the short-
term, low-cost improvements shown in Figure 66 could be implemented at the same time in future
data center improvement projects. As shown in Figure 67 and Figure 68, the changes made during
this pilot were low-cost (- $300), and were not difficult to complete.
Figure 67: Example of the changes Figure 68: Example of the changes (installation
(installation of egg crates above CRAC and of egg crates in hot aisles and chimneys on
blanking racks) made during Phase I and II CRACs) made during Phase I and II
This phased approach allowed the team to differentiate the effects both the equipment (CRAC) and
non-equipment changes had on the airflow, which will be discussed in the following sections. After
all of the short-term, low-cost improvements were implemented, one CRAC unit (CRAC B) was
turned off. The affects this shut down had on air temperatures, airflow, and electricity usage will be
discussed in the following sections. Of particular interest was whether the changes and elimination
of one CRAC would cause the IT equipment to overheat due to a lack of cold air from the raised
floor.
As shown in Figure 70, the airflow was much higher than it needs to be to remove the heat generated
by the IT equipment (2.76kW per rack) in the pilot data center. Even after one CRAC was shut off,
the airflow per perforated tile (average of 742 CFM) was much more than required. This indicates
that "CRAC fighting" was occurring when all four CRACs were turned on, and there were probably
small "cyclones" of cold air remaining in the raised floor rather than reaching the racks.
TOpW Wfh
c
.
Baseline Pae
Lack2. 7 6 kW
equired =
.1 1 435 CFM
0 100 200 300 400 500 600 700 800 900 1000
(47.2] [94.4] [141.6] [1868] [236.01 [283.2] [330A) [377.6] [424.8 [471.9]
Figure 70: Comparison of actual and required airflow for pilot data center
As shown in Figure 71, the air temperatures after all changes were implemented (Post Phase II) were
lower than the baseline air temperatures at almost every point. It's quite remarkable that even after
one entire CRAC unit was turned off (Post Phase II), the average temperature in the cold aisle was
1-F lower than the baseline condition. This is an indication that there was probably "CRAC
fighting" occurring before the changes were made. The lower temperature is also an indication that
the hot air from IT equipment is leaving the room via the ceiling plenum, rather than returning to the
cold aisle and mixing with the cold air.
In addition to taking measurements in the cold aisle, air temperatures were taken in the ceiling
plenum, six inches below the ceiling, and near two of the CRACs. As shown in
Figure 72, there was a large increase in the air temperatures in the ceiling plenum. This large increase
was due to the installation of perforated ceiling tiles, or "egg crates", which replaced the solid ceiling
tiles that existed prior to the Phase I changes. These egg crates allowed hot air to flow from the data
center to the ceiling plenum, and then flow from the plenum to the CRAC units. However, as
indicated by the increase of the air temperature in the ceiling plenum above CRAC C, not all of the
hot air was returning to the CRACs after the Phase I changes were implemented.
Figure 72: Diagram showing the average and change from baseline air temperature after the
Phase I changes had been implemented within the pilot data center.
However, as shown in Figure 73-Figure 76, once the chimneys were installed onto the CRACs in
Phase II, the temperature of the air returning to the CRACs increased to about the same
temperatures observed in the ceiling plenum above the racks shown in Figure 72. This increase in
hot return air resulted in an increase in the efficiency of the heat removal equipment, which will be
discussed in section 5.3.3.
Also, as shown in Figure 73, mixing of hot and cold air that was evident above CRAC A was
eliminated once the chimney was installed, while the mixing continued to occur above CRAC B as
shown in Figure 74, since no chimney was installed. Another somewhat surprising result that was
after CRAC B was turned off, the air temperatures in the raised floor, which were measured below
racks near the applicable rack, decreased, as shown in Figure 74-Figure 76. This is a clear indication
that CRAC B and CRAC C were involved in "CRAC fighting" prior to the elimination of CRAC B.
In addition to illustrating "CRAC fighting", Figure 74 shows that a certain amount of suction of hot
air from the plenum was occurring prior to the shut off CRAC B. This indicates that if installing
chimneys is not possible, perforated ceiling tiles above CRACs help somewhat. However, as
discussed previously, a benefit of the CRAC chimney is the elimination of mixing above the CRAC.
Raised floor near
CRACshowed lower
No temperatures
Chimney CRACB
Chimney CRACB
Installed Turned off
Installed Turned off
Figure 73: Air temperatures near CRAC A Figure 74: Air temperatures near CRAC B
after Phase II changes were implemented after Phase II changes were implemented
Figure 75: Air temperatures near CRAC C Figure 76: Air temperatures near CRAC D
after Phase II changes were implemented after Phase II changes were implemented
In addition to the perforated ceiling tiles, the blanking panels and blanking racks that were installed
during Phase I probably helped to reduce the suction of hot air into the cold aisle, which in turn
allowed more hot air to return to the CRAC units, leading to a higher efficiency. To understand the
affect of the blanking panels and blanking racks, the air temperatures surrounding two racks were
collected. As shown in Figure 78, the cold aisle exhibited a more consistent temperature range after
the blanking rack (shown in Figure 77) was installed. For example, Rack 33, which originally had a
blank space beside it, showed a range of air temperatures in the cold aisle of 73-77oF prior to the
installation of the blanking rack. After the blanking rack was installed, the range of air temperatures
in the cold aisle was 69-72oF. Both the overall temperatures and the range were reduced. In
addition, the air temperatures in the hot aisle for Rack 33 increased from a range of 79-83oF to 77-
87-F after the blanking rack was installed. These changes are evidence of the reduction in mixing
blanking racks can have on a data center. Rack 36, which never had an empty space beside it,
showed less change from the baseline condition, as shown in Figure 78.
. sii
Ale
. . .. . .lankingRack.
F Raised
Floor EBot Mid 3Top U Plenum
As shown in Figure 79 and Figure 80, a drastic reduction in the electricity took place during the pilot.
In addition, the cyclical nature of the load was reduced so that the electrical load of the CRACs
became almost constant. The total reduction in electricity required by the CRACs was 68% for the
pilot data center.
rIICti.D
O.)1/1VIrms RMS Ch.D 11/04/09 - 11/5-/Oc
114f pftet 7 After Turning Off One CRAC
~ F rU 'i eliinating I? Afler L
r
eliinating
rehea
____ ___18L
0 0
Figure 79: Current of the data center CRAC Figure 80: Current of the data center C R.AC
Fgunitsbere an af the rat funtin as units before and after one of the units was turned
units before and after the reheat function was off. In addition, dehumidification an d
disabled on all four in room CRAC units. humidification was turned off
As shown in Table 12, the electricity consumed by the CRACs changed significantly during the
implementation of the improvement process. Since only three of the four CRAC units served the
data center upon completion of the pilot, the cooling capacity was reduced 25% from 500kW to 372
kW. Even after shutting down one CRAC, there is still 2.6 times the amount of cooling capacity
needed in the pilot data center, which is must closer to the n+1 condition desirable in critical data
centers.
In addition to measuring the electricity required by the CRACs, the chilled water valves on the units
were monitored, as shown in Table 13. The measurements showed that the chilled water required an
average of 51.2 tons. Therefore, the electricity required by the chilled water system was reduced by
22% (27 kW versus the baseline value of 35 kW), and the total electricity required by the heat
removal system is 50 kW/ton (a 54% reduction).
Post Phase II, post turning one 23 556 1,186 -68% 432,820
off I
----
------
- -.......
The data center contained 28 - 4 foot T8 lamp fixtures. Since each fixture requires 96 W of
electricity, the baseline total load for the lighting - which was turned on 24 hours a day, 7 days a
week - in the pilot data center was 2.7 kW. However, from interviews with the data center manager,
the room is normally unoccupied 20 hours a day. It was assumed that the lights would be turned off
during those hours, so that the electricity required for lighting the data center during use was 0.5 kW.
The improvements made a substantial improvement to the data center efficiency, as shown by
comparing Figure 81 and Figure 82. Because of the 23% reduction in electricity consumption of the
data center, the PUE of the pilot data center after the improvement process was implemented 1.39, a
23% increase from the baseline condition, which yielded a PUE of 1.82. As discussed earlier, the
average PUE for industry data centers was found to be 2.5, and data centers with the most efficient
equipment and no over provisioning of capacity have been able to reach a PUE of 1.6.113 Even
though the pilot data center had a PUE of 1.39 after the pilot was complete, the same caveat exists -
there is no UPS in the pilot data center. The UPS in data centers can consume as much as 25% of
the total electricity required for the data center. 1 4
KLight
Transformer Loss
3%
0.2%
Transformer Loss
Lgtng 2%
1%
Figure 81: Baseline electricity usage for the Figure 82: Post pilot electricity usage for the
pilot data center pilot data center
Additional potential savings not achieved during the pilot were estimated by assuming that 20% of
the electricity consumed by the racks could be eliminated if power management software was utilized
(as discussed in Section 4.2.2.2). This assumption seems reasonable since the constant load of the
building containing the data center shows a decrease in the constant load every December during the
Raytheon holiday break, indicating that racks can be powered off. This additional savings, as shown
in Figure 83, could increase the overall annual savings in the pilot data center to approximately
$83.9k.
7000 !90
-6000 --
$83 9k$89
b 80
$70
~*
i
5000
4000$50 3
Ughtig
0Trns. Loss
3000 $40
+3000 - -
4000
$10
10020
0
Baseline Post Phase I Post Phase Il Past Phase III
(Assumed)
Figure 83: Comparison of electricity used by the pilot data center and annual savings
achieved during different phases of the pilot program
Table 15: Issues identified by data nalysis in the pilot data center
Description Root Cause Short-term Long-term Change
Solution Required
Install light switch Implement lighting
24/7 Lighting No lighting controls and occupancy controls in DESIGN
sensors PHASE
Air is reheat, humidified, Default settings of Disable reheat,
and dehumidified CRACs were used, humidify, and Right size during the
unnecessarily even though the room dehumidify settings DESIGN PHASE
is a closed system on CRAC
Heat Removal Equipment Room was Turn off 1-2 Right size during the
Overcapacity (3.5 times overdesigned CRACs DESIGN PHASE
the IT equipment load)
Cold Air/Hot Air Mixing Poor segregation of Evacuate hot air Improve DESIGN
ColdAir/HoAirMixing hot and cold aisle into plenum PHASE; Intensify
Path hot air must take PAE nesf
Hot Air not returning to Patot tae Create vacuum from DATA CENTER
CRAC for mixing plenum to CRAC MANAGEMENT
In the case of the pilot data center, the room was designed to meet the maximum electrical
requirements of the IT equipment. This is a very common, because the cost of downtime due to a
data center has soared in recent years. 116 To ensure that the data center experiences no downtime,
facility designers simply design the data center to the worst-case scenario - all pieces of IT equipment
consuming the maximum amount of electricity possible. Two solutions exist to correct
overdesigning - rightsizing with provisions and integrated design teams.
Da -ue smiry
IAbbtc ofcearkg acepit dan th i
1 tag
( pier e40) M A20 )
n
118( t iader 0I5 ) - 1-1 1 1
119 (hGtheab
Figure 84: Survey results of what is keeping Figure 85: Example of Staff Alignment for
data centers managers "up at night"'1 8 Energy Efficiency 19
In the case of the pilot data center, the room was designed and built by the facilities organization;
while the day-to-day activities, including purchasing and setting up IT equipment, is managed by the
Therefore, to set up accountability and the correct incentive structure, submetering of high-energy
users, such as data centers, should be implemented to whatever extent possible. The submetering
can be temporary, as it was with the pilot data center, or permanent. The electricity data should be
used to charge the high-energy users for their portion of the electricity bill and to set-up incentives,
such as performance management metrics that affect future salaries and bonuses, to make sure that
those with the ability to create change are motivated to do so. If it is too difficult because of existing
In summary, in order to increase the energy efficiency of a data center over the life of the equipment,
the IT and facilities departments should be better integrated through organizational alignment. In
addition, incentives should drive accountability and a culture of continuous improvement year-to-
year to ensure that right decisions are made not only during the design of the new data centers, but
also during the purchasing of new equipment for existing data centers.
As discussed in the previous section, those who manage the data centers at IIS have no visibility of
the electricity consumed by the data center and are not incentivized to reduce the electricity
consumed. Rather, their performance is based on the amount of downtime and the speed at which
they can get new IT equipment running. However, when the project team - consisting of only
facilities personnel - chose the pilot data center, the personnel responsible for the room showed a
remarkable amount of engagement during the improvement process. The purpose of this section is
to discuss how a high level of employee engagement was achieved, and how the process can be
replicated in other data centers.
The first step of engaging the pilot data center's manager was to raise awareness of data center energy
efficiency by discussing the current state of the pilot data center. In addition, the team provided the
data center manager the total monthly and annual bill for electricity at the Garland IIS site, and how
the electricity bill affected the overhead cost and therefore the ability of IIS to win new contracts. In
addition, the environmental impact of Garland's energy usage was discussed. In addition, benefits of
improving the energy efficiency of the pilot data center, such as additional capacity in the data center
and increased reliability/redundancy of the heat removal equipment, were discussed. Finally, rather
than simply stating what was wrong with the pilot data center, the team - including facilities and the
data center manager - walked through the data center and discussed possible improvement
opportunities. By making the data center manager part of the team, it was much easier to make
improvements once the pilot began. Prior to making any changes, the plan was discussed with the
data center manager in order to cultivate an extremely inclusive team environment.
The second step of achieving employee engagement was gaining public commitment. Research
suggests that a public commitment (as opposed to a one-on-one or private commitment) leads to a
reduction in energy usage. For example, homes that made a commitment to decreasing their energy
use reduced their usage 10-20%, whereas a control group that made only a private commitment
showed no change in their energy usage. 122 In the case of the pilot data center, the team gained
public commitment by holding a meeting between facilities management, the data center manager,
and the data center manager's boss. In the meeting, proposed changes were reviewed, and the team
asked for commitment. One common flaw when achieving commitment is to pressure people in to
commitment, which research suggests does not work. 123 To avoid applying pressure for the pilot
data center, the team created a very informal environment when asking for public commitment.
In order to continue cultivating employee engagement, feedback was provided to the data center
manager and his team. Feedback is essential in order to sustain change. In one study, simply
providing feedback increased recycling by more than 25%.124 However, when feedback was
combined with public commitment, the resulting increase in recycling was even higher - 40%.125 To
provide feedback during the pilot, the electricity consumed by the CRACs were provided before and
after the reheat function had been disabled, illustrating the impact that the project was having on the
data center's electricity consumption, the building's electricity consumption, and the overall Garland
site's electricity consumption.
Finally, the last step of achieving employee engagement is to create a competitive environment. Most
companies and sites have multiple data centers. While the team was unable to create a
comprehensive competitive environment in only six months, one measure that could be used to
create a competitive environment among data center managers is the power usage effectiveness
(PUE). Not only could the PUE be used to create competition, but it also allows a way to prioritize
the implementation of the improvement process across multiple data centers.
What is different between the Garland site 1) Data center attributes like age, layout,
and other site/companies where the pilot number/site, etc vary from site to site
will be implemented? 2) Emphasis on energy conservation
From the answers to the above questions, a three step process was developed that could be followed
to implement the pilot process in both Raytheon and also other companies to ensure that the results
achieved during the pilot could be replicated:
1) Gain buy-in and commitment from leadership for data center improvements
2) Create Rapid Results Teams (RRTs) to implement the improvement process
3) Allow feedback and increased knowledge to be shared amongst the RRTs
Each of these steps will be discussed in more detail in the following sections.
6.2.1 Gain buy-in and commitment from leadership for data center improvements
When reflecting on why the team was so successful in implementing the pilot process, the first thing
that comes to mind is leadership buy-in. Because the leadership team from Facilities, IT, and the
Program agreed upon the need to improve the energy efficiency of data centers, the team was
empowered to make changes quickly, without repeatedly asking for permission. The method for
achieving leadership buy-in was to communicate the importance of data center energy efficiency
through anecdotes and relatable information. As shown in Figure 86, prior to the start of the project,
the leadership was shown the extent to which data centers were affecting their operations cost. To
make the data relatable, a car, in this case the Nissan Versa was used to illustrate how much money
was spent daily on electricity for Garland data centers. In addition, once the pilot was complete, the
information shown in Figure 87 was provided to the leadership team. The savings information not
only recognized the team's success, but also provided the leadership team with a strong incentive -
almost $500k in savings - to continue the implementation of the data center energy efficiency
process.
Figure 86: Example of slide shown to Facility Figure 87: Example of slide show to Facility
Leadership to gain buy-in of importance of data Leadership to keep momentum of data center
centers projects
6.2.3 Allow feedback and increased knowledge to be shared amongst the RRTs
One very important capability required to create a high performing, high velocity organization is
knowledge sharing. 128 In the case of RRTs, the best approach to share information and further refine
the improvement process is through a Lean tool called jishuken (pronounced "jee-shoe-ken"').
Jishuken, an example of which is shown in Figure 88, is the process of first asking RRTs from
different sites to perfect data center improvement processes specific to the data centers at their site.
The impact of the low-cost, systematic approach for increasing the energy efficiency of data centers
was a 23% reduction in electricity consumption, leading to annual savings of over $53k. If other
low-cost measures are implemented, the total reduction could be 34% and the annual savings could
be over $80k. Raytheon and other organizations should roll-out this low-cost process across all
existing data centers to not only increase their energy efficiency, but also to change the energy
consuming behavior of their professional staff.
While these results are impressive, they highlight two problems that many data centers have - poor
initial design that does not include an emphasis on energy efficiency and poor day-to-day monitoring
of the data center's electricity consumption. In order to remedy these problems, management must
utilize incentives and performance management to drive behavior change during the design process.
In addition, similar management techniques must be used to engage employees and set continuous
improvement goals so that the electricity consumption of data centers decreases over time. During
this project, the improvement process was broken down into three phases. This phased approach
allowed the team to test a behavior change process on the IT organization responsible for the data
center. We first highlighted the overall problem (low energy efficiency). Then, we helped the
employees brainstorm improvement ideas, and encouraged them to implement their suggestions in a
manner that allowed us to monitor each action's affects. Finally, we reinforced their positive
behavior by illustrating the impact to both the employees and to the employees' management team.
This simple, well established behavior change model worked to motivate the employees to action.
The next steps for any company that would like to increase the energy efficiency of their data centers
is to roll-out the piloted process systematically using rapid results team who are empowered by IT
and Facilities leadership to make changes. Lean techniques such as jishuken could be used to ensure
feedback is integrated into the piloted process and to maximize the organization's return on data
center energy efficiency in both the short and long-term. Because the improvement process outlined
in this thesis is low-cost, it provides a foundation for any organization that wants to start improving
their data center's energy efficiency. I would encourage organizations that utilize the piloted low-cost
process to not simply replicate the process, but rather to work to continuously improve the pilot
process to yield more impressive results.
8 Bibliography
American Power Conversion (APC). Solutions For: Server Consolidation.
http://www.apc.com/solutions/csegment.cfm?segmentID=3&cid=10 (accessed 12 14, 2009).
ASHRAE. 2008 ASHRAE EnvironmentalGuidelinesfor DatacomEquipment. Refrigerating, and Air-
Conditioning Engineers, Inc. American Society of Heating. August 1, 2008.
http://tc99.ashraetcs.org/documents/ASHRAEExtendedEnvironmentalEnvelopeFinalAug_1
_2008.pdf (accessed December 21, 2009).
Atwood, Don, and John Miner. Reducing Data Center Cost with an Air EconomiZer.Intel Corporation,
IT@Intel Brief, 2008.
Bailey, M, M Eastwood, T Grieser, L Borovick, V Turner, and R C Gray. SpecialStudy: Data Centers of
the Future. New York, NY: IDC, 2007.
Beers, Michael C. "The Strategy that Wouldn't Travel." HarvardBusiness Review, November-December
1996: 18-31.
Blackburn, Mark. "Five Ways to Reduce Data Center Sever Power Consumption." White Paper #7,
The Green Grid, 2008.
Blazek, Michele, Huimin Chong, Woonsien Loh, and Jonathan G. Koomey. "Data Centers Revisited:
Assessment of the Energy Impact of Retrofits and Technology Trends in a High-Density Computing
Facility."Journalof Infrastructure System (ASCE), September 2004: 98-104.
Brealey, Richard, Stewart Myers, and Franklin Allen. Principlesof CorporateFinance. 9th Edition. New
York, NY: McGrawHill/Irwin, 2008.
Carr, Nicholas G. "The End of Corporate Computing." MIT Sloan Management Review 46, no. 3
(2005): 66-73.
Deleon, Ig, and R Fuqua. "The Effects of Public Commitment and Group Feedback on Curbside
Recylcing." Environment and Behavior27, no. 2 (March 1995): 230-250.
Dranove, David, and Neil Gandal. "The DVD vs. DIVX Standard War: Empirical Evidence of
Vaporware." Edited by UC Berkeley: Competition Policy Center. 2000.
Dunlap, Kevin. CoolingAuditfor Identing PotentialCooling Problems in Data Centers.White Paper #40,
American Power Conversion (APC), 2006.
Eaton, Robert J. Getting the Most out of EnvironmentalMetrics. 5 1, 2009.
http://www.nae.edu/nae/bridgecom.nsf/weblinks/NAEW-4NHM88 (accessed 5 1, 2009).
Energy Information Administration. US Climate Zonesfor 2003 for CBECS. 2003.
http://www.eia.doe.gov/emeu/cbecs/climatezones.html (accessed 11 05, 2010).
EPA. Climate Leaders Basic Information. 2010. http://www.epa.gov/stateply/basic/index.html (accessed
01 26, 2010).
-. "The Lean and Energy Toolkit." www.epa.gov/ lean.
http://www.epa.gov/lean/toolkit/LeanEnergyToolkit.pdf (accessed 06 25, 2009).
Evans, Tom. Fundamental PrinaplesofAir Conditionersfor Information Technology. White Paper #57,
American Power Conversion (APC), 2004.
Evans, Tom. The Diferent Types ofAir ConditioningEqupmentfor IT Environments. White Paper #59,
American Power Conversion (APC), 2004.
Fanara, Andrew. "Enterprise Storage Announcement." Letter, ENERGY STAR Program, U.S.
Environmental Protection Agency, Climate Protection Partnerships Division , Washington, 2009.
Google Maps. http://www.maps.google.com (accessed 06 17, 2009).
Google. The CaseforEnergy - ProportionalComputing. Luiz Andre Barroso and Urs Holzle. December
2007.
http://dsonline.computer.org/portal/site/computer/index.jsp?pageID=computer-levell&path=co
mputer/homepage/Dec07&file=feature.xml&xsl=article.xsl (accessed December 14, 2009).
Greenburg, Steve, Evan Mills, Bill Tshudi, Peter Rumsey, and Bruce Myatt. "Best Practices for Data
Centers: Lessons Learned from Benchmarking 22 Data Centers." Proceedingsof thE ACEEE Summer
Study on Energy Effiaeng in Buildings in Asilomar,CA. Asilomar: ACEEE, 2007. 76-87.
Hennessy, John, and David Patterson. ComputerArchitecture:A QuantitativeApproach,. 4th Edition. San
Francisco, CA: Morgan Kauffman, 2007.
Hirstius, Andreas, Sverre Jarp, and Andrzej Nowak. Strategiesforincreasing data centrepower efien.
CERN Openlab, 2008.
HP. HP Power Advisor Utiliy:a toolfor estimatingpower requirementsforHP Proilantserver systems.
Technology Brief, HP, 2009.
Hughes, Ron. CaliforniaData Center Design Group. Data Center Journal Online. May 2005.
http://www.cdcdg.com/articles/9/ (accessed December 7, 2009).
Hwang, Douglas. "Data Center Power Consumption Reductions Achievable with Liquid Cooling: A
Case Study." Thesis, Mechanical Engineering, Tufts University, Somerville, 2009.
Joint Air Force - Army - Navy. "Phsyical Security Standards for Special Access Program Facilities."
Manual, 2004, 50.
jme~r
Baseline
"Constant
Constan Dil PT
P lug
Iu7g Loadt
L
Appendix II - Summary of Alternative Efficiency Scenario Assumptions1 3 0
IBM Calculator for the blades and modular product lines (x86
architecture server systems and rack storage), the tool can
be downloaded at:
http://www-03.ibm.com/systems/bladecenter/resources/
powerconfig/
.......... 1..*
IF,
4c
-- tC
To understand the sensitivity of the assumed chiller electricity consumption per ton of chilled water, a
sensitivity analysis was performed. As seen in Figure 89 and Table 17, the building's electricity
consumption decreased by 133 kWh from the week prior to the start of the improvement process to the
time in which all improvements had been made. However, when only the constant load, which includes
data centers, is analyzed, a drop from about 925 kWh to 825 kWh is observed. Therefore, a value of 100
kWh can be assumed to be the amount of electricity avoided in the data center due to the improvement
process.
1200
S Phase II
Baseline Changes
Completed
Total Electricity 167,977 145,6541
Consumed (kWh)
200 Phase I Phase I PhaseIl Phase Ir Average Outside Air 62.7
Changes Changes Changes Changes
63.6
Temp (F)
Made Completed in-process Compleled
0 Total Reduction of n/a 22,323
Electricity (kWh)
It is notable that when excluding the chiller's contribution, the Daily Reduction in n/a 3,189
pilot data center accounted for 222 kWh in the "Baseline" time Electricity (kWh)
period (about 24% of the total constant load) and 170 kWh in the Reduction in
"Phase II Changes Completed" time period (about 21% of the
total cnstant llod)
Electricity (kqh) n/a 133
I Missing data on electricity consumption for
Figure 89: Building Electricity before and after Saturday was assumed to be the same as Sunday
Phase I changes were implemented
Of the 100 kWh avoided, 52.2 kWh can be attributed to the CRAC and the lighting improvements, which
leaves 47.8 kWh of avoided electricity that is not accounted for. As mentioned previously, the data center
owner copied the improvement process in other data centers during the completion of the pilot, which
would have led to some of the decrease. Also, it is reasonable to assume the "typical value of 0.53
kW/ton" is too low for the Garland, TX site. As shown in Table 18, when the assumed 0.53 kWh/ton is
used, the savings is only 7.7 kWh. To explore what chiller electricity rate is required to account for the 47.8
kWh of avoided electricity, Excel Solver was used to determine what value would be required to achieve
47.8 kWh. A value of 3.30 kWh/ton was calculated, which is unreasonably large.The correct value is
probably between 0.60-0.85 kWh/ton. While 0.53 kW/ton is too low, it was nonetheless used in the
analysis in order to be conservative with the savings calculations.
Table 18: Calculations to determine sensitivity of Chiller Electricity Consumption
Excluding Chiller Electricity
Baseline excluding Post Pilot excluding Change
Chiller Chiller
Electrical Load (kW) 221.9 169.5
Cost of $0.102
Electricity/kW
Building Electricity Usage Breakdown for 2008 Total Electricity in 2008 for Building
1,000.000 HVAC
900.000
19%
800.000
700.000
600,000
500.00
I
400,000
100.000
0 EEEEEEEEE
UCI0t Plug
L~d P~gLoad
M370606. MHVAC
30,000 1300
-250 E
25,000 - 00
12006
20,000. 9
00
1~k.
150 O 1100
15000
100
906
Figure 92: Building's electricity consumption as Figure 93: Building's electricity for two
a function of weather from 2006-2009 weeks in 2009
Appendix VII - Specific lessons learned and implementation plan for Raytheon
Lessons Learned
- No one agrees on data center "ownership" within closed 11S areas
- Program owns compuLing equipment
- Facilities owns layout and HVAC equipment
- IT owns standardizaion
2. Leadership team assigns anprovement team for every IS data centers, comprised of contact by
01/3112010
- Fee
- Progr Teams
6. Set up surveying schedule to ensure improvements do not deteriorate over time by 01/01/2011