Transcript Slide 1

PG&E and Altera
Data Center
Energy Efficiency Project
PG&E and Altera:
A History of Energy Efficiency
• After hours cooling project
• Chiller VFD retrofit
• CDA compressor replaced with
VFD CDA compressor
• Data Center Efficiency Project
PG&E and Altera:
A History of Energy Efficiency
VFD CDA
Compressor Retrofit
Project Cost: $110k
ROI: 3 years
PG&E Rebate: $15k
Annual Savings: $32k
After Hours
Cooling Project
Project Cost: $78k
ROI: 1.1 years
PG&E Rebate: $36k
Annual Savings: $39k
Chiller VFD Retrofit
Project Cost: $139k
ROI: 3.8 years
PG&E Rebate: $31k
Annual Savings: $29k
Altera Data Center Energy
Efficiency Project Objectives
• Keep servers between
68° and 77° F (ASHRAE)
• Reduce energy use
• Accommodate server growth
• Increase server and data center
reliability
Data Center Layout
Data Center Before Improvements
Portable AC unit
Two Interests Meet
• Altera had a temporary cooling unit in
place to serve a ‘hot spot’ and was
looking at ways to handle planned
load increases.
• In PG&E’s territory, improving data
center energy efficiency by 15%
would save 100 GWh of electricity –
the equivalent of powering 15,000
homes for a year or taking almost
8,000 cars off the road.
Why is Airflow a
Problem in Data Centers?
•
•
•
•
Wasted energy
Wasted money
Less capacity
Less reliability
Why Do These Problems Exist?
In a typical data center…
• Only 40% of AC air is used to cool servers
Robert 'Dr. Bob' Sullivan, Ph.D. Uptime Institute
• HVAC systems are 2.6 times what is actually
needed
Robert 'Dr. Bob' Sullivan, Ph.D. Uptime Institute
• Not a big concern when power density was
low but it continues to grow (about 150w/sqft)
• The culprit: allowing hot and cold air
to mix
First Step:
Assess Current Situation
• Data loggers placed in the inlet
and discharge air streams of each
cooling unit.
• Four loggers placed in each cold aisle.
• Current transformers installed on the
electrical distribution circuits feeding the
two roof-top condensing units.
• Total rack kW load was recorded to
establish baseline.
Findings
• Temperature across cooling units
ranged from 12° to 18°F.
• Temperature variance of up to
14° degrees from one server to
another.
• Approximately 45 kW could be saved
in theory if air flow was ideal
Second Step: Implement
Ideas from Meeting with PG&E
• ALL servers must be in hot/cold
aisles (HACA)
• Blanking plates between servers
• Strip curtains at ends of aisles
• Remove perforated tiles from
hot aisles
• Partitions above racks
Third Step:
Altera Adopts Changes
• APC in-row coolers installed
• Temporary cooling unit removed
• Blanking plates added
• Installed strip curtains to separate the
computing racks from the telecom area
• Shutting off CRAC unit
Addition of APC IRCs’
Simple changes, big benefits
Simple Changes, big benefit
Final Measurement and Review
• Even after Altera made all of these
changes, excess cooling capacity
still existed.
• PG&E recommended shutting down a
second CRAC unit, thus putting all
primary cooling on chilled water units.
Altera’s New and
Improved Data Center
• Temporary mobile cooling unit gone
• Two CRAC units shut off
• Server temperature variance a
mere 2° F
• Net electricity reduction of 44.9 kW
• Annual energy savings of 392.9 MWh
• Overall energy savings: 25%
Moral of the Story?
• Improving airflow is a safe and
sensible strategy to simultaneously
make data centers greener, more
reliable, higher capacity, and more
economical to operate.
• To achieve results such as Altera’s, it
takes teamwork between IT, Facilities,
HVAC experts, and PG&E.