2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

66

description

Agility and cost control are critical issues in today’s one-size-does-not-fit-all data centers. This presentation from Emerson Network Power and DPR Construction, presented during the Feb. 22, 2011 Data Center Pulse modularity design summit, is focused on the latest developments in modularity, multi-tier, free cooling and other important components of a flexible data design.

Transcript of 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Page 1: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design
Page 2: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Agenda

• Introduction• Project Delivery • Cooling• Power• Economics of Efficiency• The Right Team

Page 3: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Project Delivery

Page 4: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Project Delivery

• Local Team• Schedule• Cutting Edge, Collaborative, Conceptual Precon• Base Building Coordination• LEAN Manufacturing and Construction• Sustainability

Page 5: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Project Delivery

• Local Team– No one understands the market and the

municipalities better.– Quickest decisions, (entire team can be face-to-face

at any time).– Currently performing the Base Build.– Best choice for sustaining work/future phases.– Team has Nationwide reach and offices.

Page 6: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Schedule

• APS Power (Ready to Roll Now)• Design Process (now-Dec 1st)• Permitting (Dec 1st-Feb 2nd)• Central Plant (Jan 22nd- April 15th)• Data Center Buildout (Jan 22nd-April 29th)• Commissioning (April 18th-May 25th)• All Complete by May 26th

Page 7: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Schedule

• Design Process– Immediate Design Charette with eBay– Conceptual/Schematic Design (five weeks)– Design Development (three weeks)– Construction Documents (four weeks)– Ready to submit to City for Review- Dec. 1st.

Page 8: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Cutting Edge, Collaborative, Conceptual Precon

Customer Service

PRECON

Communication

Conceptual Alternatives

Competitive Procurement

Collaboration

•Innovative•Flexible•Scalable•Modular•Capital Expense•Oper. Expense•Efficient•Clean•Sustainable•Modern•Available•Maintainable

Page 9: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Base Building Coordination

• This is the Team already performing the Base Build.• Coordination has already been on-going.

– Example: Power needs for Data Center being coordinated now to eliminate rework.

• No one understands the permitting and inspection process like we do.

Page 10: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

LEAN

• Manufacturing LEAN Six Sigma driven to reduce all waste– Materials, energy, time, manpower– Process optimization, automation, and control– Quality systems

• Construction LEAN – Last Planner – Collaboration– Just in Time Delivery

Page 11: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Sustainability

• Emerson’s environmental stewardship efforts include: – Developing technologies that help address environmental challenges– Applying cost-effective and energy-efficient manufacturing processes

• Innovative Technologies for a Greener World– Enabling efficient use of energy– Supporting cleaner forms of energy production– Providing technologies that monitor and limit emissions

• Energy Saving Operating Initiatives– Replacing shop lighting with fluorescent, add motion sensors, turn off sodium lights in Weld Shop– Replacing Bondrite cleaner with NT-1 in Powder Paint– Using heat recovery from compressor for powder paint area make up air– Adding VSD to 70HP compressor, replacing standard electric motors with premium efficiency

• Environmental Operating Compliance– Annually assess the environmental compliance of each Emerson facility– Adhere to industry standards and initiatives - WEEE, RoHS, EPEAT, CSCIhttp://www.emerson.com/en-US/about/corporate-citizenship/csr-sustainability/Pages/default.aspx

Page 12: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Sustainability

• DPR builds ALL projects as if they were LEED projects. It’s simply the right thing to do.

• Unmatched LEED Data Center Experience• Maximum construction recycling• APS Energy Efficiency Incentives- currently

coordinating with APS.• 28 Local LEED AP, over 350 nationally• ASU Poly Local LEED Gold Certified (DPR/ESD)

Page 13: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Cooling

Page 14: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Central Plant

• Variable Primary System design• N+1 Redundancy• Two cooling loops (55°F Loop & 78°F Loop)• Initially have cooling capacity of 2MW of IT

load in each loop• Could be increased to total of 5MW of IT

load

Page 15: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Mechanical Roof Plan

Page 16: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Central Plant Floor Plan

Page 17: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Mechanical Flow Diagram

Page 18: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Cooling System 55°F Loop

• Utilizes variable speed chiller• Chiller at 0.440 kW/Ton at peak cooling load (Max.

summer temperature)– COP of Central Plant = 6.25

• Chiller at 0.225 kW/Ton at peak cooling load (Mild outdoor condition)– COP of Central Plant = 10.2

• 34% of time on free cooling mode (Chiller will be off)– COP of Central Plant = 28.8

Page 19: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Cooling System 78°F Loop

• Utilizes Cooling Tower/Heat Exchanger• 97% of time can provide 78°F water or lower

– COP of Central Plant = 28.8• 3% of time need assistance from 55°F loop

– COP of Central Plant = 21.3

Page 20: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Design Highlights

• Liebert® XD™ platform (Liebert® XDS™ server cooling rack) enables direct liquid cooling

• Total site can support above 4 MW of IT load with high density cooling– First floor utilizes Liebert® XDR™ rack door cooling modules

and Liebert XDS, enabling 4 MW of IT load– Additional containers on the roof can push IT Load towards

6MW• PUE of 1.11 for best scenario- All Liebert XDS• For more information on all Liebert XD systems, visit

our website

Page 21: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

0102030405060708090

100

0 10 20 30 40 50 60 70 80 90 100

% H

P a

nd

To

rqu

e

% Speed

HP

Torque

Flow

Base Cooling Provided By Liebert® CW™ 181 CRAC Units

• Features and Benefits– Energy efficiency

• EC Plug Fans• Team work mode prevents

units from competing with each other

• Variable Speed Drive

– Free Cooling – Air Economizer Option

– For more information: Liebert CW and air economizer

Page 22: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert CW 181 - Team Work Mode

In Team Work mode up to 32 units can be connected to each other and controlled by a master unit.

Page 23: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert Monitoring Interface with SetpointModbus 485BACnet IP (to Delta Controls)

XDF

Extender Card

Liebert CW181

Extender Card

BMS (Delta Controls)

XDF

Liebert® MPX™ Adaptive Rack PDU

Liebert UPS

SNMP

Liebert XDP Liebert CRV

Page 24: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

• This control logic uses Artificial Intelligence techniques including “fuzzy logic” and “expert systems” methods– This control is similar to the actions that a human operator would take to

maintain precise, stable control. • This control will maintain precise, stable control and increase reliability by

reducing component cycles and repositions– Eliminating oscillations extends valve seat life

Reliability Through Intelligent Control

Page 25: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert XD – Provides a Migration Path for the Future

Base Infrastructure

(160 kW)

Embedded Cooling

Tested 80 kW (100% redundant)

Capable over 100 kW

Rear Door

Cooling Modules

10-35 kW Future Product Configurations

Liebert XDRLiebert XDS20

Direct Cooling

Technology

(microchannel coldplate)

Liebert XDS40

Liebert XDC or Liebert XDP

Page 26: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert® XDR™ Rear Door Cooling Module• Rated at 10, 20 & 32 kW

– Adds 6” to rack depth• Makes the rack room neutral• Relies on the server fans for airflow across the

coils – no additional fans are required• NO electrical connections• Energy efficient!

– TOTAL fan / pump energy 100W/ 20kW cabinet• Fluid hinge technology allows 120° door swing,

easy access to rack components• Allows for cascading configuration

Page 27: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert XDR at Sandia National Labs

30°

27.5°

25°

22.5°

20°

Temperature (°C)

Used for Projects for the National Renewable Energy Lab (NREL)

70 racks of Sun HPC equipment generating 32kW each Very efficient power-wise: 1.27 PUE “Red Sky should really be called Green Sky. This

machine is the most energy efficient HPC system we have deployed to date.”

- John Zepper, Sr. Manager of Computing Systems

Actual Red Sky temperature readingsAir leaving the last rack is colder than that

entering the first rack (on the left).

Dave Martinez, Facilities CoordinatorSandia National Labs

Page 28: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert XDS with Clustered Systems Technology

• Rack is 45U tall by 800mm wide by 1200mm deep.

• 36 - 1U server slots with cold plate cooling (tested to 600W per 1U slot).– Slots are on 2 inch ( 50mm) centers.

• Three non-cooled server slots are available for low power devices like switches, etc.

• Rack can be configured for a cooling capacity of 20kW.

• Could be made to be “Rack and Roll” compatible.

• View the Liebert XDS video

Page 29: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Energy Consumption Comparisons

Liebert XDS consumes less power than the server fans alone.“Free-er than Free”

0.000

0.050

0.100

0.150

0.200

0.250

0.300

0.350

0.400

Free Cooling Liebert XDS w/water Econ

Kw co

olin

g pe

r kw

of I

T lo

ad (p

roce

ssin

g) Cooling Power - PhoenixCW - 60°F

Liebert XD modules Liebert XDP

CRAH fans

Chiller Condenser PumpsCooling Tower (kW)Chiller Pumps (kW)Chiller

Server fans

Assume any infrastructure to exchange the air.

Page 30: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

First Floor Future Scenario A – Liebert XDR

• With approximately 3,000 SF of white space available for IT equipment, the IT infrastructure deploying the – Liebert XDR20 would allow for :

• Approximately 185 IT racks • A total cooling capacity of 3,700 kW (1,233

W/SF). • Cooling system would consist of 185 Liebert XDR

units with 24 Liebert XDP units. • Corresponding PUE would be 1.16.

– Liebert XDR32 would allow for:• Same number of racks (185) with 37 Liebert XDP

units.• A total cooling capacity of 5,920 kW (1,973

W/SF). • Corresponding PUE would be 1.16.

Page 31: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

First Floor Future Scenario B – Liebert XDS20

• With approximately 3,000 SF of white space available for IT equipment the IT infrastructure:– Deploying the Liebert XDS20

would allow for approximately 155 IT racks.

– Total cooling capacity of 3,100 kW (1,033 W/SF).

– Cooling system will consist of 155 Liebert XDS units with 20 Liebert XDP units.

– Corresponding PUE would be 1.11.

Page 32: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

First Floor Future Scenario C – Liebert XDS40

• With approximately 3,000 SF of white space available for IT equipment the IT infrastructure:– Deploying the Liebert XDS40

would allow for approximately 155 IT racks.

– Total cooling capacity of 6,200 kW (2,067 W/SF).

– Cooling system will consist of 155 Liebert XDS units with 39 Liebert XDP units.

– Corresponding PUE would be 1.11.

Page 33: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Two IT Container Alternatives – Modular Expansion

eBay can pick and choose between the following depending on their application requirements

1. Emerson SmartMod – Liebert XDS - Uses 78°F Chilled Water– Most efficient, Tier 2 Cooling (No redundancy) – Requires server OEM commercialization

2. Emerson SmartMod – Liebert XDR – Uses 55°F Chilled Water– Tier 3 (N+1) Redundancy– Compatible with any server manufactures rack– Uses Liebert XDR and Liebert CRV for N+1 Redundancy

Page 34: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Emerson SmartMod – Liebert XDS

• Single Container 42’ x 11’• Based on 78°F chilled water from chiller plant• 10 Racks at 20 kw each. • Qty 1 - 800 mm network rack• PUE = 1.11

– Cooling plant is cooling tower and circulating pumps– Liebert XDP is only 1.1 kw each– IT Load is 200 Kw

Page 35: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert® CRV™ Row-Based Cooling (In Containers)

• Temperature and Humidity control• Energy efficient load matching

– Digital scroll compressor– 20-100% cooling capacity modulation– Variable speed EC plug fans

• Multiple cooling configurations– Air-cooled, 20 & 35 kW– Water-cooled, 20 & 33 kW– Chilled water, 40kW

• Refrigerant– R410A

• Controls – Liebert® iCOM®• Applied as “Row-based” cooling

– Rack width – 24”– Footprint: 600w x 1100d x 2000h (mm)– 24w x 43d x 78h (in)

35

Page 36: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Emerson SmartMod – Liebert XDR

• Single Container 43’ x 11’• Based on 55°F Chilled Water from chiller plant• 10 Racks 20 kw Average, 23 KW peak N+1 redundancy

– Redundancy supplied by Liebert CRV units– Liebert CRV units are cooled with 55°F water

• PUE = 1.24 (115° F day)– PUE based on Chiller plant COP of 6.25– Chiller plant has an integral fluid economizer providing some free cooling 34.0%– Liebert XDP units are only 1.3 kw each. The Liebert CRV units are 35 KW– IT Load is 200 KW

• View standard configurations

Page 37: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

CFD Analysis of Emerson SmartMod- Liebert XDR

Page 38: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Water-Based Rear Cooling Doors

• We did not quote a water based solution for a variety of reasons.– Water door solution limits future

growth to 30KW per rack, maximum.

– Minimal energy savings by not having a secondary loop.

– Leaks with no secondary loop can be catastrophic.

• Refrigerant-based infrastructure provides a migration path to 40KW and beyond.

Knurr CoolDoor

Page 39: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

39

- Phoenix Data CenterBrad Brindley - Liebert – Sr. Western Region Manager

Page 40: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Electrical System Demand

• Initial IT Load: 2 MW• Total Estimated Load: 2.8 MW• At 0.8 pf (Emergency Generator Criteria): 3.5 MVA

~ 4200A, 277/480V

• Shell Building (SES-1) 2.5 MVA, 3000A, 277/480V– Add 2 MW / 2.5 MVA Diesel Generator

Page 41: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Modularity and Flexibility

• Modularity– 2ND system (SES-2)– 2.5 MVA, 3000A, 277/480V with – 2 MW / 2.5 MVA diesel generator

• Flexibility– Spare Capacity: 1.2 MW / 1.5 MVA

Page 42: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

UPS Sizing

• Initial – 2 MW of IT Load– Split the load between (2) systems

– UPS -1 1350 KW / 1500 KVA (2) 675 kW / 750 kVA UPS Modules

– UPS -2 1350 KW / 1500 KVA (2) 675 KW / 750 KVA UPS

Modules

Page 43: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Full Buildout

• Full Buildout– 4 MW+ IT Load– 5.6 MW Total Load– 7 MVA (Utility Company Limit)

• Modularity– (3) 2500 kVA Systems each with:

• 2500 KVA 12.47 – 480 / 277V Transformer – 7.5 MVA Total• 3000A, 277 / 480V SES• 2 MW / 2500 KVA Diesel Generator – 6 MW / 7.5 MVA Total• 3000A ATS• 1350 KW / 1500 kVA UPS – 4.05 MW

Page 44: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Flexibility of the Proposed Configuration

Page 45: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Agenda - Liebert UPS / Liebert® MB™ Busway

• Liebert 2mW-6mW UPS Solution• Advantages of transformer-based UPS• Intelligent Eco-Mode™ for greater efficiency• Liebert MB Modular Busway• Service and Support

Page 46: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

46

System Voltage Pros and ConsPROS CONS

480 – 208/120600 – 208/120

• Handles 120V & 208V loads• Uses standard 240V panel board &

breakers• Reduced wiring cost, 3W vs 4W• Transformer provides harmonic, noise

and fault current reduction

• 1.5-2% transformation energy loss• 208V requires 2 pole breaker• Reduces the number of poles

480 – 240/139600 – 240/139

• No neutral required• Uses standard 240V panelboard &

breakers• Higher energy efficiency

• 2-3% transformation energy loss• Can’t power 120V equipment• 240V requires 2 pole breaker• Reduces the number of poles

480 – 480480 – 480/277

• No transformation energy losses• No neutral required (unless 277V

loads)

• Can’t power 120V or 208V loads• Limited number of loads operate at

480V or 277V• Requires 600V panelboard & breakers• Higher fault current at the load: 35-

150kAIC

480 – 415/240 • No transformation energy losses• 240V load requires 1 pole breaker• More useable pole spaces• Higher energy efficiency

• Can’t power 120V equipment• Requires 480V panelboard& breakers• Higher fault current• Requires UPS Maint Bypass Xfmr

415 – 415/240 • No transformation energy losses• 240V load requires 1 pole breaker• More useable pole spaces• Higher energy efficiency

• Can’t power 120V equipment• Requires 480V panelboard& breakers• Higher fault current

Page 47: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

47

TCO Tradeoff ConsiderationsUPS / DISTRIBUTION VOLTAGE

Attribute 600- 208/120 480-208/120 415-415/240 480 – 240/139 600 – 240/139

Installation Costs Smallest conductor sizeNo neutral requirement

No neutral requirement

Neutral requiredSpecial building TX . UPS bypass, or Auto TX requiredPDUs do not require transformers

No neutral requirement

No neutral requirementSmallest conductor size

Utilization of infrastructure

Best utilization of infrastructure

Good utilization of infrastructure

Least utilization of infrastructure

Good utilization of infrastructure

Best utilization of infrastructure

PDU Distribution Standard low voltage (250V) panels and breakers

Standard low voltage (250V) panels and breakers

Requires larger (600V) panel and breakers at higher first cost

Standard low voltage (250V) panels and breakers

Standard low voltage (250V) panels and breakers

Maintenance Costs Better utilization may reduce quantity of systems to meet mission

Better utilization may reduce quantity of systems to meet mission

No PDU TX to maintain (may be elsewhere?)More complex maintenance practices

Better utilization may reduce quantity of systems to meet mission

Better utilization may reduce quantity of systems to meet mission

Operating Efficiency( Combined PS & TX)

Baseline Baseline Best ~2.0% improvement

Better~ .6% improvement

Better~ .6% improvement

Page 48: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

1+N Parallel System3 Wire Bypass

3 Wire Maintenance Bypass3 Wire Load

Optional BypassInductor

Optional BypassInductor

AC

BackfeedDevice*

3P

3P

3P

3P

3P

3P

3P

3P

3P* *

AC

3P

AC

3P

3P

E

E

MBB

3PMIB

optionload bank brkr

inv onlyload bank

CriticalLoad

Refer to LNArectifer dwgs forinput and rectifier

section details

Refer to LNArectifer dwgs forinput and rectifier

section details

Tout

Tout

Lout

Lout

Lbyp

Lbyp

CB2

CB2

MOB -x

MOB -x

BFB-xBFB-x

Liebert® NXL™ System Configuration

Single-Module Parallel System (Distributed Static Switch)

Page 49: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Design Philosophy - Transformer-Based UPS

• Emerson Network power offers both transformer-free and transformer-based UPS

– We feel there are appropriate applications for both– Our UPS systems for the SMB space are transformer-free– Our UPS systems for the enterprise space are transformer-based

• We made the decision to go transformer-based on our enterprise UPS for availability– Galvanic isolation at the input and output for DC fault isolation– Output isolation safely protects the critical load and makes for robust fault

management– Transformer-based UPS rectifiers use ultra-reliable and efficient SCRs and

simple inverters in open (or closed) delta configurations• A comparable transformer-free UPS uses high voltage DC and has a lot more

semiconductors due to the need for an additional buck-boost charger in the rectifier and neutral management in the inverter

• The semiconductors in a transformer-free design see more of their operating limit than a transformer-based design

49

Page 50: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Transformer-Based UPSMIB

OutputMBB

3P

CB2

BFB

E

BIB

EG

FBO

AC

FBO

AC

GEC

MBJ

N N

EG

A

Trap Disconnect

CB1

12P isolated

12P non isolatedTo Batteries

- Two separate input AC sources for higher reliability and maintainability. A transformer-based UPS can be used to create and maintain a separately derived source, meaning poor grounding and long neutral conductors are eliminated. -Common mode noise/voltages are significantly reduced via isolation. -Safety and noise grounding and references all the way back at the service entrance especially with switched neutrals are maintained. -Better ground fault protection/coordination with transformer-based UPS.-Lower arc flash energy of faulted transformer based systems.-An input isolation transformer-based UPS provides better personnel safety for open rack battery applications.- Less Component Count – Higher Reliability-Pulse paralleling advantages with two separate input AC sources

No Fusing

• Bypass can be connected to separate utility source• HRG Option

Input Isolation

Battery and DC Bus

Isolation

Neutral-Gnd Management;Low Common Mode Noise;

Separately Derived Source

OthersUse Y

Page 51: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert NXL Intelligent Eco-Mode™

Normal Mode

51

Batt

Rectifier Inverter Load

Bypass

Intelligent Eco-Mode– STANDARD Firmware intelligence

built into every Liebert NXL Single Module System (SMS)

– Increases efficiency by idling the inverter while the load is fed by the bypass.

– If poor quality conditions are detected, unit automatically switches to full double conversion mode.

Multi-Module Systems– Expected availability- Fall 2010

– Existing 1+N and N+1 systems will be flash upgradeable Eco Mode

Batt

Rectifier Load

Bypass

Inverter

Page 52: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert NXL Intelligent Eco-Mode™

52

Key Features of Intelligent Eco-Mode– User can set acceptable voltage, frequency and slew rate limits– User determines schedule of Intelligent Eco-mode operations on a daily or hourly

basis– Eco-mode will automatically restart based on user defined delay– Remote suspension of Intelligent Eco-mode available through Remote Contact

interface– Special HMI shows potential energy savings by using Eco-Mode– Efficiencies up to 97% possible

90.0%

91.0%

92.0%

93.0%

94.0%

95.0%

96.0%

97.0%

98.0%

99.0%

100.0%

30.0% 40.0% 50.0% 60.0% 70.0% 80.0% 90.0% 100.0%

750kVA NXL Efficiency

Eff EcoMode Efficiency

Page 53: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert® NXL™ UPS Intelligent Paralleling™

• Intelligent Paralleling– Firmware intelligence built into all

future SCC systems.– Increases the efficiency of an N+1

and 1+N system by turning off redundant module inverters

– Maintains user programmed redundancy and distributes off-time equally between modules

– Maintains battery charging through the rectifier at all times

• What about existing units?– Flash upgradeable on previously sold

1+N and 1+N systems

53

System Control Cabinet (SCC)

Rectifier Inverter

Batt

Rectifier

Batt

Rectifier

Batt

Rectifier Inverter

Batt

Bypass

Inverter

Inverter

Load

Page 54: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert® MB™ Modular Busway - Overview Key Features:

– UL 857 - 400A bus rated 35kAIC– 5’ &10’ Plug-in (12 plugs per 10ft)– Cost effective for low amperage– 200% Neutral– Comprehensive offering of standard fittings

Construction:– 2 Piece steel enclosure– Insulated Tin Plated Copper bus bars– Snap together joints (Maintenance-free)– IR windows for easy scanning of joints

Available Ratings:– 100, 225, & 400Amp (Tin Plated Copper)– Fusible Bus Plugs: 30A – 100A– Circuit Breaker Bus Plugs: 15A – 100A– Receptacle Bus Plugs: Standard & Custom

Page 55: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

55

Liebert® MDC20 – Liebert® XDS™ Electrical

Page 56: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert Global Services: Field Organization- 637 Associates

- 422 Engineers throughout the US providing direct UPS/Power factory service- 215 Associates providing backup technical support and customer response

Las Vegas-1 CE

Phoenix- 8 CE’s

Service and Support

Page 57: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

PUE Worst Case and Typical Month

115 DegF Design

DayPhoenix

MildPhoenix

EconoPhoenix Average

115 DegF Design

DayPhoenix

MildPhoenix Average

115 DegF Design

DayPhoenix

MildPhoenix

EconoPhoenix Average

Percent of Time 5% 61% 34% 3% 97% 5% 61% 34%

Equivalent IT Load (KW) 2000 2000 2000 2000 2000 2000 2000 2000 2000 2000 2000

Chiller Plant COP 6.3 10.2 28.8 16.3 21.3 28.8 28.6 6.3 10.2 28.8 16.3Chiller Plant Electrical (KW) 320.0 196.1 69.4 159.2 93.9 69.4 70.2 320.0 196.1 69.4 159.2CW181 (KW) 180.6 180.6 180.6 180.6

XDS Container Cooling (KW) 31 31 31XDR Container Cooling (KW) 31 31 31 31

UPS Losses (Econo) (KW) 80 80 80 80 80 80 80 80 80 80 80

Distribution Losses (KW) 40 40 40 40 40 40 40 40 40 40 40

Total Energy Consumption (KW) 2471 2347 2220 2310 2245 2220 2221 2621 2497 2370 2460

PUE 1.24 1.17 1.11 1.16 1.12 1.11 1.11 1.31 1.25 1.19 1.23

Note:Calculations Use BIN Data for Phoenix, Arizona

Liebert MDS-XDR Liebert MDS-XDS Liebert CW181 (CRAC)

Page 58: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Total Cost of Ownership

Operational Cost Based on $0.07 /KWh

Total IT Load First Floor (KW) 1240Total IT Load Container (KW) 760Total Power- First Floor (KW) 1525 $ 935,252.64 Total Power- Containers (KW) 844 $ 517,295.52

Total PUE 1.18Total Operaional Cost / yr $ 1,452,548.16

Total Cost of project $ 17,138,046.00

TCO first year = $ 18,590,594.16

Note: TCO depreciates the entire project cost over the first year. No energy rebates are calculated.

Page 59: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

The Right Team

Page 60: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

The Right Team

• Cutting-edge Precision Cooling, Power, and Infrastructure Container Solutions

• Efficient, flexible Power and Cooling Infrastructure

• Leading Local Mission Critical Designer and Builder

• Committed to Collaborative Integrated Delivery• Flexibility, Modularity, Efficiency

Page 61: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Q & A

Page 62: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Transformer-Based UPSMIB

OutputMBB

3P

CB2

BFB

E

BIB

EG

FBO

AC

FBO

AC

GEC

MBJ

N N

EG

A

Trap Disconnect

CB1

12P isolated

12P non isolatedTo Batteries

- Two separate input AC sources for higher reliability and maintainability. A transformer-based UPS can be used to create and maintain a separately derived source, meaning poor grounding and long neutral conductors are eliminated. -Common mode noise/voltages are significantly reduced via isolation. -Safety and noise grounding and references all the way back at the service entrance especially with switched neutrals are maintained. -Better ground fault protection/coordination with transformer-based UPS.-Lower arc flash energy of faulted transformer based systems.-An input isolation transformer-based UPS provides better personnel safety for open rack battery applications.- Less Component Count – Higher Reliability-Pulse paralleling advantages with two separate input AC sources

No Fusing

• Bypass can be connected to separate utility source• HRG Option

Input Isolation

Battery and DC Bus

Isolation

Neutral-Gnd Management;Low Common Mode Noise;

Separately Derived Source

OthersUse Y

Page 63: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

63

Emerson SmartMod – XDR Electrical

Page 64: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

Liebert XDS Connections

Over head manifold suppliespumped refrigerant from Liebert XDP

Cam levers activates engagement and disengagement of serverRefrigerant connection points

at top of the rack on left side. All plumbing runs along left side.

Page 65: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

• Teamwork Mode 1– “Master” unit determines

how intensely (none, partial, full) to perform an operation

• Teamwork Mode: 2– “Master” unit defines

operation– Each unit determines how

intensely to perform the operation (none, partial, or full capacity)

CoolingHalf Capacity

Cooling Half Capacity

CoolingHalf Capacity

CoolingHalf Capacity

CR

AC AC

CR

AC

AC

Unit-to-Unit network established

CR

AC AC

CR

AC

AC

OffCooling

Half Capacity

CoolingFull Capacity

CoolingFull Capacity

CR

AC

CR

AC

CR

AC

CR

AC

AC

Unit-to-Unit network established

Energy Efficiency - Team Work

Page 66: 2011 Data Center Pulse Modularity Design Summit: Modular Data Center Design

© Clustere

d Systems Company, Inc

eBay Server Conversion

• Fans and heatsinks removed, replaced with heat risers.

• Lid treated with encapsulated Thermal Interface Material

• Fan area now available for additional CPU/memory devices