architecting the green data center - e-techservices...affording the next-gen data center legacy...
TRANSCRIPT
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
1
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 1
Architecting the Green Data Center
Bill Dufresne Application Networking Services
CSE-III, CCIE, CISSP
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 2
Political
Social
Health
Technology
Blue State, Red State Green Party Regulatory Compliance?
Trader Joe’s Whole Foods Walmart
Hybrid’s Bottled Water
Blades Scalable UPS Fuel Cells
Cost Centers Selective Accounting
YOU? Control Influence
Power Consumption Mitigation What does “Green” Mean to You?
Economics
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
2
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 3
Growth Projections what are the experts saying?
2008, 50% of today’s data centers will have insufficient power and cooling capacity to meet the demands of high-density equipment
Through 2009, energy costs will emerge as the second-highest operating cost (behind labor) in 70% of DC facilities WW
2011, power demand for high-density equipment will level off or decline
2011, in-rack and in-row cooling will be the predominant cooling strategy for high-density equipment
2011, in chasis cooling technologies will be adopted in 15% the servers
Source: Gartner; Meeting the DC power and cooling challenge
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 4
U.S. Data Center Consumption Estimate 2004
Source: EPRI/Ecos Consulting & EIA Commercial Electricity Consumption Data
USA Total Generated 2005:
405.5B KWh
Generation and Consumption “Coal Fired Packets”
Our Impact as DC Professionals
It Requires 838 lb’s of coal to power one PC for
one Year!
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
3
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 5
Power Supplied in the DC CRAC 50%
Server/Storage 26%
Conversion 11%
Network 10%
Lighting 3%
Each watt consumed by IT infrastructure carries a “burden factor” of 1.8 to 2.5 for power consumption associated with cooling, conversion/distribution and lighting* Source: APC
Where does the Power Go? Losses in power generate heat
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 6
Cooling – Supply or Distribution? Current designs being specified to cool up to 30kW per rack
• Cisco Partner tested at 30KW and were able to cool in a normal server rack • Tested different configurations and had very different cooling airflow results • Targeted cooling utilizing modular hot/cold segregation essential to “future-proof” designs
Supply? Distribution?
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
4
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 7
A Connected Office with Converged Network Solutions Uses Less Energy and Requires Less Space Cooling per Person
CCRE Converged IT & Building Solution: Reduced Energy and Greenhouse Gases Typically 25% of Total Office Electrical Demand is for IT Equipment …
Heating, Cooling
and Ventilation
58%
Lighting 11%
IT Equipment 25%
“… it takes twice as much energy to remove the heat generated … as it takes to run the equipment.”
Other 6%
© 2006 Cisco Systems, Inc. All rights reserved. Cisco Confidential Presentation_ID 7
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 8
Cooling Strategies by Rack Density
0
5
10
15
20
25
Pow
er p
er R
ack
(kW
)
Vent rack exhausts, install closed cabinets and blanks to prevent recirculation
More than 2 tiles needed per rack. Consider Spot Cooling
Fully ducted exhaust into hot aisle to prevent recirculation
Consider In- row cooling or liquid-cooled racks
Infrastructure Examples
Cooling Options
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
5
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 9
Enabling Hot Aisle/Cold Aisle Designs with High Density 6509 and 9513 Chassis
Example: Panduit Cabinet 45RU (32”W x 40”D x 84”H)
Up to 20kW/cabinet heat rejection capability
3 6509’s or 3 9513’s per Rack
Front to back airflow into Hot Aisles
Integrated Cable Management
Modular design to support future air handlers or spot cooling
Part # CN4-1 and CN4-2 for MDS 9513 and # CN4-3 for the Catalyst 6509E
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 10
Affording the Next-Gen Data Center Legacy Server High-Density Server
Power per Server 2-3 kW per rack > 20 kW per rack
Power per Floor Space 30-40 W/ft² 700-800 W/ft²
Cooling Needs—chilled airflow 200-300 cfm 3,000 cfm
Source: Gartner 2006
20,000 ft²
800kW
+33%
100-200 Racks
Annual Operating Expense = $800k
Annual Operating Expense = $4.6M*
*Peripheral DC costs considered
Legacy DC designed to accommodate 2-3kW per Rack
Introducing 1/3 high-density infrastructure into a legacy facility is cost prohibitive
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
6
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 12
Data Center Evolution
Consolidation • Reduce Operating Expense
• Can Strain Power Grids, HVAC Systems, and Control Planes
• Also an Opportunity to “Audit” Power and Cooling Design
Power & Cooling • Reduce Component Count
• Increases Density
• Assessment Needed to Determine OPEX for site
• Assessment may be Needed to Audit Existing CPI
Best Practices • Choose Efficient
Components (Systems)
• Balance Density/Space
• Assess new Technology
• Targeted, Incremental Changes for Efficiency
• Virtualize to Increase Utilization
• Share Practices
Evolving to Green
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 13
Data Center Growth Lifecycle Alignment Facilities - 80KW
Network - 400 Ports
Server - 200 Units
Storage - 220 TB
Infrastructure Starting Capacity
Infrastructure Lifecycles (years)
Facilities 15
Network 7
Storage 5
Server 4
Annual Growth Estimates
Facilities 10%
Network 6%
Storage 22%
Server 11%
Business Growth?
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
7
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 14
Cisco’s Positive Impact on DC Efficiency
Catalyst 6500 Power Supply Efficiency has improved from 70%-80% since introduction in 1999. High end power supplies are better than 90% efficient.
6502 6704 6708 6816 6516a 6724-SFP
6748-SFP
CSM ACE 9016 9124 9148
Catalyst 6500 10GbE w/ DFC Catalyst 6500 GbE w/ DFC
Catalyst 6500 SLB MDS9500
Cisco Will Continue to Reduce Power Per Work Unit Performed – Port Per Service Component Efficiency
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 15
Cisco ACE with FWSM Reduces Power by 85% Component/Conversion Point Reduction
Incremental Power Required (W)
• 85% power reduction with virtualized, integrated modules ~ 11kW • Rack space saved by using virtualized, integrated modules ~30RU • Additional savings from reduced cabling, port consumption and support costs
11,400 11,300 13,300
1,820
Design Efficiency
Performance Requirement
• 10 Gbps load balancing • 20 Gbps Firewall • 10 Virtual Contexts • High availability
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
8
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 16
Example: 10 Servers per SLB/FW Context
ACE Module = 220W Incremental for SSL Termination + SLB
FWSModule = 172W Incremental for FireWalling
ACE + FWSM Handle 200 SLB/FW Contexts
392W Incremental Total
Reduce Power Consumption Through Service Density
Design Efficiency
Support for 200
contexts
Catalyst 6500
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 17
Appliance Model Repeats itself 700Watts at a time Design Inefficiency
Example: 10 Servers per SLB/FW Context/Group
1) FW Appliance = 700W Incremental
2) SLB Appliance = 300W Incremental
3) App Firewall Appliance = 400W Incremental
700 Watts Incremental with each new Server Group
(Items 2 and 3)
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
9
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 18
Appliance Model Repeats itself 700Watts at a time
Design Inefficiency
Example: 10 Servers per SLB/FW Context/Group
700 Watts Incremental with
EACH new Server Group
392 Watts Incremental One Time Supports 200 Groups
Design Efficiency
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 19
Appliance Model Deployed in Redundant Config Repeats itself 1400Watts at a time
Design Inefficiency Example: 10 Servers per SLB/FW Context/Group
1400 Watts Incremental with EACH new Server
Group (x N)
800 Watts Incremental One Time Supports 200 Server
Groups
Redundant
Configuration
Design Efficiency
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
10
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 20
Reduce Power Consumption Through Service Density
Reduce complexity
Increase manageability
Reduce latency
Eliminate single points of failure
Support for 200
contexts
The ACE and FWSM deployed in a Catalyst
6500 provide these services within the
network fabric, eliminating the
appliances and their associated load
Design Efficiency
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 21
Competitor’s Solution Cisco MDS Solution Devices Required 5 1
Stranded (wasted) Ports and Slots
Yes - 90 ports plus 9 slots None
Max. Number of Isolated SANs Available
3 256
Routing Bandwidth between SANs
Up to 5x4G ISLs: 20G Backplane BW: 48G / slot
Power per Host Port 11.98 watts 3.23 watts
Power Breakdown 2 routers (600w each)
3 Storage Directors:
2 CPs @ 100w each
3 32-port modules @ 90w each
2 16-port modules @ 90w each
3 Chassis with fans @ 100w each
1 Cisco MDS 9513:
2 48-port modules @ 130w each
2 12-port modules @ 92w each
2 Sup-2 @ 88w each
2 Crossbars @ 44w each
1 Chassis with fans @ 223w
Total Power Required
3450 Watts 931 Watts
Cisco MDS is more Power Efficient than the closest Competitor Apples to Apples Design Efficiency
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
11
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 22
3rd Party Validation
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 23
Cisco’s Virtualization Solutions Industry Leading Results
HR SAN
Sales SAN
Common Physical Fabric
Marketing SAN
Tape SAN
MS
MS MS
Taking a tape subsystem off-line can save $3,800 in power and cooling per year
Step1: Build a network solution using Virtual Fabrics
Step2 :Storage Virtualization
• Increase disk utilization to ~70%
• Consolidate usable space in the actual Storage devices
• Greater power savings and reduced footprint
Design Efficiency
© 2006, Cisco Systems, Inc. All rights reserved. Presentation_ID.scr
12
© 2006 Cisco Systems, Inc. All rights reserved. Presentation_ID 25
In Closing…
Cisco’s virtualization solutions slow the growth of power demand through increased utilization while reducing component count
Cisco’s next-generation products and solutions will further reduce power consumption at the systems level
Cisco is focused on environmental concerns from executive direction to individual product design
Cisco’s CA group provides services for Power and Cooling auditing as a “Network Readiness Assessment” to assist with facilities support planning