northwestern university dialogue on october 26, 2007 eyp mission critical facilities, inc. 200 west...
TRANSCRIPT
Northwestern UniversityDialogue on October 26, 2007
EYP Mission Critical Facilities, Inc.200 West Adams Street, Suite 2750Chicago, IL 60606312-846-8500
Albany – Atlanta – Chicago – Dallas – London – Los Angeles – Middletown – New York City – San Francisco – Washington DC – White Plains
Agenda
| Firm Overview and Experience
| Research University Data Center Trends
| Q&A
Firm Overview
Firm Overview| EYP A&E founded in 1972; EYP MCF becomes stand-alone firm in
2001
| 300-person international MEP/FP Engineering Design, MEP/FP Systems Operations and IT Consulting firm
| 95% of current projects are data centers
| Best-in-class experience in planning, design, construction, migration and project management of HPC and Tiered Data Centers
| HPC and high-density cooling thought leadership
| Designed fifteen 15+MW data centers, including five 35+MW data centers
| In-house building performance simulation tools to develop cost/benefit strategies and reduce unnecessary costs
| CFD Modeling
| Reliability Modeling
| Energy Modeling
| Cost Modeling
EYP MCF Firm Overview| Designed 32 million sq. ft. of data centers | Planned and designed 50+ Greenfield data
centers since 2001| 800 MW power/back-up power systems design
since 2001| 200,000 tons of critical cooling design in the last
5 years| 23 million sq. ft. of data center risk, reliability &
feasibility assessments| Commissioned 15 million sq. ft. of critical
facilities| Breadth of experience to bring multiple options
to the table | Designing in flexibility & scalability| We know higher costs don’t equate with higher
reliability & efficiency| Team very driven by client satisfaction, trust,
communication, and long-term relationships
| Collaborating with IBM, HP and Dell, we understand the evolution taking place in computing technologies – also assessing & designing this trio’s own data centers
| LBNL data center energy consumption benchmarking study
| Close working relationship with the Uptime Institute/Computer Site Engineering
| Vendor, contractor and real estate relationships
| Industry and International Publications
EYP MCF Strategic Industry Leadership
EYP MCF Chicago Office — est. January 2004׀ Office at 200 West Adams Street, one block east of Sears Tower
׀ EYP MCF Center of Excellence re:
׀ Strategic Consulting for IT and facilities infrastructure
׀ Greenfield data centers
׀ High-performance computing
׀ Sustainability/energy efficiency/LEED
׀ Building performance simulation
׀ EYP MCF Chicago’s accomplishments over the last 3 years:
׀ 200+ projects equating to $2+ Billion in construction
׀ 60 corporate and institutional clients
׀ 40 talented professionals on staff
׀ 30 invitations from national organizations to present and publish our thought-leader expertise in the design of high-reliability and high-performance facilities
Representative Clients
Representative Clients – Higher Education
Representative Clients – HPC
Representative Experience
| Providing strategic consulting services
| Determining the best options for data center consolidation
| Determining the requirements for a purpose-built data center to accommodate future enterprise computing needs
| Schematic design and ROM cost estimates
University of Illinois
Urbana-Champaign, IL
| New Greenfield 81,000 sq. ft. building housing 28,200 sq. ft. of scalable HPC machine rooms
| Master planning, feasibility study, programming, data center layout, development of power and cooling loads and high-level design concepts, and ROM construction cost estimating
| Requires in excess of 8,000 tons of chilled water and in excess of 25 MW of power
NCSA/UIUC
׀ 80,000 sq. ft. data center part of new Greenfield Cyber Infrastructure Building complex
׀ Provided programming, conceptual design, order of magnitude cost estimating, design development-level drawings, construction documents peer reviews, and construction administration services
׀ 20,000 sq. ft. of research computing raised floor for academic research computing
׀ 10,000 sq. ft. of Tier III raised floor for enterprise data center
׀ Anticipated load densities of 100-300 W/sq. ft.
׀ Also, data center risk assessments and feasibility studies of three existing sites to support future research computing expansion scenarios
Indiana University
Bloomington, IN
Children’s Memorial Hospital
Chicago, IL
| Overall Technology Infrastructure Consultant for a new 1 million square foot replacement hospital building
| Performing a variety of services throughout the planning, design, construction, and migration phases
| Delivering IT/network and medical technology consulting
| Conceptual designs
| Final designs
| Migration planning/execution
| Project management of the entire technology infrastructure build-out and the eventual migration efforts
| 13,800 sq ft supercomputing machine room
| Data Center Assessment, CFD Modeling, and Master Planning for multiple upgrade/expansion scenarios
| Analysis of load increase from 100 W/sq ft to 150 W/sq ft to 300+ W/sq ft
San Diego, CA
San Diego Supercomputer Center at UCSD
׀ Ranked #7 on TOP500 June 2007 list
׀ New 70TFlops Computational Center for Nanotechnology Innovation (CCNI)
׀ Conversion of existing manufacturing building into a new, state-of-the-art computing facility with BlueGene/L machine and blade servers
׀ 5,000 sq. ft. of 48” raised floor area at 250-300 W/sq. ft. (zones approaching 600 W/sq. ft.
׀ 7,500 SF of office/hoteling area and 7,500 SF of mechanical/electrical support space
Rensselaer Polytechnic Institute (opened September 2007)
Troy, NY
׀ Goal is to be in Top 3 on TOP500 June 2008 list
׀ New Interim Supercomputing Support Facility (ISSF) within existing high-bay lab building
׀ Supports a BlueGene/P supercomputer with 44 kW per rack loads
׀ ISSF will house a 100 TFlops system, expandable to 500TFlops, then 1PFlop
׀ Master planning, feasibility study, and conceptual design for new Greenfield 160,000 sq. ft. Theory & Computational Sciences facility to house 10sPFlop system
Argonne National Laboratory (opened October 10, 2007)
Lawrence Berkeley Laboratory Computational Research and Theory Building – LEED Silver*
Research University Data Center Trends
University Benchmarking
Academic Research Computing ׀ Quality of research computing facilities is increasingly
a point of separation for top institutions.׀ Existing data centers cannot accommodate the
projected high growth rates especially for power and cooling .
׀ New researchers often require research computing resources immediately. (ND Engineering college stated 50% of their new hires require new research computing resources.)
׀ Top researchers also bring funding opportunities if computing facilities are available.
University Benchmarking
University Conceptual Scenario 1
Administrative and other non-HPC (~100 Watts/sq ft)
׀ Usually combined into a single data center
׀ Administrative space utilizes more traditional corporate enterprise computing reliability model
׀ Slower growth rates
University Benchmarking
University Conceptual Scenario 2
HPC for research only use (> 200 Watts/sq ft)
׀ Separate purpose-built HPC data center facility only
׀ Very high power and cooling density with lower reliability goals
׀ Usually associated with a specific NSF grant (e.g., Track 1 and Track 2 grants)
University Benchmarking
University Conceptual Scenario 3Combined general University and HPC facility (~100 to 200+
Watts/sq ft) ׀ Usually combined into a single data center׀ Typically a blended reliability/power density
infrastructure addressing both general university and HPC needs
׀ More cost-effective versus stand-alone facilities׀ Requires separation of general university and research
raised floor areas for security, accessibility and reliability
University Trends Matrix (All AAU Member Institutions)
Institution Administrative (ft²) Research (ft²) Power (watts/ft²) Redundancy
UIUC (Planned) 13,200 7500 130 N + 1/ N
Indiana University 10,000 20,000
HPC Standalone
166 Tier 3/Tier 2
University of Chicago 6,000 3,200 Unknown Tier 2/Tier 1
University of Texas 15,000
Regional DC
10,000 125 Tier 3/Tier 1
University of Minnesota 6,000 4,000 150 - 200 Tier 3/Tier 1
Stanford 17,000 33,000 Unknown Unknown
Cornell 8,500 20,000 Unknown Tier 3
Princeton 5.000 10,000 Unknown Unknown
University of Washington 6,000 12,000 150 Tier2 /Tier1 or NT
Holistic View ofData Center Design
20012002
20032004
20052006
20072008
20092010
0
25
50
75
100
125
150
175
CP
U P
ow
er
- w
att
s
02.557.51012.51517.52022.52527.53032.535
Thousa
nds
Rack
Pow
er
- w
att
s
CPU PowerRack Power
The Growth of Power Consumption
Source: IBM Corporation
This is where we are now
Four Tier System
Tier 1 – Basic Non-Redundant Data Center
| Single path for power and cooling distribution without redundant components
Tier 2 – Basic Redundant Data Center
| Single path for power and cooling distribution with redundant components
Tier 3 – Concurrently Maintainable Data Center
|Multiple paths for power and cooling distribution with only one path active and with redundant components
Tier 4 – Fault Tolerant Data Center
|Multiple active power and cooling distribution paths with redundant components and fault tolerant
Site AvailabilityTIER 1 TIER 2 TIER 3 TIER 4
SITE AVAILABILITY 99.67% 99.75% 99.98% 99.99%
OUTAGE OVER 5 YEARS 144 HOURS 110 HOURS 8 HOURS 4 HOURS
Sample - Rough Order of Magnitude Cost Estimate*
INFRASTRUCTURE CONCEPT Tier 1 Tier 2 Tier 3
7500 FT² Day 1
with 2500 FT² Shell
80 WATTS / FT²
600 KW Total
$8,200,000 $8,800,000 $14,200,000
Ultimate Build Out
10,000 FT²
120 WATTS / FT²
1200 KW Total
$14,200,000 $15,400,000 $26,200,000
* Utilizing the Uptime Institute data center cost estimation methodology
* Actual data center cost may vary widely (more than 20%) based on final design
Public Law 109-431
To study and promote the use of energy-efficient computer servers in the United States
Questions/Discussion
Michael [email protected]
Bill Kosik, PE, CEM, LEED [email protected]
Tom Kutz [email protected]
End Slide