performance evaluation of amazon ec2 for nasa hpc applications

21
National Aeronautics and Space Administration www.nasa.gov Performance Evaluation of Amazon EC2 for NASA HPC Applications Piyush Mehrotra J. Djomehri, S. Heistand, R. Hood, H. Jin, A. Lazanoff, S. Saini, R. Biswas NASA Advanced Supercomputing (NAS) Division NASA Ames Research Center [email protected]

Upload: independent

Post on 15-Nov-2023

0 views

Category:

Documents


0 download

TRANSCRIPT

National Aeronautics and Space Administration

www.nasa.gov

Performance Evaluation of Amazon EC2 for NASA HPC Applications!

Piyush Mehrotra!!

J. Djomehri, S. Heistand, R. Hood, H. Jin, A. Lazanoff, !S. Saini, R. Biswas !

NASA Advanced Supercomputing (NAS) Division!NASA Ames Research Center!

[email protected]!!

NASA High End Computing Capability!

Cloud Computing!

•  Cloud computing is becoming ubiquitous!-  Allows organizations to outsource hardware acquisition and

maintenance!-  Satisfies burst computational demands at the cost of performance!-  Provides support for customized execution environments!-  Increasing catalog of useful software services!

•  Extensively being used for enterprise/business tasks!•  Increasing use for scientific and engineering (S&E) tasks, e.g.,

bioinformatics, oil exploration, data analytics, etc.!-  Embarrassingly parallel or loosely coupled parallel tasks!

•  To attract the High Performance Computing (HPC) market Amazon EC2 provides HPC instance types!

2

Are Clouds ready for NASA’s HPC applications?

NASA High End Computing Capability!

Outline!

• Clouds!• Systems tested!• Benchmarks and Applications used!• Performance achieved!• Conclusions!

3

NASA High End Computing Capability!

Systems Tested!•  Pleiades: NASAʼs primary supercomputer!-  SGI Altix ICE – 4 generations of Intel Xeon processors: Harpertown (4c), Nehalem (4c),

Westmere (6c), Sandy Bridge (8c) – total 11,776 nodes, 126,720 cores, peak 1.75 Pflops!

•  Amazon EC2: premier public cloud computing environment!

4

4

Pleiades Amazon HPC cc1.4xlarge

CPU Intel Xeon X5570 Nehalem

Intel Xeon X5570 Nehalem

Clock Speed 2.93 GHz 2.93 GHz No. of cores/node 8 8 RAM 24 GB 23 GB Hyper-threading Enabled Enabled Network InfiniBand QDR 10 GigE Interconnect Topology Hypercube Unknown OS Type SLES11SP1 SLES11SP1 Virtualization No Yes Multi-tenancy No No Compiler Intel 12.0.4 Intel 12.0.4 MPI Library MPT 2.04 Open MPI 1.4.4 File Structure Lustre NFS

Pleiades Amazon HPC cc1.4xlarge

CPU Intel Xeon X5570 Nehalem

Intel Xeon X5570 Nehalem

Clock Speed 2.93 GHz 2.93 GHz No. of cores/node 8 8 RAM 24 GB 23 GB Hyper-threading Enabled Enabled Network InfiniBand QDR 10 GigE Interconnect Topology Hypercube Unknown OS Type SLES11SP1 SLES11SP1 Virtualization No Yes Multi-tenancy No No Compiler Intel 12.0.4 Intel 12.0.4 MPI Library MPT 2.04 Open MPI 1.4.4 File Structure Lustre NFS

NASA High End Computing Capability!

Benchmarks/Applications!•  HPCC Benchmarks!-  DGEMM – floating point rate

measurement!-  Bandwidth and latency tests!

•  NPBs:!-  Kernel benchmarks!

•  MG: multi-grid on a sequence of meshes, long- and short-distance communication, memory intensive!

•  FT: discrete 3D FFTs, all-to-all communication!

•  IS: integer sort, random memory access!

•  CG: conjugate gradient, irregular memory access and communication!

•  EP: embarrassingly parallel with some communication!

-  Compact Application benchmarks!•  BT: block tri-diagonal solver!•  SP: scalar penta-diagonal solver!•  LU: lower-upper Gauss Seidel!

!

5

•  Applications:!-  Cart3D: high-fidelity inviscid CFD

code; multiple cartesian grids; cpu-intensive and memory bound!

-  DDSCAT: embarrassingly parallel code to calculate the absorption and scattering properties oftargets!

-  ENZO: Simulation of cosmological structure formulation using adaptive mesh refinement!

-  MITGcm: Global ocean circulation model; memory bound; latency bound large number of cores!

NASA High End Computing Capability!

Performance Results!

6

NASA High End Computing Capability!

HPCC results (64 cores)!

7

Pleiades Amazon HPC

EP DGEMM N=5773 (Gflops) 11.417 10.758

Single DGEMM N=5773 (Gflops) 12.252 11.062

Max Ping Pong Latency (usec) 2.574 73.999

Min Ping Pong Latency (usec) 0.238 0.626

Avg Ping Pong Latency (usec) 1.954 61.535

Naturally Ordered Ring Latency (usec) 1.597 58.293

Randomly Ordered Ring Latency (usec) 4.466 180.786

Max Ping Pong Bandwidth (Gbytes) 5.044 6.982

Min Ping Pong Bandwidth (Gbytes) 1.538 0.406

Avg Ping Pong Bandwidth (Gbytes) 2.171 0.952

Naturally Ordered Ring Bandwidth (Gbytes) 1.768 0.299

Randomly Ordered Ring Bandwidth (Gbytes) 0.283 0.070

Pleiades Amazon HPC

EP DGEMM N=5773 (Gflops) 11.417 10.758

Single DGEMM N=5773 (Gflops) 12.252 11.062

Max Ping Pong Latency (usec) 2.574 73.999

Min Ping Pong Latency (usec) 0.238 0.626

Avg Ping Pong Latency (usec) 1.954 61.535

Naturally Ordered Ring Latency (usec) 1.597 58.293

Randomly Ordered Ring Latency (usec) 4.466 180.786

Max Ping Pong Bandwidth (Gbytes) 5.044 6.982

Min Ping Pong Bandwidth (Gbytes) 1.538 0.406

Avg Ping Pong Bandwidth (Gbytes) 2.171 0.952

Naturally Ordered Ring Bandwidth (Gbytes) 1.768 0.299

Randomly Ordered Ring Bandwidth (Gbytes) 0.283 0.070

Pleiades Amazon HPC

EP DGEMM N=5773 (Gflops) 11.417 10.758

Single DGEMM N=5773 (Gflops) 12.252 11.062

Max Ping Pong Latency (usec) 2.574 73.999

Min Ping Pong Latency (usec) 0.238 0.626

Avg Ping Pong Latency (usec) 1.954 61.535

Naturally Ordered Ring Latency (usec) 1.597 58.293

Randomly Ordered Ring Latency (usec) 4.466 180.786

Max Ping Pong Bandwidth (Gbytes) 5.044 6.982

Min Ping Pong Bandwidth (Gbytes) 1.538 0.406

Avg Ping Pong Bandwidth (Gbytes) 2.171 0.952

Naturally Ordered Ring Bandwidth (Gbytes) 1.768 0.299

Randomly Ordered Ring Bandwidth (Gbytes) 0.283 0.070

NASA High End Computing Capability!

NPB Results!

8

NASA High End Computing Capability!

NPB Results (MG)!

9

EC2-HPC relative to Pleiades

NASA High End Computing Capability!

NPB Results (contd.)!

10

NASA High End Computing Capability!

Application Performance!

11

NASA High End Computing Capability!

CART3D Performance!

12 12

EC2-HPC relative to Pleiades

NASA High End Computing Capability!

DDSCAT Performance!

EC2-HPC relative to Pleiades

13 13

NASA High End Computing Capability!

EC2-HPC Performance Relative to Pleiades!

14

NASA High End Computing Capability!

Communication Impact!

15

NASA High End Computing Capability!

Communication Impact (2)!

16

NASA High End Computing Capability!

Virtualization Overhead!

17

Pleiades Amazon HPC Perf. Diff.

EP DGEMM N=5773 (Gflops) 11.417 10.758 5.8%

Single DGEMM N=5773 (Gflops) 12.252 11.062 9.7%

NASA High End Computing Capability!

Performance Variability!

18

NASA High End Computing Capability!

Summary!

• Amazon EC2-HPC performance lags Pleiades supercomputer for most NASA HPC applications!- Network technology!- Virtualization overhead!

• Other issues with clouds for NASA scientists!- Cost!- Security!- Usability!- Reliability!

19

Cloud computing currently not a viable solution for NASA HPC applications

… but they are catching up fast!!!

NASA High End Computing Capability!

Characteristics Cloud Environments NASA HPC Center

On demand self-service For users with a credit card

For approved users with allocation

Broad network access Worldwide Worldwide

(subject to NASA security constraints)

Resource Pooling Yes Across the Agency

Rapid Elasticity Extreme Users can provision

multiple thousands of cores

Measured Service Users credit cards

charged according to usage

Users allocations charged according to

usage Service Model Mainly IaaS PaaS Deployment Model Public Private

20

NASA’s HECC facility is a private, PaaS, HPC cloud

NASA High End Computing Capability! 21