nutanix - expert session - metro availability
TRANSCRIPT
Nutanix Metro-Availability
Christian Johannsen, Senior SE Nutanix
Nutanix – Technology Review
3
Nutanix Virtual Computing Platform
4
Convergence21
Dat
a15
Metadata8
Clo
ud
4
VM Mobility3
Control Plane10
VD
I2
Map
Red
uce
2
Security1Support1
An
alyt
ics1
• Shared-nothing storage controller for virtualization environments.
• Method for networking converged shared-nothing storage for high availability.
• I/O and storage for a virtualization environment with multiple hypervisor types.
• Performing hot-swap of a storage device in a converged architecture.
Key Patents
Top Categories
Web-scale Foundation Platform
22 patents filed
Powerful Control Plane
10 patents filed
Scale-out Data Plane
15 patents filed
47 Patents
Patent Distribution
Nutanix Patent Portfolio
5
Nutanix Distributed File System (NDFS)
Virtual Storage Control Virtual Storage Control
Virtual Machine/Virtual Disk
Flash HDD
Enterprise Storage
• Data Locality• Tiering and Caching• Compression• Deduplication• Shadow Clones• Snapshots and Clones
Data Protection
• Converged Backups• Integrated DR• Cloud Connect• Metro Availability• 3rd party Backup
Solutions
Resiliency
• Tunable Redundancy• Data Path Redundancy• Data Integrity Checks• Availability Domains
Security
• Data at Rest Encryption• Nutanix Security DL• Cluster Shield• Two-factor Auth
Nutanix – Data Protection
7
Stay covered for Critical Workloads
RTORPONutanix Offers
Minutes Minutes Time Stream
Hours Hours Cloud Connect
Near -zero Minutes Metro Availability
Minutes Minutes Remote Replication
Minor
incidents
Major
incidents
Time between backups Maximum tolerable outage
8
Time Stream
Set retention policy for local
and remote snapshots
Set snapshot schedule for a
protection domain
Time-Based backup (Storage Snapshot) with local and remote retention
Imagine the snapshots beside the integrated replication
Application consistent snapshot possible
9
Nutanix Cloud Connect
Datacenter Cloud
Backup and recovery of VMs from Nutanix cluster to the public cloud
VMCaliber and WAN optimized
Fully integrated management experience with Prism
Quick restore and state recovery
10
Async DR
VM-centric workflows
Granular VM based snapshots and policies, better than LUN based
Space efficient sub block-level snapshots (redirect-on-write)
N-Way master-master model for more than on site
VM and application level crash consistency
11
Introducing Nutanix Metro Availability
Geographic distributed High Availability – covers the entire infrastructure stack
Covers entire infrastructure stack
Leverage existing network
Deploy in minutes through Nutanix Prism with minimal change management
Mix and match models to workloads
Customer
Network
12
Network
<=5ms RTT
< 400 KMs between two sites
Bandwidth depends on ‘data change rate’
Recommended: redundant physical networks b/w sites
General
2 Nutanix clusters, one on each site
Mixing hardware models allowed
Hypervisor
ESXi in NOS 4.1
Hyper-V/KVM in the future (Q1 CY 2015)
Requirements
13
Architecture
Synchronous storage replication
Datastore stretched over both Nutanix cluster in a single Hypervisor cluster (vMotion, HA)
In conjunction with existing data management features, compression, deduplication, and
tiering
Standby containers are unavailable for direct virtual machine traffic (first release)
14
Nutanix I/O Path
I/O Path
1. OpLog acts as a write buffer (Random Writes)
2. Data is replicated to other nodes sync.
3. Sequentially drained to Extent Store
4. ILM (Information Lifecycle Management)
chooses the right target or the data
5. Deduplicated read cache (Content Cache) spans
Memory and SSD
6. VM accessing the same data on just one copy
(deduplicated)
1. IF data not in Content Cache it will be promoted
per ILM
2. Extensible Platform for future I/O patterns
1515
1. Write IO
2a. Written to local OpLog (RF) and remote
replication to remote OpLog
2b. Local Replication in remote OpLog (RF)
3a. Write IO Ack in local OpLog (RF)
3b. Write IO Ack in remote OpLog (RF)
3c. Write IO Ack from remote OpLog
4. Write IO Ack from local OpLog to the hypervisor
Write Anatomy
1616
1. Write IO
2. Write IO forwarded to Active Container
3a. Written to local OpLog (RF) and remote
replication to remote OpLog
3b. Local Replication in remote OpLog (RF)
4a. Write IO Ack in local OpLog (RF)
4b. Write IO Ack in remote OpLog (RF)
4c. Write IO Ack from remote OpLog
5. Write IO Ack from local OpLog to the remote
OpLog
6. Write IO Ack from local OpLog to the hypervisor
Write Anatomy (vMotion, Recovery)
1717
1. Read Request
2. Read Request forwarded to Active Container
3. Data returned from the Active Container
4. Data sent to the VM
Read Anatomy (vMotion, Recovery)
1818
Scenarios
1919
Scenarios
Network failure between sites
Manual or Automatic (seconds)
2020
Scenarios
Site Failure
Demo Time!
https://drive.google.com/a/nutanix.com/file/d/0B3sqKkY-Et4deF9Db2NPdlYzMmM/view
Thank You