carrier grade nfv titanium server · 2015-09-04 · © 2015 wind river. all rights reserved....
TRANSCRIPT
© 2015 Wind River. All Rights Reserved.
Carrier Grade NFV Titanium Server Bruno Rouchouse
2 © 2015 Wind River. All Rights Reserved.
Business Challenge #1 How Do We Carry Network Traffic Profitably?
Exponential increase in network traffic
– Driven by growth in video
Ongoing increase in network equipment costs
– To address subscriber bandwidth needs
Flat to declining per-user revenue (ARPU)
Network cost increases are outpacing subscriber revenue growth
3 © 2015 Wind River. All Rights Reserved.
Business Challenge #2 Who‟s Extracting the Most Value from Network Traffic?
OTT projected to cost service providers $14B in lost revenue in 2015
– Up 26% from 2014
Social media, IM and VoIP are primary sources of lost revenue
– User loyalty and stickiness
All about speed, agility, ease-of-use and innovative services
– Providing value to subscribers
4 © 2015 Wind River. All Rights Reserved.
Business Challenge #3 IoT: A Threat or an Opportunity?
Exponential growth in devices connected over service provider networks
– Corresponding increase IoT in traffic
Critical demands on network capability
– Security
– Latency
– Availability
How will service providers monetize IoT?
– Analytics, contextual-based services etc.
5 © 2015 Wind River. All Rights Reserved.
NFV Architecture Concept
To NFV…
Network Virtualization Software Platform
Industry-Standard Servers
Virtual
Firewall
Virtual
Gateway
Virtual
EPC
Other
VNFs
Service Orchestration
From This…
Firewall Gateway EPC
6 © 2015 Wind River. All Rights Reserved.
Business Benefits Promised by NFV OPEX Reductions Plus Top-line Revenue Growth
Reduce OPEX
Improved resource utilization through consolidation
Simplified network management
Boost ARPU and Top-line Revenue
Accelerated service introduction: Service agility
Innovative, value-added services
7 © 2015 Wind River. All Rights Reserved.
Four Key Business Challenges for NFV
Protecting top-line revenue during the transition to NFV
Minimizing OPEX to fully leverage the benefits of NFV
Gaining market share through accelerated deployments of new services
Maximizing flexibility through adoption of open solutions
8 © 2015 Wind River. All Rights Reserved.
Protecting Top-Line Revenue During the Transition
Even with traditional physical infrastructure, network outages cause major revenue impact for service providers
$15B per year (1% to 5% of revenues)
– Customer churn
– plus operational expenses
– plus SLA penalties
Thousands of dollars per minute, per server
10 © 2015 Wind River. All Rights Reserved.
NFV Has the Potential to Make This Much Worse
Traditional physical infrastructure typically delivers six 9s reliability (99.9999% uptime)
– Carrier grade reliability developed and proven over many years and thousands of deployments
NFV adds significant risk factors
– New, virtualized network applications
– Dynamic reallocation of VMs across servers, racks and data centers
– Complex, hard-to-debug traffic flows
Need to implement carrier grade reliability in NFV deployments
– Significant risk to service providers‟ revenues if network uptime not maintained
11 © 2015 Wind River. All Rights Reserved.
Enterprise Platforms Don‟t Provide Telco Reliability
Network Reliability Is Key to Commercial Success of NFV
Enterprise IT Platform Capability
Carrier Grade Telecom Infrastructure Requirements
Fault Detection (in hardware or virtualization layer)
~ 1 Minute Sub-second
Detection of Failed VMs > 1 Minute Sub-second
Detection of Failed Platform Management Services
No Support < 10s
vSwitch Performance 1-2 Gbps Line Rate With Minimum
Core Utilization
Network Link Failure Detection for Compute Nodes
Depends on Linux distribution 50ms
Live Migration for DPDK-based VMs No Support Full Support
12 © 2015 Wind River. All Rights Reserved.
Application-Level High Availability Is Not the Solution Several Approaches Exist for Application-level HA
Active / active
Active / standby
N-way active with load balancing
None of These Meet System-level Requirements for Reliability or Resiliency
No awareness of underlying system resources: Could deploy both instances on same server
No guarantee of deterministic, consistent VNF performance (e.g. NUMA awareness)
No awareness of service chains
No capability of automated recovery from system-level failures
No support for platform-level security
OSS / BSS
NFV Infrastructure Must Deliver Six-nines Network Reliability
VNF Cannot Guarantee Service Reliability
VNF VNF VNF
NFV Orchestrators
VNF
VM VM VM VM
Service Platforms
13 © 2015 Wind River. All Rights Reserved.
TITANIUM SERVER
Architecture Overview
14 © 2015 Wind River. All Rights Reserved.
Titanium Server Architecture Details
Wind River Titanium Server Software
Virtual Network
Functions (VNFs)
OSS / BSS NFV Orchestrators
Carrier Grade Accelerated vSwitch Accelerated Data Plane
vNICs Intel® DPDK
VM
vEPC
VM
vBRAS
VM
vCPE
VM
Other
KVM
Standard Server Platform (Multi-Core IA Hardware)
OpenStack
AVP AVP AVP AVP
Accelerated Virtual Port (AVP) vNIC
Driver
Based on standard open source components
Add critical real-time performance enhancements to KVM
Add Carrier Grade accelerated vSwitch
Add Carrier Grade management and telco middleware functions
Add accelerated virtual NIC driver for Virtual Network Functions
Host any guest Operating System
Carrier Grade Linux
Running any guest OS
Carrier Grade Management and Telco Middleware
Fault Management
Software Management
Performance Management
OAMP
VM High Availability Management
Carrier Grade KVM
Real-Time Extensions
Low Latency
15 © 2015 Wind River. All Rights Reserved.
System Architecture – Physical View
VM VM VM
Compute
Control
Storage
VM VM VM
Compute
Control
Storage
Storage
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
Storage
Control
Control
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
Storage
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
VM VM VM
Compute
Storage
Control
Control
16 © 2015 Wind River. All Rights Reserved.
NFV Architecture Mapped to Titanium Server
API to VNF: Health check, shutdown, event notification, active / standby, sparing
API into OSS / BSS: Fault, performance, upgrades, patching, power management, security
API to orchestrator: System inventory, VM management, service chain definitions
Computing Hardware
Storage Hardware
Network Hardware
VI-Ha
Hardware Sources
OSS / BSS
Service, VNF and Infrastructure Description
Tit
an
ium
OS
S A
PI’
s
VIM
Control and manage NFVI
Collect and forward FM / PM
VNFM WR
Provided or
Ecosystem
Titanium Guest API’s
Virtual Computing
Virtual Storage
Virtual
Network
NFVI
Orchestrator
Titanium Cloud API’s
VNF Manager
VNF lifecycle manager
VNF Manager(s)
EMS 1 EMS 2 EMS 3
VNF 1 VNF 2 VNF 3
Virtualization Layer
Virtualized Infrastructure
Manager(s)
17 © 2015 Wind River. All Rights Reserved.
NFV Functional Layers
1Source: ETSI NFV MANO WI
Network Functions Virtualization Orchestrator (NFVO)
A functional block that manages the Network Service (NS) lifecycle and coordinates the management of NS lifecycle, VNF lifecycle (supported by the VNFM) and NFVI resources (supported by the VIM) to ensure an optimized allocation of the necessary resources and connectivity.
NFVO NFVO
NFVI-Node
Physical devices deployed and managed as a single
entity, providing the NFVI Functions required to support
the execution environment for VNFs.
NFVI OpenStack
Virtualized Infrastructure Manager (VIM)
A functional block that is responsible for controlling and
managing the NFVI compute, storage and network resources,
usually within one operator‟s Infrastructure Domain.
Virtual Infrastructure Manager
VIM
Virtualized Network Function Manager (VNFM)
A functional block that is responsible for the lifecycle
management of VNF.
VNF Managers
Vendor X VNFM
Vendor Y VNFM
18 © 2015 Wind River. All Rights Reserved.
Titanium Server VIM and VNFM
Implement discrete components for VIM and optional VNFM
VNFM component is optional. Customer can choose to leverage or provide another vendors
– Both could co-exist in the same deployment
Communication between VIM and 3rd party VNFM would be via a Titanium Server provided plugin
– Standard is not at the level of maturity where the detailed interface spec is defined (e.g., a set of REST API‟s) therefore the plugin is required
Provide a plugin to allow a NFVO to talk to a Titanium Server VNFM
Flexible Open Spec Compliant Implementation
VNF Managers
Vendor X VNFM
Vendor Y VNFM
VNFM Plugin
VNFM Plugin
NFVO NFVO NFVO Plugin
NFVO Plugin
WRS VNFM
Virtual Infrastructure Manager
VIM
NFVI OpenStack
OpenStack Extensions
TIS Components
19 © 2015 Wind River. All Rights Reserved.
TITANIUM SERVER
Carrier Grade Availability
20 © 2015 Wind River. All Rights Reserved.
Key Pillars of a Carrier Grade Solution
BUILT ON OPEN SOURCE AND OPEN STANDARDS WITH OPEN APIs
Carrier Grade
Management
Highest Performance
Highest Availability
1. Management
• Hitless upgrades and patching
• Faults, alarms, metrics collection
• OAM - Configuration, Fault and Performance
2. Performance and scalability
• Accelerated vSwitch: 20Gb/s guest throughput
• 10µs interrupt-latency real-time virtualization
• Minimum latency on VM failover (~500ms)
3. Reliability and availability
• No single point of failure
• Less than 1 second host/VM Fault Detection
• Full automatic recovery of VM or node
• Live VM migration (with DPDK guests)
21 © 2015 Wind River. All Rights Reserved.
Design for Carrier Grade Availability 1. Predictive Modeling
• Modeling (Markov) based on reference HW design, FIT / FMECA assessment, reference HA models (failover, live migration)
2. Design Methodology
• TL 9000 certified methodology end to end from architecture to release
• Quality system - defect density prediction, phase containment, automated testing, RCA
3. Carrier Grade Testing
• Large system stress / overload testing (link, CPU / memory, system)
• Fault insertion testing under load (links, controllers, computes, VM's)
• Long duration load soaks (72 hours)
• Systems engineering assessment (speeds / feeds, memory)
4. Performance Tracking
• TL 9000 practices, monthly tracking, allocation based on agreed deployment and scope parameters
5. Issue Responsiveness
• Fast response to mitigate impacts, corrective and preventative actions, rigorous RCA
22 © 2015 Wind River. All Rights Reserved.
Carrier-Grade Cloud Controller Services
1:1 Cloud Controller Services
On 2 controller nodes
Controller cluster management manages active / standby status of „controller‟ services
Synchronized File Systems
Synchronized DBs
Split-brain prevention mechanisms Computer / Networking
App VM
Computer / Networking
App VM
App VM
Computer / Networking
App VM
App VM
Computer / Networking
App VM
Cloud Services
Maintenance Services
Inventory Services
Titanium Server Controller Node 0
Titanium Servicer Controller Node 1
Controller Cluster Management
Controller Cluster Management
Cloud Services
Maintenance Services
Inventory Services
23 © 2015 Wind River. All Rights Reserved.
Automatic VM recovery on compute node failure – With fast failover
Automatic VM recovery on KVM / QEMU failure – With improved KVM/QEMU failure detection times
– Enhanced VM monitoring optionally tied into application health checks
– Within guest VM
Live migration of VMs – For maintenance and/or orchestration procedures
– Note: Includes support for VMs running Intel-DPDK
Graceful VM shutdowns
VM server groups – e.g., ensures VMs of same server group (e.g. 1:1 pair) are created
on different compute nodes and/or different availability zones
Traditional TELCO HA Models (1:1, 1:N) and enterprise load balancing approaches supported – HA_Unaware / cold standby
– HA_Aware / application-provided HA-framework
Carrier-Grade VM Management
Computer / Networking
App VM
Computer / Networking
App VM
App VM
Computer / Networking
App VM
App VM
Computer / Networking
App VM
Controller Controller
24 © 2015 Wind River. All Rights Reserved.
Service Manager vs. Pacemaker / Corosync
OpenStack controller high availability is managed by Pacemaker / Corosynch
Titanium server provides service manager for controller high availability
– Pacemaker / Corosynch does not scale and causes significant outages
Titanium Server – Service Manager Pacemaker / Corosync
Asynchronous process death notification Polling / Monitoring detection of process death
5 second interval
Faster Failover performance (on order of 1-2 seconds)
Failover times on order of ~5 seconds
Restart of „Service Manager‟ without impacting monitored services
Restart of Pacemaker tears down all services.
25 © 2015 Wind River. All Rights Reserved.
VM Management – Server Groups
Server group entity introduced to implement group level behaviors across multiple VM‟s to facilitate more intelligent scheduling decisions
– VM placement rules – numa, anti-numa etc.
– Minimum number of instances during mtce activities
REST APIs exposed to provision server group and add / remove a VM instance to a group
Auto-scaling to add / remove instances from a server groups as it sizes up / down a stack instance
Host1 VM1a
VM1d
VM1d
VM2a VM2c
VM3a
VM2b
VM3c
Host2
Host3
Server Group A Server Group D
VM1d
VM3b
26 © 2015 Wind River. All Rights Reserved.
Carrier Grade Availability API‟s
VM health checking / heart beating
– VM health monitoring – heart beat
– Various levels of integration available
From simple validation of kernel scheduling to
Application-specific health checks
VM event acknowledgement / notification
– VM are notified and have opportunity to reject a VM event,
– Where VM event: VM shutdown, migrate, reboot, pause and suspend request
VM peer state notification
– This API informs a guest of a state change of any VM within its server group
VM guest scaling
– API to trigger the add of vCPU‟s to a running VM
Health Checking
Send
ACPI
Signal
Shutdown Request
Guest VM
VM Lifecycle API
Titanium Server Compute Node
28 © 2015 Wind River. All Rights Reserved.
Intel hyper-threading allows a single physical core of a processor to behave like two logical cores
– The two logical cores are referred to as “sibling” cores
Using hyper-threading doesn‟t double performance
Titanium server provides a mechanism to specify which VM Instances share siblings of the same physical core(s)
Configured via server groups and a new policy: “Affinity-hyper-thread”
– Enables explicit pairing of VM‟s to sibling cores
– Enables a full server to share HT modes
– Previously, only granularity was at the server level
For guaranteed performance don‟t use hyper-threading
– There is only one pipeline
– Can result in lower performance and higher latency
Hyper-Threading Affinity
Source: Adapted from G. Waines, TiS_L2_Training_hyperthreading.pptx”, 2014.
1 0
Used Used
3 2
Used Used
5 4
Rsvd Used
7 6
Free Free
VM1 VM2
29 © 2015 Wind River. All Rights Reserved.
Inventory and Resource Management End-to-end node / server management
– Install, configure, monitor and maintain nodes / servers of cloud
Detailed inventory collection and monitoring of node resources
– Beyond just CPU cores, RAM and disk space
– Physical NICs (# and bandwidth), crypto / compression hardware, etc.
– Presented to the user with real time updates
Configuration of:
– Nodes‟ / servers‟ network interfaces, core usage, disk storage, etc.
– System OAM IP address, SNMP, DNS servers, NTP servers, etc.
Support of administrative operations on nodes / servers
– Lock / unlock, SWACT, reboot / reset, power-on / -off, re-install, etc.
Status reporting of nodes / servers
– Admin state, operational state, availability state, uptime, and real-time command execution report/progress (e.g. booting, testing…)
30 © 2015 Wind River. All Rights Reserved.
Titanium Server: Management Interfaces CLI
– Centralized system level CLI
– Powerful for using to build scripts
GUI
– Present all available parameters
Minimize required CLI commands
– Consistent GUI interface design for each module
Focus GUI on accomplishing tasks with a smooth flow of info
– Presentation of data consistent across all screens,
User feedback during operations
– Clear response / error messaging,
– Busy indicators, progress indicators for long-running commands
REST API
– Fully supported REST API for off-switch management applications
– Backwards compatible with vanilla OpenStack REST APIs
– But extended with titanium server-specific capabilities
SNMPv2c
– Mandatory subset of Agent MIBs / Tables (System, SNMP, Community…)
– Wind River Enterprise Alarm MIB (Active Alarm List, Historical Alarm List, Alarm TRAPS)
31 © 2015 Wind River. All Rights Reserved.
Titanium Server Guest API‟s and SDK
Standard README files
Key Libraries provided
Documentation on how to use the SDK
Comprehensive API guide
32 © 2015 Wind River. All Rights Reserved.
Titanium Server Open API‟s
Open Integration
Performance
Scaling
Reliability
Flexibility
Optional
Easy to integrate
AVP-KMOD
– For Linux guest
AVP-DPDK
– For DPDK guest
Guest Drivers
Optional
Easy to integrate
Event notification
Heart beat
Health check
Scaling
Guest API’s
OS / APP
AVP
vNIC
OS / APP
API
KVM
33 © 2015 Wind River. All Rights Reserved.
TITANIUM SERVER
Networking and Performance
34 © 2015 Wind River. All Rights Reserved.
Accelerated vSwitch
Source: 2 cores for AVS; 1 core for Guest VM (Guest is bottleneck)
Virtual Switching 20G Line Rate Performance on 2 Cores
Data Path App
Red
Hat
Data Path App
Cen
tOS
Data Path App
Ub
un
tu
Data Path App
WR
L
Virtio
App
vNIC
AVP
App
vNIC
AVP
App
vNIC
DPDK
App
vNIC
AVP
DPDK
DPDK
AVS Linux / KVM
Intel Server
Nova Neutron
Titanium Server OpenStack Plug-in
For VNFs using kernel-based-networking
– Supports unmodified VNFs via Virtio NIC drivers
e.g. VNFs written for OVS
– Supports accelerated (~7x) networking via simple KLM containing Accelerated Virtual Port (AVP) NIC drivers
e.g. NO application change!
Supports VNFs using DPDK-accelerated- networking for BEST (~50x) performance
– Simple 3-line integration of poll-Mode version of Accelerated Virtual Port (AVP) NIC drivers
Enables multi-core data path independent of number of vNICs in Guest
– Multi-queue support in AVP NIC drivers; up to 8 queues supported per vNIC
35 © 2015 Wind River. All Rights Reserved.
Carrier Grade Networking Failover
50ms Link Fail Over Time
Enterprise – 2 Sec Detection
Server
VM VM VM VM
Port
AVS
Port
Failed Link
36 © 2015 Wind River. All Rights Reserved.
Key Networking Data Points
Support for various segment identification protocols:
– Flat (physical), VLAN and VXLAN
vNIC presented to the guest application ease of adoption and migration from physical
Support for Live VM migration and VM failovers
VLAN separation optimized to minimize overhead
– VLAN tagging in the guest is also supported
Guest networking modes supported with high performance:
– Unmodified guest – with 10X performance stock kernel
– KNI mode – 10G line rate with 1Kbyte packets
– Guest VM w/DPDK Data Path – 10G line rate
0.005.00
10.0015.0020.0025.00
Th
rou
gh
pu
t (G
bp
s)
Frame Size
Guest Throughput (Gbps)
Virtio
Kernel AVP
PMD AVP
37 © 2015 Wind River. All Rights Reserved.
KVM ++ Predictable Performance
40X Reduction in Max Interrupt Latency
After (Scale 50X smaller than above graph)
Before
38 © 2015 Wind River. All Rights Reserved.
TITANIUM SERVER
Security
39 © 2015 Wind River. All Rights Reserved.
CG Security: Integrity and Confidentiality Integrity Titanium Server
Node Infrastructure
‒ Protect availability and reliability of platform infrastructure nodes
Nodal Critical Process Monitor and Recovery
Nodal Resource Monitor
Cluster
‒ Protect the connectivity and availability of platform
1:1 HA Controller Node Cluster
Real-time Monitoring and Recovery of Connectivity to Compute Nodes
Virtual Machine
‒ Protect the availability and reliability of the Hypervisor Real-time Monitoring and Recovery of KVM / QEMU Hosting Container
Program Store
‒ Protect and ensure the authenticity of Titanium server code base (BIOS, Kernel, KVM, OpenStack, Titanium Services)
Controller Services Program Store Signature Validation
Roadmap: Chain of Trust
‒ Program store authenticity checks of BIOS, Kernel and KVM
‒ Program store authenticity checks on compute nodes
Tenant Data
‒ Control who access to tenant data and prevent tampering
Guest VM Volumes Accessible Only thru Private Closed Network
Guest VM Volumes Mountable by One and Only One VM
Confidentiality Titanium Server
Infrastructure Database Encryption
– Protect internal titanium information privacy
Keyring Database for Storage of Encrypted Passwords
Roadmap: Encryption of Titanium / OpenStack Database
Tenant Data Encryption (service)
– Protect tenant information privacy Roadmap: Encryption of Guest VM Volumes
Tenant Data Localization / Filtering (Country and Hosting Rules)
– Ensure tenant data control and protect tenant information privacy Roadmap: Support of Data Localization for Guest VM Volumes
Network Connectivity
– Ensure authenticity and privacy of tenant and infrastructure communication
ACL Filters for Authenticity of Connectivity to Guest VMs (for both external-to-VM and VM-to-VM)
Optional VM Source MAC Filtering (anti-MAC-spoofing)
QOS for Protection of Connectivity to Guest VMs
Roadmap:
‒ ARP spoofing
‒ IPSEC / SSL encryption of connectivity to Guest VMs
‒ IPSEC / SSL encryption of internal platform connectivity
40 © 2015 Wind River. All Rights Reserved.
CG Security: AAA AAA Titanium Server
AAA Architecture
Linux Interfaces
‒ Role-based-access-control provided and configurable
‒ Strict Linux access attributes are enforced
GUI Interfaces, OpenStack CLIs and REST APIs
‒ WEB Server that hosts dashboard / Horizon is configured for SSL to ensure data is encrypted
‒ Role-based-access-control provided and configurable
‒ Leverage keystone key management infra-structure
Linux Access Attributes
LDAP for Centralized Management of Linux User Accounts
‒ User account distribution to all nodes
Support for „Secure‟ Passwords (min. length, upper/lower characters, etc.)
Support for Password Aging
Restricted Access to „Root‟ Account After Initial Installation
Restricted Shell for Non-‟Root‟ Accounts
Configurable Pre-login and Post-login Warning Messages
Auto-logout of Local, SSH and GUI After Configurable Period of Inactivity
Account Usage is Logged for Liability Purposes
No Store of Clear Text Passwords Passwords Stored Only in Encrypted Key Ring Database
Encrypted LDAP User Database
41 © 2015 Wind River. All Rights Reserved.
TITANIUM SERVER
Storage Configuration
42 © 2015 Wind River. All Rights Reserved.
Storage
Scalable storage solution
Small scale on control nodes – up to 10 TB
Medium – Ceph based cluster – up to 100TB
Large – partner with SAN provider – over 100TB
43 © 2015 Wind River. All Rights Reserved.
Storage Technology Considerations
Cinder Recommended for Scalability and Reliability
Cinder
Block storage for booting or standard block device
Backbends options:
– LVM on controller over iSCSI
– Ceph Storage node over RDB
Image created up front
Images used to boot
Volume survives VM failure
No disk copies for live migration
Fast live migrations
Globally available or per tenant
Used for block storage
Ephemeral Storage
Ephemeral shared
Ephemeral local storage in Release 2
– If VM fails storage is lost as well
– Live migration forces a disk copy – long migration times
Image is created at boot time
44 © 2015 Wind River. All Rights Reserved.
Carrier Grade VM Storage Solution
Controller Services
Controller 0
Controller Services
Controller 1
Compute Services
Controller #
Guest VM
Guest VM
Guest VM
Compute Services
Controller #
Guest VM
Guest VM
Guest VM
Compute Services
Controller #
Guest VM
Guest VM
Guest VM
Small Scale
VM volumes on controller
Synchronized between controllers
Access by VMs over internal network (optionally dedicated 10G network for improved disk performance)
Volumes survive VM restarts, migrations, etc.
Volume survive computer node failures
Volumes survive controller features / switchover
Local storage on computer node is planned for release 2
45 © 2015 Wind River. All Rights Reserved.
Carrier Grade VM Storage Solution
Medium Scale
Dedicated storage nodes
CEPH highly-scalable highly-available highly-performant distributed storage solution
Efficient flexible management of backup copies of VM volumes on CEPH storage nodes
Accessed by VMs over dedicated 10G network
– Volumes survive VM restarts, migrations, etc.
– Volumes survive compute node failures
– Volumes survive controller failures / switchovers
– Volumes survive storage node failures / replacements
Controller Services
Controller 0
Controller Services
Controller 1
Compute Services
Controller #
Guest VM
Guest VM
Guest VM
Compute Services
Controller #
Guest VM
Guest VM
Guest VM
Compute Services
Controller # Storage Services
Storage 0
CEPH
Storage Services
Storage N
CEPH
Guest VM
Guest VM
Guest VM
CEPH CEPH
46 © 2015 Wind River. All Rights Reserved.
TITANIUM SERVER
Patching and Upgrades
47 © 2015 Wind River. All Rights Reserved.
Patching and Upgrades
Ability to deploy patches to each node
Comprehensive patch status for each node and at the system level
Patch roll back supported
Rolling upgrades across the Titanium server nodes
Back up and restore of Titanium server
48 © 2015 Wind River. All Rights Reserved.
Moving from PoCs to Commercial Deployments Open Standards Are the Key Enabler
Technology
Evaluation
Commercial
Network
Deployment
Product
Development
Field
Trials
Lab
Trials
Titanium Server
Commercial NFVI Software
Guaranteed Carrier Grade reliability with SLA support
e.g. Open Network Platform (ONP) Reference Software
Accelerates evaluations and
start of product development
Fully Compatible With Open Standards
49 © 2015 Wind River. All Rights Reserved.
Summary
TITANIUM SERVER
50 © 2015 Wind River. All Rights Reserved.
Summary of Carrier Grade Proof Points
Attribute Titanium
Server
Standard OpenStack
and Enterprise Linux
Detection of Failed VM 500 msec > 1 Minute
Compute Node Failure Detection ~ 1 sec 1 Minute or Longer
Controller Node Failure Recovery Sub 25 sec Requires Custom
Development to Enable
Live Migrate DPDK Apps 200 msec Not Possible
vSwitch Throughput 40Gbps @ 256 byte packets 1.5Gbps
Detection of a Failed Link 50 msec Unknown
51 © 2015 Wind River. All Rights Reserved.
Delivering Carrier Grade Reliability for NFV
Live Migration
Sub-200ms outage time
Live migration of DPDK-based VMs
Minimum Latency for Virtualized Applications
10µs deterministic interrupt latency (60x faster than enterprise software)
Automatic Recovery of Failed VMs
Detect failed controllers, hosts or VMs 60x faster than enterprise Linux, with automatic recovery
High-performance Switching for VM-to-VM Traffic
50x performance of Open vSwitch
No compromises from pass-through or SRIOV
Sophisticated VM Scale-up and Scale-down
Enhanced resource monitoring (port, CPU, memory, crypto acceleration, NICs, bandwidth)
Telecom Grade Security
Comprehensive AAA security with VM isolation and full QoS
52 © 2015 Wind River. All Rights Reserved.
Summary
Titanium Server
– Flexible but simple architecture
– HW independent
10G line rate switching supported in SW
Transparent support maintained for interfaces into carrier infrastructure
Carrier grade availability built in – not bolt on
Open API‟s and easy integration models
© 2015 Wind River. All Rights Reserved.
Thank You