nutanix - expert session - metro availability

Post on 15-Jul-2015

407 Views

Category:

Software

1 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Nutanix Metro-Availability

Christian Johannsen, Senior SE Nutanix

Nutanix – Technology Review

3

Nutanix Virtual Computing Platform

4

Convergence21

Dat

a15

Metadata8

Clo

ud

4

VM Mobility3

Control Plane10

VD

I2

Map

Red

uce

2

Security1Support1

An

alyt

ics1

• Shared-nothing storage controller for virtualization environments.

• Method for networking converged shared-nothing storage for high availability.

• I/O and storage for a virtualization environment with multiple hypervisor types.

• Performing hot-swap of a storage device in a converged architecture.

Key Patents

Top Categories

Web-scale Foundation Platform

22 patents filed

Powerful Control Plane

10 patents filed

Scale-out Data Plane

15 patents filed

47 Patents

Patent Distribution

Nutanix Patent Portfolio

5

Nutanix Distributed File System (NDFS)

Virtual Storage Control Virtual Storage Control

Virtual Machine/Virtual Disk

Flash HDD

Enterprise Storage

• Data Locality• Tiering and Caching• Compression• Deduplication• Shadow Clones• Snapshots and Clones

Data Protection

• Converged Backups• Integrated DR• Cloud Connect• Metro Availability• 3rd party Backup

Solutions

Resiliency

• Tunable Redundancy• Data Path Redundancy• Data Integrity Checks• Availability Domains

Security

• Data at Rest Encryption• Nutanix Security DL• Cluster Shield• Two-factor Auth

Nutanix – Data Protection

7

Stay covered for Critical Workloads

RTORPONutanix Offers

Minutes Minutes Time Stream

Hours Hours Cloud Connect

Near -zero Minutes Metro Availability

Minutes Minutes Remote Replication

Minor

incidents

Major

incidents

Time between backups Maximum tolerable outage

8

Time Stream

Set retention policy for local

and remote snapshots

Set snapshot schedule for a

protection domain

Time-Based backup (Storage Snapshot) with local and remote retention

Imagine the snapshots beside the integrated replication

Application consistent snapshot possible

9

Nutanix Cloud Connect

Datacenter Cloud

Backup and recovery of VMs from Nutanix cluster to the public cloud

VMCaliber and WAN optimized

Fully integrated management experience with Prism

Quick restore and state recovery

10

Async DR

VM-centric workflows

Granular VM based snapshots and policies, better than LUN based

Space efficient sub block-level snapshots (redirect-on-write)

N-Way master-master model for more than on site

VM and application level crash consistency

11

Introducing Nutanix Metro Availability

Geographic distributed High Availability – covers the entire infrastructure stack

Covers entire infrastructure stack

Leverage existing network

Deploy in minutes through Nutanix Prism with minimal change management

Mix and match models to workloads

Customer

Network

12

Network

<=5ms RTT

< 400 KMs between two sites

Bandwidth depends on ‘data change rate’

Recommended: redundant physical networks b/w sites

General

2 Nutanix clusters, one on each site

Mixing hardware models allowed

Hypervisor

ESXi in NOS 4.1

Hyper-V/KVM in the future (Q1 CY 2015)

Requirements

13

Architecture

Synchronous storage replication

Datastore stretched over both Nutanix cluster in a single Hypervisor cluster (vMotion, HA)

In conjunction with existing data management features, compression, deduplication, and

tiering

Standby containers are unavailable for direct virtual machine traffic (first release)

14

Nutanix I/O Path

I/O Path

1. OpLog acts as a write buffer (Random Writes)

2. Data is replicated to other nodes sync.

3. Sequentially drained to Extent Store

4. ILM (Information Lifecycle Management)

chooses the right target or the data

5. Deduplicated read cache (Content Cache) spans

Memory and SSD

6. VM accessing the same data on just one copy

(deduplicated)

1. IF data not in Content Cache it will be promoted

per ILM

2. Extensible Platform for future I/O patterns

1515

1. Write IO

2a. Written to local OpLog (RF) and remote

replication to remote OpLog

2b. Local Replication in remote OpLog (RF)

3a. Write IO Ack in local OpLog (RF)

3b. Write IO Ack in remote OpLog (RF)

3c. Write IO Ack from remote OpLog

4. Write IO Ack from local OpLog to the hypervisor

Write Anatomy

1616

1. Write IO

2. Write IO forwarded to Active Container

3a. Written to local OpLog (RF) and remote

replication to remote OpLog

3b. Local Replication in remote OpLog (RF)

4a. Write IO Ack in local OpLog (RF)

4b. Write IO Ack in remote OpLog (RF)

4c. Write IO Ack from remote OpLog

5. Write IO Ack from local OpLog to the remote

OpLog

6. Write IO Ack from local OpLog to the hypervisor

Write Anatomy (vMotion, Recovery)

1717

1. Read Request

2. Read Request forwarded to Active Container

3. Data returned from the Active Container

4. Data sent to the VM

Read Anatomy (vMotion, Recovery)

1818

Scenarios

1919

Scenarios

Network failure between sites

Manual or Automatic (seconds)

2020

Scenarios

Site Failure

Demo Time!

https://drive.google.com/a/nutanix.com/file/d/0B3sqKkY-Et4deF9Db2NPdlYzMmM/view

Thank You

top related