vmworld 2013: just because you could, doesn't mean you should: lessons learned in storage best...

45
Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0) Patrick Carmichael, VMware STO4791 #STO4791

Upload: vmworld

Post on 04-Jul-2015

25 views

Category:

Technology


2 download

DESCRIPTION

VMworld 2013 Patrick Carmichael, VMware Learn more about VMworld and register at http://www.vmworld.com/index.jspa?src=socmed-vmworld-slideshare

TRANSCRIPT

Page 1: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

Just Because You Could, Doesn't Mean You Should:

Lessons Learned in Storage Best Practices (v2.0)

Patrick Carmichael, VMware

STO4791

#STO4791

Page 2: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

2 2

Just Because You Could, Doesn’t Mean You SHOULD.

Storage Performance and Technology

Interconnect vs IOP

Command Sizing.

PDU Sizing

SSD vs Magnetic Disk

Tiering

SSD vs Spinning Media, Capacity vs Performance

Thin Provisioning

Architecting for Failure

Planning for the failure from the initial design

Complete Site Failure

Backup RTO

DR RTO

The low-cost DR site

Page 3: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

3 3

Storage Performance – Interconnect vs IOP

Significant advances in interconnect performance

FC 2/4/8/16GB

iSCSI 1G/10G/40G

NFS 1G/10G/40G

Differences in performance between technologies

• None – NFS, iSCSI and FC are effectively interchangeable

• IO Footprint from array perspective.

Despite advances, performance limit is still hit at the media itself

90% of storage performance cases seen by GSS that are not config related,

are media related

Payload (throughput) is fundamentally different from IOP (cmd/s)

IOP performance is always lower than throughput

Page 4: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

4 4

Factors That Affect Performance

Performance versus Capacity

Disk performance does not scale with drive size

Larger drives generally equate lower performance

IOPS(I/Os per second) is crucial

How many IOPS does this number of disks provide?

How many disks are required to achieve a

required number of IOPS?

More spindles generally equals greater performance

Page 5: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

5 5

Factors That Affect Performance - I/O Workload and RAID

Understanding workload is a crucial consideration when designing

for optimal performance

Workload is characterized by IOPS and write % vs read %

Design choice is usually a question of:

How many IOPs can I achieve with a given number of disks?

• Total Raw IOPS = Disk IOPS * Number of disks

• Functional IOPS = (Raw IOPS * Write%)/(Raid Penalty) + (Raw IOPS * Read %)

How many disks are required to achieve a required IOPS value?

• Disks Required = ((Read IOPS) + (Write IOPS*Raid Penalty))/ Disk IOPS

Page 6: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

6 6

IOPS Calculations – Fixed Number of disks

Calculating IOPS for a given number of disks

• 8 x 146GB 15K RPM SAS drives

• ~150 IOPS per disk

• RAID 5

• 150 * 8 = 1200 Raw IOPS

• Workload is 80% Write, 20% Read

• (1200*0.8)/4 + (1200*0.2) = 480 Functional IOPS

Raid Level IOPS(80%Read 20%Write) IOPS(20%Read 80%Write)

5 1020 480

6 1000 400

10 1080 720

Page 7: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

7 7

IOPS Calculations – Minimum IOPS Requirement

Calculating number of disks for a required IOPS value

1200 IOPS required

15K RPM SAS drives. ~150 IOPS per disk

Workload is 80% Write, 20% Read

RAID 5

Disks Required = (240 + (960*4))/150 IOPS

27 Disks required

Raid Level Disks(80%Read 20%Write) Disks (20%Read 80%Write)

5 13 27

6 16 40

10 10 14

Page 8: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

8 8

Storage Performance – Interconnect vs IOP – contd.

While still true for 90% of workloads, this is starting to shift

SCSI commands passed at native sizes

• Windows 2k8/2k12 (8MB copy)

• Linux (native applications, up to 8MB+)

• NoSQL Databases (Especially Key/Value or Document stores)

• Cassandra (128k)

• Mongo (64-256k)

ESX will gladly pass whatever size command the guest requests.

New applications are starting to run into some protocol limitations

• iSCSI PDU (64k/256k)

• NFS Buffer/Request Size

• Write Amplification!

Offloading these processes offers significant performance

improvements

Page 9: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

9 9

Storage Performance – Interconnect vs IOP – contd.

0.00%

20.00%

40.00%

60.00%

80.00%

100.00%

120.00%

140.00%

160.00%

180.00%

200.00%

4k 8k 16k 32k 128k 1M 2M 4M 8M

Efficiency/Size - Read

Efficiency/Size - Read

Page 10: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

10 10

Tiering

Very old concept

• Primarily a manual process in the past

• Based on RAID groups / Disk types

• Second generation arrays offer batch mode / scheduled auto-movement

system of blocks under IO pressure

Page 11: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

11 11

Why Tiering Works, and Matters

Page 12: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

12 12

Tiering – contd.

New arrays from several partners now offer near real-time

block movement

• Dot Hill AssuredSAN

• X-IO ISE2

Or, batch mode

• Compellent

• VNX

Auto-tiering will get you better performance, faster

• ~2 hours for 150G working set to be fully moved to SSD.

However…

Page 13: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

13 13

Tiering – contd.

Remember – Just because you could, doesn’t mean you should!

There are several places where an auto-tiering array may not be

able to help

• Working Set

• Total Data > SSD Space

• Second tier of cache to worry about

• Fully utilized

• May not have spare IOPS with which to move data between layers.

• Rapidly changing datasets

• VDI

• In-memory databases.

• Often hard to predict when they’re fully utilized

• “Never ending” performance bucket.

Page 14: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

14 14

Tiering – contd.

These platforms have amazing and very capable technology

Have the vendor show you, with your workload, what their platform

is capable of

Workloads differ massively between users and vendors – validate!

Page 15: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

15 15

SSD vs Spinning Disk

SSDs offer extremely high performance per single “spindle”

• Exceptionally great for “L2 Cache”, especially on auto-tiering arrays.

• Capable of 1,000,000+ IOPS in pure SSD arrays

• Exceptionally dense consolidation ratios (Excellent for VDI/HPC).

However…

Ultra-high performance arrays have other limitations

• Surrounding infrastructure

• Switching (packets/frames a second)

• CPU (packaging frames in software initiators)

• Command sizing

• Some SSDs do not work well outside of certain command sizes

• Architectural limitations

• Overhead loss for scratch space

• Garbage Collect

Page 16: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

16 16

VMFS/VMDK Updates for the Next Release

VMFS5 (ESX5) introduced support for LUNs larger than 2TB

• Still limited to 2TB VMDK, regardless of LUN sizing

• 64TB RDM

Major update in this release

• Max size: 64TB for both VMDK and VMFS volumes.

• Lets you avoid spanning disks inside of a guest, or using RDMs for large

workloads

Page 17: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

17 17

VMDK Contd.

VMFS5, in combination with ATS, will

allow consolidation of ever-larger

number of VMs onto single VMFS

datastores

Alternatively, a couple of very large

VMs can occupy a single lun

• While potentially easier for management, this

means that significantly more data is now

stored in one location.

• When defining a problem space, you’ve now

expanded greatly the amount of data stored

in one place.

Just because you could, does that

mean you should?

Page 18: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

18 18

Thin Provisioning

Thin provisioning offers very unique opportunities to manage your

storage “after” provisioning

• Workloads that require a certain amount of space, but don’t actually use it

• Workloads that may grow over time, and can be managed/moved if they do

• Providing space for disparate groups that have competing requirements

Many arrays ~only~ offer Thin Provisioning, by design

The question is, where should you thin provision, and what are the

ramifications?

Page 19: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

19 19

Thin Provisioning – VM Level

VM disk is created @ 0b, until used, and then grows @ VMFS Block

size as needed

Minor performance penalty for provisioning / zeroing

Disk cannot currently be shrunk – once grown, it stays grown

• There are some workarounds for this, but they are not guaranteed

What happens when you finally run out of space?

• All VMs stun until space is created

• Production impacting, but potentially a quick fix (shut down VMs, adjust

memory reservation, etc)

• Extremely rare to see data loss of any kind

Page 20: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

20 20

Thin Provisioning – LUN Level

Allows your array to seem bigger than it actually is

Allows you to share resources between groups (the whole goal

of virtualization)

Some groups may not use all or much of what they’re allocated,

allowing you to utilize the space they’re not using

Standard sized or process defined luns may waste significant

amounts of space, and space being wasted is $$ being wasted

Significant CapEX gains can be seen with thin luns

Page 21: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

21 21

Thin Provisioning – LUN Level - contd

What happens when you finally run out of space?

• New VAAI T10 primitives, for compatible arrays, will let ESX know that the

underlying storage has no free blocks (related to NULL_UNMAP)

• If VAAI works, and your array is compatible, and you’re on a supported version of

ESX, this will result in the same as a thin VMDK running out of space – All VMs will

stun (that are waiting on blocks). VMs not waiting on blocks will continue as normal.

• Cleanup will require finding additional space on the array, as VSWP files / etc will

already be allocated blocks at the lun level. Depending on your utilization, this may

not be possible, unless UNMAP also works (very limited support at this time).

• If NULL_UNMAP is not available for your environment, or does not work

correctly, then what?

• On a good day, the VMs will simply crash with a write error, or the application inside

will fail (depends on array and how it handles a full filesystem).

• Databases and the like are worst affected, will most likely require rebuild/repair.

• And on a bad day?

Page 22: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

22 22

Thin Provisioning LUN Level – contd.

Page 23: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

23 23

Thin Provisioning

There are many reasons to use Thin Provisioning, at both the VM

and the LUN level

Thin provisioning INCREASES the management workload of

maintaining your environment. You cannot just ignore it

Page 24: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

24 24

Freeing space on thin provisioned LUNs

ESX 5.0 introduced NULL_UNMAP, a T10 command to mark blocks

as “freed” for an underlying storage device.

• Disabled in P01 due to major performance issues

• Re-enabled in U1, manual only, to prevent sudden spikes in load, or corruption

from array firmware issues.

Vmkfstools –y command, set amount of space to try and reclaim.

*VALIDATE IF VMFSTOOLS / Auto unmap made it into 5.5*

Page 25: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

25 25

Just Because You Could, Doesn’t Mean You Should

Everything we’ve covered so far is based on new technologies

What about the existing environments?

The ultimate extension of “Just because you could, doesn’t mean

you should,” is what I call “Architecting for Failure”

Page 26: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

26 26

Architecting for Failure

The ultimate expression of “Just because you could, doesn’t mean

you should.”

• Many core infrastructure designs are built with tried and true hardware and

software, and people assume that things will always work

• We all know this isn’t true – Murphy’s law

Architect for the failure.

• Consider all of your physical infrastructure.

• If any component failed, how would you recover?

• If everything failed, how would you recover?

• Consider your backup/DR plan as well

Page 27: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

27 27

Black Box Testing

Software engineering concept

• Consider your design, all of the inputs, and all of the expected outputs

• Feed it good entries, bad entries, and extreme entries, find the result, and

make sure it is sane

• If not, make it sane

This can be applied before, or after, you build your environment

Page 28: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

28 28

Individual Component Failure

Black box: consider each step your IO takes, from VM to physical

media

Physical hardware components are generally easy to compensate

for

• VMware HA and FT both make it possible for a complete system to fail with

little/no downtime to the guests in question

• Multiple hardware components add redundancy and eliminate single points of

failure

• Multiple NICs

• Multiple storage paths

• Traditional hardware (multiple power supplies, etc)

• Even with all of this, many environments are not taking advantage of these

features

Sometimes, the path that IO takes passes through a single point of

failure that you don’t realize is one

Page 29: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

29 29

What About a Bigger Problem?

You’re considering all the different ways to make sure individual

components don’t ruin your day

What if your problem is bigger?

Page 30: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

30 30

Temporary Total Site Loss

Consider your entire infrastructure during a temporary complete

failure

• What would happen if you had to bootstrap it cold?

• This actually happens more often than would be expected

• Hope for the best, prepare for the worst

• Consider what each component relies on – do you have any circular dependencies?

• Also known as the “Chicken and the Egg” problem, these can increase your

RTO significantly

• Example: Storage mounted via DNS, all DNS servers on the same storage

devices. Restoring VC from backup when all networking is via DVS

• What steps are required to bring your environment back to life?

• How long will it take? Is that acceptable?

Page 31: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

31 31

Permanent Site Loss

Permanent site loss is not always an “Act of God” type event

• Far more common is a complete loss of a major, critical component

• Site infrastructure (power, networking, etc) may be intact, but your data is not

• Array failures (controller failure, major filesystem corruption, RAID failure)

• Array disaster (thermal event, fire, malice)

• Human error – yes, it happens!

• Multiple recovery options – which do you have?

• Backups

• Tested and verified?

• What’s your RTO for a full failure?

• Array based replication

• Site Recovery Manager

• Manual DR

• How long for a full failover?

• Host based replication

Page 32: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

32 32

Permanent Site Loss – contd.

Consider the RTO of your choice of disaster recovery technology

• It equates directly to the amount of time you will be without your virtual

machines

• How long can you, and your business, be without those services?

• A perfectly sound backup strategy is useless, if it cannot return you to

operation quickly enough

Architect for the Failure – make sure every portion of your

environment can withstand a total failure, and recovery is possible

in a reasonable amount of time

Page 33: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

33 33

Let’s Be Honest for a Moment…

We would all love to have a DR site that is equally capable as your

production site.

• I’d also like a horsey. Neither of us have all the budget we’d really like for that.

There’s a key DR strategy that is often missed: Data existence is

far more important than full operation

• Many places decline to implement a DR strategy due to budget, time, or

believing that they have to build an equal DR site as their production

environment.

• For many of these places, having key services ~only~ online, with the rest of

the data being protected at the DR site and present if needed, is more than

enough.

Page 34: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

34 34

How to Do It On a Budget of Almost $0

VMware GSS set out to prove that you can build a “protection” site

with a lot less than you’d expect

Services required were true tier 1 apps:

• Store front end (based on DVD Store)

• Exchange

• SQL

• Oracle RAC

• JVM

Page 35: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

35 35

Hardware

8x Dell D630 Laptops (4g, 150G internal drives)

2x Dell Inspiron 5100 Laptops (8g, 200G internal drives)

4x Dell Precision 390 Workstations (8G/16G, 5x 1TB internal drives)

1 Intel Atom Supermicro (2G / 5x1TB Internal Drives)

• Found by the side of the road

• No joke

1 Dumb Ethernet Switch (Netgear), purchased from Best Buy

Duct tape, 1 roll (boot drive for one of the 390s is taped to the

chassis)

4x Broadcom dual-port 1G cards (borrowed from R710 servers)

Page 36: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

36 36

Software

ESX 5.1 U1

SRM 5.1.1

Ubuntu + knfsd (low performance storage)

• Host based replication target

Datacore San Symphony V

• 4x Nodes, high performance, array based replication

Starwind iSCSI

• 2x Nodes, high performance, management software.

Lefthand VSA

• 4x Nodes, medium/high performance, array based replication

Page 37: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

37 37

End results

While we were certainly unable to bring up the full environment on

the DR site, we were able to replicate all 6TB of data successfully.

Critical services for managing a disaster-caused outage were

online

• 1/5 of the Exchange mailboxes (executives and IT personnel)

• Web front end

• SQL database for minimal customer queries/interactions

• No major processing

• RAC

• Online for queries, no processing

• JVM

• Offline (RAC processing engine).

This certainly isn’t ideal, but everything was intact, we could

“manage” the situation, and the “company” hadn’t vanished.

Page 38: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

38 38

You Didn’t Believe Me…

Page 39: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

39 39

Really!

Page 40: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

40 40

Conclusion

You can make a DR site out of almost anything

• Retired servers

• Spare parts

• Out of warranty equipment

You don’t need high end equipment – hardware a generation or two

old is more than enough for basic, or even partial operation

Existing software solutions make it very easy to turn servers alone

into extremely capable shared storage solutions

Page 41: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

41 41

The End.

Questions?

Page 42: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

42 42

Other VMware Activities Related to This Session

HOL:

HOL-SDC-1304

vSphere Performance Optimization

Group Discussions:

STO1000-GD

Overall Storage with Patrick Carmichael

STO4791

Page 43: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

THANK YOU

Page 44: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)
Page 45: VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)

Just Because You Could, Doesn't Mean You Should:

Lessons Learned in Storage Best Practices (v2.0)

Patrick Carmichael, VMware

STO4791

#STO4791