an alternative to all - flash arrays - stormagic.com · 100. 1000. 10000. 21 mb 110 gb 221 gb ......
Post on 16-Jul-2019
262 Views
Preview:
TRANSCRIPT
AN ALTERNATIVE TO ALL-FLASH ARRAYS:
PREDICTIVE STORAGE CACHING
THE EASIEST WAY TO INCREASE PERFORMANCE AND LOWER STORAGE
COSTS
Bruce Kornfeld, Chief Marketing Officer, StorMagicLuke Pruen, Technical Services Director, StorMagic
Peter Smith, System Admin, Harris Corporation
STORMAGIC SVSAN: BRIEF OVERVIEW
SvSAN turns the internal disk, SSD and memory of two or more servers into highly available shared storage.
2
ALL-FLASH ARRAYS ARE TEMPTING• SSDs vs. HDD
• 10x-20x more performance
• 10x more expensive
• Lots of hype from flash array vendors
• Not all workloads need all-flash
• A balanced approach with advanced caching algorithms can meet your needs
3
Source: Statistica
SSDs are still 10x more costly than HDD.
SSD: $0.50 - $1.00 per GB
HDD: $0.05 - $0.10 per GB
PREDICTIVE STORAGE CACHINGOver 400% performance improvement with a patent-pending method for predicting when
data will become ‘hot’
4
AUTOMATEDTwo caches – system
memory and SSD
Patent-pending predictive algorithms
Automation tracks ‘hot’ data and places in the right cache
BUILT FOR PERFORMANCE
Lower latency – less waiting for spinning disks
Data pinning
Solves the virtual server ‘I/O blender effect’
COST EFFECTIVEUse fewer and less expensive
disk drives (7.2K)
Lower CAPEX – server memory is inexpensive
Lower OPEX – power, cooling and maintenance
AN ALTERNATIVE TO ALL-FLASH ARRAYS:
PREDICTIVE STORAGE CACHING
THE EASIEST WAY TO INCREASE PERFORMANCE AND LOWER STORAGE
COSTS
Luke Pruen, Technical Services Director, StorMagic
TODAY’S STORAGE OPTIONS
• The performance gap between CPU and storage
• Disk only• High capacity, low
performance workloads
• All-flash• Performance at any cost
workloads
• Hybrid• Most workloads fit here
6
THE IMPORTANCE OF CACHING
• Virtualized environments suffer from the ‘I/O blender effect’
• Working sets of data change over time
• Advanced caching can solve both problems
7
WRITE CACHING
Stage 1• All new data written to SSD
• Data marked ‘dirty’ – not committed to HDDs
Stage 2• Write operation is acknowledged immediately to the server
Stage 3• ‘Dirty’ data is reordered and grouped based on disk locality
• Data de-staged and written to HDD, sequentially as possible
Stage 4• SSD cache notified when data successfully written to HDD
• Cached data marked ‘clean’, remains in cache until space needed
8
WRITE CACHING
Stage 1• All new data written to SSD
• Data marked ‘dirty’ – not committed to HDDs
Stage 2• Write operation is acknowledged immediately to the server
Stage 3• ‘Dirty’ data is reordered and grouped based on disk locality
• Data de-staged and written to HDD, sequentially as possible
Stage 4• SSD cache notified when data successfully written to HDD
• Cached data marked ‘clean’, remains in cache until space needed
8
READ AHEAD & DATA PINNING
Read ahead mode• I/O blender effect aware!
• Identifies sequential interleaved I/Os
• Detects sequential read streams
• Pre-fetches data into memory
9
Data pinning mode• Pin specific data/workloads in memory
• Delivers most efficient read performance
• DBs, VDI, frequently repeated operations
• Manage multiple pin groups
PREDICTIVE CACHING
Predictive approach• Tracker module has 7 levels• All read I/Os monitored and
analyzed• Most frequently used – higher levels• Cache placement based on levels
Flexible storage options• RAM: Most frequently accessed
data• SSD/flash: Next most frequently
accessed data• HDD: Infrequently accessed data –
‘cold’ data
10
Advanced, patent-pending method for predicting ‘hot’ data, placing it on the best storage cache available
Sizing• Assign cache sizes to meet
requirements
• Grow caches as working sets change
• Use any combination of memory, SSD/flash and disk
Play to the strengths• Play to the strengths of all mediums
• Memory highest IOPS
• SSD/flash magnetic drives providing low price per GB
REAL CUSTOMER DATA (BEFORE CACHING)
Workload• 12 Virtual Machines
• 78 applications
11
Performance characteristics of their existing workloads – measured by StorMagic
0
500
1000
1500
2000
2500
3000
3500
18:3
921
:25
00:1
102
:57
05:4
308
:29
11:1
514
:01
16:4
719
:33
22:1
901
:05
03:5
106
:37
09:2
312
:09
14:5
517
:41
20:2
723
:13
01:5
904
:45
07:3
110
:17
13:0
315
:49
18:3
5
IOPs
Time of Day (UTC)
Throughput IOPs
Read
Write
0.0010.01
0.11
10100
100010000
21 M
B11
0 G
B22
1 G
B33
2 G
B44
3 G
B55
4 G
B66
5 G
B77
6 G
B88
7 G
B99
8 G
B1.
1 TB
1.2
TB1.
3 TB
1.4
TB1.
5 TB
1.6
TB1.
7 TB
1.8
TB2.
0 TB
2.1
TB2.
2 TB
2.3
TB2.
4 TB
2.5
TB2.
6 TB
Num
ber o
f acc
esse
s (lo
garit
hmic
sca
le)
Thou
sand
s
Locality of access
Read
Write
0
5,000,000
10,000,000
15,000,000
20,000,000
25,000,000
1KB
2KB
4KB
8KB
16KB
32KB
64KB
128KB
256KB
512KB
1MB
2MB
4MB
Hit
Cou
nt
Block Size
Block Size Distribution
Writes
Read
Read Write
Read/Write % 77% 23%
Sequential % 49% 39%
Average Per Day 991 GB 294 GB
Average Block Size 58 KB 54 KB
Average IOPS 212 138
REAL CUSTOMER DATA (WITH CACHING)
HDD Only (no caching)
• 1 x RAID5 = 3 x 1.2TB 10K SAS disks
HDD & Memory (with caching)
• 12GB of memory per host for caching
• 1 x RAID5 = 3 x 1.2TB 10K SAS disks
HDD & SSD (with caching)
• 1 x 200GB Samsung SSD
• 1 x RAID5 = 3 x 1.2TB 10K SAS disks
HDD, SSD & Memory (with caching)
• 12GB of memory per host for caching
• 1 x 200GB Samsung SSD
• 1 x RAID5 = 3 x 1.2TB 10K SAS disks
12
Impact of Predictive Storage Caching to increase performance
2400
12341
6165
13061
0 2000 4000 6000 8000 10000 12000 14000
HDD Only
HDD & Memory
HDD & SSD
HDD, SSD & Memory
Total IOPS
21.86
4.94
9.37
5.21
0 5 10 15 20 25
HDD Only
HDD & Memory
HDD & SSD
HDD, SSD & Memory
AVG Latency (ms)
HDD Only HDD & Memory HDD & SSD HDD, SSD & MemoryTotal IOPS 2400 12341 6165 13061AVG Latency (ms) 21.86 4.94 9.37 5.21
With Caching
SYNTHETIC ‘HERO’ NUMBERS
HDD Config
• 1 x RAID5 = 3 x 1.2TB 10K SAS disks
Tiered Config
• 1 x RAID5 = 3 x 1.2TB 10K SAS disks
• 1 x 200GB Samsung SSD
• 32GB of memory per host for caching
13
High performance achievement in a controlled lab environment
4263
2004
1832
7729
8728
207117
38196
79738
85803
15651
4KB 100% RND RD
4KB 100% RND WR
4KB 70/30 RD/WR 80% RND
32KB 100% SEQ RD
32KB 100% SEQ WR
IOPS
HDD
Tiered
60.04
127.17
147.36
33.62
29.33
1.24
6.70
3.21
2.98
16.36
4KB 100% RND RD
4KB 100% RND WR
4KB 70/30 RD/WR 80% RND
32KB 100% SEQ RD
32KB 100% SEQ WR
AVG Latency (ms)
HDD
Tiered
Test
• vSphere 6.5
• 2 x Windows VMs
• IOmeter
• VMDK on VMFS
Total IOPS Average Response (ms)
Workload HDD Tiered HDD Tiered4KB 100% Random Read 4263 207117 60.04 1.244KB 100% Random Write 2004 38196 127.17 6.704KB 70/30 Read Write 80% Random 1832 79738 147.36 3.2132KB 100% Sequential Read 7729 85803 33.62 2.9832KB 100% Sequential Write 8728 15651 29.33 16.36
SUMMARY: PREDICTIVE STORAGE CACHING
Increased performance – without the high cost of all-flash arrays
Flexible configurations – we’ll help ‘right-size’ the server hardware
Lower CAPEX – use slower drives for capacity, low-cost system memory and some SSD for performance
Lower OPEX – less power, cooling and maintenance
14
Q&A – HARRIS CORPORATIONPeter SmithSystems AdministratorNetwork, Platforms and Security GroupHarris Corporation
15
NEXT STEPSFurther reading:
An overview of SvSAN – http://stormagic.com/svsan/
SvSAN data sheet – http://stormagic.com/svsan-data-sheet/
Predictive Storage Caching – http://stormagic.com/svsan/predictive-storage-caching/
16
Download your free trial of SvSAN
stormagic.com/trial
SvSAN Product InformationProduct Options
SvSAN license 2, 6, 12 and unlimited TBs
License entitlement 2 mirrored servers
Maintenance and support Platinum - 24x7 / Gold - 9x5
For further information, please contact: sales@stormagic.com
top related