how port density of a dc lan switch impacts scalability and tco

12
How the Port Density of a Data Center LAN Switch Impacts Scalability and Total Cost of Ownership June 4, 2012

Upload: venugopal-athiur-ramachandran

Post on 21-Jul-2016

8 views

Category:

Documents


0 download

DESCRIPTION

DC port density

TRANSCRIPT

Page 1: How Port Density of a DC LAN Switch Impacts Scalability and TCO

How the Port Density of a Data Center

LAN Switch Impacts Scalability and

Total Cost of Ownership

June 4, 2012

Page 2: How Port Density of a DC LAN Switch Impacts Scalability and TCO

2

Introduction

As data centers are forced to accommodate rapidly

growing volumes of information, their capacities to

process, store, and transfer data are having difficulty

keeping pace. A Gartner survey1 in November 2010

found that 47% of representatives from 1,004 large

enterprises from eight countries ranked data growth as

one of their top three challenges. This was followed by

system performance and scalability at 37%, and network

congestion and connectivity architecture at 36%.

Growing demand based on data proliferation, new

applications, and federated applications are exhausting

the capacities of many data centers. According to the

Uptime Institute’s 2011 Data Center Industry Survey 2

,

more than a third (36%) of data center facilities will run

out of space, power, or cooling between 2011 and 2012.

These findings were based on a survey of 525 data

center executives. Of the respondents reaching full

capacity, 40% plan to build a new data center, while

62% plan to address the problem via server

consolidation. Server virtualization is the primary

mechanism for implementing server consolidation, but

server virtualization introduces new sources of server-

to-server traffic that pose challenges for the scalability

of the data center network. Server virtualization also

requires that the bandwidth of each physical server’s

network connection scale in proportion to the increased

demands for network I/O from multiple virtual

machines sharing the same physical server port(s).

As IT managers re-architect or upgrade their data

centers their focus will be on assuring the ability to

scale computing performance and networking capacity

within the facility’s constraints of space, power, and

cooling capacities.

The goal of this white paper is help data center owners

and designers quantify the value of 40 Gigabit Ethernet

switch port density and scalability as part of a Total

Cost of Ownership (TCO) approach to evaluating

alternative data center network designs and 1 http://www.informationweek.com/news/hardware/data_centers/229600034 2 http://gigaom2.files.wordpress.com/2011/06/inaugural-uptime-institute-

annual-data-center-survey-fact-sheet.pdf

Definition of Symbols

P: The number of 40 GbE ports per

aggregation switch

M: The effective over-subscription

Ratio

S: The number of aggregation switches

Ccore: The number of 40 GbE ports per

aggregation switch that are used to connect to the LAN

coreCagg: The number of 40 GbE ports per aggregation switch used to connect

to other aggregation switches

Cacc: The number of connections

between TOR switches

P – Ccore – Cagg: The number of 40 GbE ports per aggregation switch

available for connections to the access

layer

4 x m x (P-Ccore-Cagg): The number

of 10 GbE access layer ports that are

available for server connection per aggregation

4 x S x m x (P-Ccore-Cagg): For two tier LAN design with multiple

aggregation switches, the number of

available server ports

Page 3: How Port Density of a DC LAN Switch Impacts Scalability and TCO

3

implementations with different brands of switching products. This white paper is part of an

on-going series of documents and webinars from Extreme Networks that focus on quantifying

the value of the key attributes of a data center LAN switch.

The Impact of Server Virtualization on Data Center LANs

The widespread adoption of server virtualization within the data center is introducing new

traffic patterns that must be accommodated by the data center network. In addition to the

traditional client/server traffic flows (“north-south” traffic), virtual machine migrations and

access to networked storage give rise to a significant amount of server-to-server (“east-west”)

traffic within PODs and even across the entire data center. Virtual machine migrations among

physical servers are being used to support a number of critical data center functions, including

load balancing, expanding capacity, disaster recovery, and ensuring high availability. Some of

the impact that server virtualization has had on the data center network includes:

Flatter networks

Virtual machine migrations are typically performed within the boundaries of a VLAN.

Extending VLANs across wider segments of the data center requires the elimination or

circumvention of Layer 3 boundaries between server PODs. Flatter networks can help

lower costs by reducing configuration complexity and by reducing the number of

devices that must be managed. Another benefit of flatter data center networks is

reduced end-to-end latency due to fewer hops between source and destination nodes.

Higher Bandwidth Server Connectivity

Servers with multi-core processors can support a large and growing number of virtual

machines, increasing the demand for network I/O capacity. The trend to multi-core

and multi-virtual machine (VM) servers is accelerating the need to transition server

connectivity from GbE to 10 GbE. Adding to that pressure is the fact that the next

generation of servers emerging in 2012 is expected to feature dual port 10 GbE LANs

on the Motherboard (LOM) based on 10GBASE-T. 10GbE server connectivity will

also provide support for unified storage networking based on NAS, iSCSI, and FCoE.

I/O Virtualization

I/O virtualization supports a number of virtual Ethernet connections over a single

physical port. This allows the assignment of a virtual GbE port to each VM, a

recommended best practice by VMware.

The combination of flatter network designs and a transition of server connectivity from GbE

to dual 10 GbE will result in a much greater emphasis of the scalability and port density of 10

GbE and 40 GbE data center switches.

Page 4: How Port Density of a DC LAN Switch Impacts Scalability and TCO

4

Scalability of Popular Two Tier Data Center LAN Designs

The scalability of a LAN architecture is determined by the number of server ports that can be

supported with a given level of redundancy and over-subscription at different points within

the LAN topology. Many data center LANs being deployed today are based on a two tier

design that provides high levels of redundancy and low over-subscription levels for server-to-

server traffic. Two tier LAN designs are frequently implemented with Top of Rack (TOR)

access switches in conjunction with chassis based aggregation switches. The aggregation

switches are connected to the LAN core and to the Internet, but all the server-to-server traffic

within the data center flows only through the two tiers of access and aggregation switches.

Figure 1 shows a general model for two tier switched LANs that takes into account both

connections for redundancy and connections to the LAN core. It is assumed that all servers

are attached to the access/TOR switches via 10 GbE ports. Any inter-switch links at the

access layer are assumed to be 10 GbE, and all other inter-switch links (i.e., inter-aggregation,

access-to-aggregation and aggregation-to-core) are assumed to be 40 GbE. If a given model of

switch does not yet support 40 GbE, a LAG with four 10 GbE member links could be

substituted. It should be noted, however, that a 40 GbE link is preferable to a LAG of four 10

GbE links because having a single 40 GbE link avoids the issues that can occur when

attempting to load balance traffic that consists of a small number of high volume flows.

Access Layer

Aggregation Layer

Ccore

Cagg

Cacc

P=ports/switch

P-Ccore-Cagg

P-Ccore-Cagg

4m(P-Ccore-Cagg)

Servers

Figure 1: Scalability Model for Two Tier Data Center LANs

Page 5: How Port Density of a DC LAN Switch Impacts Scalability and TCO

5

This model can be applied equally well to two tier LANs based on multi-chassis LAGs and

two tier fat trees. The model focuses on P, the number of 40 GbE ports per aggregation switch

and the number of ports required to make connections both within and among network tiers.

In the model, Ccore is the number of 40 GbE ports per aggregation switch that are used to

connect to the LAN core, Cagg is the number of 40 GbE ports per aggregation switch that are

used to connect to other aggregation switches (e. g., for ISL/VSL). There may also be 10 GbE

inter-switch links within the access/TOR tier to support virtual switch/router functions such as

multi-chassis LAG (MLAG) or VRRP.

The access/TOR switches may be oversubscribed with more switch bandwidth allocated to

server connections vs. the amount of bandwidth that is provided from the access tier to the

aggregation tier. The over-subscription ratio is given by the following ratio:

The amount of bandwidth allocated to server access/The amount of bandwidth

allocated to access-to-aggregation connectivity)

Throughout this white paper, the over-subscription ratio will be referred to by the letter m.

A typical high density TOR switch has 48 10 GbE ports for server connectivity and four 40

GbE ports for inter-switch connectivity. Where servers are single-attached to these TOR

switches, m is equal to (48 x 10)/(4 x 40) = 3. Where the servers are dual-attached to a pair of

TOR switches with active-passive redundancy, m = 3, but the effective over-subscription ratio

is 1.5:1 because only one of the pair of server ports is active at any given time. Where the

servers are dual-attached to a pair of TOR switches with active-active MLAG redundancy, the

requirement for inter-switch connections (Cacc) between the TOR switches means there are

two fewer 10 GbE ports per TOR switch available for server connectivity and the over-

subscription ratio is equal to m = (46 x 10)/(4 x 40) = 2.88

As shown in Figure 1, the number of 40 GbE ports per aggregation switch that is available for

connections to the access layer is equal to P-Ccore-Cagg and the number of 10 GbE access

layer ports that are available for server connection per aggregation is equal to 4 x m x (P-

Core-Cagg). For a two tier LAN design with multiple aggregation switches, the number of

available server ports is 4 x S x m x (P-Core-Cagg), where S is the number of aggregation

switches.

It should be noted that the model presented in Figure 1 is based on having a single

aggregation switch, and the factor S needs to be included to account for an aggregation tier

with multiple aggregation switches. For fat trees, the number of aggregation switches, or

spine switches, is limited by the equal cost forwarding capabilities (16 paths is a typical limit),

as well as the port density P. The port configuration of the access/TOR switch also imposes

some limitations on the number of aggregation/spine switches that can be configured. For

example, for a TOR switch with 48 10 GbE ports and four 40 GbE ports the number of 40

GbE aggregation switches is limited to four. Scaling beyond S=4, requires both a denser

access switch with more 40 GbE ports and more 10 GbE port as well to maintain a desired

Page 6: How Port Density of a DC LAN Switch Impacts Scalability and TCO

6

maximum over-subscription ratio. The ultimate fat tree scalability is attained where the 10

GbE/40 GbE access switch has same switching capacity as the aggregation/spine switches.

With these caveats, the model takes into account redundancy and scalability for various Layer

2 and Layer 3 two-tier network designs as summarized in Table 1.

Parameter 2 Tier L2

2 Tier Layer 2

MLAG

2 Tier Layer 2

Fat Tree

2 Tier Layer 3

Fat Tree

Redundancy none full full Full

Ccore variable variable variable variable

Cagg 0 ISL/VSL

2 per agg

switch

0 0

Cacc 0 active/passive

server access: 0

active/active:

2 per TOR

active/passive

server access: 0

active/active:

2 per TOR

active/passive:

2 per TOR

active/active:

2 per TOR

Max 10 GbE

server ports

4Sm(P-Ccore-

Cagg)

S=1

4Sm(P-Ccore-

Cagg)

S=2

4Sm(P-Ccore-

Cagg)

S = # of

aggregation

switches

4Sm(P-Ccore-

Cagg)

S = # of

aggregation

switches

Scaling Larger P, m Larger P, m Larger P,m,S Larger P,m,S

Table 1: Scalability of Two Tier 10/40 GbE Data Center LANs

As highlighted in Table 1, the only way that the scalability of the data center LAN can be

increased is by increasing the:

Number of aggregation switches

Number of 40 GbE ports per aggregation switch

Level of over-subscription

As stated earlier, a typical initial design process might start from identifying the required

number of server ports, the required redundancy, and an upper limit on the over-subscription

ratio. As shown in Figure 2, calculating the required number of 40 GbE ports per aggregation

switch to meet these requirements is accomplished by inverting the scaling formula. The way

this formula can be used is for an IT organization to:

1. Determine required number of server ports

2. Select the desired network type from Table 1. This will determine Cagg

3. Select an access/TOR switch model. This together with the network type will determine

Cacc and m.

Page 7: How Port Density of a DC LAN Switch Impacts Scalability and TCO

7

4. Select the desired Ccore. This will determine over-subscription ratio for client/server

traffic via the core

5. Calculate the required port density of the aggregation switch

P = ((# of server ports)/4 x S x m) + Ccore + Cagg

Figure 2: Required Aggregation Switch Port Density

for Two Tier 10/40 GbE Data Center LANs

To exemplify the formula shown in Figure 2, consider the following network parameters:

The number of servers ports = 4512

M = 3

S = 2

Ccore = 2

Cagg = 2

The formula in Figure indicates that in order to support the indicated network parameters, an

access switch with 192 40 GbE ports is required.

Figure 3 shows an example of a data center network that provides fully redundant Layer 2

server-to-server connectivity based on 94 TOR switches, each having 48 10 GbE ports and 4

40 GbE ports plus a pair of high density aggregation switches with 192 40 GbE ports each.

The topology is an MLAG Layer 2 network with oversubscribed TOR switches. Each of the

2,256 servers is connected to two TOR switches in an active/passive mode. The same

configuration could also support 4,512 single-attached servers. With active/passive

redundancy, the over-subscription of access switches for server-to-server traffic is 1.5:1.

Two AggregationSwitches with 192 40 GbE

ports each

4,512 10GbE server ports

M-LAG with 2 40 GbE links to

each Aggregation Switch

ISL/VSL

94 TOR switches

2,256 servers dual attached active/passive --oversubscription 1.5:1or 4,512 single attached servers --oversubscription 3:1

Four 40 GbE links to LAN

core

Figure 3: Redundant Two Tier Network Configuration

Page 8: How Port Density of a DC LAN Switch Impacts Scalability and TCO

8

For active-active server connectivity, each pair of TOR switches would need to be configured

as a virtual switch with a pair of inter-TOR 10 GbE links for the ISL/VSL connectivity

required for the virtual switch, as shown in Figure 4. This would reduce the number of servers

per TOR switch from 24 to 23 and the number of dual-attached servers to 2,072. With

active/active redundant MLAG server connectivity, the over-subscription ratio for server-to-

server traffic is 2.88:1.

Two AggregationSwitches with 192 40 GbE

ports each

4,144 10GbE server ports

M-LAG with 2 40 GbE links to

each Aggregation Switch

ISL/VSL

94 TOR switches

2,072 servers dual-attached active/actiive –over-subscription 2.88:1

Four 40 GbE links to LAN

core

Figure 4: Redundant Two-Tier Network Configuration

Building a comparable network with essentially the same number of 10 GbE server ports and

similar over-subscription ratios using similar TOR switches and an aggregation switch with

half the density (i.e., 96 40 GbE ports) requires some design changes. Comparing the two

designs provides an illustration of the effect that the density of the aggregation switch can

have on the network design and the resulting TCO.

One possibility would be to build a Layer 2 fat tree network using four aggregation switches

in the spine/aggregation layer and the same number of TOR switches (94) as the leaves/access

switches. However, most TOR switches do not yet support Layer 2 equal cost multi-path

forwarding alternatives other than with some form of MLAG. One workaround is to move the

Layer 3 boundary from the aggregation switch to the TOR switch and build a Layer 3 fat tree

with OSPF ECMP providing the multi-path functionality. Figure 5 shows what this could look

like. Here the ISL links are at the TOR level rather than the aggregation level and the server

connection can be made active/active without affecting the topology. With active/passive

redundancy, the over-subscription of aggregation switches for server-to-server traffic is

1.44:1, while with active/active redundant server connectivity, the over-subscription ratio is

2.88:1. Note that a Layer 2 and Layer 3 fat trees based on switches with the same port

densities at the aggregation and access levels have the same physical topology.

Page 9: How Port Density of a DC LAN Switch Impacts Scalability and TCO

9

Four Aggregation

Switches with 96 40 GbE

ports each

4,144 10GbE server ports

one 40 GbE link to each

Aggregation Switch

2,072 servers dual-attached active/passive --oversubscription 1.44:1or 2.072 dual-attached active/active servers or 4,144 single attached servers –over-subscription 2.88:1

Eight 40 GbE links to LAN

core

94 TOR switches

Figure 5: Redundant Two-Tier, Layer 3 Fat Tree

If a TCO comparison is made of the two networks shown in Figures 4 and 5, some of the

differences to consider are the:

Capex and Opex differences with four switches vs. two at the aggregation level, including

switch cost, power capacity requirements, rack space requirements, annual power, annual

routine admin, and annual service contract costs

Difference in the number of server ports per TOR

Differences in over-subscription ratios to the core

Eight links vs. four links to the LAN core needed for redundancy

Administrative cost and complexity differences with 98 Layer 3 devices vs. two Layer 3

devices

In addition, in a Layer 3 fat tree, there is a need for pre-standard VXLAN, NV-GRE or some

other virtual networking solution to enable VM migration across Layer 3 boundaries.

This example shows some of the complexities that can be encountered in comparing the

TCOs of competing data center switching solutions that are based on switches of different

port densities, as well as somewhat different functionality.

A Data Center LAN Cost Model

A typical design process for a data center LAN would include an analysis of the client/server

and the server-to-server traffic flows, resulting in a preferred topology as well as the required

number of server access ports, permissible over-subscription, required redundancy, and

required connectivity of the data center to the campus LAN and/or Internet.

Page 10: How Port Density of a DC LAN Switch Impacts Scalability and TCO

10

Adding redundancy at the server access level (e.g., by taking advantage of dual port 10 GbE

LOMs) and connecting each server to two physical access switches doubles the number of

server access ports required. Similarly, adding redundancy at other layers of the hierarchy can

essentially double the number of ports required at those levels.

The formula in the preceding section can be used to estimate the total number of 10/40 GbE

ports that are required by the design and the number of aggregation and access switches.

These estimates can then be used to develop Total Cost of Ownership (TCO) comparisons

between different models of switch or different two tier switch topologies.

The equations in Figure 6 provide a simple example of a TCO analysis for a greenfield data

center network requiring all new switches, floor space, and power/cooling facilities. This

analysis focuses on the comparison of competing models of switch in a similar topology. For

simplicity, costs of cabling and some other costs are not considered.

CAPEX (Aggregation) = [(power/port) (# of ports)+ (power/chassis) (# of switches)]

x ($cost/kW power/cooling capacity) +

[(racks/switch) x (# of switches) x 18 sq ft/rack] x $cost/sq ft +

[(racks/switch) x (# of switches)] x $cost/rack +

[# of switches x $cost/switch]

OPEX (Aggregation) = [(# of switches) x (kw/switch) x (8760 hrs/year) x

($cost/kWhr)] +

[ (# of switches) x ($cost of annual routine admin/switch)] +

[ (# of switches) x ($ annual service contract/switch)]

CAPEX (TOR) = [(power/switch) (# of switches)] x

($cost/kW power/cooling capacity) +

[(racks/switch) x (# of switches) x 18 sq ft/rack] x $cost/sq ft +

[(racks/switch) x (# of switches)] x $cost/rack +

[# of switches x $cost/switch]

OPEX (TOR) = [ (# of switches) x (kw/switch) x (8760 hrs/year) x ($cost/kWhr)] +

[ (# of switches) x ($cost of annual routine admin/switch)] +

[ (# of switches) x ($ annual service contract/switch)]

TCO = CAPEX (Aggregation) + CAPEX TOR) +

NPV [OPEX (Aggregation) + OPEX (TOR)]

Figure 6: A Simple TCO Model

According to the Uptime Institute3, the cost of provisioning a kilowatt of UPC power/cooling

for a Tier II data center is $12,500, while the corresponding figure for a Tier III data center is

$23,000. Their estimate for the corresponding cost per square foot of rack space is $300. The

cost a data center rack is estimated to be $3,500. Cost per kWhr is estimated at $0.073, and

3 Uptime Institute White Paper: Cost Model $/kW and $/sq ft of Computer Floor

Page 11: How Port Density of a DC LAN Switch Impacts Scalability and TCO

11

the cost of annual routine maintenance per Layer 2 switch is estimated at $360. The cost of

annual routine maintenance per Layer 3 switch is estimated at $720.

Based on the TCO model presented in Figure 6 and the above cost parameters, a TCO

comparison was made between two data center LAN designs. One design is the network

shown in Figure 4 and implemented with two 8-slot Extreme Black Diamond X-8 switches

with 192 wire speed 40 GbE ports and 94 Extreme Summit X670V TOR switches. The

second design is shown in Figure 5 and is based on a competing aggregation switch (Switch

X) with 96 wire speed 40 GbE ports and TOR switches of the same port density as the

X670V. The comparison is based on a greenfield data center network with a requirement of

4,144 total server ports, active-active dual 10 GbE server connectivity and an over-

subscription ratio for server-to-server traffic of 3 or less.

Number of

switches

Cost/

switch

44U Racks/

switch

Power/

port

Power/

Chassis

Annual

Service

BD X-8 2 $802,900? 14.5/44 W 2,52 kW

Summit

X670V

94 $25,000 1/44 257W

Switch X

Aggregation

4 $ 25/44 W 2.82 kW

Switch X

TOR L3

94 $44,000 1/44 199W

Table 2: TCO Parameters for Comparison of Two Network Solutions for 4.144 Server

Ports

Switches Racks Tier II

Power

Capacity

Space Annual Power

Routine Admin

Service TCO

Extreme

Switch X

Difference

Table 3: Results of TCO Comparison of Two Tier Network with Aggregation Switches

of Different Densities

In Table 3, the TCO was reflects OPEX NPV calculated based on a 5 year lifetime and an

interest rate of 4%.

Summary

An estimation of the capital cost of a new or upgraded data center LAN needs to take into

account more factors than simply the cost per port of LAN switches being considered. A

preferred approach is to perform a TCO analysis based on a preliminary network design

taking into account the total number of server access ports required, required limits of over-

subscription for server-to-server traffic, the port density of the switches, and the total number

of switches and switch ports required.

Page 12: How Port Density of a DC LAN Switch Impacts Scalability and TCO

12

As has been shown, the port density of the switch can have a significant effect on the LAN

topology, the level of over-subscription, the total number of switch ports, and the network

power consumed in order to support a required number of server access ports. This

underscores the fact that IT organizations that are updating their data center LANs should

place greater emphasis on the port density of LAN switches as a differentiating factor among

competing products.

Beyond switch cost, other significant CAPEX factors include the cost of power/cooling

capacity required to support the switches deployed and the physical density of the switches

which affects the rack space and square footage consumed by network hardware.

The port density of switch, together with the number of switches required, can also directly

affect OPEX in terms of routine administrative costs, power consumption, and maintenance

contract costs.