Download - Intel Open Network Platform
Intel Open network platformBrian Skerry, Sr. SW ArchitectNetwork Platforms Group, Intel
2
SDN and NFV are Driving Network Transformation
TEM/OEM Proprietary OS
ASIC, DSP, FPGA, ASSP
Intel Xeon processor
ChipsetAcceleration
SwitchSilicon
NICSilicon
Open Source
SDN/NFV
VM: Firewal
l
VM: VPN
Single Application on Dedicated Hardware
Firewall VPN Intrusion Detection System
VM: NAT
VM: DPI
VM: LB
NFV M
anagement
and OrchestrationSDN/NFV Infrastructure
Innovation
Enabling the server to become the new networking platform
3
Enable the TransformationAdvance Open Source and
Standards
Deliver Open Reference Architecture
Enable Open Ecosystem on IA
Collaborate on Trials and Deployments
4
Intel® Open Network Platform (Intel® ONP)
Intel® ONP Software Ingredients Based on
Open Source and Open Standards
Industry Standard Server Based on Intel Architecture
What is Intel® ONP Reference Architecture?*
Reference Architecture that brings together hardware and open source software ingredients
Optimized server architecture for SDN/NFV in Telco, Enterprise
and CloudVehicle to drive development and
to showcase solutions for SDN/NFV based on IA
*Not a commercial product
VMVIRTUAL SWITCH
HWOFFLOA
D
LINUX/
KVMDPD
K
ETSI NFV Goals• Improved CAPEX via COTS (instead
of dedicated hardware) • Flexibility in assigning VNFs to
hardware• Rapid service innovation• Improved OPEX from automation• Reduced power usage by migrating
workloads (so unused hardware can be powered down)
• Standardized and open interfaces between VNF and NFVI (to enable multi-vendor solutions)Adapted from: http://www.etsi.org/deliver/etsi_gs/nfv/001_099/002/01.01.01_60/gs_nfv002v010101p.pdf
5
VNF 1 VNF 2 VNF 3
NFVI
NFVI Hardware
NFVI Software
Key Requirements for Network Function Virtualization Data Plane
Programmability
Scalability
Efficiency
High Performance
Security
• Software Programmable & Flexible• Control Plane & Policy Controls
• Scales across multi-core CPU options• Generational Scalability• Best-In-Class Perf/Watt/$$• Seamless integration of platform
accelerators• 40G/100G/nx100G line rate over time• Real time latency & jitter characteristics• Data Security encompassing Platform
Security, Network Security, Storage Security, Trust & Attestation
VNF Virtual Network Interface Options
7
VNF A
virtio
KernelStack
Network
App
StockvSwitch
AnyNIC
VNF BDPDK
virtio
Network
App
AnyNIC
VNF CDPDK
virtio
Network
App
DPDKvSwitch
NIC
VNF DDPDK
IVSHMEM
Network
App
NIC
VNF E
SR-IOVNIC
Performance
Flexibility, VNF-NFVI Independence
VNF F
NIC VF Driver
KernelStack
Network
App
SR-IOVNIC
StockvSwitch
DPDKvSwitch
DPDK
NIC VFPMD
Network
App
vSwitch Acceleration is the most optimal solution for a scalable NFVI
8
Open Network Plaform
8
EMSEMS
OpenStack Plug
in
Enhancements
OSS/BSS EMS
OCP NodeOCP Node Linux / KVM
vCPE vBRAS vEPC
ONP
Open vSwitch with DPDK
NIC
VNF Manager
(Service) Orchestrator
Open FlowOVSDB Other
OCP NodeOCP Node Linux / KVM
ServerApp vFW vADC
ONP
Open vSwitch with DPDK
NIC
Enhancements
9
Open vSwitch
ovs-switchd
NICovs kernel module
qemu
VMvirtio
kernel packet
processing
User Space Forwarding
socketTAP netdev
User Space
External
OpenDaylight
ovsdb OF
ovsdb server
ovs-switchd
DPIF
10
Open vSwitch with DPDK
ovs-switchd
NIC
DPDK Libraries PMD
DPDKnetdev
ovs kernel module
vHost User
VMvirtio
kernel packet
processing
User Space Forwarding
socketTAP netdev
User Space
External
OpenDaylight
ovsdb OF
ovsdb server
ovs-switchd
DPIF
Available on openvswitch.org
Tunnels
OpenvSwitch 2.4Platform Performance Configuration
Item DescriptionServer Platform Intel® Server Board S2600WT2 DP (Formerly Wildcat Pass)
2 x 1GbE integrated LAN portsTwo processors per platform
Chipset Intel® C610 series chipset (Formerly Wellsburg)Processor Intel® Xeon® Processor E5-2697 v3 (Formerly Haswell)
Speed and power: 2.60 GHz, 145 WCache: 35 MB per processorCores: 14 cores, 28 hyper-threaded cores per processor for 56 total hyper-threaded coresQPI: 9.6 GT/s Memory types: DDR4-1600/1866/2133, Reference: http://ark.intel.com/products/81059/Intel-Xeon-Processor-E5-2697-v3-35M-Cache-2_60-GHz
Memory Micron 16 GB 1Rx4 PC4-2133MHz, 16 GB per channel, 8 Channels, 128 GB TotalLocal Storage 500 GB HDD Seagate SATA Barracuda 7200.12 (SN:9VMKQZMT)PCIe Port 3a and Port 3c x8NICs 2 x Intel® Ethernet CAN X710-DA2 Adapter (Total: 4 x 10GbE ports)
(Formerly Fortville)BIOS Version: SE5C610.86B.01.01.0008.021120151325
Date: 02/11/2015
OpenvSwitch 2.4Phy-OVS-Phy Performance
Disclaimer: For more complete information about performance and benchmark results, visit www.intel.com/benchmarks and https://download.01.org/packet-processing/ONPS1.5/Intel_ONP_Server_Release_1.5_Performance_Test_Report_Rev1.2.pdf
OpenvSwitch 2.4Phy-VM-Phy Performance
Aggregate Switching Rate
Disclaimer: For more complete information about performance and benchmark results, visit www.intel.com/benchmarks and https://download.01.org/packet-processing/ONPS1.5/Intel_ONP_Server_Release_1.5_Performance_Test_Report_Rev1.2.pdf
Need for an Efficient Data Plane for NFV
Server AVNF1 VNF1
vSwitch
NIC
VNFnServer B
VNF1 VNF1 VNFn
vSwitch
NIC
Server CVNF1 VNF1 VNFn
vSwitch
NICNSH VXLAN-GPE Encapsulatio
n
NSH Forwarding L2/L3
Forwarding
L2/L3 Routing
NSH: Network Services HeaderVNF: Virtual Network Function
ServiceChain #1
ServiceChain #2
A Programmable, Scalable, Efficient & High Performance Data Plane is a key requirement for NFV deployments
OpenvSwitch 2.4Phy-OVS Tunnel-Phy Performance
Aggregate Switching Rate
Disclaimer: For more complete information about performance and benchmark results, visit www.intel.com/benchmarks and
https://download.01.org/packet-processing/ONPS1.5/Intel_ONP_Server_Release_1.5_Performance_Test_Report_Rev1.2.pdf
System SettingsSystem Capability Version
Host Operating System Fedora 21 x86_64 (Server version)
Kernel version: 3.17.4-301.fc21.x86_64VM Operating System Fedora 21 (Server version)
Kernel version: 3.17.4-301.fc21.x86_64libvirt libvirt-1.2.9.3-2.fc21.x86_64QEMU QEMU-KVM version 2.2.1
http://wiki.qemu-project.org/download/qemu-2.2.1.tar.bz2DPDK DPDK 2.0.0
http://www.dpdk.org/browse/dpdk/snapshot/dpdk-2.0.0.tar.gzOVS with DPDK-netdev Open vSwitch 2.4.0
http://openvswitch.org/releases/openvswitch-2.4.0.tar.gz
System Capability DescriptionHost Boot Settings HugePage size = 1 G; no. of HugePages = 16
HugePage size = 2 MB; no. of HugePages = 2048
intel_iommu=off
Hyper-threading disabled: isolcpus = 1-13,15-27
Hyper-threading enabled: isolcpus = 1-13,15-27,29-41,43-55VM Kernel Boot Parameters GRUB_CMDLINE_LINUX="rd.lvm.lv=fedora-server/root
rd.lvm.lv=fedora-server/swap default_hugepagesz=1G hugepagesz=1G hugepages=1 hugepagesz=2M hugepages=1024 isolcpus=1,2 rhgb quiet"
System Capability ConfigurationDPDK Compilation CONFIG_RTE_BUILD_COMBINE_LIBS=y
CONFIG_RTE_LIBRTE_VHOST=y
CONFIG_RTE_LIBRTE_VHOST_USER=y
DPDK compiled with "-Ofast -g"OVS Compilation OVS configured and compiled as follows:
#./configure --with-dpdk=<DPDK SDK PATH>/x86_64-native-linuxapp \
CFLAGS="-Ofast -g"
make CFLAGS="-Ofast -g -march=native"DPDK Forwarding Applications
Build L3fwd: (in l3fwd/main.c)
#define RTE_TEST_RX_DESC_DEFAULT 2048
#define RTE_TEST_TX_DESC_DEFAULT 2048
Build L2fwd: (in l2fwd/main.c)
#define NB_MBUF 16384
#define RTE_TEST_RX_DESC_DEFAULT 2048
#define RTE_TEST_TX_DESC_DEFAULT 2048
Build testpmd: (in test-pmd/testpmd.c)
#define RTE_TEST_RX_DESC_DEFAULT 2048
#define RTE_TEST_TX_DESC_DEFAULT 2048
System Settings System Capability SettingsLinux OS Services Settings
# systemctl disable NetworkManager.service
# chkconfig network on
# systemctl restart network.service
# systemctl stop NetworkManager.service
# systemctl stop firewalld.service
# systemctl disable firewalld.service
# systemctl stop irqbalance.service
# killall irqbalance
# systemctl disable irqbalance.service
# service iptables stop
# echo 0 > /proc/sys/kernel/randomize_va_space
# SELinux disabled
# net.ipv4.ip_forward=0Uncore Frequency Settings
Set the uncore frequency to the max ratio.
PCI Settings # setpci –s 00:03.0 184.l
0000000
# setpci –s 00:03.2 184.l
0000000
# setpci –s 00:03.0 184.l=0x1408
# setpci –s 00:03.2 184.l=0x1408Linux Module Settings # rmmod ipmi_msghandler
# rmmod ipmi_si
# rmmod ipmi_devintf
18
DPDK Acceleration Enhancements
DPDK – ArchitectureFocus to Date DPDK-AE (Acceleration Enhancements)
Data Plane Development Kit (DPDK) APICryptoDevice
DPIDevice
Classification
DeviceFutureDevice
AES-NI
QAT
Hyperscan RRC
SoCs*
SoCPMD
externalmemorymanager
Network StacksStorage and file systems
PktgenTraffic generator
Example Applications
Light weightthreads
Futurefeatures Event system
EAL
MALLOC
MBUF
MEMPOOL
RING
TIMER
Core Libraries
KNI
POWERIVSHME
M
Platform
LPMClassificatio
n
ACL
Classify
e1000
ixgbe
bonding
af_pkt
i40e
fm10k
Packet Access (PMD)
ETHDEV
xenvirt
enic
ring
METER
SCHED
QoS
cxgbe
vmxnet3 virtio 3rd Party
3rd Party
PIPELINE
mlx4 memnic
others
HASH
Utilities
IP Frag
CMDLINE JOBSTAT
KVARGSREORDER
TABLE 3rd Party
NBT
Simple SOC model
18
Key Intel Development Areas
Cost Reduction & Efficiency
• Application metadata catalog for intelligent scheduling• Storage Policies & erasure codes• VxLAN support for vSwitch• Capacity and bandwidth monitoring
High Availability
• Probable Root Cause Analysis, continuous analytics• Platform status monitoring• Live migration, Host evacuation
Trust & Compliance
• Trusted Compute Pools, including bare metal• Boundary Control or Geotagging• Role-based access control• Enabling Firewall as a Service
Key Focus Areas User Needs
Technology Alignment
Intel® VT, ASA
Intel® TXT
AES-NI, AVX, DPDK, Intel® QuickAssist
Performance• Accelerated packet processing (Open vSwitch with DPDK)• Intelligent scheduling through enhanced platform awareness
(CPU features, PCI Express* Accelerators, SR-IOV etc.)
Node Manager, Cache/Memory
QoS
Deployability &
Stability
• Improved installation & upgradability• Disaster recovery capabilities• User experience and scalability
Intel® RSA, Intel® AMT, Intel® vPro
20
The Need for Enhanced Platform AwarenessPort
Degradation: UP to 50%
System Degradation
up to 2.5 time
Source: Telefonica and Intel testing
21
Example: Intel contributions to OpenStack Kilo
HaswellSocket 1
VIM
Haswell Socket 0
CORE CORE CORE CORE
CORE CORE CORE COREApplication
Process
Application Process
Application Process
Application Process
Mem
ory
Mem
ory
GrantleyNIC
Enhance Platform Awareness(EPA) leading to improves SLA• Non-uniform memory access (NUMA) topology filter for memory proximity • NUMA I/O Awareness
NICNIC
* Other names and brands may be claimed as the property of others.
MANO EPA Features
22
Non-Uniform Memory Architecture (NUMA) CPU & Memory configuration (co-located memory and socket)NUMA I/O Device Locality configuration (co-located PCI device and socket)CPU PinningHuge Page Support (2MB/1GB)QATTXT, Trusted Compute PoolsAES-NI, AVX, SSE4.2, RD RAND (Instruction Set Extensions)CPU Model (explicit model match)CPU LLC(cache size)vSwitches (type, capability) - OVS specified, with or without DPDKLLC utilizationCPU ddio (direct i/o)CAT (cache allocation)
23
OpenDaylight Lithium Release
24
Service Function Chaining Use CaseOpenStack
Neutron Nova
OpenDaylightSFC GBP
Open vSwitch
OVSDB / OF
ServiceFunction
1
ServiceFunction
2
ServiceFunction
3
PolicyTackerGBP?
SFF Classifier
25
ONP 1.5 Deliverables and IngredientsDocument Description
Intel® ONP Server 1.5 Release Notes System description and solution ingredients list, new features description, system limitation, and installation instructions
Intel® ONP Server 1.5 Reference Architecture Guide
Scripts for the integration of the Intel® ONP Server Release 1.5 Reference Architecture. Content includes high-level architecture, setup and configuration procedures, and integration learnings.
Intel® ONP Server 1.5 Benchmark Performance Test Report
Performance characterization and baseline performance data based on ONP release 1.5 software.
Intel® ONP Server 1.5 Application Note Integration activities were done on the configuration in the ONP Server Release 1.5 Reference Architecture Guide. Benchmarking activities were done on the configuration in the ONP Server Release 1.5 Benchmark Performance Test Report. The Application Note contains information on the differences between these two configurations.
Reference Architecture
Intel® Xeon E5-2600 V3Intel® Ethernet Controller XL710
Fedora
DPDK
Intel ONP Server 1.5
Kilo 2015.1.1
Lithium SR1
v2.4.90
v2.0
2.3.0.5
Fedora v21
Industry SHVS
Integrated Software
26
ONP Key Take-aways
• Fully open source, aligned with major upstream projects• Quarterly cadence, allows for rapid turnaround on complex use
cases• Optimized for latest Intel Platforms• Benchmarks showing platform and ingredient improvements• Ingredients aligned with a subset of OPNFV projects• Available now at:
• https://01.org/packet-processing/intel%C2%AE-onp-servers
Thank You
27
Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. Configurations: [describe config + what test used + who did testing]. For more information go to http://www.intel.com/performance
No computer system can provide absolute security. Requires an enabled Intel® processor, enabled chipset, firmware, software and may require a subscription with a capable service provider (may not be available in all countries). Intel assumes no liability for lost or stolen data and/or systems or any other damages resulting thereof. Consult your system or service provider for availability and functionality.
© 2015 Intel Corporation. All rights reserved. Intel and the Intel logo are trademarks of Intel Corporation in the U.S. and/or other countries. *Other names and brands may be claimed as the property of others.