diff options
author | Maciek Konstantynowicz <mkonstan@cisco.com> | 2019-01-28 21:52:02 +0000 |
---|---|---|
committer | Maciek Konstantynowicz <mkonstan@cisco.com> | 2019-03-20 13:55:04 +0000 |
commit | 8231ea666c43c0d514708b41cf52667f5e8d0311 (patch) | |
tree | 77fdb174c8773b280cc07e972533c5d3c0c7a8d0 /docs | |
parent | 6f082130e508e151bb8e92ce459a10b05b3c82b8 (diff) |
ietf draft: nfv service density rev. -00
Change-Id: I0490db919b4da198545f4a332fd722855bac84a8
Signed-off-by: Maciek Konstantynowicz <mkonstan@cisco.com>
Diffstat (limited to 'docs')
-rw-r--r-- | docs/ietf/draft-mkonstan-nf-service-density-00.md | 874 |
1 files changed, 874 insertions, 0 deletions
diff --git a/docs/ietf/draft-mkonstan-nf-service-density-00.md b/docs/ietf/draft-mkonstan-nf-service-density-00.md new file mode 100644 index 0000000000..a3216d06dd --- /dev/null +++ b/docs/ietf/draft-mkonstan-nf-service-density-00.md @@ -0,0 +1,874 @@ +--- +title: NFV Service Density Benchmarking +# abbrev: nf-svc-density +docname: draft-mkonstan-nf-service-density-00 +date: 2019-03-11 + +ipr: trust200902 +area: ops +wg: Benchmarking Working Group +kw: Internet-Draft +cat: info + +coding: us-ascii +pi: # can use array (if all yes) or hash here +# - toc +# - sortrefs +# - symrefs + toc: yes + sortrefs: # defaults to yes + symrefs: yes + +author: + - + ins: M. Konstantynowicz + name: Maciek Konstantynowicz + org: Cisco Systems + role: editor + email: mkonstan@cisco.com + - + ins: P. Mikus + name: Peter Mikus + org: Cisco Systems + role: editor + email: pmikus@cisco.com + +normative: + RFC2544: + RFC8174: + +informative: + RFC8204: + TST009: + target: https://www.etsi.org/deliver/etsi_gs/NFV-TST/001_099/009/03.01.01_60/gs_NFV-TST009v030101p.pdf + title: "ETSI GS NFV-TST 009 V3.1.1 (2018-10), Network Functions Virtualisation (NFV) Release 3; Testing; Specification of Networking Benchmarks and Measurement Methods for NFVI" + date: 2018-10 + BSDP: + target: https://fd.io/wp-content/uploads/sites/34/2019/03/benchmarking_sw_data_planes_skx_bdx_mar07_2019.pdf + title: "Benchmarking Software Data Planes Intel® Xeon® Skylake vs. Broadwell" + date: 2019-03 + draft-vpolak-mkonstan-bmwg-mlrsearch: + target: https://tools.ietf.org/html/draft-vpolak-mkonstan-bmwg-mlrsearch-00 + title: "Multiple Loss Ratio Search for Packet Throughput (MLRsearch)" + date: 2018-11 + draft-vpolak-bmwg-plrsearch: + target: https://tools.ietf.org/html/draft-vpolak-bmwg-plrsearch-00 + title: "Probabilistic Loss Ratio Search for Packet Throughput (PLRsearch)" + date: 2018-11 + LFN-FDio-CSIT: + target: https://wiki.fd.io/view/CSIT + title: "Fast Data io, Continuous System Integration and Testing Project" + date: 2019-03 + CNCF-CNF-Testbed: + target: https://github.com/cncf/cnf-testbed/ + title: "Cloud native Network Function (CNF) Testbed" + date: 2019-03 + TRex: + target: https://github.com/cisco-system-traffic-generator/trex-core + title: "TRex Low-Cost, High-Speed Stateful Traffic Generator" + date: 2019-03 + CSIT-1901-testbed-2n-skx: + target: https://docs.fd.io/csit/rls1901/report/introduction/physical_testbeds.html#node-xeon-skylake-2n-skx + title: "FD.io CSIT Test Bed" + date: 2019-03 + CSIT-1901-test-enviroment: + target: https://docs.fd.io/csit/rls1901/report/vpp_performance_tests/test_environment.html + title: "FD.io CSIT Test Environment" + date: 2019-03 + CSIT-1901-nfv-density-methodology: + target: https://docs.fd.io/csit/rls1901/report/introduction/methodology_nfv_service_density.html + title: "FD.io CSIT Test Methodology: NFV Service Density" + date: 2019-03 + CSIT-1901-nfv-density-results: + target: https://docs.fd.io/csit/rls1901/report/vpp_performance_tests/nf_service_density/index.html + title: "FD.io CSIT Test Results: NFV Service Density" + date: 2019-03 + CNCF-CNF-Testbed-Results: + target: https://github.com/cncf/cnf-testbed/blob/master/comparison/doc/cncf-cnfs-results-summary.md + title: "CNCF CNF Testbed: NFV Service Density Benchmarking" + date: 2018-12 + +--- abstract + +Network Function Virtualization (NFV) system designers and operators +continuously grapple with the problem of qualifying performance of +network services realised with software Network Functions (NF) running +on Commercial-Off-The-Shelf (COTS) servers. One of the main challenges +is getting repeatable and portable benchmarking results and using them +to derive deterministic operating range that is production deployment +worthy. + +This document specifies benchmarking methodology for NFV services that +aims to address this problem space. It defines a way for measuring +performance of multiple NFV service instances, each composed of multiple +software NFs, and running them at a varied service “packing” density on +a single server. + +The aim is to discover deterministic usage range of NFV system. In +addition specified methodology can be used to compare and contrast +different NFV virtualization technologies. + +--- middle + +# Terminology + +* NFV - Network Function Virtualization, a general industry term + describing network functionality implemented in software. +* NFV service - a software based network service realized by a topology + of interconnected constituent software network function applications. +* NFV service instance - a single instantiation of NFV service. +* Data-plane optimized software - any software with dedicated threads + handling data-plane packet processing e.g. FD.io VPP (Vector Packet + Processor), OVS-DPDK. + +# Motivation + +## Problem Description + +Network Function Virtualization (NFV) system designers and operators +continuously grapple with the problem of qualifying performance of +network services realised with software Network Functions (NF) running +on Commercial-Off-The-Shelf (COTS) servers. One of the main challenges +is getting repeatable and portable benchmarking results and using them +to derive deterministic operating range that is production deployment +worthy. + +Lack of well defined and standardised NFV centric performance +methodology and metrics makes it hard to address fundamental questions +that underpin NFV production deployments: + +1. What NFV service and how many instances can run on a single compute + node? +2. How to choose the best compute resource allocation scheme to maximise + service yield per node? +3. How do different NF applications compare from the service density + perspective? +4. How do the virtualisation technologies compare e.g. Virtual Machines, + Containers? + +Getting answers to these points should allow designers to make a data +based decision about the NFV technology and service design best suited +to meet requirements of their use cases. Equally, obtaining the +benchmarking data underpinning those answers should make it easier for +operators to work out expected deterministic operating range of chosen +design. + +## Proposed Solution + +The primary goal of the proposed benchmarking methodology is to focus on +NFV technologies used to construct NFV services. More specifically to i) +measure packet data-plane performance of multiple NFV service instances +while running them at varied service “packing” densities on a single +server and ii) quantify the impact of using multiple NFs to construct +each NFV service instance and introducing multiple packet processing +hops and links on each packet path. + +The overarching aim is to discover a set of deterministic usage ranges +that are of interest to NFV system designers and operators. In addition, +specified methodology can be used to compare and contrast different NFV +virtualisation technologies. + +In order to ensure wide applicability of the benchmarking methodology, +the approach is to separate NFV service packet processing from the +shared virtualisation infrastructure by decomposing the software +technology stack into three building blocks: + + +-------------------------------+ + | NFV Service | + +-------------------------------+ + | Virtualization Technology | + +-------------------------------+ + | Host Networking | + +-------------------------------+ + + Figure 1. NFV software technology stack. + +Proposed methodology is complementary to existing NFV benchmarking +industry efforts focusing on vSwitch benchmarking [RFC8204], [TST009] +and extends the benchmarking scope to NFV services. + +This document does not describe a complete benchmarking methodology, +instead it is focusing on system under test configuration part. Each of +the compute node configurations identified by (RowIndex, ColumnIndex) is +to be evaluated for NFV service data-plane performance using existing +and/or emerging network benchmarking standards. This may include +methodologies specified in [RFC2544], [TST009], +[draft-vpolak-mkonstan-bmwg-mlrsearch] and/or +[draft-vpolak-bmwg-plrsearch]. + +# NFV Service + +It is assumed that each NFV service instance is built of one or more +constituent NFs and is described by: topology, configuration and +resulting packet path(s). + +Each set of NFs forms an independent NFV service instance, with multiple +sets present in the host. + +## Topology + +NFV topology describes the number of network functions per service +instance, and their inter-connections over packet interfaces. It +includes all point-to-point virtual packet links within the compute +node, Layer-2 Ethernet or Layer-3 IP, including the ones to host +networking data-plane. + +Theoretically, a large set of possible NFV topologies can be realised +using software virtualisation topologies, e.g. ring, partial -/full- +mesh, star, line, tree, ladder. In practice however, only a few +topologies are in the actual use as NFV services mostly perform either +bumps-in-a-wire packet operations (e.g. security filtering/inspection, +monitoring/telemetry) and/or inter-site forwarding decisions (e.g. +routing, switching). + +Two main NFV topologies have been identified so far for NFV service +density benchmarking: + +1. Chain topology: a set of NFs connect to host data-plane with minimum + of two virtual interfaces each, enabling host data-plane to + facilitate NF to NF service chain forwarding and provide connectivity + with external network. + +2. Pipeline topology: a set of NFs connect to each other in a line + fashion with edge NFs homed to host data-plane. Host data-plane + provides connectivity with external network. + +Both topologies are shown in figures below. + +NF chain topology: + + +-----------------------------------------------------------+ + | Host Compute Node | + | | + | +--------+ +--------+ +--------+ | + | | S1NF1 | | S1NF2 | | S1NFn | | + | | | | | .... | | Service1 | + | | | | | | | | + | +-+----+-+ +-+----+-+ + + +-+----+-+ | + | | | | | | | | | Virtual | + | | |<-CS->| |<-CS->| |<-CS->| | Interfaces | + | +-+----+------+----+------+----+------+----+-+ | + | | | CS: Chain | + | | | Segment | + | | Host Data-Plane | | + | +-+--+----------------------------------+--+-+ | + | | | | | | + +-----------------------------------------------------------+ + | | | | Physical + | | | | Interfaces + +---+--+----------------------------------+--+--------------+ + | | + | Traffic Generator | + | | + +-----------------------------------------------------------+ + + Figure 2. NF chain topology forming a service instance. + +NF pipeline topology: + + +-----------------------------------------------------------+ + | Host Compute Node | + | | + | +--------+ +--------+ +--------+ | + | | S1NF1 | | S1NF2 | | S1NFn | | + | | +--+ +--+ .... +--+ | Service1 | + | | | | | | | | + | +-+------+ +--------+ +------+-+ | + | | | Virtual | + | |<-Pipeline Edge Pipeline Edge->| Interfaces | + | +-+----------------------------------------+-+ | + | | | | + | | | | + | | Host Data-Plane | | + | +-+--+----------------------------------+--+-+ | + | | | | | | + +-----------------------------------------------------------+ + | | | | Physical + | | | | Interfaces + +---+--+----------------------------------+--+--------------+ + | | + | Traffic Generator | + | | + +-----------------------------------------------------------+ + + Figure 3. NF pipeline topology forming a service instance. + + +## Configuration + +NFV configuration includes all packet processing functions in NFs +including Layer-2, Layer-3 and/or Layer-4-to-7 processing as appropriate +to specific NF and NFV service design. L2 sub- interface encapsulations +(e.g. 802.1q, 802.1ad) and IP overlay encapsulation (e.g. VXLAN, IPSec, +GRE) may be represented here too as appropriate, although in most cases +they are used as external encapsulation and handled by host networking +data-plane. + +NFV configuration determines logical network connectivity that is +Layer-2 and/or IPv4/IPv6 switching/routing modes, as well as NFV service +specific aspects. In the context of NFV density benchmarking methodology +the initial focus is on the former. + +Building on the two identified NFV topologies, two common NFV +configurations are considered: + +1. Chain configuration: + * Relies on chain topology to form NFV service chains. + * NF packet forwarding designs: + * IPv4/IPv6 routing. + * Requirements for host data-plane: + * L2 switching with L2 forwarding context per each NF chain + segment, or + * IPv4/IPv6 routing with IP forwarding context per each NF chain + segment or per NF chain. + +2. Pipeline configuration: + * Relies on pipeline topology to form NFV service pipelines. + * Packet forwarding designs: + * IPv4/IPv6 routing. + * Requirements for host data-plane: + * L2 switching with L2 forwarding context per each NF pipeline + edge link, or + * IPv4/IPv6 routing with IP forwarding context per each NF pipeline + edge link or per NF pipeline. + +## Packet Path(s) + +NFV packet path(s) describe the actual packet forwarding path(s) used +for benchmarking, resulting from NFV topology and configuration. They +are aimed to resemble true packet forwarding actions during the NFV +service lifecycle. + +Based on the specified NFV topologies and configurations two NFV packet +paths are taken for benchmarking: + +1. Snake packet path + * Requires chain topology and configuration. + * Packets enter the NFV chain through one edge NF and progress to the + other edge NF of the chain. + * Within the chain, packets follow a zigzagging "snake" path entering + and leaving host data-plane as they progress through the NF chain. + * Host data-plane is involved in packet forwarding operations between + NIC interfaces and edge NFs, as well as between NFs in the chain. + +2. Pipeline packet path + * Requires pipeline topology and configuration. + * Packets enter the NFV chain through one edge NF and progress to the + other edge NF of the pipeline. + * Within the chain, packets follow a straight path entering and + leaving subsequent NFs as they progress through the NF pipeline. + * Host data-plane is involved in packet forwarding operations between + NIC interfaces and edge NFs only. + +Both packet paths are shown in figures below. + +Snake packet path: + + +-----------------------------------------------------------+ + | Host Compute Node | + | | + | +--------+ +--------+ +--------+ | + | | S1NF1 | | S1NF2 | | S1NFn | | + | | | | | .... | | Service1 | + | | XXXX | | XXXX | | XXXX | | + | +-+X--X+-+ +-+X--X+-+ +X X+ +-+X--X+-+ | + | |X X| |X X| |X X| |X X| Virtual | + | |X X| |X X| |X X| |X X| Interfaces | + | +-+X--X+------+X--X+------+X--X+------+X--X+-+ | + | | X XXXXXXXXXX XXXXXXXXXX XXXXXXXXXX X | | + | | X X | | + | | X Host Data-Plane X | | + | +-+X-+----------------------------------+-X+-+ | + | |X | | X| | + +----X--------------------------------------X---------------+ + |X | | X| Physical + |X | | X| Interfaces + +---+X-+----------------------------------+-X+--------------+ + | | + | Traffic Generator | + | | + +-----------------------------------------------------------+ + + Figure 4. Snake packet path thru NF chain topology. + + +Pipeline packet path: + + +-----------------------------------------------------------+ + | Host Compute Node | + | | + | +--------+ +--------+ +--------+ | + | | S1NF1 | | S1NF2 | | S1NFn | | + | | +--+ +--+ .... +--+ | Service1 | + | | XXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXX | | + | +--X-----+ +--------+ +-----X--+ | + | |X X| Virtual | + | |X X| Interfaces | + | +-+X--------------------------------------X+-+ | + | | X X | | + | | X X | | + | | X Host Data-Plane X | | + | +-+X-+----------------------------------+-X+-+ | + | |X | | X| | + +----X--------------------------------------X---------------+ + |X | | X| Physical + |X | | X| Interfaces + +---+X-+----------------------------------+-X+--------------+ + | | + | Traffic Generator | + | | + +-----------------------------------------------------------+ + + Figure 5. Pipeline packet path thru NF pipeline topology. + +In all cases packets enter NFV system via shared physical NIC interfaces +controlled by shared host data-plane, are then associated with specific +NFV service (based on service discriminator) and subsequently are cross- +connected/switched/routed by host data-plane to and through NF +topologies per one of above listed schemes. + +# Virtualization Technology + +NFV services are built of composite isolated NFs, with virtualisation +technology providing the workload isolation. Following virtualisation +technology types are considered for NFV service density benchmarking: + +1. Virtual Machines (VMs) + * Relying on host hypervisor technology e.g. KVM, ESXi, Xen. + * NFs running in VMs are referred to as VNFs. +2. Containers + * Relying on Linux container technology e.g. LXC, Docker. + * NFs running in Containers are referred to as CNFs. + +Different virtual interface types are available to VNFs and CNFs: + +1. VNF + * virtio-vhostuser: fully user-mode based virtual interface. + * virtio-vhostnet: involves kernel-mode based backend. +2. CNF + * memif: fully user-mode based virtual interface. + * af_packet: involves kernel-mode based backend. + * (add more common ones) + +# Host Networking + +Host networking data-plane is the central shared resource that underpins +creation of NFV services. It handles all of the connectivity to external +physical network devices through physical network connections using +NICs, through which the benchmarking is done. + +Assuming that NIC interface resources are shared, here is the list of +widely available host data-plane options for providing packet +connectivity to/from NICs and constructing NFV chain and pipeline +topologies and configurations: + +* Linux Kernel-Mode Networking. +* Linux User-Mode vSwitch. +* Virtual Machine vSwitch. +* Linux Container vSwitch. +* SRIOV NIC Virtual Function - note: restricted support for chain and + pipeline topologies, as it requires hair-pinning through the NIC and + oftentimes also through external physical switch. + +Analysing properties of each of these options and their Pros/Cons for +specified NFV topologies and configurations is outside the scope of this +document. + +From all listed options, performance optimised Linux user-mode vswitch +deserves special attention. Linux user-mode switch decouples NFV service +from the underlying NIC hardware, offers rich multi-tenant functionality +and most flexibility for supporting NFV services. But in the same time +it is consuming compute resources and is harder to benchmark in NFV +service density scenarios. + +Following sections focus on using Linux user-mode vSwitch, focusing on +its performance benchmarking at increasing levels of NFV service +density. + +# NFV Service Density Matrix + +In order to evaluate performance of multiple NFV services running on a +compute node, NFV service instances are benchmarked at increasing +density, allowing to construct an NFV Service Density Matrix. Table +below shows an example of such a matrix, capturing number of NFV service +instances (row indices), number of NFs per service instance (column +indices) and resulting total number of NFs (values). + + NFV Service Density - NF Count View + + SVC 001 002 004 006 008 00N + 001 1 2 4 6 8 1*N + 002 2 4 8 12 16 2*N + 004 4 8 16 24 32 4*N + 006 6 12 24 36 48 6*N + 008 8 16 32 48 64 8*N + 00M M*1 M*2 M*4 M*6 M*8 M*N + + RowIndex: Number of NFV Service Instances, 1..M. + ColumnIndex: Number of NFs per NFV Service Instance, 1..N. + Value: Total number of NFs running in the system. + +In order to deliver good and repeatable network data-plane performance, +NFs and host data-plane software require direct access to critical +compute resources. Due to a shared nature of all resources on a compute +node, a clearly defined resource allocation scheme is defined in the +next section to address this. + +In each tested configuration host data-plane is a gateway between the +external network and the internal NFV network topologies. Offered packet +load is generated and received by an external traffic generator per +usual benchmarking practice. + +It is proposed that initial benchmarks are done with the offered packet +load distributed equally across all configured NFV service instances. +This could be followed by various per NFV service instance load ratios +mimicking expected production deployment scenario(s). + +Following sections specify compute resource allocation, followed by +examples of applying NFV service density methodology to VNF and CNF +benchmarking use cases. + +# Compute Resource Allocation + +Performance optimized NF and host data-plane software threads require +timely execution of packet processing instructions and are very +sensitive to any interruptions (or stalls) to this execution e.g. cpu +core context switching, or cpu jitter. To that end, NFV service density +methodology treats controlled mapping ratios of data plane software +threads to physical processor cores with directly allocated cache +hierarchies as the first order requirement. + +Other compute resources including memory bandwidth and PCIe bandwidth +have lesser impact and as such are subject for further study. For more +detail and deep-dive analysis of software data plane performance and +impact on different shared compute resources is available in [BSDP]. + +It is assumed that NFs as well as host data-plane (e.g. vswitch) are +performance optimized, with their tasks executed in two types of +software threads: + +* data-plane - handling data-plane packet processing and forwarding, + time critical, requires dedicated cores. To scale data-plane + performance, most NF apps use multiple data-plane threads and rely on + NIC RSS (Receive Side Scaling), virtual interface multi-queue and/or + integrated software hashing to distribute packets across the data + threads. + +* main-control - handling application management, statistics and + control-planes, less time critical, allows for core sharing. For most + NF apps this is a single main thread, but often statistics (counters) + and various control protocol software are run in separate threads. + +Core mapping scheme described below allocates cores for all threads of +specified type belonging to each NF app instance, and separately lists +number of threads to a number of logical/physical core mappings for +processor configurations with enabled/disabled Symmetric Multi- +Threading (SMT) (e.g. AMD SMT, Intel Hyper-Threading). + +If NFV service density benchmarking is run on server nodes with +Symmetric Multi-Threading (SMT) (e.g. AMD SMT, Intel Hyper-Threading) +for higher performance and efficiency, logical cores allocated to data- +plane threads should be allocated as pairs of sibling logical cores +corresponding to the hyper-threads running on the same physical core. + +Separate core ratios are defined for mapping threads of vSwitch and NFs. +In order to get consistent benchmarking results, the mapping ratios are +enforced using Linux core pinning. + +| application | thread type | app:core ratio | threads/pcores (SMT disabled) | threads/lcores map (SMT enabled) | +|:-----------:|:-----------:|:--------------:|:-------------------------------:|:----------------------------------:| +| vSwitch-1c | data | 1:1 | 1DT/1PC | 2DT/2LC | +| | main | 1:S2 | 1MT/S2PC | 1MT/1LC | +| | | | | | +| vSwitch-2c | data | 1:2 | 2DT/2PC | 4DT/4LC | +| | main | 1:S2 | 1MT/S2PC | 1MT/1LC | +| | | | | | +| vSwitch-4c | data | 1:4 | 4DT/4PC | 8DT/8LC | +| | main | 1:S2 | 1MT/S2PC | 1MT/1LC | +| | | | | | +| NF-0.5c | data | 1:S2 | 1DT/S2PC | 1DT/1LC | +| | main | 1:S2 | 1MT/S2PC | 1MT/1LC | +| | | | | | +| NF-1c | data | 1:1 | 1DT/1PC | 2DT/2LC | +| | main | 1:S2 | 1MT/S2PC | 1MT/1LC | +| | | | | | +| NF-2c | data | 1:2 | 2DT/2PC | 4DT/4LC | +| | main | 1:S2 | 1MT/S2PC | 1MT/1LC | + +* Legend to table + * Header row + * application - network application with optimized data-plane, a + vSwitch or Network Function (NF) application. + * thread type - either "data", short for data-plane; or "main", + short for all main-control threads. + * app:core ratio - ratio of per application instance threads of + specific thread type to physical cores. + * threads/pcores (SMT disabled) - number of threads of specific + type (DT for data-plane thread, MT for main thread) running on a + number of physical cores, with SMT disabled. + * threads/lcores map (SMT enabled) - number of threads of specific + type (DT, MT) running on a number of logical cores, with SMT + enabled. Two logical cores per one physical core. + * Content rows + * vSwitch-(1c|2c|4c) - vSwitch with 1 physical core (or 2, or 4) + allocated to its data-plane software worker threads. + * NF-(0.5c|1c|2c) - NF application with half of a physical core (or + 1, or 2) allocated to its data-plane software worker threads. + * Sn - shared core, sharing ratio of (n). + * DT - data-plane thread. + * MT - main-control thread. + * PC - physical core, with SMT/HT enabled has many (mostly 2 today) + logical cores associated with it. + * LC - logical core, if more than one lc get allocated in sets of + two sibling logical cores running on the same physical core. + * SnPC - shared physical core, sharing ratio of (n). + * SnLC - shared logical core, sharing ratio of (n). + +Maximum benchmarked NFV service densities are limited by a number of +physical cores on a compute node. + +A sample physical core usage view is shown in the matrix below. + + NFV Service Density - Core Usage View + vSwitch-1c, NF-1c + + SVC 001 002 004 006 008 010 + 001 2 3 6 9 12 15 + 002 3 6 12 18 24 30 + 004 6 12 24 36 48 60 + 006 9 18 36 54 72 90 + 008 12 24 48 72 96 120 + 010 15 30 60 90 120 150 + + RowIndex: Number of NFV Service Instances, 1..10. + ColumnIndex: Number of NFs per NFV Service Instance, 1..10. + Value: Total number of physical processor cores used for NFs. + +# NFV Service Density Benchmarks + +To illustrate defined NFV service density applicability, following +sections describe three sets of NFV service topologies and +configurations that have been benchmarked in open-source: i) in +[LFN-FDio-CSIT], a continuous testing and data-plane benchmarking +project, and ii) as part of CNCF CNF Testbed initiative +[CNCF-CNF-Testbed]. + +In both cases each NFV service instance definition is based on the same +set of NF applications, and varies only by network addressing +configuration to emulate multi-tenant operating environment. + +## Test Methodology - MRR Throughput + +Initial NFV density throughput benchmarks have been performed using +Maximum Receive Rate (MRR) test methodology defined and used in FD.io +CSIT. + +MRR tests measure the packet forwarding rate under the maximum load +offered by traffic generator over a set trial duration, regardless of +packet loss. Maximum load for specified Ethernet frame size is set to +the bi-directional link rate (2x 10GbE in referred results). + +Tests were conducted with two traffic profiles: i) continuous stream of +64B frames, ii) continuous stream of IMIX sequence of (7x 64B, 4x 570B, +1x 1518B), all sizes are L2 untagged Ethernet. + +NFV service topologies tested include: VNF service chains, CNF service +chains and CNF service pipelines. + +## VNF Service Chain + +VNF Service Chain (VSC) topology is tested with KVM hypervisor (Ubuntu +18.04-LTS), with NFV service instances consisting of NFs running in VMs +(VNFs). Host data-plane is provided by FD.io VPP vswitch. Virtual +interfaces are virtio-vhostuser. Snake forwarding packet path is tested +using [TRex] traffic generator, see figure. + + +-----------------------------------------------------------+ + | Host Compute Node | + | | + | +--------+ +--------+ +--------+ | + | | S1VNF1 | | S1VNF2 | | S1VNFn | | + | | | | | .... | | Service1 | + | | XXXX | | XXXX | | XXXX | | + | +-+X--X+-+ +-+X--X+-+ +-+X--X+-+ | + | |X X| |X X| |X X| Virtual | + | |X X| |X X| |X X| |X X| Interfaces | + | +-+X--X+------+X--X+------+X--X+------+X--X+-+ | + | | X XXXXXXXXXX XXXXXXXXXX XXXXXXXXXX X | | + | | X X | | + | | X FD.io VPP vSwitch X | | + | +-+X-+----------------------------------+-X+-+ | + | |X | | X| | + +----X--------------------------------------X---------------+ + |X | | X| Physical + |X | | X| Interfaces + +---+X-+----------------------------------+-X+--------------+ + | | + | Traffic Generator (TRex) | + | | + +-----------------------------------------------------------+ + + Figure 6. VNF service chain test setup. + + +## CNF Service Chain + +CNF Service Chain (CSC) topology is tested with Docker containers +(Ubuntu 18.04-LTS), with NFV service instances consisting of NFs running +in Containers (CNFs). Host data-plane is provided by FD.io VPP vswitch. +Virtual interfaces are memif. Snake forwarding packet path is tested +using [TRex] traffic generator, see figure. + + +-----------------------------------------------------------+ + | Host Compute Node | + | | + | +--------+ +--------+ +--------+ | + | | S1CNF1 | | S1CNF2 | | S1CNFn | | + | | | | | .... | | Service1 | + | | XXXX | | XXXX | | XXXX | | + | +-+X--X+-+ +-+X--X+-+ +-+X--X+-+ | + | |X X| |X X| |X X| Virtual | + | |X X| |X X| |X X| |X X| Interfaces | + | +-+X--X+------+X--X+------+X--X+------+X--X+-+ | + | | X XXXXXXXXXX XXXXXXXXXX XXXXXXXXXX X | | + | | X X | | + | | X FD.io VPP vSwitch X | | + | +-+X-+----------------------------------+-X+-+ | + | |X | | X| | + +----X--------------------------------------X---------------+ + |X | | X| Physical + |X | | X| Interfaces + +---+X-+----------------------------------+-X+--------------+ + | | + | Traffic Generator (TRex) | + | | + +-----------------------------------------------------------+ + + Figure 7. CNF service chain test setup. + +## CNF Service Pipeline + +CNF Service Pipeline (CSP) topology is tested with Docker containers +(Ubuntu 18.04-LTS), with NFV service instances consisting of NFs running +in Containers (CNFs). Host data-plane is provided by FD.io VPP vswitch. +Virtual interfaces are memif. Pipeline forwarding packet path is tested +using [TRex] traffic generator, see figure. + + +-----------------------------------------------------------+ + | Host Compute Node | + | | + | +--------+ +--------+ +--------+ | + | | S1NF1 | | S1NF2 | | S1NFn | | + | | +--+ +--+ .... +--+ | Service1 | + | | XXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXX | | + | +--X-----+ +--------+ +-----X--+ | + | |X X| Virtual | + | |X X| Interfaces | + | +-+X--------------------------------------X+-+ | + | | X X | | + | | X X | | + | | X FD.io VPP vSwitch X | | + | +-+X-+----------------------------------+-X+-+ | + | |X | | X| | + +----X--------------------------------------X---------------+ + |X | | X| Physical + |X | | X| Interfaces + +---+X-+----------------------------------+-X+--------------+ + | | + | Traffic Generator (TRex) | + | | + +-----------------------------------------------------------+ + + Figure 8. CNF service chain test setup. + +## Sample Results: FD.io CSIT + +FD.io CSIT project introduced NFV density benchmarking in release +CSIT-1901 and published results for the following NFV service topologies +and configurations: + +1. VNF Service Chains + * VNF: DPDK-L3FWD v18.10 + * IPv4 forwarding + * NF-1c + * vSwitch: VPP v19.01-release + * L2 MAC switching + * vSwitch-1c, vSwitch-2c + * frame sizes: 64B, IMIX +2. CNF Service Chains + * CNF: VPP v19.01-release + * IPv4 routing + * NF-1c + * vSwitch: VPP v19.01-release + * L2 MAC switching + * vSwitch-1c, vSwitch-2c + * frame sizes: 64B, IMIX +3. CNF Service Pipelines + * CNF: VPP v19.01-release + * IPv4 routing + * NF-1c + * vSwitch: VPP v19.01-release + * L2 MAC switching + * vSwitch-1c, vSwitch-2c + * frame sizes: 64B, IMIX + +More information is available in FD.io CSIT-1901 report, with specific +references listed below: + +* Testbed: [CSIT-1901-testbed-2n-skx] +* Test environment: [CSIT-1901-test-enviroment] +* Methodology: [CSIT-1901-nfv-density-methodology] +* Results: [CSIT-1901-nfv-density-results] + +## Sample Results: CNCF/CNFs + +CNCF CI team introduced a CNF testbed initiative focusing on benchmaring +NFV density with open-source network applications running as VNFs and +CNFs. Following NFV service topologies and configurations have been +tested to date: + +1. VNF Service Chains + * VNF: VPP v18.10-release + * IPv4 routing + * NF-1c + * vSwitch: VPP v18.10-release + * L2 MAC switching + * vSwitch-1c, vSwitch-2c + * frame sizes: 64B, IMIX +2. CNF Service Chains + * CNF: VPP v18.10-release + * IPv4 routing + * NF-1c + * vSwitch: VPP v18.10-release + * L2 MAC switching + * vSwitch-1c, vSwitch-2c + * frame sizes: 64B, IMIX +3. CNF Service Pipelines + * CNF: VPP v18.10-release + * IPv4 routing + * NF-1c + * vSwitch: VPP v18.10-release + * L2 MAC switching + * vSwitch-1c, vSwitch-2c + * frame sizes: 64B, IMIX + +More information is available in CNCF CNF Testbed github, with summary +test results presented in summary markdown file, references listed +below: + +* Results: [CNCF-CNF-Testbed-Results] + +# IANA Considerations + +No requests of IANA + +# Security Considerations + +.. + +# Acknowledgements + +Thanks to Vratko Polak of FD.io CSIT project and Michael Pedersen of the +CNCF Testbed initiative for their contributions and useful suggestions. + +--- back
\ No newline at end of file |