aboutsummaryrefslogtreecommitdiffstats
path: root/docs/ietf
diff options
context:
space:
mode:
authorVratko Polak <vrpolak@cisco.com>2021-07-12 19:52:26 +0200
committerMaciek Konstantynowicz <mkonstan@cisco.com>2021-07-12 18:18:37 +0000
commit52f109b0e14b5f192e2b7f0573e6cebb601d0651 (patch)
tree3e2bfa0d2a8c658d88693d7620786f4305612d55 /docs/ietf
parent8f9290d1594d35eb838c0c4c9d754b80a35e4a3a (diff)
IETF: Update MLRsearch draft
Change-Id: I591b76b72868697242cfbece8f569dc82128ed85 Signed-off-by: Vratko Polak <vrpolak@cisco.com>
Diffstat (limited to 'docs/ietf')
-rw-r--r--docs/ietf/draft-ietf-bmwg-mlrsearch-00.md556
-rw-r--r--docs/ietf/draft-ietf-bmwg-mlrsearch-01.md685
2 files changed, 685 insertions, 556 deletions
diff --git a/docs/ietf/draft-ietf-bmwg-mlrsearch-00.md b/docs/ietf/draft-ietf-bmwg-mlrsearch-00.md
deleted file mode 100644
index 05bc41f0fa..0000000000
--- a/docs/ietf/draft-ietf-bmwg-mlrsearch-00.md
+++ /dev/null
@@ -1,556 +0,0 @@
----
-title: Multiple Loss Ratio Search for Packet Throughput (MLRsearch)
-# abbrev: MLRsearch
-docname: draft-ietf-bmwg-mlrsearch-00
-date: 2021-02-05
-
-ipr: trust200902
-area: ops
-wg: Benchmarking Working Group
-kw: Internet-Draft
-cat: info
-
-coding: us-ascii
-pi: # can use array (if all yes) or hash here
-# - toc
-# - sortrefs
-# - symrefs
- toc: yes
- sortrefs: # defaults to yes
- symrefs: yes
-
-author:
- -
- ins: M. Konstantynowicz
- name: Maciek Konstantynowicz
- org: Cisco Systems
- role: editor
- email: mkonstan@cisco.com
- -
- ins: V. Polak
- name: Vratko Polak
- org: Cisco Systems
- role: editor
- email: vrpolak@cisco.com
-
-normative:
- RFC2544:
- RFC8174:
-
-informative:
- FDio-CSIT-MLRsearch:
- target: https://docs.fd.io/csit/rls2001/report/introduction/methodology_data_plane_throughput/methodology_mlrsearch_tests.html
- title: "FD.io CSIT Test Methodology - MLRsearch"
- date: 2020-02
- PyPI-MLRsearch:
- target: https://pypi.org/project/MLRsearch/0.3.0/
- title: "MLRsearch 0.3.0, Python Package Index"
- date: 2020-02
-
---- abstract
-
-This document proposes changes to [RFC2544], specifically to packet
-throughput search methodology, by defining a new search algorithm
-referred to as Multiple Loss Ratio search (MLRsearch for short). Instead
-of relying on binary search with pre-set starting offered load, it
-proposes a novel approach discovering the starting point in the initial
-phase, and then searching for packet throughput based on defined packet
-loss ratio (PLR) input criteria and defined final trial duration time.
-One of the key design principles behind MLRsearch is minimizing the
-total test duration and searching for multiple packet throughput rates
-(each with a corresponding PLR) concurrently, instead of doing it
-sequentially.
-
-The main motivation behind MLRsearch is the new set of challenges and
-requirements posed by NFV (Network Function Virtualization),
-specifically software based implementations of NFV data planes. Using
-[RFC2544] in the experience of the authors yields often not repetitive
-and not replicable end results due to a large number of factors that are
-out of scope for this draft. MLRsearch aims to address this challenge
-in a simple way of getting the same result sooner, so more repetitions
-can be done to describe the replicability.
-
---- middle
-
-# Terminology
-
-* Frame size: size of an Ethernet Layer-2 frame on the wire, including
- any VLAN tags (dot1q, dot1ad) and Ethernet FCS, but excluding Ethernet
- preamble and inter-frame gap. Measured in bytes.
-* Packet size: same as frame size, both terms used interchangeably.
-* Device Under Test (DUT): In software networking, "device" denotes a
- specific piece of software tasked with packet processing. Such device
- is surrounded with other software components (such as operating system
- kernel). It is not possible to run devices without also running the
- other components, and hardware resources are shared between both. For
- purposes of testing, the whole set of hardware and software components
- is called "system under test" (SUT). As SUT is the part of the whole
- test setup performance of which can be measured by [RFC2544] methods,
- this document uses SUT instead of [RFC2544] DUT. Device under test
- (DUT) can be re-introduced when analysing test results using whitebox
- techniques, but this document sticks to blackbox testing.
-* System Under Test (SUT): System under test (SUT) is a part of the
- whole test setup whose performance is to be benchmarked. The complete
- test setup contains other parts, whose performance is either already
- established, or not affecting the benchmarking result.
-* Bi-directional throughput tests: involve packets/frames flowing in
- both transmit and receive directions over every tested interface of
- SUT/DUT. Packet flow metrics are measured per direction, and can be
- reported as aggregate for both directions and/or separately
- for each measured direction. In most cases bi-directional tests
- use the same (symmetric) load in both directions.
-* Uni-directional throughput tests: involve packets/frames flowing in
- only one direction, i.e. either transmit or receive direction, over
- every tested interface of SUT/DUT. Packet flow metrics are measured
- and are reported for measured direction.
-* Packet Loss Ratio (PLR): ratio of packets received relative to packets
- transmitted over the test trial duration, calculated using formula:
- PLR = ( pkts_transmitted - pkts_received ) / pkts_transmitted.
- For bi-directional throughput tests aggregate PLR is calculated based
- on the aggregate number of packets transmitted and received.
-* Packet Throughput Rate: maximum packet offered load DUT/SUT forwards
- within the specified Packet Loss Ratio (PLR). In many cases the rate
- depends on the frame size processed by DUT/SUT. Hence packet
- throughput rate MUST be quoted with specific frame size as received by
- DUT/SUT during the measurement. For bi-directional tests, packet
- throughput rate should be reported as aggregate for both directions.
- Measured in packets-per-second (pps) or frames-per-second (fps),
- equivalent metrics.
-* Bandwidth Throughput Rate: a secondary metric calculated from packet
- throughput rate using formula: bw_rate = pkt_rate * (frame_size +
- L1_overhead) * 8, where L1_overhead for Ethernet includes preamble (8
- Bytes) and inter-frame gap (12 Bytes). For bi-directional tests,
- bandwidth throughput rate should be reported as aggregate for both
- directions. Expressed in bits-per-second (bps).
-* Non Drop Rate (NDR): maximum packet/bandwith throughput rate sustained
- by DUT/SUT at PLR equal zero (zero packet loss) specific to tested
- frame size(s). MUST be quoted with specific packet size as received by
- DUT/SUT during the measurement. Packet NDR measured in
- packets-per-second (or fps), bandwidth NDR expressed in
- bits-per-second (bps).
-* Partial Drop Rate (PDR): maximum packet/bandwith throughput rate
- sustained by DUT/SUT at PLR greater than zero (non-zero packet loss)
- specific to tested frame size(s). MUST be quoted with specific packet
- size as received by DUT/SUT during the measurement. Packet PDR
- measured in packets-per-second (or fps), bandwidth PDR expressed in
- bits-per-second (bps).
-* Maximum Receive Rate (MRR): packet/bandwidth rate regardless of PLR
- sustained by DUT/SUT under specified Maximum Transmit Rate (MTR)
- packet load offered by traffic generator. MUST be quoted with both
- specific packet size and MTR as received by DUT/SUT during the
- measurement. Packet MRR measured in packets-per-second (or fps),
- bandwidth MRR expressed in bits-per-second (bps).
-* Trial: a single measurement step. See [RFC2544] section 23.
-* Trial duration: amount of time over which packets are transmitted
- in a single measurement step.
-
-# MLRsearch Background
-
-Multiple Loss Ratio search (MLRsearch) is a packet throughput search
-algorithm suitable for deterministic systems (as opposed to
-probabilistic systems). MLRsearch discovers multiple packet throughput
-rates in a single search, with each rate associated with a distinct
-Packet Loss Ratio (PLR) criteria.
-
-For cases when multiple rates need to be found, this property makes
-MLRsearch more efficient in terms of time execution, compared to
-traditional throughput search algorithms that discover a single packet
-rate per defined search criteria (e.g. a binary search specified by
-[RFC2544]). MLRsearch reduces execution time even further by relying on
-shorter trial durations of intermediate steps, with only the final
-measurements conducted at the specified final trial duration. This
-results in the shorter overall search execution time when compared to a
-traditional binary search, while guaranteeing the same results for
-deterministic systems.
-
-In practice two rates with distinct PLRs are commonly used for packet
-throughput measurements of NFV systems: Non Drop Rate (NDR) with PLR=0
-and Partial Drop Rate (PDR) with PLR>0. The rest of this document
-describes MLRsearch for NDR and PDR. If needed, MLRsearch can be
-adapted to discover more throughput rates with different pre-defined
-PLRs.
-
-Similarly to other throughput search approaches like binary search,
-MLRsearch is effective for SUTs/DUTs with PLR curve that is continuously
-flat or increasing with growing offered load. It may not be as
-effective for SUTs/DUTs with abnormal PLR curves.
-
-MLRsearch relies on traffic generator to qualify the received packet
-stream as error-free, and invalidate the results if any disqualifying
-errors are present e.g. out-of-sequence frames.
-
-MLRsearch can be applied to both uni-directional and bi-directional
-throughput tests.
-
-For bi-directional tests, MLRsearch rates and ratios are aggregates of
-both directions, based on the following assumptions:
-
-* Traffic transmitted by traffic generator and received by SUT/DUT
- has the same packet rate in each direction,
- in other words the offered load is symmetric.
-* SUT/DUT packet processing capacity is the same in both directions,
- resulting in the same packet loss under load.
-
-# MLRsearch Overview
-
-The main properties of MLRsearch:
-
-* MLRsearch is a duration aware multi-phase multi-rate search algorithm:
- * Initial Phase determines promising starting interval for the search.
- * Intermediate Phases progress towards defined final search criteria.
- * Final Phase executes measurements according to the final search
- criteria.
- * Final search criteria are defined by following inputs:
- * PLRs associated with NDR and PDR.
- * Final trial duration.
- * Measurement resolution.
-* Initial Phase:
- * Measure MRR over initial trial duration.
- * Measured MRR is used as an input to the first intermediate phase.
-* Multiple Intermediate Phases:
- * Trial duration:
- * Start with initial trial duration in the first intermediate phase.
- * Converge geometrically towards the final trial duration.
- * Track two values for NDR and two for PDR:
- * The values are called lower_bound and upper_bound.
- * Each value comes from a specific trial measurement:
- * Most recent for that transmit rate.
- * As such the value is associated with that measurement's duration
- and loss.
- * A bound can be valid or invalid:
- * Valid lower_bound must conform with PLR search criteria.
- * Valid upper_bound must not conform with PLR search criteria.
- * Example of invalid NDR lower_bound is if it has been measured
- with non-zero loss.
- * Invalid bounds are not real boundaries for the searched value:
- * They are needed to track interval widths.
- * Valid bounds are real boundaries for the searched value.
- * Each non-initial phase ends with all bounds valid.
- * Bound can become invalid if it re-measured at a longer trial
- duration in a sub-sequent phase.
- * Search:
- * Start with a large (lower_bound, upper_bound) interval width, that
- determines measurement resolution.
- * Geometrically converge towards the width goal of the phase.
- * Each phase halves the previous width goal.
- * First measurement of the next phase will be internal search
- which always gives a valid bound and brings the width to the new goal.
- * Only one bound then needs to be re-measured with new duration.
- * Use of internal and external searches:
- * External search:
- * Measures at transmit rates outside the (lower_bound,
- upper_bound) interval.
- * Activated when a bound is invalid, to search for a new valid
- bound by multiplying (for example doubling) the interval width.
- * It is a variant of "exponential search".
- * Internal search:
- * A "binary search" that measures at transmit rates within the
- (lower_bound, upper_bound) valid interval, halving the interval
- width.
-* Final Phase:
- * Executed with the final test trial duration, and the final width
- goal that determines resolution of the overall search.
-* Intermediate Phases together with the Final Phase are called
- Non-Initial Phases.
-
-The main benefits of MLRsearch vs. binary search include:
-
-* In general MLRsearch is likely to execute more trials overall, but
- likely less trials at a set final trial duration.
-* In well behaving cases, e.g. when results do not depend on trial
- duration, it greatly reduces (>50%) the overall duration compared to a
- single PDR (or NDR) binary search over duration, while finding
- multiple drop rates.
-* In all cases MLRsearch yields the same or similar results to binary
- search.
-* Note: both binary search and MLRsearch are susceptible to reporting
- non-repeatable results across multiple runs for very bad behaving
- cases.
-
-Caveats:
-
-* Worst case MLRsearch can take longer than a binary search e.g. in case of
- drastic changes in behaviour for trials at varying durations.
-
-# Sample Implementation
-
-Following is a brief description of a sample MLRsearch implementation,
-which is a simlified version of the existing implementation.
-
-## Input Parameters
-
-1. **maximum_transmit_rate** - Maximum Transmit Rate (MTR) of packets to
- be used by external traffic generator implementing MLRsearch,
- limited by the actual Ethernet link(s) rate, NIC model or traffic
- generator capabilities.
-2. **minimum_transmit_rate** - minimum packet transmit rate to be used for
- measurements. MLRsearch fails if lower transmit rate needs to be
- used to meet search criteria.
-3. **final_trial_duration** - required trial duration for final rate
- measurements.
-4. **initial_trial_duration** - trial duration for initial MLRsearch phase.
-5. **final_relative_width** - required measurement resolution expressed as
- (lower_bound, upper_bound) interval width relative to upper_bound.
-6. **packet_loss_ratio** - maximum acceptable PLR search criterion for
- PDR measurements.
-7. **number_of_intermediate_phases** - number of phases between the initial
- phase and the final phase. Impacts the overall MLRsearch duration.
- Less phases are required for well behaving cases, more phases
- may be needed to reduce the overall search duration for worse behaving cases.
-
-## Initial Phase
-
-1. First trial measures at configured maximum transmit rate (MTR) and
- discovers maximum receive rate (MRR).
- * IN: trial_duration = initial_trial_duration.
- * IN: offered_transmit_rate = maximum_transmit_rate.
- * DO: single trial.
- * OUT: measured loss ratio.
- * OUT: MRR = measured receive rate.
- If loss ratio is zero, MRR is set below MTR so that interval width is equal
- to the width goal of the first intermediate phase.
-2. Second trial measures at MRR and discovers MRR2.
- * IN: trial_duration = initial_trial_duration.
- * IN: offered_transmit_rate = MRR.
- * DO: single trial.
- * OUT: measured loss ratio.
- * OUT: MRR2 = measured receive rate.
- If loss ratio is zero, MRR2 is set above MRR so that interval width is equal
- to the width goal of the first intermediate phase.
- MRR2 could end up being equal to MTR (for example if both measurements so far
- had zero loss), which was already measured, step 3 is skipped in that case.
-3. Third trial measures at MRR2.
- * IN: trial_duration = initial_trial_duration.
- * IN: offered_transmit_rate = MRR2.
- * DO: single trial.
- * OUT: measured loss ratio.
-
-## Non-Initial Phases
-
-1. Main loop:
- 1. IN: trial_duration for the current phase. Set to
- initial_trial_duration for the first intermediate phase; to
- final_trial_duration for the final phase; or to the element of
- interpolating geometric sequence for other intermediate phases.
- For example with two intermediate phases, trial_duration of the
- second intermediate phase is the geometric average of
- initial_trial_duration and final_trial_duration.
- 2. IN: relative_width_goal for the current phase. Set to
- final_relative_width for the final phase; doubled for each
- preceding phase. For example with two intermediate phases, the
- first intermediate phase uses quadruple of final_relative_width
- and the second intermediate phase uses double of
- final_relative_width.
- 3. IN: ndr_interval, pdr_interval from the previous main loop
- iteration or the previous phase. If the previous phase is the
- initial phase, both intervals are formed by a (correctly ordered)
- pair of MRR2 and MRR. Note that the initial phase is likely
- to create intervals with invalid bounds.
- 4. DO: According to the procedure described in point 2., either exit
- the phase (by jumping to 1.7.), or calculate new transmit rate to
- measure with.
- 5. DO: Perform the trial measurement at the new transmit rate and
- trial_duration, compute its loss ratio.
- 6. DO: Update the bounds of both intervals, based on the new
- measurement. The actual update rules are numerous, as NDR external
- search can affect PDR interval and vice versa, but the result
- agrees with rules of both internal and external search. For
- example, any new measurement below an invalid lower_bound becomes
- the new lower_bound, while the old measurement (previously acting
- as the invalid lower_bound) becomes a new and valid upper_bound.
- Go to next iteration (1.3.), taking the updated intervals as new
- input.
- 7. OUT: current ndr_interval and pdr_interval. In the final phase
- this is also considered to be the result of the whole search. For
- other phases, the next phase loop is started with the current
- results as an input.
-2. New transmit rate (or exit) calculation (for point 1.4.):
- 1. If there is an invalid bound then prepare for external search:
- * IF the most recent measurement at NDR lower_bound transmit
- rate had the loss higher than zero, then the new transmit rate
- is NDR lower_bound decreased by two NDR interval widths.
- * Else, IF the most recent measurement at PDR lower_bound
- transmit rate had the loss higher than PLR, then the new
- transmit rate is PDR lower_bound decreased by two PDR interval
- widths.
- * Else, IF the most recent measurement at NDR upper_bound
- transmit rate had no loss, then the new transmit rate is NDR
- upper_bound increased by two NDR interval widths.
- * Else, IF the most recent measurement at PDR upper_bound
- transmit rate had the loss lower or equal to PLR, then the new
- transmit rate is PDR upper_bound increased by two PDR interval
- widths.
- 2. Else, if interval width is higher than the current phase goal:
- * IF NDR interval does not meet the current phase width
- goal, prepare for internal search. The new transmit rate is a
- in the middle of NDR lower_bound and NDR upper_bound.
- * IF PDR interval does not meet the current phase width
- goal, prepare for internal search. The new transmit rate is a
- in the middle of PDR lower_bound and PDR upper_bound.
- 3. Else, if some bound has still only been measured at a lower
- duration, prepare to re-measure at the current duration (and the
- same transmit rate). The order of priorities is:
- * NDR lower_bound,
- * PDR lower_bound,
- * NDR upper_bound,
- * PDR upper_bound.
- 4. Else, do not prepare any new rate, to exit the phase.
- This ensures that at the end of each non-initial phase
- all intervals are valid, narrow enough, and measured
- at current phase trial duration.
-
-# FD.io CSIT Implementation
-
-The only known working implementation of MLRsearch is in
-the open-source code running in Linux Foundation
-FD.io CSIT project [FDio-CSIT-MLRsearch] as part of
-a Continuous Integration / Continuous Development (CI/CD) framework.
-
-MLRsearch is also available as a Python package in [PyPI-MLRsearch].
-
-## Additional details
-
-This document so far has been describing a simplified version of
-MLRsearch algorithm. The full algorithm as implemented in CSIT contains
-additional logic, which makes some of the details (but not general
-ideas) above incorrect. Here is a short description of the additional
-logic as a list of principles, explaining their main differences from
-(or additions to) the simplified description, but without detailing
-their mutual interaction.
-
-1. Logarithmic transmit rate.
- * In order to better fit the relative width goal, the interval
- doubling and halving is done differently.
- * For example, the middle of 2 and 8 is 4, not 5.
-2. Optimistic maximum rate.
- * The increased rate is never higher than the maximum rate.
- * Upper bound at that rate is always considered valid.
-3. Pessimistic minimum rate.
- * The decreased rate is never lower than the minimum rate.
- * If a lower bound at that rate is invalid, a phase stops refining
- the interval further (until it gets re-measured).
-4. Conservative interval updates.
- * Measurements above the current upper bound never update a valid upper
- bound, even if drop ratio is low.
- * Measurements below the current lower bound always update any lower
- bound if drop ratio is high.
-5. Ensure sufficient interval width.
- * Narrow intervals make external search take more time to find a
- valid bound.
- * If the new transmit increased or decreased rate would result in
- width less than the current goal, increase/decrease more.
- * This can happen if the measurement for the other interval
- makes the current interval too narrow.
- * Similarly, take care the measurements in the initial phase create
- wide enough interval.
-6. Timeout for bad cases.
- * The worst case for MLRsearch is when each phase converges to
- intervals way different than the results of the previous phase.
- * Rather than suffer total search time several times larger than pure
- binary search, the implemented tests fail themselves when the
- search takes too long (given by argument *timeout*).
-7. Pessimistic external search.
- * Valid bound becoming invalid on re-measurement with higher duration
- is frequently a sign of SUT behaving in non-deterministic way
- (from blackbox point of view). If the final width interval goal
- is too narrow compared to width of rate region where SUT
- is non-deterministic, it is quite likely that there will be multiple
- invalid bounds before the external search finds a valid one.
- * In this case, external search can be sped up by increasing interval width
- more rapidly. As only powers of two ensure the subsequent internal search
- will not result in needlessly narrow interval, a parameter *doublings*
- is introduced to control the pessimism of external search.
- For example three doublings result in interval width being multiplied
- by eight in each external search iteration.
-
-### FD.io CSIT Input Parameters
-
-1. **maximum_transmit_rate** - Typical values: 2 * 14.88 Mpps for 64B
- 10GE link rate, 2 * 18.75 Mpps for 64B 40GE NIC (specific model).
-2. **minimum_transmit_rate** - Value: 2 * 10 kpps (traffic generator
- limitation).
-3. **final_trial_duration** - Value: 30 seconds.
-4. **initial_trial_duration** - Value: 1 second.
-5. **final_relative_width** - Value: 0.005 (0.5%).
-6. **packet_loss_ratio** - Value: 0.005 (0.5%).
-7. **number_of_intermediate_phases** - Value: 2.
- The value has been chosen based on limited experimentation to date.
- More experimentation needed to arrive to clearer guidelines.
-8. **timeout** - Limit for the overall search duration (for one search).
- If MLRsearch oversteps this limit, it immediatelly declares the test failed,
- to avoid wasting even more time on a misbehaving SUT.
- Value: 600 (seconds).
-9. **doublings** - Number of dublings when computing new interval width
- in external search.
- Value: 2 (interval width is quadroupled).
- Value of 1 is best for well-behaved SUTs, but value of 2 has been found
- to decrease overall search time for worse-behaved SUT configurations,
- contributing more to the overall set of different SUT configurations tested.
-
-## Example MLRsearch Run
-
-The following table shows data from a real test run in CSIT
-(using the default input values as above).
-The first column is the phase, the second is the trial measurement performed
-(aggregate bidirectional offered load in megapackets per second,
-and trial duration in seconds).
-Each of last four columns show one bound as updated after the measurement
-(duration truncated to save space).
-Loss ratio is not shown, but invalid bounds are marked with a plus sign.
-
-| Phase | Trial | NDR lower | NDR upper | PDR lower | PDR upper |
-| ----: | ---------: | --------: | --------: | --------: | --------: |
-| init. | 37.50 1.00 | N/A | 37.50 1. | N/A | 37.50 1. |
-| init. | 10.55 1.00 | +10.55 1. | 37.50 1. | +10.55 1. | 37.50 1. |
-| init. | 9.437 1.00 | +9.437 1. | 10.55 1. | +9.437 1. | 10.55 1. |
-| int 1 | 6.053 1.00 | 6.053 1. | 9.437 1. | 6.053 1. | 9.437 1. |
-| int 1 | 7.558 1.00 | 7.558 1. | 9.437 1. | 7.558 1. | 9.437 1. |
-| int 1 | 8.446 1.00 | 8.446 1. | 9.437 1. | 8.446 1. | 9.437 1. |
-| int 1 | 8.928 1.00 | 8.928 1. | 9.437 1. | 8.928 1. | 9.437 1. |
-| int 1 | 9.179 1.00 | 8.928 1. | 9.179 1. | 9.179 1. | 9.437 1. |
-| int 1 | 9.052 1.00 | 9.052 1. | 9.179 1. | 9.179 1. | 9.437 1. |
-| int 1 | 9.307 1.00 | 9.052 1. | 9.179 1. | 9.179 1. | 9.307 1. |
-| int 2 | 9.115 5.48 | 9.115 5. | 9.179 1. | 9.179 1. | 9.307 1. |
-| int 2 | 9.243 5.48 | 9.115 5. | 9.179 1. | 9.243 5. | 9.307 1. |
-| int 2 | 9.179 5.48 | 9.115 5. | 9.179 5. | 9.243 5. | 9.307 1. |
-| int 2 | 9.307 5.48 | 9.115 5. | 9.179 5. | 9.243 5. | +9.307 5. |
-| int 2 | 9.687 5.48 | 9.115 5. | 9.179 5. | 9.307 5. | 9.687 5. |
-| int 2 | 9.495 5.48 | 9.115 5. | 9.179 5. | 9.307 5. | 9.495 5. |
-| int 2 | 9.401 5.48 | 9.115 5. | 9.179 5. | 9.307 5. | 9.401 5. |
-| final | 9.147 30.0 | 9.115 5. | 9.147 30 | 9.307 5. | 9.401 5. |
-| final | 9.354 30.0 | 9.115 5. | 9.147 30 | 9.307 5. | 9.354 30 |
-| final | 9.115 30.0 | +9.115 30 | 9.147 30 | 9.307 5. | 9.354 30 |
-| final | 8.935 30.0 | 8.935 30 | 9.115 30 | 9.307 5. | 9.354 30 |
-| final | 9.025 30.0 | 9.025 30 | 9.115 30 | 9.307 5. | 9.354 30 |
-| final | 9.070 30.0 | 9.070 30 | 9.115 30 | 9.307 5. | 9.354 30 |
-| final | 9.307 30.0 | 9.070 30 | 9.115 30 | 9.307 30 | 9.354 30 |
-
-# IANA Considerations
-
-No requests of IANA.
-
-# Security Considerations
-
-Benchmarking activities as described in this memo are limited to
-technology characterization of a DUT/SUT using controlled stimuli in a
-laboratory environment, with dedicated address space and the constraints
-specified in the sections above.
-
-The benchmarking network topology will be an independent test setup and
-MUST NOT be connected to devices that may forward the test traffic into
-a production network or misroute traffic to the test management network.
-
-Further, benchmarking is performed on a "black-box" basis, relying
-solely on measurements observable external to the DUT/SUT.
-
-Special capabilities SHOULD NOT exist in the DUT/SUT specifically for
-benchmarking purposes. Any implications for network security arising
-from the DUT/SUT SHOULD be identical in the lab and in production
-networks.
-
-# Acknowledgements
-
-Many thanks to Alec Hothan of OPNFV NFVbench project for thorough
-review and numerous useful comments and suggestions.
-
---- back
diff --git a/docs/ietf/draft-ietf-bmwg-mlrsearch-01.md b/docs/ietf/draft-ietf-bmwg-mlrsearch-01.md
new file mode 100644
index 0000000000..f904722478
--- /dev/null
+++ b/docs/ietf/draft-ietf-bmwg-mlrsearch-01.md
@@ -0,0 +1,685 @@
+---
+title: Multiple Loss Ratio Search for Packet Throughput (MLRsearch)
+# abbrev: MLRsearch
+docname: draft-ietf-bmwg-mlrsearch-01
+date: 2021-07-12
+
+ipr: trust200902
+area: ops
+wg: Benchmarking Working Group
+kw: Internet-Draft
+cat: info
+
+coding: us-ascii
+pi: # can use array (if all yes) or hash here
+# - toc
+# - sortrefs
+# - symrefs
+ toc: yes
+ sortrefs: # defaults to yes
+ symrefs: yes
+
+author:
+ -
+ ins: M. Konstantynowicz
+ name: Maciek Konstantynowicz
+ org: Cisco Systems
+ role: editor
+ email: mkonstan@cisco.com
+ -
+ ins: V. Polak
+ name: Vratko Polak
+ org: Cisco Systems
+ role: editor
+ email: vrpolak@cisco.com
+
+normative:
+ RFC2544:
+
+informative:
+ FDio-CSIT-MLRsearch:
+ target: https://docs.fd.io/csit/rls2101/report/introduction/methodology_data_plane_throughput/methodology_mlrsearch_tests.html
+ title: "FD.io CSIT Test Methodology - MLRsearch"
+ date: 2021-02
+ PyPI-MLRsearch:
+ target: https://pypi.org/project/MLRsearch/0.4.0/
+ title: "MLRsearch 0.4.0, Python Package Index"
+ date: 2021-04
+
+--- abstract
+
+This document proposes changes to [RFC2544], specifically to packet
+throughput search methodology, by defining a new search algorithm
+referred to as Multiple Loss Ratio search (MLRsearch for short). Instead
+of relying on binary search with pre-set starting offered load, it
+proposes a novel approach discovering the starting point in the initial
+phase, and then searching for packet throughput based on defined packet
+loss ratio (PLR) input criteria and defined final trial duration time.
+One of the key design principles behind MLRsearch is minimizing the
+total test duration and searching for multiple packet throughput rates
+(each with a corresponding PLR) concurrently, instead of doing it
+sequentially.
+
+The main motivation behind MLRsearch is the new set of challenges and
+requirements posed by NFV (Network Function Virtualization),
+specifically software based implementations of NFV data planes. Using
+[RFC2544] in the experience of the authors yields often not repetitive
+and not replicable end results due to a large number of factors that are
+out of scope for this draft. MLRsearch aims to address this challenge
+in a simple way of getting the same result sooner, so more repetitions
+can be done to describe the replicability.
+
+--- middle
+
+# Terminology
+
+* Frame size: size of an Ethernet Layer-2 frame on the wire, including
+ any VLAN tags (dot1q, dot1ad) and Ethernet FCS, but excluding Ethernet
+ preamble and inter-frame gap. Measured in bytes (octets).
+* Packet size: same as frame size, both terms used interchangeably.
+* Device Under Test (DUT): In software networking, "device" denotes a
+ specific piece of software tasked with packet processing. Such device
+ is surrounded with other software components (such as operating system
+ kernel). It is not possible to run devices without also running the
+ other components, and hardware resources are shared between both. For
+ purposes of testing, the whole set of hardware and software components
+ is called "system under test" (SUT). As SUT is the part of the whole
+ test setup performance of which can be measured by [RFC2544] methods,
+ this document uses SUT instead of [RFC2544] DUT. Device under test
+ (DUT) can be re-introduced when analysing test results using whitebox
+ techniques, but this document sticks to blackbox testing.
+* System Under Test (SUT): System under test (SUT) is a part of the
+ whole test setup whose performance is to be benchmarked. The complete
+ test setup contains other parts, whose performance is either already
+ established, or not affecting the benchmarking result.
+* Bi-directional throughput tests: involve packets/frames flowing in
+ both transmit and receive directions over every tested interface of
+ SUT/DUT. Packet flow metrics are measured per direction, and can be
+ reported as aggregate for both directions and/or separately
+ for each measured direction. In most cases bi-directional tests
+ use the same (symmetric) load in both directions.
+* Uni-directional throughput tests: involve packets/frames flowing in
+ only one direction, i.e. either transmit or receive direction, over
+ every tested interface of SUT/DUT. Packet flow metrics are measured
+ and are reported for measured direction.
+* Packet Loss Ratio (PLR): ratio of packets received relative to packets
+ transmitted over the test trial duration, calculated using formula:
+ PLR = ( pkts_transmitted - pkts_received ) / pkts_transmitted.
+ For bi-directional throughput tests aggregate PLR is calculated based
+ on the aggregate number of packets transmitted and received.
+* Effective loss ratio: A corrected value of measured packet loss ratio
+ chosen to avoid difficulties if SUT exhibits decreasing loss
+ with increasing load. Maximum of packet loss ratios measured at the same
+ duration on all loads smaller than (and including) the current one.
+* Target loss ratio: A packet loss ratio value acting as an imput for search.
+ The search is finding tight enough lower and upper bound in intended load,
+ so that the lower bound has smaller or equal loss ratio, and upper bound
+ has strictly larger loss ratio. For the tighterst upper bound,
+ the effective loss ratio is the same as packet loss ratio.
+ For the tightest lower bound, the effective loss ratio can be higher
+ than the packet loss ratio, but still not larger than the target loss ratio.
+* Packet Throughput Rate: maximum packet offered load DUT/SUT forwards
+ within the specified Packet Loss Ratio (PLR). In many cases the rate
+ depends on the frame size processed by DUT/SUT. Hence packet
+ throughput rate MUST be quoted with specific frame size as received by
+ DUT/SUT during the measurement. For bi-directional tests, packet
+ throughput rate should be reported as aggregate for both directions.
+ Measured in packets-per-second (pps) or frames-per-second (fps),
+ equivalent metrics.
+* Bandwidth Throughput Rate: a secondary metric calculated from packet
+ throughput rate using formula: bw_rate = pkt_rate * (frame_size +
+ L1_overhead) * 8, where L1_overhead for Ethernet includes preamble (8
+ octets) and inter-frame gap (12 octets). For bi-directional tests,
+ bandwidth throughput rate should be reported as aggregate for both
+ directions. Expressed in bits-per-second (bps).
+* Non Drop Rate (NDR): maximum packet/bandwith throughput rate sustained
+ by DUT/SUT at PLR equal zero (zero packet loss) specific to tested
+ frame size(s). MUST be quoted with specific packet size as received by
+ DUT/SUT during the measurement. Packet NDR measured in
+ packets-per-second (or fps), bandwidth NDR expressed in
+ bits-per-second (bps).
+* Partial Drop Rate (PDR): maximum packet/bandwith throughput rate
+ sustained by DUT/SUT at PLR greater than zero (non-zero packet loss)
+ specific to tested frame size(s). MUST be quoted with specific packet
+ size as received by DUT/SUT during the measurement. Packet PDR
+ measured in packets-per-second (or fps), bandwidth PDR expressed in
+ bits-per-second (bps).
+* Maximum Receive Rate (MRR): packet/bandwidth rate regardless of PLR
+ sustained by DUT/SUT under specified Maximum Transmit Rate (MTR)
+ packet load offered by traffic generator. MUST be quoted with both
+ specific packet size and MTR as received by DUT/SUT during the
+ measurement. Packet MRR measured in packets-per-second (or fps),
+ bandwidth MRR expressed in bits-per-second (bps).
+* Trial: a single measurement step. See [RFC2544] section 23.
+* Trial duration: amount of time over which packets are transmitted
+ in a single measurement step.
+
+# MLRsearch Background
+
+Multiple Loss Ratio search (MLRsearch) is a packet throughput search
+algorithm suitable for deterministic systems (as opposed to
+probabilistic systems). MLRsearch discovers multiple packet throughput
+rates in a single search, each rate is associated with a distinct
+Packet Loss Ratio (PLR) criterion.
+
+For cases when multiple rates need to be found, this property makes
+MLRsearch more efficient in terms of time execution, compared to
+traditional throughput search algorithms that discover a single packet
+rate per defined search criteria (e.g. a binary search specified by
+[RFC2544]). MLRsearch reduces execution time even further by relying on
+shorter trial durations of intermediate steps, with only the final
+measurements conducted at the specified final trial duration. This
+results in the shorter overall search execution time when compared to a
+traditional binary search, while guaranteeing the same results for
+deterministic systems.
+
+In practice two rates with distinct PLRs are commonly used for packet
+throughput measurements of NFV systems: Non Drop Rate (NDR) with PLR=0
+and Partial Drop Rate (PDR) with PLR>0. The rest of this document
+describes MLRsearch with NDR and PDR pair as an example.
+
+Similarly to other throughput search approaches like binary search,
+MLRsearch is effective for SUTs/DUTs with PLR curve that is
+non-decreasing with growing offered load. It may not be as
+effective for SUTs/DUTs with abnormal PLR curves, although
+it will always converge to some value.
+
+MLRsearch relies on traffic generator to qualify the received packet
+stream as error-free, and invalidate the results if any disqualifying
+errors are present e.g. out-of-sequence frames.
+
+MLRsearch can be applied to both uni-directional and bi-directional
+throughput tests.
+
+For bi-directional tests, MLRsearch rates and ratios are aggregates of
+both directions, based on the following assumptions:
+
+* Traffic transmitted by traffic generator and received by SUT/DUT
+ has the same packet rate in each direction,
+ in other words the offered load is symmetric.
+* SUT/DUT packet processing capacity is the same in both directions,
+ resulting in the same packet loss under load.
+
+MLRsearch can be applied even without those assumptions,
+but in that case the aggregate loss ratio is less useful as a metric.
+
+MLRsearch can be used for network transactions consisting of more than
+just one packet, or anything else that has intended load as input
+and loss ratio as output (duration as input is optional).
+This text uses mostly packet-centric language.
+
+# MLRsearch Overview
+
+The main properties of MLRsearch:
+
+* MLRsearch is a duration aware multi-phase multi-rate search algorithm:
+ * Initial Phase determines promising starting interval for the search.
+ * Intermediate Phases progress towards defined final search criteria.
+ * Final Phase executes measurements according to the final search
+ criteria.
+ * Final search criteria are defined by following inputs:
+ * Target PLRs (e.g. 0.0 and 0.005 when searching for NDR and PDR).
+ * Final trial duration.
+ * Measurement resolution.
+* Initial Phase:
+ * Measure MRR over initial trial duration.
+ * Measured MRR is used as an input to the first intermediate phase.
+* Multiple Intermediate Phases:
+ * Trial duration:
+ * Start with initial trial duration in the first intermediate phase.
+ * Converge geometrically towards the final trial duration.
+ * Track all previous trial measurement results:
+ * Duration, offered load and loss ratio are tracked.
+ * Effective loss ratios are tracked.
+ * While in practice, real loss ratios can decrease with increasing load,
+ effective loss ratios never decrease. This is achieved by sorting
+ results by load, and using the effective loss ratio of the previous load
+ if the current loss ratio is smaller than that.
+ * The algorithm queries the results to find best lower and upper bounds.
+ * Effective loss ratios are always used.
+ * The phase ends if all target loss ratios have tight enough bounds.
+ * Search:
+ * Iterate over target loss ratios in increasing order.
+ * If both upper and lower bound are in measurement results for this duration,
+ apply bisect until the bounds are tight enough,
+ and continue with next loss ratio.
+ * If a bound is missing for this duration, but there exists a bound
+ from the previous duration (compatible with the other bound
+ at this duration), re-measure at the current duration.
+ * If a bound in one direction (upper or lower) is missing for this duration,
+ and the previous duration does not have a compatible bound,
+ compute the current "interval size" from the second tightest bound
+ in the other direction (lower or upper respectively)
+ for the current duration, and choose next offered load for external search.
+ * The logic guarantees that a measurement is never repeated with both
+ duration and offered load being the same.
+ * The logic guarantees that measurements for higher target loss ratio
+ iterations (still within the same phase duration) do not affect validity
+ and tightness of bounds for previous target loss ratio iterations
+ (at the same duration).
+ * Use of internal and external searches:
+ * External search:
+ * It is a variant of "exponential search".
+ * The "interval size" is multiplied by a configurable constant
+ (powers of two work well with the subsequent internal search).
+ * Internal search:
+ * A variant of binary search that measures at offered load between
+ the previously found bounds.
+ * The interval does not need to be split into exact halves,
+ if other split can get to the target width goal faster.
+ * The idea is to avoid returning interval narrower than the current
+ width goal. See sample implementation details, below.
+* Final Phase:
+ * Executed with the final test trial duration, and the final width
+ goal that determines resolution of the overall search.
+* Intermediate Phases together with the Final Phase are called
+ Non-Initial Phases.
+* The returned bounds stay within prescribed min_rate and max_rate.
+ * When returning min_rate or max_rate, the returned bounds may be invalid.
+ * E.g. upper bound at max_rate may come from a measurement
+ with loss ratio still not higher than the target loss ratio.
+
+The main benefits of MLRsearch vs. binary search include:
+
+* In general MLRsearch is likely to execute more trials overall, but
+ likely less trials at a set final trial duration.
+* In well behaving cases, e.g. when results do not depend on trial
+ duration, it greatly reduces (>50%) the overall duration compared to a
+ single PDR (or NDR) binary search over duration, while finding
+ multiple drop rates.
+* In all cases MLRsearch yields the same or similar results to binary
+ search.
+* Note: both binary search and MLRsearch are susceptible to reporting
+ non-repeatable results across multiple runs for very bad behaving
+ cases.
+
+Caveats:
+
+* Worst case MLRsearch can take longer than a binary search, e.g. in case of
+ drastic changes in behaviour for trials at varying durations.
+ * Re-measurement at higher duration can trigger a long external search.
+ That never happens in binary search, which uses the final duration
+ from the start.
+
+# Sample Implementation
+
+Following is a brief description of a sample MLRsearch implementation,
+which is a simplified version of the existing implementation.
+
+## Input Parameters
+
+1. **max_rate** - Maximum Transmit Rate (MTR) of packets to
+ be used by external traffic generator implementing MLRsearch,
+ limited by the actual Ethernet link(s) rate, NIC model or traffic
+ generator capabilities.
+2. **min_rate** - minimum packet transmit rate to be used for
+ measurements. MLRsearch fails if lower transmit rate needs to be
+ used to meet search criteria.
+3. **final_trial_duration** - required trial duration for final rate
+ measurements.
+4. **initial_trial_duration** - trial duration for initial MLRsearch phase.
+5. **final_relative_width** - required measurement resolution expressed as
+ (lower_bound, upper_bound) interval width relative to upper_bound.
+6. **packet_loss_ratios** - list of maximum acceptable PLR search criteria.
+7. **number_of_intermediate_phases** - number of phases between the initial
+ phase and the final phase. Impacts the overall MLRsearch duration.
+ Less phases are required for well behaving cases, more phases
+ may be needed to reduce the overall search duration for worse behaving cases.
+
+## Initial Phase
+
+1. First trial measures at configured maximum transmit rate (MTR) and
+ discovers maximum receive rate (MRR).
+ * IN: trial_duration = initial_trial_duration.
+ * IN: offered_transmit_rate = maximum_transmit_rate.
+ * DO: single trial.
+ * OUT: measured loss ratio.
+ * OUT: MRR = measured receive rate.
+ Received rate is computed as intended load multiplied by pass ratio
+ (which is one minus loss ratio). This is useful when loss ratio is computed
+ from a different metric than intended load. For example, intended load
+ can be in transactions (multiple packets each), but loss ratio is computed
+ on level of packets, not transactions.
+
+ * Example: If MTR is 10 transactions per second, and each transaction has
+ 10 packets, and receive rate is 90 packets per second, then loss rate
+ is 10%, and MRR is computed to be 9 transactions per second.
+
+ If MRR is too close to MTR, MRR is set below MTR so that interval width
+ is equal to the width goal of the first intermediate phase.
+ If MRR is less than min_rate, min_rate is used.
+2. Second trial measures at MRR and discovers MRR2.
+ * IN: trial_duration = initial_trial_duration.
+ * IN: offered_transmit_rate = MRR.
+ * DO: single trial.
+ * OUT: measured loss ratio.
+ * OUT: MRR2 = measured receive rate.
+ If MRR2 is less than min_rate, min_rate is used.
+ If loss ratio is less or equal to the smallest target loss ratio,
+ MRR2 is set to a value above MRR, so that interval width is equal
+ to the width goal of the first intermediate phase.
+ MRR2 could end up being equal to MTR (for example if both measurements so far
+ had zero loss), which was already measured, step 3 is skipped in that case.
+3. Third trial measures at MRR2.
+ * IN: trial_duration = initial_trial_duration.
+ * IN: offered_transmit_rate = MRR2.
+ * DO: single trial.
+ * OUT: measured loss ratio.
+ * OUT: MRR3 = measured receive rate.
+ If MRR3 is less than min_rate, min_rate is used.
+ If step 3 is not skipped, the first trial measurement is forgotten.
+ This is done because in practice (if MRR2 is above MRR), external search
+ from MRR and MRR2 is likely to lead to a faster intermediate phase
+ than a bisect between MRR2 and MTR.
+
+## Non-Initial Phases
+
+1. Main phase loop:
+ 1. IN: trial_duration for the current phase. Set to
+ initial_trial_duration for the first intermediate phase; to
+ final_trial_duration for the final phase; or to the element of
+ interpolating geometric sequence for other intermediate phases.
+ For example with two intermediate phases, trial_duration of the
+ second intermediate phase is the geometric average of
+ initial_trial_duration and final_trial_duration.
+ 2. IN: relative_width_goal for the current phase. Set to
+ final_relative_width for the final phase; doubled for each
+ preceding phase. For example with two intermediate phases, the
+ first intermediate phase uses quadruple of final_relative_width
+ and the second intermediate phase uses double of
+ final_relative_width.
+ 3. IN: Measurement results from the previous phase (previous duration).
+ 4. Internal target ratio loop:
+ 1. IN: Target loss ratio for this iteration of ratio loop.
+ 2. IN: Measurement results from all previous ratio loop iterations
+ of current phase (current duration).
+ 3. DO: According to the procedure described in point 2:
+ 1. either exit the phase (by jumping to 1.5),
+ 2. or exit loop iteration (by continuing with next target loss ratio,
+ jumping to 1.4.1),
+ 3. or calculate new transmit rate to measure with.
+ 4. DO: Perform the trial measurement at the new transmit rate and
+ current trial duration, compute its loss ratio.
+ 5. DO: Add the result and go to next iteration (1.4.1),
+ including the added trial result in 1.4.2.
+ 5. OUT: Measurement results from this phase.
+ 6. OUT: In the final phase, bounds for each target loss ratio
+ are extracted and returned.
+ 1. If a valid bound does not exist, use min_rate or max_rate.
+2. New transmit rate (or exit) calculation (for point 1.4.3):
+ 1. If the previous duration has the best upper and lower bound,
+ select the middle point as the new transmit rate.
+ 1. See 2.5.3. below for the exact splitting logic.
+ 2. This can be a no-op if interval is narrow enough already,
+ in that case continue with 2.2.
+ 3. Discussion, assuming the middle point is selected and measured:
+ 1. Regardless of loss rate measured, the result becomes
+ either best upper or best lower bound at current duration.
+ 2. So this condition is satisfied at most once per iteration.
+ 3. This also explains why previous phase has double width goal:
+ 1. We avoid one more bisection at previous phase.
+ 2. At most one bound (per iteration) is re-measured
+ with current duration.
+ 3. Each re-measurement can trigger an external search.
+ 4. Such surprising external searches are the main hurdle
+ in achieving low overal search durations.
+ 5. Even without 1.1, there is at most one external search
+ per phase and target loss ratio.
+ 6. But without 1.1 there can be two re-measurements,
+ each coming with a risk of triggering external search.
+ 2. If the previous duration has one bound best, select its transmit rate.
+ In deterministic case this is the last measurement needed this iteration.
+ 3. If only upper bound exists in current duration results:
+ 1. This can only happen for the smallest target loss ratio.
+ 2. If the upper bound was measured at min_rate,
+ exit the whole phase early (not investigating other target loss ratios).
+ 3. Select new transmit rate using external search:
+ 1. For computing previous interval size, use:
+ 1. second tightest bound at current duration,
+ 2. or tightest bound of previous duration,
+ if compatible and giving a more narrow interval,
+ 3. or target interval width if none of the above is available.
+ 4. In any case increase to target interval width if smaller.
+ 2. Quadruple the interval width.
+ 3. Use min_rate if the new transmit rate is lower.
+ 4. If only lower bound exists in current duration results:
+ 1. If the lower bound was measured at max_rate,
+ exit this iteration (continue with next lowest target loss ratio).
+ 2. Select new transmit rate using external search:
+ 1. For computing previous interval size, use:
+ 1. second tightest bound at current duration,
+ 2. or tightest bound of previous duration,
+ if compatible and giving a more narrow interval,
+ 3. or target interval width if none of the above is available.
+ 4. In any case increase to target interval width if smaller.
+ 2. Quadruple the interval width.
+ 3. Use max_rate if the new transmit rate is higher.
+ 5. The only remaining option is both bounds in current duration results.
+ 1. This can happen in two ways, depending on how the lower bound
+ was chosen.
+ 1. It could have been selected for the current loss ratio,
+ e.g. in re-measurement (2.2) or in initial bisect (2.1).
+ 2. It could have been found as an upper bound for the previous smaller
+ target loss ratio, in which case it might be too low.
+ 3. The algorithm does not track which one is the case,
+ as the decision logic works well regardless.
+ 2. Compute "extending down" candidate transmit rate exactly as in 2.3.
+ 3. Compute "bisecting" candidate transmit rate:
+ 1. Compute the current interval width from the two bounds.
+ 2. Express the width as a (float) multiple of the target width goal
+ for this phase.
+ 3. If the multiple is not higher than one, it means the width goal
+ is met. Exit this iteration and continue with next higher
+ target loss ratio.
+ 4. If the multiple is two or less, use half of that
+ for new width if the lower subinterval.
+ 5. Round the multiple up to nearest even integer.
+ 6. Use half of that for new width if the lower subinterval.
+ 7. Example: If lower bound is 2.0 and upper bound is 5.0, and width
+ goal is 1.0, the new candidate transmit rate will be 4.0.
+ This can save a measurement when 4.0 has small loss.
+ Selecting the average (3.5) would never save a measurement,
+ giving more narrow bounds instead.
+ 4. If either candidate computation want to exit the iteration,
+ do as bisecting candidate computation says.
+ 5. The remaining case is both candidates wanting to measure at some rate.
+ Use the higher rate. This prefers external search down narrow enough
+ interval, competing with perfectly sized lower bisect subinterval.
+
+# FD.io CSIT Implementation
+
+The only known working implementation of MLRsearch is in
+the open-source code running in Linux Foundation
+FD.io CSIT project [FDio-CSIT-MLRsearch] as part of
+a Continuous Integration / Continuous Development (CI/CD) framework.
+
+MLRsearch is also available as a Python package in [PyPI-MLRsearch].
+
+## Additional details
+
+This document so far has been describing a simplified version of
+MLRsearch algorithm. The full algorithm as implemented in CSIT contains
+additional logic, which makes some of the details (but not general
+ideas) above incorrect. Here is a short description of the additional
+logic as a list of principles, explaining their main differences from
+(or additions to) the simplified description, but without detailing
+their mutual interaction.
+
+1. Logarithmic transmit rate.
+ * In order to better fit the relative width goal, the interval
+ doubling and halving is done differently.
+ * For example, the middle of 2 and 8 is 4, not 5.
+2. Timeout for bad cases.
+ * The worst case for MLRsearch is when each phase converges to
+ intervals way different than the results of the previous phase.
+ * Rather than suffer total search time several times larger than pure
+ binary search, the implemented tests fail themselves when the
+ search takes too long (given by argument *timeout*).
+3. Intended count.
+ * The number of packets to send during the trial should be equal to
+ the intended load multiplied by the duration.
+ * Also multiplied by a coefficient, if loss ratio is calculated
+ from a different metric.
+ * Example: If a successful transaction uses 10 packets,
+ load is given in transactions per second, byt loss ratio is calculated
+ from packets, the coefficient to get intended count of packets is 10.
+ * But in practice that does not work.
+ * It could result in a fractional number of packets,
+ * so it has to be rounded in a way traffic generator chooses,
+ * which may depend on the number of traffic flows
+ and traffic generator worker threads.
+4. Attempted count. As the real number of intended packets is not known exactly,
+ the computation uses the number of packets traffic generator reports as sent.
+ Unless overriden by the next point.
+5. Duration stretching.
+ * In some cases, traffic generator may get overloaded,
+ causing it to take significantly longer (than duration) to send all packets.
+ * The implementation uses an explicit stop,
+ * causing lower attempted count in those cases.
+ * The implementation tolerates some small difference between
+ attempted count and intended count.
+ * 10 microseconds worth of traffic is sufficient for our tests.
+ * If the difference is higher, the unsent packets are counted as lost.
+ * This forces the search to avoid the regions of high duration stretching.
+ * The final bounds describe the performance of not just SUT,
+ but of the whole system, including the traffic generator.
+6. Excess packets.
+ * In some test (e.g. using TCP flows) Traffic generator reacts to packet loss
+ by retransmission. Usually, such packet loss is already affecting loss ratio.
+ If a test also wants to treat retransmissions due to heavily delayed packets
+ also as a failure, this is once again visible as a mismatch between
+ the intended count and the attempted count.
+ * The CSIT implementation simply looks at absolute value of the difference,
+ so it offes the same small tolerance before it start marking a "loss".
+7. For result processing, we use lower bounds and ignore upper bounds.
+
+### FD.io CSIT Input Parameters
+
+1. **max_rate** - Typical values: 2 * 14.88 Mpps for 64B
+ 10GE link rate, 2 * 18.75 Mpps for 64B 40GE NIC (specific model).
+2. **min_rate** - Value: 2 * 9001 pps (we reserve 9000 pps
+ for latency measurements).
+3. **final_trial_duration** - Value: 30.0 seconds.
+4. **initial_trial_duration** - Value: 1.0 second.
+5. **final_relative_width** - Value: 0.005 (0.5%).
+6. **packet_loss_ratios** - Value: 0.0, 0.005 (0.0% for NDR, 0.5% for PDR).
+7. **number_of_intermediate_phases** - Value: 2.
+ The value has been chosen based on limited experimentation to date.
+ More experimentation needed to arrive to clearer guidelines.
+8. **timeout** - Limit for the overall search duration (for one search).
+ If MLRsearch oversteps this limit, it immediatelly declares the test failed,
+ to avoid wasting even more time on a misbehaving SUT.
+ Value: 600.0 (seconds).
+9. **expansion_coefficient** - Width multiplier for external search.
+ Value: 4.0 (interval width is quadroupled).
+ Value of 2.0 is best for well-behaved SUTs, but value of 4.0 has been found
+ to decrease overall search time for worse-behaved SUT configurations,
+ contributing more to the overall set of different SUT configurations tested.
+
+
+## Example MLRsearch Run
+
+
+The following list describes a search from a real test run in CSIT
+(using the default input values as above).
+
+* Initial phase, trial duration 1.0 second.
+
+Measurement 1, intended load 18750000.0 pps (MTR),
+measured loss ratio 0.7089514628479618 (valid upper bound for both NDR and PDR).
+
+Measurement 2, intended load 5457160.071600716 pps (MRR),
+measured loss ratio 0.018650817320118702 (new tightest upper bounds).
+
+Measurement 3, intended load 5348832.933500009 pps (slightly less than MRR2
+in preparation for first intermediate phase target interval width),
+measured loss ratio 0.00964383362905351 (new tightest upper bounds).
+
+* First intermediate phase starts, trial duration still 1.0 seconds.
+
+Measurement 4, intended load 4936605.579021453 pps (no lower bound,
+performing external search downwards, for NDR),
+measured loss ratio 0.0 (valid lower bound for both NDR and PDR).
+
+Measurement 5, intended load 5138587.208637197 pps (bisecting for NDR),
+measured loss ratio 0.0 (new tightest lower bounds).
+
+Measurement 6, intended load 5242656.244044665 pps (bisecting),
+measured loss ratio 0.013523745379347257 (new tightest upper bounds).
+
+* Both intervals are narrow enough.
+* Second intermediate phase starts, trial duration 5.477225575051661 seconds.
+
+Measurement 7, intended load 5190360.904111567 pps (initial bisect for NDR),
+measured loss ratio 0.0023533920869969953 (NDR upper bound, PDR lower bound).
+
+Measurement 8, intended load 5138587.208637197 pps (re-measuring NDR lower bound),
+measured loss ratio 1.2080222912800403e-06 (new tightest NDR upper bound).
+
+* The two intervals have separate bounds from now on.
+
+Measurement 9, intended load 4936605.381062318 pps (external NDR search down),
+measured loss ratio 0.0 (new valid NDR lower bound).
+
+Measurement 10, intended load 5036583.888432355 pps (NDR bisect),
+measured loss ratio 0.0 (new tightest NDR lower bound).
+
+Measurement 11, intended load 5087329.903232804 pps (NDR bisect),
+measured loss ratio 0.0 (new tightest NDR lower bound).
+
+* NDR interval is narrow enough, PDR interval not ready yet.
+
+Measurement 12, intended load 5242656.244044665 pps (re-measuring PDR upper bound),
+measured loss ratio 0.0101174866190136 (still valid PDR upper bound).
+
+* Also PDR interval is narrow enough, with valid bounds for this duration.
+* Final phase starts, trial duration 30.0 seconds.
+
+Measurement 13, intended load 5112894.3238511775 pps (initial bisect for NDR),
+measured loss ratio 0.0 (new tightest NDR lower bound).
+
+Measurement 14, intended load 5138587.208637197 (re-measuring NDR upper bound),
+measured loss ratio 2.030389804256833e-06 (still valid PDR upper bound).
+
+* NDR interval is narrow enough, PDR interval not yet.
+
+Measurement 15, intended load 5216443.04126728 pps (initial bisect for PDR),
+measured loss ratio 0.005620871287975237 (new tightest PDR upper bound).
+
+Measurement 16, intended load 5190360.904111567 (re-measuring PDR lower bound),
+measured loss ratio 0.0027629971184465604 (still valid PDR lower bound).
+
+* PDR interval is also narrow enough.
+* Returning bounds:
+* NDR_LOWER = 5112894.3238511775 pps; NDR_UPPER = 5138587.208637197 pps;
+* PDR_LOWER = 5190360.904111567 pps; PDR_UPPER = 5216443.04126728 pps.
+
+# IANA Considerations
+
+No requests of IANA.
+
+# Security Considerations
+
+Benchmarking activities as described in this memo are limited to
+technology characterization of a DUT/SUT using controlled stimuli in a
+laboratory environment, with dedicated address space and the constraints
+specified in the sections above.
+
+The benchmarking network topology will be an independent test setup and
+MUST NOT be connected to devices that may forward the test traffic into
+a production network or misroute traffic to the test management network.
+
+Further, benchmarking is performed on a "black-box" basis, relying
+solely on measurements observable external to the DUT/SUT.
+
+Special capabilities SHOULD NOT exist in the DUT/SUT specifically for
+benchmarking purposes.Any implications for network security arising
+from the DUT/SUT SHOULD be identical in the lab and in production
+networks.
+
+# Acknowledgements
+
+Many thanks to Alec Hothan of OPNFV NFVbench project for thorough
+review and numerous useful comments and suggestions.
+
+--- back