diff options
author | Maciek Konstantynowicz <mkonstan@cisco.com> | 2018-10-13 17:32:41 +0100 |
---|---|---|
committer | Maciek Konstantynowicz <mkonstan@cisco.com> | 2018-10-23 16:40:49 +0000 |
commit | d350b2d24cb116e14e991f784dc11b3462def56c (patch) | |
tree | 21e0de6f16d9a32a2e1513d7ec7c7572eb76ab70 | |
parent | db62f8849fd40bf3ee812eb87534e1a3c358b711 (diff) |
Initial version of PLRsearch ietf draft in kramdown-rfc2629 format.
Change-Id: I4f26dfc6c981255d0178cadf9b48a09e10d89f03
Signed-off-by: Maciek Konstantynowicz <mkonstan@cisco.com>
-rw-r--r-- | docs/ietf/draft-vpolak-plrsearch-00.md | 228 |
1 files changed, 228 insertions, 0 deletions
diff --git a/docs/ietf/draft-vpolak-plrsearch-00.md b/docs/ietf/draft-vpolak-plrsearch-00.md new file mode 100644 index 0000000000..e71b527919 --- /dev/null +++ b/docs/ietf/draft-vpolak-plrsearch-00.md @@ -0,0 +1,228 @@ +--- +title: Probabilistic Loss Ratio Search for Packet Throughput (PLRsearch) +# abbrev: PLRsearch +docname: draft-vpolak-plrsearch-00 +date: 2018-10-22 + +ipr: trust200902 +area: ops +wg: Benchmarking Working Group +kw: Internet-Draft +cat: info + +coding: us-ascii +pi: # can use array (if all yes) or hash here +# - toc +# - sortrefs +# - symrefs + toc: yes + sortrefs: # defaults to yes + symrefs: yes + +author: + - + ins: M. Konstantynowicz + name: Maciek Konstantynowicz + org: Cisco Systems + role: editor + email: mkonstan@cisco.com + - + ins: V. Polak + name: Vratko Polak + org: Cisco Systems + role: editor + email: vrpolak@cisco.com + +normative: + RFC2544: + RFC8174: + +informative: + +--- abstract + +This document addresses challenges while applying methodologies +described in [RFC2544] to benchmarking NFV (Network Function +Virtualization) over an extended period of time, sometimes referred to +as "soak testing". More specifically to benchmarking software based +implementations of NFV data planes. Packet throughput search approach +proposed by this document assumes that system under test is +probabilistic in nature, and not deterministic. + +--- middle + +# Motivation + +Network providers are interested in throughput a device can sustain. + +RFC 2544 assumes loss ratio is given by a deterministic function of +offered load. But NFV software devices are not deterministic (enough). +This leads for deterministic algorithms (such as MLRsearch with single +trial) to return results, which when repeated show relatively high +standard deviation, thus making it harder to tell what "the throughput" +actually is. + +We need another algorithm, which takes this indeterminism into account. + +# Model + +Each algorithm searches for an answer to a precisely formulated +question. When the question involves indeterministic systems, it has to +specify probabilities (or prior distributions) which are tied to a +specific probabilistic model. Different models will have different +number (and meaning) of parameters. Complicated (but more realistic) +models have many parameters, and the math involved can be very +complicated. It is better to start with simpler probabilistic model, and +only change it when the output of the simpler algorithm is not stable or +useful enough. + +TODO: Refer to packet forwarding terminology, such as "offered load" and +"loss ratio". + +TODO: Mention that no packet duplication is expected (or is filtered +out). + +TODO: Define critical load and critical region earlier. + +This document is focused on algorithms related to packet loss count +only. No latency (or other information) is taken into account. For +simplicity, only one type of measurement is considered: dynamically +computed offered load, constant within trial measurement of +predetermined trial duration. + +Also, running longer trials (in some situations) could be more efficient, +but in order to perform trial at multiple offered loads withing critical region, +trial durations should be kept as short as possible. + +# Poisson Distribution + +TODO: Give link to more officially published literature about Poisson +distribution. + +Note-1: that the algorithm makes an assumption that packet traffic +generator detects duplicate packets on receive detection, and reports +this as an error. + +Note-2: Binomial distribution is a better fit compared to Poisson +distribution (acknowledging that the number of packets lost cannot be +higher than the number of packets offered), but the difference tends to +be relevant in loads far above the critical region, so using Poisson +distribution helps the algorithm focus on critical region better. + +When comparing different offered loads, the average loss per second is +assumed to increase, but the (deterministic) function from offered load +into average loss rate is otherwise unknown. + +Given a loss target (configurable, by default one packet lost per +second), there is an unknown offered load when the average is exactly +that. We call that the "critical load". If critical load seems higher +than maximum offerable load, we should use the maximum offerable load to +make search output more stable. + +Of course, there are great many increasing functions. The offered load +has to be chosen for each trial, and the computed posterior distribution +of critical load can change with each trial result. + +To make the space of possible functions more tractable, some other +simplifying assumption is needed. As the algorithm will be examining +(also) loads close to the critical load, linear approximation to the +function (TODO: name the function) in the critical region is important. +But as the search algorithm needs to evaluate the function also far +away from the critical region, the approximate function has to be well- +behaved for every positive offered load, specifically it cannot predict +non-positive packet loss rate. + +Within this document, "fitting function" is the name for such well-behaved +function which approximates the unknown function in the critical region. + +Results from trials far from the critical region are likely to affect +the critical rate estimate negatively, as the fitting function does not +need to be a good approximation there. Instead of discarding some +results, or "suppressing" their impact with ad-hoc methods (other than +using Poisson distribution instead of binomial) is not used, as such +methods tend to make the overall search unstable. We rely on most of +measurements being done (eventually) within the critical region, and +overweighting far-off measurements (eventually) for well-behaved fitting +functions. + +# Fitting Function Coefficients Distribution + +To accomodate systems with different behaviours, the fitting function is +expected to have few numeric parameters affecting its shape (mainly +affecting the linear approximation in the critical region). + +The general search algorithm can use whatever increasing fitting +function, some specific functions can be described later. + +TODO: Describe sigmoid-based and erf-based functions. + +It is up to implementer to chose a fitting function and prior +distribution of its parameters. The rest of this document assumes each +parameter is independently and uniformly distributed over common +interval. Implementers are to add non-linear transformations into their +fitting functions if their prior is different. + +TODO: Move the following sentence into more appropriate place. + +Speaking about new trials, each next trial will be done at offered load +equal to the current average of the critical load. + +Exit condition is either critical load stdev becoming small enough, or +overal search time becoming long enough. + +The algorithm should report both avg and stdev for critical load. If the +reported averages follow a trend (without reaching equilibrium), avg and +stdev should refer to the equilibrium estibated based on the trend, not +to immediate posterior values. + +TODO: Explicitly mention the iterative character of the search. + +# Algorithm Formulas + +## Integration + +The posterior distributions for fitting function parameters will not be +integrable in general. + +The search algorithm utilises the fact that trial measurement takes some +time, so this time can be used for numeric integration (using suitable +method, such as Monte Carlo) to achieve sufficient precision. + +## Optimizations + +After enough trials, the posterior distribution will be concentrated in +a narrow area of parameter space. The integration method could take +advantage of that. + +Even in the concentrated area, the likelihood can be quite small, so the +integration algorithm should track the logarithm of the likelihood, and +also avoid underflow errors bu ther means. + +# Known Implementations + +The only known working implementatin of Probabilistic Loss Ratio Search +for Packet Throughput is in Linux Foundation FD.io CSIT project. https://wiki.fd.io/view/CSIT. https://git.fd.io/csit/. + +## FD.io CSIT Implementation Specifics + +In a sample implemenation in FD.io CSIT project, there is around 0.5 +second delay between trials due to restrictons imposed by packet traffic +generator in use (T-Rex), avoiding that delay is out of scope of this +document. + +TODO: Describe how the current integration algorithm finds the +concentrated area. + +# IANA Considerations + +.. + +# Security Considerations + +.. + +# Acknowledgements + +.. + +--- back
\ No newline at end of file |