diff options
Diffstat (limited to 'resources')
23 files changed, 4994 insertions, 7531 deletions
diff --git a/resources/tools/presentation/__init__.py b/resources/tools/presentation/__init__.py index a3b7344358..67fa7230d6 100644 --- a/resources/tools/presentation/__init__.py +++ b/resources/tools/presentation/__init__.py @@ -16,3 +16,19 @@ __init__ file for directory presentation This makes the presentation a part of the great CSIT resources package. """ + +from .pal_errors import PresentationError +from .pal_utils import archive_input_data, classify_anomalies, \ + convert_csv_to_pretty_txt, execute_command, get_files, \ + get_last_completed_build_number, get_last_successful_build_nr, \ + get_rst_title_char, mean, relative_change, relative_change_stdev, stdev +from .environment import Environment +from .specification_parser import Specification +from .input_data_parser import InputData +from .generator_tables import generate_tables +from .generator_plots import generate_plots +from .generator_files import generate_files +from .static_content import prepare_static_content +from .generator_report import generate_report +from .generator_cpta import generate_cpta +from .generator_alerts import Alerting diff --git a/resources/tools/presentation/conf.py b/resources/tools/presentation/conf.py index ae781bdac8..66dc46952c 100644 --- a/resources/tools/presentation/conf.py +++ b/resources/tools/presentation/conf.py @@ -1,25 +1,26 @@ # -*- coding: utf-8 -*- -# -# CSIT report documentation build configuration file -# -# This file is execfile()d with the current directory set to its -# containing dir. -# -# Note that not all possible configuration values are present in this -# autogenerated file. -# -# All configuration values have a default; values that are commented out -# serve to show the default. -# If extensions (or modules to document with autodoc) are in another directory, -# add these directories to sys.path here. If the directory is relative to the -# documentation root, use os.path.abspath to make it absolute, like shown here. +"""CSIT report documentation build configuration file + +This file is execfile()d with the current directory set to its +containing dir. + +Note that not all possible configuration values are present in this +autogenerated file. + +All configuration values have a default; values that are commented out +serve to show the default. + +If extensions (or modules to document with autodoc) are in another directory, +add these directories to sys.path here. If the directory is relative to the +documentation root, use os.path.abspath to make it absolute, like shown here. +""" import os import sys -sys.path.insert(0, os.path.abspath('.')) +sys.path.insert(0, os.path.abspath(u'.')) # -- General configuration ------------------------------------------------ @@ -30,22 +31,22 @@ sys.path.insert(0, os.path.abspath('.')) # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. -extensions = ['sphinxcontrib.programoutput', - 'sphinx.ext.ifconfig'] +extensions = [u'sphinxcontrib.programoutput', + u'sphinx.ext.ifconfig'] # Add any paths that contain templates here, relative to this directory. -templates_path = ['_templates'] +templates_path = [u'_templates'] # The suffix(es) of source file names. # You can specify multiple suffix as a list of string: # -source_suffix = ['.rst', '.md'] +source_suffix = [u'.rst', u'.md'] # The master toctree document. -master_doc = 'index' +master_doc = u'index' # General information about the project. -report_week = '38' +report_week = u'38' project = u'FD.io CSIT-1908.{week}'.format(week=report_week) copyright = u'2019, FD.io' author = u'FD.io CSIT' @@ -59,7 +60,7 @@ author = u'FD.io CSIT' # The full version, including alpha/beta/rc tags. # release = u'' -rst_epilog = """ +rst_epilog = u""" .. |release-1| replace:: {prev_release} .. |srelease| replace:: {srelease} .. |csit-release| replace:: CSIT-{csitrelease} @@ -108,35 +109,35 @@ rst_epilog = """ .. _CSIT Testbed Setup: https://git.fd.io/csit/tree/resources/tools/testbed-setup/README.md?h={release} .. _K8s configuration files: https://github.com/FDio/csit/tree/{release}/resources/templates/kubernetes .. _VPP startup.conf: https://git.fd.io/vpp/tree/src/vpp/conf/startup.conf?h=stable/{srelease}&id={vpp_release_commit_id} -""".format(release='rls1908', +""".format(release=u'rls1908', report_week=report_week, - prev_release='rls1904', - srelease='1908', - csitrelease='1908', - csit_prev_release='1904', - vpprelease='19.08', - vpp_prev_release='19.04', - dpdkrelease='19.05', - sdpdkrelease='1905', - trex_version='v2.54', - csit_ubuntu_ver='csit-ubuntu-16.04.1_2019-03-27_2.4', - csit_centos_ver='csit-centos-7.6-1810_2019-05-20_2.5', - vpp_release_commit_id='1c586de48cc76fc6eac50f5d87003e2a80aa43e7') + prev_release=u'rls1904', + srelease=u'1908', + csitrelease=u'1908', + csit_prev_release=u'1904', + vpprelease=u'19.08', + vpp_prev_release=u'19.04', + dpdkrelease=u'19.05', + sdpdkrelease=u'1905', + trex_version=u'v2.54', + csit_ubuntu_ver=u'csit-ubuntu-16.04.1_2019-03-27_2.4', + csit_centos_ver=u'csit-centos-7.6-1810_2019-05-20_2.5', + vpp_release_commit_id=u'1c586de48cc76fc6eac50f5d87003e2a80aa43e7') # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. -language = 'en' +language = u'en' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This patterns also effect to html_static_path and html_extra_path -exclude_patterns = ['_build', 'Thumbs.db', '.DS_Store'] +exclude_patterns = [u'_build', u'Thumbs.db', u'.DS_Store'] # The name of the Pygments (syntax highlighting) style to use. -pygments_style = 'sphinx' +pygments_style = u'sphinx' # If true, `todo` and `todoList` produce output, else they produce nothing. todo_include_todos = False @@ -146,40 +147,41 @@ todo_include_todos = False # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # -html_theme = 'sphinx_rtd_theme' +html_theme = u'sphinx_rtd_theme' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. # html_theme_options = { - 'canonical_url': '', - 'analytics_id': '', - 'logo_only': False, - 'display_version': True, - 'prev_next_buttons_location': 'bottom', - 'style_external_links': False, + u'canonical_url': u'', + u'analytics_id': u'', + u'logo_only': False, + u'display_version': True, + u'prev_next_buttons_location': u'bottom', + u'style_external_links': False, # Toc options - 'collapse_navigation': True, - 'sticky_navigation': True, - 'navigation_depth': 3, - 'includehidden': True, - 'titles_only': False + u'collapse_navigation': True, + u'sticky_navigation': True, + u'navigation_depth': 3, + u'includehidden': True, + u'titles_only': False } # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". -html_theme_path = ['env/lib/python2.7/site-packages/sphinx_rtd_theme'] +html_theme_path = [u'env/lib/python2.7/site-packages/sphinx_rtd_theme'] # html_static_path = ['_build/_static'] -html_static_path = ['_tmp/src/_static'] +html_static_path = [u'_tmp/src/_static'] html_context = { - 'css_files': [ - '_static/theme_overrides.css', # overrides for wide tables in RTD theme - ], - } + u'css_files': [ +# overrides for wide tables in RTD theme + u'_static/theme_overrides.css' + ] +} # If false, no module index is generated. html_domain_indices = True @@ -192,59 +194,58 @@ html_split_index = False # -- Options for LaTeX output --------------------------------------------- -latex_engine = 'pdflatex' +latex_engine = u'pdflatex' latex_elements = { - # The paper size ('letterpaper' or 'a4paper'). - # - 'papersize': 'a4paper', - - # The font size ('10pt', '11pt' or '12pt'). - # - #'pointsize': '10pt', - - # Additional stuff for the LaTeX preamble. - # - 'preamble': r''' - \usepackage{pdfpages} - \usepackage{svg} - \usepackage{charter} - \usepackage[defaultsans]{lato} - \usepackage{inconsolata} - \usepackage{csvsimple} - \usepackage{longtable} - \usepackage{booktabs} - ''', - - # Latex figure (float) alignment - # - 'figure_align': 'H', - - # Latex font setup - # - 'fontpkg': r''' - \renewcommand{\familydefault}{\sfdefault} - ''', - - # Latex other setup - # - 'extraclassoptions': 'openany', - 'sphinxsetup': r''' - TitleColor={RGB}{225,38,40}, - InnerLinkColor={RGB}{62,62,63}, - OuterLinkColor={RGB}{225,38,40}, - shadowsep=0pt, - shadowsize=0pt, - shadowrule=0pt - ''' + # The paper size ('letterpaper' or 'a4paper'). + # + u'papersize': u'a4paper', + + # The font size ('10pt', '11pt' or '12pt'). + # + #'pointsize': '10pt', + + # Additional stuff for the LaTeX preamble. + # + u'preamble': r''' + \usepackage{pdfpages} + \usepackage{svg} + \usepackage{charter} + \usepackage[defaultsans]{lato} + \usepackage{inconsolata} + \usepackage{csvsimple} + \usepackage{longtable} + \usepackage{booktabs} + ''', + + # Latex figure (float) alignment + # + u'figure_align': u'H', + + # Latex font setup + # + u'fontpkg': r''' + \renewcommand{\familydefault}{\sfdefault} + ''', + + # Latex other setup + # + u'extraclassoptions': u'openany', + u'sphinxsetup': r''' + TitleColor={RGB}{225,38,40}, + InnerLinkColor={RGB}{62,62,63}, + OuterLinkColor={RGB}{225,38,40}, + shadowsep=0pt, + shadowsize=0pt, + shadowrule=0pt + ''' } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ - (master_doc, 'csit.tex', u'CSIT REPORT', - u'', 'manual'), + (master_doc, u'csit.tex', u'CSIT REPORT', u'', u'manual'), ] # The name of an image file (relative to this directory) to place at the top of @@ -263,7 +264,7 @@ latex_show_pagerefs = True # If true, show URL addresses after external links. # -latex_show_urls = 'footnote' +latex_show_urls = u'footnote' # Documents to append as an appendix to all manuals. # diff --git a/resources/tools/presentation/conf_cpta/conf.py b/resources/tools/presentation/conf_cpta/conf.py index 4eb51d884e..c20a5e2996 100644 --- a/resources/tools/presentation/conf_cpta/conf.py +++ b/resources/tools/presentation/conf_cpta/conf.py @@ -18,7 +18,7 @@ import os import sys -sys.path.insert(0, os.path.abspath('.')) +sys.path.insert(0, os.path.abspath(u'.')) # -- General configuration ------------------------------------------------ @@ -29,19 +29,19 @@ sys.path.insert(0, os.path.abspath('.')) # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. -extensions = ['sphinxcontrib.programoutput', - 'sphinx.ext.ifconfig'] +extensions = [u'sphinxcontrib.programoutput', + u'sphinx.ext.ifconfig'] # Add any paths that contain templates here, relative to this directory. -templates_path = ['_templates'] +templates_path = [u'_templates'] # The suffix(es) of source filenames. # You can specify multiple suffix as a list of string: # -source_suffix = ['.rst', '.md'] +source_suffix = [u'.rst', u'.md'] # The master toctree document. -master_doc = 'index' +master_doc = u'index' # General information about the project. project = u'FD.io CSIT' @@ -62,15 +62,15 @@ author = u'FD.io CSIT' # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. -language = 'en' +language = u'en' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This patterns also effect to html_static_path and html_extra_path -exclude_patterns = ['_build', 'Thumbs.db', '.DS_Store'] +exclude_patterns = [u'_build', u'Thumbs.db', u'.DS_Store'] # The name of the Pygments (syntax highlighting) style to use. -pygments_style = 'sphinx' +pygments_style = u'sphinx' # If true, `todo` and `todoList` produce output, else they produce nothing. todo_include_todos = False @@ -80,7 +80,7 @@ todo_include_todos = False # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # -html_theme = 'sphinx_rtd_theme' +html_theme = u'sphinx_rtd_theme' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the @@ -91,14 +91,14 @@ html_theme = 'sphinx_rtd_theme' # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". -html_theme_path = ['env/lib/python2.7/site-packages/sphinx_rtd_theme'] +html_theme_path = [u'env/lib/python2.7/site-packages/sphinx_rtd_theme'] # html_static_path = ['_build/_static'] -html_static_path = ['../_tmp/src/_static'] +html_static_path = [u'../_tmp/src/_static'] html_context = { - 'css_files': [ - '_static/theme_overrides.css', # overrides for wide tables in RTD theme + u'css_files': [ + u'_static/theme_overrides.css', # overrides for wide tables in RTD theme ], } diff --git a/resources/tools/presentation/doc/pal_lld.rst b/resources/tools/presentation/doc/pal_lld.rst index d3a9782681..ec4c97b454 100644 --- a/resources/tools/presentation/doc/pal_lld.rst +++ b/resources/tools/presentation/doc/pal_lld.rst @@ -841,7 +841,7 @@ VPP HTTP server performance in a box chart with pre-defined data - type: "plot" title: "VPP HTTP Server Performance" - algorithm: "plot_http_server_performance_box" + algorithm: "plot_http_server_perf_box" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/http-server-performance-cps" data: @@ -1263,7 +1263,7 @@ The model specifies: - type: "table" title: "Performance comparison" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file-ext: ".csv" output-file: "{DIR[DTR,PERF,VPP,IMPRV]}/vpp_performance_comparison" reference: @@ -1480,7 +1480,7 @@ At first, the .csv tables are generated (only the table for 1t1c is shown): - type: "table" title: "Performance trending dashboard" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c" data: "plot-performance-trending-all" @@ -1503,7 +1503,7 @@ Then, html tables stored inside .rst files are generated: - type: "table" title: "HTML performance trending dashboard 1t1c" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c.rst" diff --git a/resources/tools/presentation/environment.py b/resources/tools/presentation/environment.py index 7cddb0279c..ea4d94ad70 100644 --- a/resources/tools/presentation/environment.py +++ b/resources/tools/presentation/environment.py @@ -21,7 +21,7 @@ import os import shutil import logging -from errors import PresentationError +from pal_errors import PresentationError class Environment: @@ -60,34 +60,37 @@ class Environment: """ if self._force: - logging.info("Removing old build(s) ...") - for directory in self._env["build-dirs"]: - dir_to_remove = self._env["paths"][directory] + logging.info(u"Removing old build(s) ...") + for directory in self._env[u"build-dirs"]: + dir_to_remove = self._env[u"paths"][directory] if os.path.isdir(dir_to_remove): try: shutil.rmtree(dir_to_remove) - logging.info(" Removed: {}".format(dir_to_remove)) + logging.info(f" Removed: {dir_to_remove}") except OSError: - raise PresentationError("Cannot remove the directory " - "'{}'".format(dir_to_remove)) - logging.info("Done.") + raise PresentationError( + f"Cannot remove the directory {dir_to_remove}" + ) + logging.info(u"Done.") - logging.info("Making directories ...") + logging.info(u"Making directories ...") - for directory in self._env["make-dirs"]: - dir_to_make = self._env["paths"][directory] + for directory in self._env[u"make-dirs"]: + dir_to_make = self._env[u"paths"][directory] try: if os.path.isdir(dir_to_make): - logging.warning("The directory '{}' exists, skipping.". - format(dir_to_make)) + logging.warning( + f"The directory {dir_to_make} exists, skipping." + ) else: os.makedirs(dir_to_make) - logging.info(" Created: {}".format(dir_to_make)) + logging.info(f" Created: {dir_to_make}") except OSError: - raise PresentationError("Cannot make the directory '{}'". - format(dir_to_make)) + raise PresentationError( + f"Cannot make the directory {dir_to_make}" + ) - logging.info("Done.") + logging.info(u"Done.") def set_environment(self): """Set the environment. @@ -101,28 +104,26 @@ def clean_environment(env): :param env: Environment specification. :type env: dict - :raises: PresentationError if it is not possible to remove a directory. """ - logging.info("Cleaning the environment ...") + logging.info(u"Cleaning the environment ...") - if not env["remove-dirs"]: # None or empty - logging.info(" No directories to remove.") + if not env[u"remove-dirs"]: # None or empty + logging.info(u" No directories to remove.") return - for directory in env["remove-dirs"]: - dir_to_remove = env["paths"][directory] - logging.info(" Removing the working directory {} ...". - format(dir_to_remove)) + for directory in env[u"remove-dirs"]: + dir_to_remove = env[u"paths"][directory] + logging.info(f" Removing the working directory {dir_to_remove} ...") if os.path.isdir(dir_to_remove): try: shutil.rmtree(dir_to_remove) except OSError as err: - logging.warning("Cannot remove the directory '{}'". - format(dir_to_remove)) + logging.warning( + f"Cannot remove the directory {dir_to_remove}" + ) logging.debug(str(err)) else: - logging.warning("The directory '{}' does not exist.". - format(dir_to_remove)) + logging.warning(f"The directory {dir_to_remove} does not exist.") - logging.info("Done.") + logging.info(u"Done.") diff --git a/resources/tools/presentation/generator_alerts.py b/resources/tools/presentation/generator_alerts.py index 3a9b5ddfb6..10c6734aad 100644 --- a/resources/tools/presentation/generator_alerts.py +++ b/resources/tools/presentation/generator_alerts.py @@ -11,6 +11,13 @@ # See the License for the specific language governing permissions and # limitations under the License. +"""Generator of alerts: +- failed tests +- regressions +- progressions +""" + + import smtplib import logging @@ -19,8 +26,8 @@ from email.mime.multipart import MIMEMultipart from os.path import isdir from collections import OrderedDict -from utils import get_last_completed_build_number -from errors import PresentationError +from pal_utils import get_last_completed_build_number +from pal_errors import PresentationError class AlertingError(PresentationError): @@ -34,7 +41,7 @@ class AlertingError(PresentationError): - relevant data if there are any collected (optional parameter details). """ - def __init__(self, msg, details='', level="CRITICAL"): + def __init__(self, msg, details=u'', level=u"CRITICAL"): """Sets the exception message and the level. :param msg: Short description of the encountered problem. @@ -48,13 +55,13 @@ class AlertingError(PresentationError): :type level: str """ - super(AlertingError, self).__init__( - "Alerting: {0}".format(msg), details, level) + super(AlertingError, self).__init__(f"Alerting: {msg}", details, level) def __repr__(self): return ( - "AlertingError(msg={msg!r},details={dets!r},level={level!r})". - format(msg=self._msg, dets=self._details, level=self._level)) + f"AlertingError(msg={self._msg!r},details={self._details!r}," + f"level={self._level!r})" + ) class Alerting: @@ -69,57 +76,58 @@ class Alerting: """ # Implemented alerts: - self._ALERTS = ("failed-tests", ) + self._implemented_alerts = (u"failed-tests", ) self._spec = spec try: self._spec_alert = spec.alerting except KeyError as err: - raise AlertingError("Alerting is not configured, skipped.", - repr(err), - "WARNING") + raise AlertingError(u"Alerting is not configured, skipped.", + repr(err), + u"WARNING") - self._path_failed_tests = spec.environment["paths"]["DIR[STATIC,VPP]"] + self._path_failed_tests = spec.environment[u"paths"][u"DIR[STATIC,VPP]"] # Verify and validate input specification: - self.configs = self._spec_alert.get("configurations", None) + self.configs = self._spec_alert.get(u"configurations", None) if not self.configs: - raise AlertingError("No alert configuration is specified.") - for config_type, config_data in self.configs.iteritems(): - if config_type == "email": - if not config_data.get("server", None): - raise AlertingError("Parameter 'server' is missing.") - if not config_data.get("address-to", None): - raise AlertingError("Parameter 'address-to' (recipient) is " - "missing.") - if not config_data.get("address-from", None): - raise AlertingError("Parameter 'address-from' (sender) is " - "missing.") - elif config_type == "jenkins": - if not isdir(config_data.get("output-dir", "")): - raise AlertingError("Parameter 'output-dir' is " - "missing or it is not a directory.") - if not config_data.get("output-file", None): - raise AlertingError("Parameter 'output-file' is missing.") + raise AlertingError(u"No alert configuration is specified.") + for config_type, config_data in self.configs.items(): + if config_type == u"email": + if not config_data.get(u"server", None): + raise AlertingError(u"Parameter 'server' is missing.") + if not config_data.get(u"address-to", None): + raise AlertingError(u"Parameter 'address-to' (recipient) " + u"is missing.") + if not config_data.get(u"address-from", None): + raise AlertingError(u"Parameter 'address-from' (sender) is " + u"missing.") + elif config_type == u"jenkins": + if not isdir(config_data.get(u"output-dir", u"")): + raise AlertingError(u"Parameter 'output-dir' is " + u"missing or it is not a directory.") + if not config_data.get(u"output-file", None): + raise AlertingError(u"Parameter 'output-file' is missing.") else: - raise AlertingError("Alert of type '{0}' is not implemented.". - format(config_type)) + raise AlertingError( + f"Alert of type {config_type} is not implemented." + ) - self.alerts = self._spec_alert.get("alerts", None) + self.alerts = self._spec_alert.get(u"alerts", None) if not self.alerts: - raise AlertingError("No alert is specified.") - for alert, alert_data in self.alerts.iteritems(): - if not alert_data.get("title", None): - raise AlertingError("Parameter 'title' is missing.") - if not alert_data.get("type", None) in self._ALERTS: - raise AlertingError("Parameter 'failed-tests' is missing or " - "incorrect.") - if not alert_data.get("way", None) in self.configs.keys(): - raise AlertingError("Parameter 'way' is missing or incorrect.") - if not alert_data.get("include", None): - raise AlertingError("Parameter 'include' is missing or the " - "list is empty.") + raise AlertingError(u"No alert is specified.") + for alert_data in self.alerts.values(): + if not alert_data.get(u"title", None): + raise AlertingError(u"Parameter 'title' is missing.") + if not alert_data.get(u"type", None) in self._implemented_alerts: + raise AlertingError(u"Parameter 'failed-tests' is missing or " + u"incorrect.") + if not alert_data.get(u"way", None) in self.configs.keys(): + raise AlertingError(u"Parameter 'way' is missing or incorrect.") + if not alert_data.get(u"include", None): + raise AlertingError(u"Parameter 'include' is missing or the " + u"list is empty.") def __str__(self): """Return string with human readable description of the alert. @@ -127,8 +135,7 @@ class Alerting: :returns: Readable description. :rtype: str """ - return "configs={configs}, alerts={alerts}".format( - configs=self.configs, alerts=self.alerts) + return f"configs={self.configs}, alerts={self.alerts}" def __repr__(self): """Return string executable as Python constructor call. @@ -136,19 +143,19 @@ class Alerting: :returns: Executable constructor call. :rtype: str """ - return "Alerting(spec={spec})".format( - spec=self._spec) + return f"Alerting(spec={self._spec})" def generate_alerts(self): """Generate alert(s) using specified way(s). """ - for alert, alert_data in self.alerts.iteritems(): - if alert_data["way"] == "jenkins": + for alert_data in self.alerts.values(): + if alert_data[u"way"] == u"jenkins": self._generate_email_body(alert_data) else: - raise AlertingError("Alert with way '{0}' is not implemented.". - format(alert_data["way"])) + raise AlertingError( + f"Alert with way {alert_data[u'way']} is not implemented." + ) @staticmethod def _send_email(server, addr_from, addr_to, subject, text=None, html=None): @@ -169,29 +176,29 @@ class Alerting: """ if not text and not html: - raise AlertingError("No text/data to send.") + raise AlertingError(u"No text/data to send.") - msg = MIMEMultipart('alternative') - msg['Subject'] = subject - msg['From'] = addr_from - msg['To'] = ", ".join(addr_to) + msg = MIMEMultipart(u'alternative') + msg[u'Subject'] = subject + msg[u'From'] = addr_from + msg[u'To'] = u", ".join(addr_to) if text: - msg.attach(MIMEText(text, 'plain')) + msg.attach(MIMEText(text, u'plain')) if html: - msg.attach(MIMEText(html, 'html')) + msg.attach(MIMEText(html, u'html')) smtp_server = None try: - logging.info("Trying to send alert '{0}' ...".format(subject)) - logging.debug("SMTP Server: {0}".format(server)) - logging.debug("From: {0}".format(addr_from)) - logging.debug("To: {0}".format(", ".join(addr_to))) - logging.debug("Message: {0}".format(msg.as_string())) + logging.info(f"Trying to send alert {subject} ...") + logging.debug(f"SMTP Server: {server}") + logging.debug(f"From: {addr_from}") + logging.debug(f"To: {u', '.join(addr_to)}") + logging.debug(f"Message: {msg.as_string()}") smtp_server = smtplib.SMTP(server) smtp_server.sendmail(addr_from, addr_to, msg.as_string()) except smtplib.SMTPException as err: - raise AlertingError("Not possible to send the alert via email.", + raise AlertingError(u"Not possible to send the alert via email.", str(err)) finally: if smtp_server: @@ -242,12 +249,12 @@ class Alerting: :rtype: tuple(str, str, int, int, OrderedDict) """ - directory = self.configs[alert["way"]]["output-dir"] + directory = self.configs[alert[u"way"]][u"output-dir"] failed_tests = OrderedDict() - file_path = "{0}/{1}.txt".format(directory, test_set) - version = "" + file_path = f"{directory}/{test_set}.txt" + version = u"" try: - with open(file_path, 'r') as f_txt: + with open(file_path, u'r') as f_txt: for idx, line in enumerate(f_txt): if idx == 0: build = line[:-1] @@ -262,36 +269,69 @@ class Alerting: failed = line[:-1] continue try: - test = line[:-1].split('-') - nic = test[0] - framesize = test[1] - cores = test[2] - name = '-'.join(test[3:-1]) + test = line[:-1].split(u'-') + name = u'-'.join(test[3:-1]) except IndexError: continue if failed_tests.get(name, None) is None: failed_tests[name] = dict(nics=list(), framesizes=list(), cores=list()) - if nic not in failed_tests[name]["nics"]: - failed_tests[name]["nics"].append(nic) - if framesize not in failed_tests[name]["framesizes"]: - failed_tests[name]["framesizes"].append(framesize) - if cores not in failed_tests[name]["cores"]: - failed_tests[name]["cores"].append(cores) + if test[0] not in failed_tests[name][u"nics"]: + failed_tests[name][u"nics"].append(test[0]) + if test[1] not in failed_tests[name][u"framesizes"]: + failed_tests[name][u"framesizes"].append(test[1]) + if test[2] not in failed_tests[name][u"cores"]: + failed_tests[name][u"cores"].append(test[2]) except IOError: - logging.error("No such file or directory: {file}". - format(file=file_path)) + logging.error(f"No such file or directory: {file_path}") return None, None, None, None, None if sort: sorted_failed_tests = OrderedDict() - keys = [k for k in failed_tests.keys()] - keys.sort() - for key in keys: + for key in sorted(failed_tests.keys()): sorted_failed_tests[key] = failed_tests[key] return build, version, passed, failed, sorted_failed_tests - else: - return build, version, passed, failed, failed_tests + + return build, version, passed, failed, failed_tests + + def _list_gressions(self, alert, idx, header, re_pro): + """Create a file with regressions or progressions for the test set + specified by idx. + + :param alert: Files are created for this alert. + :param idx: Index of the test set as it is specified in the + specification file. + :param header: The header of the list of [re|pro]gressions. + :param re_pro: 'regression' or 'progression'. + :type alert: dict + :type idx: int + :type header: str + :type re_pro: str + """ + + if re_pro not in (u"regressions", u"progressions"): + return + + in_file = ( + f"{self.configs[alert[u'way']][u'output-dir']}/" + f"cpta-{re_pro}-{alert[u'urls'][idx].split(u'/')[-1]}.txt" + ) + out_file = ( + f"{self.configs[alert[u'way']][u'output-dir']}/" + f"trending-{re_pro}.txt" + ) + + try: + with open(in_file, u'r') as txt_file: + file_content = txt_file.read() + with open(out_file, u'a+') as reg_file: + reg_file.write(header) + if file_content: + reg_file.write(file_content) + else: + reg_file.write(f"No {re_pro}") + except IOError as err: + logging.warning(repr(err)) def _generate_email_body(self, alert): """Create the file which is used in the generated alert. @@ -300,120 +340,97 @@ class Alerting: :type alert: dict """ - if alert["type"] != "failed-tests": - raise AlertingError("Alert of type '{0}' is not implemented.". - format(alert["type"])) + if alert[u"type"] != u"failed-tests": + raise AlertingError( + f"Alert of type {alert[u'type']} is not implemented." + ) - config = self.configs[alert["way"]] - - text = "" - for idx, test_set in enumerate(alert.get("include", [])): + text = u"" + for idx, test_set in enumerate(alert.get(u"include", [])): build, version, passed, failed, failed_tests = \ self._get_compressed_failed_tests(alert, test_set) if build is None: ret_code, build_nr, _ = get_last_completed_build_number( - self._spec.environment["urls"]["URL[JENKINS,CSIT]"], - alert["urls"][idx].split('/')[-1]) + self._spec.environment[u"urls"][u"URL[JENKINS,CSIT]"], + alert[u"urls"][idx].split(u'/')[-1]) if ret_code != 0: - build_nr = '' - text += "\n\nNo input data available for '{set}'. See CSIT " \ - "build {link}/{build} for more information.\n".\ - format(set='-'.join(test_set.split('-')[-2:]), - link=alert["urls"][idx], - build=build_nr) + build_nr = u'' + text += ( + f"\n\nNo input data available for " + f"{u'-'.join(test_set.split('-')[-2:])}. See CSIT build " + f"{alert[u'urls'][idx]}/{build_nr} for more information.\n" + ) continue - text += ("\n\n{topo}-{arch}, " - "{failed} tests failed, " - "{passed} tests passed, " - "CSIT build: {link}/{build}, " - "VPP version: {version}\n\n". - format(topo=test_set.split('-')[-2], - arch=test_set.split('-')[-1], - failed=failed, - passed=passed, - link=alert["urls"][idx], - build=build, - version=version)) - regression_hdr = ("\n\n{topo}-{arch}, " - "CSIT build: {link}/{build}, " - "VPP version: {version}\n\n" - .format(topo=test_set.split('-')[-2], - arch=test_set.split('-')[-1], - link=alert["urls"][idx], - build=build, - version=version - )) - max_len_name = 0 - max_len_nics = 0 - max_len_framesizes = 0 - max_len_cores = 0 - for name, params in failed_tests.items(): - failed_tests[name]["nics"] = ",".join(sorted(params["nics"])) - failed_tests[name]["framesizes"] = \ - ",".join(sorted(params["framesizes"])) - failed_tests[name]["cores"] = ",".join(sorted(params["cores"])) - if len(name) > max_len_name: - max_len_name = len(name) - if len(failed_tests[name]["nics"]) > max_len_nics: - max_len_nics = len(failed_tests[name]["nics"]) - if len(failed_tests[name]["framesizes"]) > max_len_framesizes: - max_len_framesizes = len(failed_tests[name]["framesizes"]) - if len(failed_tests[name]["cores"]) > max_len_cores: - max_len_cores = len(failed_tests[name]["cores"]) + text += ( + f"\n\n{test_set.split('-')[-2]}-{test_set.split('-')[-1]}, " + f"{failed} tests failed, " + f"{passed} tests passed, CSIT build: " + f"{alert[u'urls'][idx]}/{build}, VPP version: {version}\n\n" + ) + + class MaxLens(): + """Class to store the max lengths of strings displayed in + failed tests list. + """ + def __init__(self, tst_name, nics, framesizes, cores): + """Initialisation. + + :param tst_name: Name of the test. + :param nics: NICs used in the test. + :param framesizes: Frame sizes used in the tests + :param cores: Cores used in th test. + """ + self.name = tst_name + self.nics = nics + self.frmsizes = framesizes + self.cores = cores + + max_len = MaxLens(0, 0, 0, 0) for name, params in failed_tests.items(): - text += "{name} {nics} {frames} {cores}\n".format( - name=name + " " * (max_len_name - len(name)), - nics=params["nics"] + - " " * (max_len_nics - len(params["nics"])), - frames=params["framesizes"] + " " * - (max_len_framesizes - len(params["framesizes"])), - cores=params["cores"] + - " " * (max_len_cores - len(params["cores"]))) + failed_tests[name][u"nics"] = u",".join(sorted(params[u"nics"])) + failed_tests[name][u"framesizes"] = \ + u",".join(sorted(params[u"framesizes"])) + failed_tests[name][u"cores"] = \ + u",".join(sorted(params[u"cores"])) + if len(name) > max_len.name: + max_len.name = len(name) + if len(failed_tests[name][u"nics"]) > max_len.nics: + max_len.nics = len(failed_tests[name][u"nics"]) + if len(failed_tests[name][u"framesizes"]) > max_len.frmsizes: + max_len.frmsizes = len(failed_tests[name][u"framesizes"]) + if len(failed_tests[name][u"cores"]) > max_len.cores: + max_len.cores = len(failed_tests[name][u"cores"]) + for name, params in failed_tests.items(): + text += ( + f"{name + u' ' * (max_len.name - len(name))} " + f"{params[u'nics']}" + f"{u' ' * (max_len.nics - len(params[u'nics']))} " + f"{params[u'framesizes']}" + f"{u' ' * (max_len.frmsizes-len(params[u'framesizes']))} " + f"{params[u'cores']}" + f"{u' ' * (max_len.cores - len(params[u'cores']))}\n" + ) + + gression_hdr = ( + f"\n\n{test_set.split(u'-')[-2]}-{test_set.split(u'-')[-1]}, " + f"CSIT build: {alert[u'urls'][idx]}/{build}, " + f"VPP version: {version}\n\n" + ) # Add list of regressions: - file_name = "{0}/cpta-regressions-{1}.txt".\ - format(config["output-dir"], alert["urls"][idx].split('/')[-1]) - try: - with open(file_name, 'r') as txt_file: - file_content = txt_file.read() - reg_file_name = "{dir}/trending-regressions.txt". \ - format(dir=config["output-dir"]) - with open(reg_file_name, 'a+') as reg_file: - reg_file.write(regression_hdr) - if file_content: - reg_file.write(file_content) - else: - reg_file.write("No regressions") - except IOError as err: - logging.warning(repr(err)) + self._list_gressions(alert, idx, gression_hdr, u"regressions") # Add list of progressions: - file_name = "{0}/cpta-progressions-{1}.txt".\ - format(config["output-dir"], alert["urls"][idx].split('/')[-1]) - try: - with open(file_name, 'r') as txt_file: - file_content = txt_file.read() - pro_file_name = "{dir}/trending-progressions.txt". \ - format(dir=config["output-dir"]) - with open(pro_file_name, 'a+') as pro_file: - pro_file.write(regression_hdr) - if file_content: - pro_file.write(file_content) - else: - pro_file.write("No progressions") - except IOError as err: - logging.warning(repr(err)) - - text += "\nFor detailed information visit: {url}\n".\ - format(url=alert["url-details"]) - file_name = "{0}/{1}".format(config["output-dir"], - config["output-file"]) - logging.info("Writing the file '{0}.txt' ...".format(file_name)) + self._list_gressions(alert, idx, gression_hdr, u"progressions") + + text += f"\nFor detailed information visit: {alert[u'url-details']}\n" + file_name = f"{self.configs[alert[u'way']][u'output-dir']}/" \ + f"{self.configs[alert[u'way']][u'output-file']}" + logging.info(f"Writing the file {file_name}.txt ...") try: - with open("{0}.txt".format(file_name), 'w') as txt_file: + with open(f"{file_name}.txt", u'w') as txt_file: txt_file.write(text) except IOError: - logging.error("Not possible to write the file '{0}.txt'.". - format(file_name)) + logging.error(f"Not possible to write the file {file_name}.txt.") diff --git a/resources/tools/presentation/generator_CPTA.py b/resources/tools/presentation/generator_cpta.py index eec401bd1f..3003557696 100644 --- a/resources/tools/presentation/generator_CPTA.py +++ b/resources/tools/presentation/generator_cpta.py @@ -16,28 +16,29 @@ import logging import csv -import prettytable -import plotly.offline as ploff -import plotly.graph_objs as plgo -import plotly.exceptions as plerr from collections import OrderedDict from datetime import datetime from copy import deepcopy -from utils import archive_input_data, execute_command, classify_anomalies +import prettytable +import plotly.offline as ploff +import plotly.graph_objs as plgo +import plotly.exceptions as plerr + +from pal_utils import archive_input_data, execute_command, classify_anomalies # Command to build the html format of the report -HTML_BUILDER = 'sphinx-build -v -c conf_cpta -a ' \ - '-b html -E ' \ - '-t html ' \ - '-D version="{date}" ' \ - '{working_dir} ' \ - '{build_dir}/' +HTML_BUILDER = u'sphinx-build -v -c conf_cpta -a ' \ + u'-b html -E ' \ + u'-t html ' \ + u'-D version="{date}" ' \ + u'{working_dir} ' \ + u'{build_dir}/' # .css file for the html format of the report -THEME_OVERRIDES = """/* override table width restrictions */ +THEME_OVERRIDES = u"""/* override table width restrictions */ .wy-nav-content { max-width: 1200px !important; } @@ -91,19 +92,20 @@ THEME_OVERRIDES = """/* override table width restrictions */ } """ -COLORS = ["SkyBlue", "Olive", "Purple", "Coral", "Indigo", "Pink", - "Chocolate", "Brown", "Magenta", "Cyan", "Orange", "Black", - "Violet", "Blue", "Yellow", "BurlyWood", "CadetBlue", "Crimson", - "DarkBlue", "DarkCyan", "DarkGreen", "Green", "GoldenRod", - "LightGreen", "LightSeaGreen", "LightSkyBlue", "Maroon", - "MediumSeaGreen", "SeaGreen", "LightSlateGrey", - "SkyBlue", "Olive", "Purple", "Coral", "Indigo", "Pink", - "Chocolate", "Brown", "Magenta", "Cyan", "Orange", "Black", - "Violet", "Blue", "Yellow", "BurlyWood", "CadetBlue", "Crimson", - "DarkBlue", "DarkCyan", "DarkGreen", "Green", "GoldenRod", - "LightGreen", "LightSeaGreen", "LightSkyBlue", "Maroon", - "MediumSeaGreen", "SeaGreen", "LightSlateGrey" - ] +COLORS = [ + u"SkyBlue", u"Olive", u"Purple", u"Coral", u"Indigo", u"Pink", + u"Chocolate", u"Brown", u"Magenta", u"Cyan", u"Orange", u"Black", + u"Violet", u"Blue", u"Yellow", u"BurlyWood", u"CadetBlue", u"Crimson", + u"DarkBlue", u"DarkCyan", u"DarkGreen", u"Green", u"GoldenRod", + u"LightGreen", u"LightSeaGreen", u"LightSkyBlue", u"Maroon", + u"MediumSeaGreen", u"SeaGreen", u"LightSlateGrey", + u"SkyBlue", u"Olive", u"Purple", u"Coral", u"Indigo", u"Pink", + u"Chocolate", u"Brown", u"Magenta", u"Cyan", u"Orange", u"Black", + u"Violet", u"Blue", u"Yellow", u"BurlyWood", u"CadetBlue", u"Crimson", + u"DarkBlue", u"DarkCyan", u"DarkGreen", u"Green", u"GoldenRod", + u"LightGreen", u"LightSeaGreen", u"LightSkyBlue", u"Maroon", + u"MediumSeaGreen", u"SeaGreen", u"LightSlateGrey" +] def generate_cpta(spec, data): @@ -116,35 +118,35 @@ def generate_cpta(spec, data): :type data: InputData """ - logging.info("Generating the Continuous Performance Trending and Analysis " - "...") + logging.info(u"Generating the Continuous Performance Trending and Analysis " + u"...") ret_code = _generate_all_charts(spec, data) cmd = HTML_BUILDER.format( - date=datetime.utcnow().strftime('%Y-%m-%d %H:%M UTC'), - working_dir=spec.environment["paths"]["DIR[WORKING,SRC]"], - build_dir=spec.environment["paths"]["DIR[BUILD,HTML]"]) + date=datetime.utcnow().strftime(u'%Y-%m-%d %H:%M UTC'), + working_dir=spec.environment[u'paths'][u'DIR[WORKING,SRC]'], + build_dir=spec.environment[u'paths'][u'DIR[BUILD,HTML]']) execute_command(cmd) - with open(spec.environment["paths"]["DIR[CSS_PATCH_FILE]"], "w") as \ + with open(spec.environment[u'paths'][u'DIR[CSS_PATCH_FILE]'], u'w') as \ css_file: css_file.write(THEME_OVERRIDES) - with open(spec.environment["paths"]["DIR[CSS_PATCH_FILE2]"], "w") as \ + with open(spec.environment[u'paths'][u'DIR[CSS_PATCH_FILE2]'], u'w') as \ css_file: css_file.write(THEME_OVERRIDES) - if spec.configuration.get("archive-inputs", True): + if spec.configuration.get(u"archive-inputs", True): archive_input_data(spec) - logging.info("Done.") + logging.info(u"Done.") return ret_code def _generate_trending_traces(in_data, job_name, build_info, - show_trend_line=True, name="", color=""): + show_trend_line=True, name=u"", color=u""): """Generate the trending traces: - samples, - outliers, regress, progress @@ -173,27 +175,27 @@ def _generate_trending_traces(in_data, job_name, build_info, xaxis = list() for idx in data_x: date = build_info[job_name][str(idx)][0] - hover_str = ("date: {date}<br>" - "value: {value:,}<br>" - "{sut}-ref: {build}<br>" - "csit-ref: mrr-{period}-build-{build_nr}<br>" - "testbed: {testbed}") - if "dpdk" in job_name: + hover_str = (u"date: {date}<br>" + u"value: {value:,}<br>" + u"{sut}-ref: {build}<br>" + u"csit-ref: mrr-{period}-build-{build_nr}<br>" + u"testbed: {testbed}") + if u"dpdk" in job_name: hover_text.append(hover_str.format( date=date, value=int(in_data[idx]), - sut="dpdk", - build=build_info[job_name][str(idx)][1].rsplit('~', 1)[0], - period="weekly", + sut=u"dpdk", + build=build_info[job_name][str(idx)][1].rsplit(u'~', 1)[0], + period=u"weekly", build_nr=idx, testbed=build_info[job_name][str(idx)][2])) - elif "vpp" in job_name: + elif u"vpp" in job_name: hover_text.append(hover_str.format( date=date, value=int(in_data[idx]), - sut="vpp", - build=build_info[job_name][str(idx)][1].rsplit('~', 1)[0], - period="daily", + sut=u"vpp", + build=build_info[job_name][str(idx)][1].rsplit(u'~', 1)[0], + period=u"daily", build_nr=idx, testbed=build_info[job_name][str(idx)][2])) @@ -210,14 +212,14 @@ def _generate_trending_traces(in_data, job_name, build_info, anomalies_colors = list() anomalies_avgs = list() anomaly_color = { - "regression": 0.0, - "normal": 0.5, - "progression": 1.0 + u"regression": 0.0, + u"normal": 0.5, + u"progression": 1.0 } if anomaly_classification: - for idx, (key, value) in enumerate(data_pd.iteritems()): + for idx, (key, value) in enumerate(data_pd.items()): if anomaly_classification[idx] in \ - ("outlier", "regression", "progression"): + (u"outlier", u"regression", u"progression"): anomalies[key] = value anomalies_colors.append( anomaly_color[anomaly_classification[idx]]) @@ -228,21 +230,21 @@ def _generate_trending_traces(in_data, job_name, build_info, trace_samples = plgo.Scatter( x=xaxis, - y=[y for y in data_y], # Was: y.avg - mode='markers', + y=data_y, + mode=u"markers", line={ - "width": 1 + u"width": 1 }, showlegend=True, legendgroup=name, - name="{name}".format(name=name), + name=f"{name}", marker={ - "size": 5, - "color": color, - "symbol": "circle", + u"size": 5, + u"color": color, + u"symbol": u"circle", }, text=hover_text, - hoverinfo="text" + hoverinfo=u"text" ) traces = [trace_samples, ] @@ -250,57 +252,59 @@ def _generate_trending_traces(in_data, job_name, build_info, trace_trend = plgo.Scatter( x=xaxis, y=avgs, - mode='lines', + mode=u"lines", line={ - "shape": "linear", - "width": 1, - "color": color, + u"shape": u"linear", + u"width": 1, + u"color": color, }, showlegend=False, legendgroup=name, - name='{name}'.format(name=name), - text=["trend: {0:,}".format(int(avg)) for avg in avgs], - hoverinfo="text+name" + name=f"{name}", + text=[f"trend: {int(avg):,}" for avg in avgs], + hoverinfo=u"text+name" ) traces.append(trace_trend) trace_anomalies = plgo.Scatter( - x=anomalies.keys(), + x=list(anomalies.keys()), y=anomalies_avgs, - mode='markers', - hoverinfo="none", + mode=u"markers", + hoverinfo=u"none", showlegend=False, legendgroup=name, - name="{name}-anomalies".format(name=name), + name=f"{name}-anomalies", marker={ - "size": 15, - "symbol": "circle-open", - "color": anomalies_colors, - "colorscale": [[0.00, "red"], - [0.33, "red"], - [0.33, "white"], - [0.66, "white"], - [0.66, "green"], - [1.00, "green"]], - "showscale": True, - "line": { - "width": 2 + u"size": 15, + u"symbol": u"circle-open", + u"color": anomalies_colors, + u"colorscale": [ + [0.00, u"red"], + [0.33, u"red"], + [0.33, u"white"], + [0.66, u"white"], + [0.66, u"green"], + [1.00, u"green"] + ], + u"showscale": True, + u"line": { + u"width": 2 }, - "colorbar": { - "y": 0.5, - "len": 0.8, - "title": "Circles Marking Data Classification", - "titleside": 'right', - "titlefont": { - "size": 14 + u"colorbar": { + u"y": 0.5, + u"len": 0.8, + u"title": u"Circles Marking Data Classification", + u"titleside": u"right", + u"titlefont": { + u"size": 14 }, - "tickmode": 'array', - "tickvals": [0.167, 0.500, 0.833], - "ticktext": ["Regression", "Normal", "Progression"], - "ticks": "", - "ticklen": 0, - "tickangle": -90, - "thickness": 10 + u"tickmode": u"array", + u"tickvals": [0.167, 0.500, 0.833], + u"ticktext": [u"Regression", u"Normal", u"Progression"], + u"ticks": u"", + u"ticklen": 0, + u"tickangle": -90, + u"thickness": 10 } } ) @@ -308,8 +312,8 @@ def _generate_trending_traces(in_data, job_name, build_info, if anomaly_classification: return traces, anomaly_classification[-1] - else: - return traces, None + + return traces, None def _generate_all_charts(spec, input_data): @@ -323,29 +327,40 @@ def _generate_all_charts(spec, input_data): def _generate_chart(graph): """Generates the chart. + + :param graph: The graph to be generated + :type graph: dict + :returns: Dictionary with the job name, csv table with results and + list of tests classification results. + :rtype: dict """ logs = list() - logs.append(("INFO", " Generating the chart '{0}' ...". - format(graph.get("title", "")))) + logs.append( + (u"INFO", f" Generating the chart {graph.get(u'title', u'')} ...") + ) - job_name = graph["data"].keys()[0] + job_name = list(graph[u"data"].keys())[0] csv_tbl = list() res = dict() # Transform the data - logs.append(("INFO", " Creating the data set for the {0} '{1}'.". - format(graph.get("type", ""), graph.get("title", "")))) + logs.append( + (u"INFO", + f" Creating the data set for the {graph.get(u'type', u'')} " + f"{graph.get(u'title', u'')}." + ) + ) data = input_data.filter_data(graph, continue_on_error=True) if data is None: - logging.error("No data.") - return + logging.error(u"No data.") + return dict() chart_data = dict() chart_tags = dict() - for job, job_data in data.iteritems(): + for job, job_data in data.items(): if job != job_name: continue for index, bld in job_data.items(): @@ -354,8 +369,8 @@ def _generate_all_charts(spec, input_data): chart_data[test_name] = OrderedDict() try: chart_data[test_name][int(index)] = \ - test["result"]["receive-rate"] - chart_tags[test_name] = test.get("tags", None) + test[u"result"][u"receive-rate"] + chart_tags[test_name] = test.get(u"tags", None) except (KeyError, TypeError): pass @@ -363,15 +378,15 @@ def _generate_all_charts(spec, input_data): for tst_name, tst_data in chart_data.items(): tst_lst = list() for bld in builds_dict[job_name]: - itm = tst_data.get(int(bld), '') + itm = tst_data.get(int(bld), u'') # CSIT-1180: Itm will be list, compute stats. tst_lst.append(str(itm)) - csv_tbl.append("{0},".format(tst_name) + ",".join(tst_lst) + '\n') + csv_tbl.append(f"{tst_name}," + u",".join(tst_lst) + u'\n') # Generate traces: traces = list() index = 0 - groups = graph.get("groups", None) + groups = graph.get(u"groups", None) visibility = list() if groups: @@ -380,52 +395,53 @@ def _generate_all_charts(spec, input_data): for tag in group: for tst_name, test_data in chart_data.items(): if not test_data: - logs.append(("WARNING", - "No data for the test '{0}'". - format(tst_name))) + logs.append( + (u"WARNING", f"No data for the test {tst_name}") + ) continue - if tag in chart_tags[tst_name]: - message = "index: {index}, test: {test}".format( - index=index, test=tst_name) - try: - trace, rslt = _generate_trending_traces( - test_data, - job_name=job_name, - build_info=build_info, - name='-'.join(tst_name.split('.')[-1]. - split('-')[2:-1]), - color=COLORS[index]) - except IndexError: - message = "Out of colors: {}".format(message) - logs.append(("ERROR", message)) - logging.error(message) - index += 1 - continue - traces.extend(trace) - visible.extend([True for _ in range(len(trace))]) - res[tst_name] = rslt + if tag not in chart_tags[tst_name]: + continue + message = f"index: {index}, test: {tst_name}" + try: + trace, rslt = _generate_trending_traces( + test_data, + job_name=job_name, + build_info=build_info, + name=u'-'.join(tst_name.split(u'.')[-1]. + split(u'-')[2:-1]), + color=COLORS[index]) + except IndexError: + logs.append( + (u"ERROR", f"Out of colors: {message}") + ) + logging.error(f"Out of colors: {message}") index += 1 - break + continue + traces.extend(trace) + visible.extend([True for _ in range(len(trace))]) + res[tst_name] = rslt + index += 1 + break visibility.append(visible) else: for tst_name, test_data in chart_data.items(): if not test_data: - logs.append(("WARNING", "No data for the test '{0}'". - format(tst_name))) + logs.append( + (u"WARNING", f"No data for the test {tst_name}") + ) continue - message = "index: {index}, test: {test}".format( - index=index, test=tst_name) + message = f"index: {index}, test: {tst_name}" try: trace, rslt = _generate_trending_traces( test_data, job_name=job_name, build_info=build_info, - name='-'.join(tst_name.split('.')[-1].split('-')[2:-1]), + name=u'-'.join( + tst_name.split(u'.')[-1].split(u'-')[2:-1]), color=COLORS[index]) except IndexError: - message = "Out of colors: {}".format(message) - logs.append(("ERROR", message)) - logging.error(message) + logs.append((u"ERROR", f"Out of colors: {message}")) + logging.error(f"Out of colors: {message}") index += 1 continue traces.extend(trace) @@ -435,101 +451,99 @@ def _generate_all_charts(spec, input_data): if traces: # Generate the chart: try: - layout = deepcopy(graph["layout"]) + layout = deepcopy(graph[u"layout"]) except KeyError as err: - logging.error("Finished with error: No layout defined") + logging.error(u"Finished with error: No layout defined") logging.error(repr(err)) - return + return dict() if groups: show = list() for i in range(len(visibility)): visible = list() - for r in range(len(visibility)): - for _ in range(len(visibility[r])): - visible.append(i == r) + for vis_idx, _ in enumerate(visibility): + for _ in range(len(visibility[vis_idx])): + visible.append(i == vis_idx) show.append(visible) buttons = list() buttons.append(dict( - label="All", - method="update", - args=[{"visible": [True for _ in range(len(show[0]))]}, ] + label=u"All", + method=u"update", + args=[{u"visible": [True for _ in range(len(show[0]))]}, ] )) for i in range(len(groups)): try: - label = graph["group-names"][i] + label = graph[u"group-names"][i] except (IndexError, KeyError): - label = "Group {num}".format(num=i + 1) + label = f"Group {i + 1}" buttons.append(dict( label=label, - method="update", - args=[{"visible": show[i]}, ] + method=u"update", + args=[{u"visible": show[i]}, ] )) - layout['updatemenus'] = list([ + layout[u"updatemenus"] = list([ dict( active=0, - type="dropdown", - direction="down", - xanchor="left", - yanchor="bottom", + type=u"dropdown", + direction=u"down", + xanchor=u"left", + yanchor=u"bottom", x=-0.12, y=1.0, buttons=buttons ) ]) - name_file = "{0}-{1}{2}".format(spec.cpta["output-file"], - graph["output-file-name"], - spec.cpta["output-file-type"]) + name_file = ( + f"{spec.cpta[u'output-file']}-{graph[u'output-file-name']}" + f"{spec.cpta[u'output-file-type']}") - logs.append(("INFO", " Writing the file '{0}' ...". - format(name_file))) + logs.append((u"INFO", f" Writing the file {name_file} ...")) plpl = plgo.Figure(data=traces, layout=layout) try: ploff.plot(plpl, show_link=False, auto_open=False, filename=name_file) except plerr.PlotlyEmptyDataError: - logs.append(("WARNING", "No data for the plot. Skipped.")) + logs.append((u"WARNING", u"No data for the plot. Skipped.")) for level, line in logs: - if level == "INFO": + if level == u"INFO": logging.info(line) - elif level == "ERROR": + elif level == u"ERROR": logging.error(line) - elif level == "DEBUG": + elif level == u"DEBUG": logging.debug(line) - elif level == "CRITICAL": + elif level == u"CRITICAL": logging.critical(line) - elif level == "WARNING": + elif level == u"WARNING": logging.warning(line) - return {"job_name": job_name, "csv_table": csv_tbl, "results": res} + return {u"job_name": job_name, u"csv_table": csv_tbl, u"results": res} builds_dict = dict() - for job in spec.input["builds"].keys(): + for job in spec.input[u"builds"].keys(): if builds_dict.get(job, None) is None: builds_dict[job] = list() - for build in spec.input["builds"][job]: - status = build["status"] - if status != "failed" and status != "not found" and \ - status != "removed": - builds_dict[job].append(str(build["build"])) + for build in spec.input[u"builds"][job]: + status = build[u"status"] + if status not in (u"failed", u"not found", u"removed"): + builds_dict[job].append(str(build[u"build"])) # Create "build ID": "date" dict: build_info = dict() - tb_tbl = spec.environment.get("testbeds", None) + tb_tbl = spec.environment.get(u"testbeds", None) for job_name, job_data in builds_dict.items(): if build_info.get(job_name, None) is None: build_info[job_name] = OrderedDict() for build in job_data: - testbed = "" - tb_ip = input_data.metadata(job_name, build).get("testbed", "") + testbed = u"" + tb_ip = input_data.metadata(job_name, build).get(u"testbed", u"") if tb_ip and tb_tbl: - testbed = tb_tbl.get(tb_ip, "") + testbed = tb_tbl.get(tb_ip, u"") build_info[job_name][build] = ( - input_data.metadata(job_name, build).get("generated", ""), - input_data.metadata(job_name, build).get("version", ""), + input_data.metadata(job_name, build).get(u"generated", u""), + input_data.metadata(job_name, build).get(u"version", u""), testbed ) @@ -537,36 +551,38 @@ def _generate_all_charts(spec, input_data): # Create the header: csv_tables = dict() - for job_name in builds_dict.keys(): + for job_name in builds_dict: if csv_tables.get(job_name, None) is None: csv_tables[job_name] = list() - header = "Build Number:," + ",".join(builds_dict[job_name]) + '\n' + header = u"Build Number:," + u",".join(builds_dict[job_name]) + u'\n' csv_tables[job_name].append(header) build_dates = [x[0] for x in build_info[job_name].values()] - header = "Build Date:," + ",".join(build_dates) + '\n' + header = u"Build Date:," + u",".join(build_dates) + u'\n' csv_tables[job_name].append(header) versions = [x[1] for x in build_info[job_name].values()] - header = "Version:," + ",".join(versions) + '\n' + header = u"Version:," + u",".join(versions) + u'\n' csv_tables[job_name].append(header) - for chart in spec.cpta["plots"]: + for chart in spec.cpta[u"plots"]: result = _generate_chart(chart) + if not result: + continue - csv_tables[result["job_name"]].extend(result["csv_table"]) + csv_tables[result[u"job_name"]].extend(result[u"csv_table"]) - if anomaly_classifications.get(result["job_name"], None) is None: - anomaly_classifications[result["job_name"]] = dict() - anomaly_classifications[result["job_name"]].update(result["results"]) + if anomaly_classifications.get(result[u"job_name"], None) is None: + anomaly_classifications[result[u"job_name"]] = dict() + anomaly_classifications[result[u"job_name"]].update(result[u"results"]) # Write the tables: for job_name, csv_table in csv_tables.items(): - file_name = spec.cpta["output-file"] + "-" + job_name + "-trending" - with open("{0}.csv".format(file_name), 'w') as file_handler: + file_name = spec.cpta[u"output-file"] + u"-" + job_name + u"-trending" + with open(f"{file_name}.csv", u"w") as file_handler: file_handler.writelines(csv_table) txt_table = None - with open("{0}.csv".format(file_name), 'rb') as csv_file: - csv_content = csv.reader(csv_file, delimiter=',', quotechar='"') + with open(f"{file_name}.csv", u"rt") as csv_file: + csv_content = csv.reader(csv_file, delimiter=u',', quotechar=u'"') line_nr = 0 for row in csv_content: if txt_table is None: @@ -580,37 +596,38 @@ def _generate_all_charts(spec, input_data): pass try: txt_table.add_row(row) + # PrettyTable raises Exception except Exception as err: - logging.warning("Error occurred while generating TXT " - "table:\n{0}".format(err)) + logging.warning( + f"Error occurred while generating TXT table:\n{err}" + ) line_nr += 1 - txt_table.align["Build Number:"] = "l" - with open("{0}.txt".format(file_name), "w") as txt_file: + txt_table.align[u"Build Number:"] = u"l" + with open(f"{file_name}.txt", u"w") as txt_file: txt_file.write(str(txt_table)) # Evaluate result: if anomaly_classifications: - result = "PASS" - for job_name, job_data in anomaly_classifications.iteritems(): - file_name = "{0}-regressions-{1}.txt".\ - format(spec.cpta["output-file"], job_name) - with open(file_name, 'w') as txt_file: - for test_name, classification in job_data.iteritems(): - if classification == "regression": - txt_file.write(test_name + '\n') - if classification == "regression" or \ - classification == "outlier": - result = "FAIL" - file_name = "{0}-progressions-{1}.txt".\ - format(spec.cpta["output-file"], job_name) - with open(file_name, 'w') as txt_file: - for test_name, classification in job_data.iteritems(): - if classification == "progression": - txt_file.write(test_name + '\n') + result = u"PASS" + for job_name, job_data in anomaly_classifications.items(): + file_name = \ + f"{spec.cpta[u'output-file']}-regressions-{job_name}.txt" + with open(file_name, u'w') as txt_file: + for test_name, classification in job_data.items(): + if classification == u"regression": + txt_file.write(test_name + u'\n') + if classification in (u"regression", u"outlier"): + result = u"FAIL" + file_name = \ + f"{spec.cpta[u'output-file']}-progressions-{job_name}.txt" + with open(file_name, u'w') as txt_file: + for test_name, classification in job_data.items(): + if classification == u"progression": + txt_file.write(test_name + u'\n') else: - result = "FAIL" + result = u"FAIL" - logging.info("Partial results: {0}".format(anomaly_classifications)) - logging.info("Result: {0}".format(result)) + logging.info(f"Partial results: {anomaly_classifications}") + logging.info(f"Result: {result}") return result diff --git a/resources/tools/presentation/generator_files.py b/resources/tools/presentation/generator_files.py index ee2fc1221c..00a78cc99a 100644 --- a/resources/tools/presentation/generator_files.py +++ b/resources/tools/presentation/generator_files.py @@ -17,17 +17,18 @@ import logging -from utils import get_files, get_rst_title_char +from pal_utils import get_files, get_rst_title_char -RST_INCLUDE_TABLE = ("\n.. only:: html\n\n" - " .. csv-table::\n" - " :header-rows: 1\n" - " :widths: auto\n" - " :align: center\n" - " :file: {file_html}\n" - "\n.. only:: latex\n\n" - "\n .. raw:: latex\n\n" - " \csvautolongtable{{{file_latex}}}\n\n") + +RST_INCLUDE_TABLE = (u"\n.. only:: html\n\n" + u" .. csv-table::\n" + u" :header-rows: 1\n" + u" :widths: auto\n" + u" :align: center\n" + u" :file: {file_html}\n" + u"\n.. only:: latex\n\n" + u"\n .. raw:: latex\n\n" + u" \\csvautolongtable{{{file_latex}}}\n\n") def generate_files(spec, data): @@ -39,14 +40,20 @@ def generate_files(spec, data): :type data: InputData """ - logging.info("Generating the files ...") + generator = { + u"file_test_results": file_test_results + } + + logging.info(u"Generating the files ...") for file_spec in spec.files: try: - eval(file_spec["algorithm"])(file_spec, data) - except NameError as err: - logging.error("Probably algorithm '{alg}' is not defined: {err}". - format(alg=file_spec["algorithm"], err=repr(err))) - logging.info("Done.") + generator[file_spec[u"algorithm"]](file_spec, data) + except (NameError, KeyError) as err: + logging.error( + f"Probably algorithm {file_spec[u'algorithm']} is not defined: " + f"{repr(err)}" + ) + logging.info(u"Done.") def _tests_in_suite(suite_name, tests): @@ -61,7 +68,7 @@ def _tests_in_suite(suite_name, tests): """ for key in tests.keys(): - if suite_name == tests[key]["parent"]: + if suite_name == tests[key][u"parent"]: return True return False @@ -77,59 +84,63 @@ def file_test_results(file_spec, input_data): :type input_data: InputData """ - file_name = "{0}{1}".format(file_spec["output-file"], - file_spec["output-file-ext"]) - rst_header = file_spec["file-header"] + file_name = f"{file_spec[u'output-file']}{file_spec[u'output-file-ext']}" + rst_header = file_spec[u"file-header"] - logging.info(" Generating the file {0} ...".format(file_name)) + logging.info(f" Generating the file {file_name} ...") - table_lst = get_files(file_spec["dir-tables"], ".csv", full_path=True) - if len(table_lst) == 0: - logging.error(" No tables to include in '{0}'. Skipping.". - format(file_spec["dir-tables"])) - return None + table_lst = get_files(file_spec[u"dir-tables"], u".csv", full_path=True) + if not table_lst: + logging.error( + f" No tables to include in {file_spec[u'dir-tables']}. Skipping." + ) + return - logging.info(" Writing file '{0}'".format(file_name)) + logging.info(f" Writing file {file_name}") - logging.info(" Creating the 'tests' data set for the {0} '{1}'.". - format(file_spec.get("type", ""), file_spec.get("title", ""))) + logging.info( + f" Creating the tests data set for the " + f"{file_spec.get(u'type', u'')} {file_spec.get(u'title', u'')}." + ) tests = input_data.filter_data(file_spec) tests = input_data.merge_data(tests) - logging.info(" Creating the 'suites' data set for the {0} '{1}'.". - format(file_spec.get("type", ""), file_spec.get("title", ""))) - file_spec["filter"] = "all" - suites = input_data.filter_data(file_spec, data_set="suites") + logging.info( + f" Creating the suites data set for the " + f"{file_spec.get(u'type', u'')} {file_spec.get(u'title', u'')}." + ) + file_spec[u"filter"] = u"all" + suites = input_data.filter_data(file_spec, data_set=u"suites") suites = input_data.merge_data(suites) suites.sort_index(inplace=True) - with open(file_name, "w") as file_handler: + with open(file_name, u"w") as file_handler: file_handler.write(rst_header) - for suite_longname, suite in suites.iteritems(): - if len(suite_longname.split(".")) <= file_spec["data-start-level"]: + for suite_longname, suite in suites.items(): + if len(suite_longname.split(u".")) <= \ + file_spec[u"data-start-level"]: continue - if not ("-ndrpdr" in suite["name"] or - "-mrr" in suite["name"] or - "-func" in suite["name"] or - "-device" in suite["name"]): - file_handler.write("\n{0}\n{1}\n".format( - suite["name"], get_rst_title_char( - suite["level"] - file_spec["data-start-level"] - 1) * - len(suite["name"]))) - - if _tests_in_suite(suite["name"], tests): - file_handler.write("\n{0}\n{1}\n".format( - suite["name"], get_rst_title_char( - suite["level"] - file_spec["data-start-level"] - 1) * - len(suite["name"]))) - file_handler.write("\n{0}\n".format( - suite["doc"].replace('|br|', '\n\n -'))) + title_line = \ + get_rst_title_char( + suite[u"level"] - file_spec[u"data-start-level"] - 1 + ) * len(suite[u"name"]) + if not (u"-ndrpdr" in suite[u"name"] or + u"-mrr" in suite[u"name"] or + u"-func" in suite[u"name"] or + u"-device" in suite[u"name"]): + file_handler.write(f"\n{suite[u'name']}\n{title_line}\n") + + if _tests_in_suite(suite[u"name"], tests): + file_handler.write(f"\n{suite[u'name']}\n{title_line}\n") + file_handler.write( + f"\n{suite[u'doc']}\n".replace(u'|br|', u'\n\n -') + ) for tbl_file in table_lst: - if suite["name"] in tbl_file: + if suite[u"name"] in tbl_file: file_handler.write( RST_INCLUDE_TABLE.format( file_latex=tbl_file, - file_html=tbl_file.split("/")[-1])) + file_html=tbl_file.split(u"/")[-1])) - logging.info(" Done.") + logging.info(u" Done.") diff --git a/resources/tools/presentation/generator_plots.py b/resources/tools/presentation/generator_plots.py index 3cbd35c430..dda5196008 100644 --- a/resources/tools/presentation/generator_plots.py +++ b/resources/tools/presentation/generator_plots.py @@ -17,23 +17,25 @@ import re import logging + +from collections import OrderedDict +from copy import deepcopy + import pandas as pd import plotly.offline as ploff import plotly.graph_objs as plgo from plotly.exceptions import PlotlyError -from collections import OrderedDict -from copy import deepcopy -from utils import mean, stdev +from pal_utils import mean, stdev -COLORS = ["SkyBlue", "Olive", "Purple", "Coral", "Indigo", "Pink", - "Chocolate", "Brown", "Magenta", "Cyan", "Orange", "Black", - "Violet", "Blue", "Yellow", "BurlyWood", "CadetBlue", "Crimson", - "DarkBlue", "DarkCyan", "DarkGreen", "Green", "GoldenRod", - "LightGreen", "LightSeaGreen", "LightSkyBlue", "Maroon", - "MediumSeaGreen", "SeaGreen", "LightSlateGrey"] +COLORS = [u"SkyBlue", u"Olive", u"Purple", u"Coral", u"Indigo", u"Pink", + u"Chocolate", u"Brown", u"Magenta", u"Cyan", u"Orange", u"Black", + u"Violet", u"Blue", u"Yellow", u"BurlyWood", u"CadetBlue", u"Crimson", + u"DarkBlue", u"DarkCyan", u"DarkGreen", u"Green", u"GoldenRod", + u"LightGreen", u"LightSeaGreen", u"LightSkyBlue", u"Maroon", + u"MediumSeaGreen", u"SeaGreen", u"LightSlateGrey"] REGEX_NIC = re.compile(r'\d*ge\dp\d\D*\d*-') @@ -47,22 +49,32 @@ def generate_plots(spec, data): :type data: InputData """ - logging.info("Generating the plots ...") + generator = { + u"plot_nf_reconf_box_name": plot_nf_reconf_box_name, + u"plot_perf_box_name": plot_perf_box_name, + u"plot_lat_err_bars_name": plot_lat_err_bars_name, + u"plot_tsa_name": plot_tsa_name, + u"plot_http_server_perf_box": plot_http_server_perf_box, + u"plot_nf_heatmap": plot_nf_heatmap + } + + logging.info(u"Generating the plots ...") for index, plot in enumerate(spec.plots): try: - logging.info(" Plot nr {0}: {1}".format(index + 1, - plot.get("title", ""))) - plot["limits"] = spec.configuration["limits"] - eval(plot["algorithm"])(plot, data) - logging.info(" Done.") + logging.info(f" Plot nr {index + 1}: {plot.get(u'title', u'')}") + plot[u"limits"] = spec.configuration[u"limits"] + generator[plot[u"algorithm"]](plot, data) + logging.info(u" Done.") except NameError as err: - logging.error("Probably algorithm '{alg}' is not defined: {err}". - format(alg=plot["algorithm"], err=repr(err))) - logging.info("Done.") + logging.error( + f"Probably algorithm {plot[u'algorithm']} is not defined: " + f"{repr(err)}" + ) + logging.info(u"Done.") -def plot_service_density_reconf_box_name(plot, input_data): - """Generate the plot(s) with algorithm: plot_service_density_reconf_box_name +def plot_nf_reconf_box_name(plot, input_data): + """Generate the plot(s) with algorithm: plot_nf_reconf_box_name specified in the specification file. :param plot: Plot to generate. @@ -72,13 +84,15 @@ def plot_service_density_reconf_box_name(plot, input_data): """ # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) + logging.info( + f" Creating the data set for the {plot.get(u'type', u'')} " + f"{plot.get(u'title', u'')}." + ) data = input_data.filter_tests_by_name( - plot, params=["result", "parent", "tags", "type"]) + plot, params=[u"result", u"parent", u"tags", u"type"] + ) if data is None: - logging.error("No data.") + logging.error(u"No data.") return # Prepare the data for the plot @@ -87,14 +101,14 @@ def plot_service_density_reconf_box_name(plot, input_data): for job in data: for build in job: for test in build: - if y_vals.get(test["parent"], None) is None: - y_vals[test["parent"]] = list() - loss[test["parent"]] = list() + if y_vals.get(test[u"parent"], None) is None: + y_vals[test[u"parent"]] = list() + loss[test[u"parent"]] = list() try: - y_vals[test["parent"]].append(test["result"]["time"]) - loss[test["parent"]].append(test["result"]["loss"]) + y_vals[test[u"parent"]].append(test[u"result"][u"time"]) + loss[test[u"parent"]].append(test[u"result"][u"loss"]) except (KeyError, TypeError): - y_vals[test["parent"]].append(None) + y_vals[test[u"parent"]].append(None) # Add None to the lists with missing data max_len = 0 @@ -103,54 +117,58 @@ def plot_service_density_reconf_box_name(plot, input_data): if len(val) > max_len: max_len = len(val) nr_of_samples.append(len(val)) - for key, val in y_vals.items(): + for val in y_vals.values(): if len(val) < max_len: val.extend([None for _ in range(max_len - len(val))]) # Add plot traces traces = list() - df = pd.DataFrame(y_vals) - df.head() - for i, col in enumerate(df.columns): - tst_name = re.sub(REGEX_NIC, "", - col.lower().replace('-ndrpdr', ''). - replace('2n1l-', '')) - tst_name = "-".join(tst_name.split("-")[3:-2]) - name = "{nr}. ({samples:02d} run{plural}, packets lost average: " \ - "{loss:.1f}) {name}".format( - nr=(i + 1), - samples=nr_of_samples[i], - plural='s' if nr_of_samples[i] > 1 else '', - name=tst_name, - loss=mean(loss[col])) - - traces.append(plgo.Box(x=[str(i + 1) + '.'] * len(df[col]), - y=[y if y else None for y in df[col]], - name=name, - hoverinfo="y+name")) + df_y = pd.DataFrame(y_vals) + df_y.head() + for i, col in enumerate(df_y.columns): + tst_name = re.sub(REGEX_NIC, u"", + col.lower().replace(u'-ndrpdr', u''). + replace(u'2n1l-', u'')) + + traces.append(plgo.Box( + x=[str(i + 1) + u'.'] * len(df_y[col]), + y=[y if y else None for y in df_y[col]], + name=( + f"{i + 1}. " + f"({nr_of_samples[i]:02d} " + f"run{u's' if nr_of_samples[i] > 1 else u''}, " + f"packets lost average: {mean(loss[col]):.1f}) " + f"{u'-'.join(tst_name.split(u'-')[3:-2])}" + ), + hoverinfo=u"y+name" + )) try: # Create plot - layout = deepcopy(plot["layout"]) - layout["title"] = "<b>Time Lost:</b> {0}".format(layout["title"]) - layout["yaxis"]["title"] = "<b>Implied Time Lost [s]</b>" - layout["legend"]["font"]["size"] = 14 - layout["yaxis"].pop("range") + layout = deepcopy(plot[u"layout"]) + layout[u"title"] = f"<b>Time Lost:</b> {layout[u'title']}" + layout[u"yaxis"][u"title"] = u"<b>Implied Time Lost [s]</b>" + layout[u"legend"][u"font"][u"size"] = 14 + layout[u"yaxis"].pop(u"range") plpl = plgo.Figure(data=traces, layout=layout) # Export Plot - file_type = plot.get("output-file-type", ".html") - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], file_type)) - ploff.plot(plpl, show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], file_type)) + file_type = plot.get(u"output-file-type", u".html") + logging.info(f" Writing file {plot[u'output-file']}{file_type}.") + ploff.plot( + plpl, + show_link=False, + auto_open=False, + filename=f"{plot[u'output-file']}{file_type}" + ) except PlotlyError as err: - logging.error(" Finished with error: {}". - format(repr(err).replace("\n", " "))) + logging.error( + f" Finished with error: {repr(err)}".replace(u"\n", u" ") + ) return -def plot_performance_box_name(plot, input_data): - """Generate the plot(s) with algorithm: plot_performance_box_name +def plot_perf_box_name(plot, input_data): + """Generate the plot(s) with algorithm: plot_perf_box_name specified in the specification file. :param plot: Plot to generate. @@ -160,13 +178,14 @@ def plot_performance_box_name(plot, input_data): """ # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) + logging.info( + f" Creating data set for the {plot.get(u'type', u'')} " + f"{plot.get(u'title', u'')}." + ) data = input_data.filter_tests_by_name( - plot, params=["throughput", "parent", "tags", "type"]) + plot, params=[u"throughput", u"parent", u"tags", u"type"]) if data is None: - logging.error("No data.") + logging.error(u"No data.") return # Prepare the data for the plot @@ -174,25 +193,24 @@ def plot_performance_box_name(plot, input_data): for job in data: for build in job: for test in build: - if y_vals.get(test["parent"], None) is None: - y_vals[test["parent"]] = list() + if y_vals.get(test[u"parent"], None) is None: + y_vals[test[u"parent"]] = list() try: - if test["type"] in ("NDRPDR", ): - if "-pdr" in plot_title.lower(): - y_vals[test["parent"]].\ - append(test["throughput"]["PDR"]["LOWER"]) - elif "-ndr" in plot_title.lower(): - y_vals[test["parent"]]. \ - append(test["throughput"]["NDR"]["LOWER"]) - else: - continue - elif test["type"] in ("SOAK", ): - y_vals[test["parent"]].\ - append(test["throughput"]["LOWER"]) + if (test[u"type"] in (u"NDRPDR", ) and + u"-pdr" in plot.get(u"title", u"").lower()): + y_vals[test[u"parent"]].\ + append(test[u"throughput"][u"PDR"][u"LOWER"]) + elif (test[u"type"] in (u"NDRPDR", ) and + u"-ndr" in plot.get(u"title", u"").lower()): + y_vals[test[u"parent"]]. \ + append(test[u"throughput"][u"NDR"][u"LOWER"]) + elif test[u"type"] in (u"SOAK", ): + y_vals[test[u"parent"]].\ + append(test[u"throughput"][u"LOWER"]) else: continue except (KeyError, TypeError): - y_vals[test["parent"]].append(None) + y_vals[test[u"parent"]].append(None) # Add None to the lists with missing data max_len = 0 @@ -201,62 +219,66 @@ def plot_performance_box_name(plot, input_data): if len(val) > max_len: max_len = len(val) nr_of_samples.append(len(val)) - for key, val in y_vals.items(): + for val in y_vals.values(): if len(val) < max_len: val.extend([None for _ in range(max_len - len(val))]) # Add plot traces traces = list() - df = pd.DataFrame(y_vals) - df.head() + df_y = pd.DataFrame(y_vals) + df_y.head() y_max = list() - for i, col in enumerate(df.columns): - tst_name = re.sub(REGEX_NIC, "", - col.lower().replace('-ndrpdr', ''). - replace('2n1l-', '')) - name = "{nr}. ({samples:02d} run{plural}) {name}".\ - format(nr=(i + 1), - samples=nr_of_samples[i], - plural='s' if nr_of_samples[i] > 1 else '', - name=tst_name) - - logging.debug(name) - traces.append(plgo.Box(x=[str(i + 1) + '.'] * len(df[col]), - y=[y / 1000000 if y else None for y in df[col]], - name=name, - hoverinfo="y+name")) + for i, col in enumerate(df_y.columns): + tst_name = re.sub(REGEX_NIC, u"", + col.lower().replace(u'-ndrpdr', u''). + replace(u'2n1l-', u'')) + traces.append( + plgo.Box( + x=[str(i + 1) + u'.'] * len(df_y[col]), + y=[y / 1000000 if y else None for y in df_y[col]], + name=( + f"{i + 1}. " + f"({nr_of_samples[i]:02d} " + f"run{u's' if nr_of_samples[i] > 1 else u''}) " + f"{tst_name}" + ), + hoverinfo=u"y+name" + ) + ) try: - val_max = max(df[col]) - except ValueError as err: + val_max = max(df_y[col]) + if val_max: + y_max.append(int(val_max / 1000000) + 2) + except (ValueError, TypeError) as err: logging.error(repr(err)) continue - if val_max: - y_max.append(int(val_max / 1000000) + 2) try: # Create plot - layout = deepcopy(plot["layout"]) - if layout.get("title", None): - layout["title"] = "<b>Throughput:</b> {0}". \ - format(layout["title"]) + layout = deepcopy(plot[u"layout"]) + if layout.get(u"title", None): + layout[u"title"] = f"<b>Throughput:</b> {layout[u'title']}" if y_max: - layout["yaxis"]["range"] = [0, max(y_max)] + layout[u"yaxis"][u"range"] = [0, max(y_max)] plpl = plgo.Figure(data=traces, layout=layout) # Export Plot - file_type = plot.get("output-file-type", ".html") - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], file_type)) - ploff.plot(plpl, show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], file_type)) + logging.info(f" Writing file {plot[u'output-file']}.html.") + ploff.plot( + plpl, + show_link=False, + auto_open=False, + filename=f"{plot[u'output-file']}.html" + ) except PlotlyError as err: - logging.error(" Finished with error: {}". - format(repr(err).replace("\n", " "))) + logging.error( + f" Finished with error: {repr(err)}".replace(u"\n", u" ") + ) return -def plot_latency_error_bars_name(plot, input_data): - """Generate the plot(s) with algorithm: plot_latency_error_bars_name +def plot_lat_err_bars_name(plot, input_data): + """Generate the plot(s) with algorithm: plot_lat_err_bars_name specified in the specification file. :param plot: Plot to generate. @@ -266,13 +288,14 @@ def plot_latency_error_bars_name(plot, input_data): """ # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) + plot_title = plot.get(u"title", u"") + logging.info( + f" Creating data set for the {plot.get(u'type', u'')} {plot_title}." + ) data = input_data.filter_tests_by_name( - plot, params=["latency", "parent", "tags", "type"]) + plot, params=[u"latency", u"parent", u"tags", u"type"]) if data is None: - logging.error("No data.") + logging.error(u"No data.") return # Prepare the data for the plot @@ -281,12 +304,11 @@ def plot_latency_error_bars_name(plot, input_data): for build in job: for test in build: try: - logging.debug("test['latency']: {0}\n". - format(test["latency"])) + logging.debug(f"test[u'latency']: {test[u'latency']}\n") except ValueError as err: logging.warning(repr(err)) - if y_tmp_vals.get(test["parent"], None) is None: - y_tmp_vals[test["parent"]] = [ + if y_tmp_vals.get(test[u"parent"], None) is None: + y_tmp_vals[test[u"parent"]] = [ list(), # direction1, min list(), # direction1, avg list(), # direction1, max @@ -295,31 +317,30 @@ def plot_latency_error_bars_name(plot, input_data): list() # direction2, max ] try: - if test["type"] in ("NDRPDR", ): - if "-pdr" in plot_title.lower(): - ttype = "PDR" - elif "-ndr" in plot_title.lower(): - ttype = "NDR" - else: - logging.warning("Invalid test type: {0}". - format(test["type"])) - continue - y_tmp_vals[test["parent"]][0].append( - test["latency"][ttype]["direction1"]["min"]) - y_tmp_vals[test["parent"]][1].append( - test["latency"][ttype]["direction1"]["avg"]) - y_tmp_vals[test["parent"]][2].append( - test["latency"][ttype]["direction1"]["max"]) - y_tmp_vals[test["parent"]][3].append( - test["latency"][ttype]["direction2"]["min"]) - y_tmp_vals[test["parent"]][4].append( - test["latency"][ttype]["direction2"]["avg"]) - y_tmp_vals[test["parent"]][5].append( - test["latency"][ttype]["direction2"]["max"]) + if test[u"type"] not in (u"NDRPDR", ): + logging.warning(f"Invalid test type: {test[u'type']}") + continue + if u"-pdr" in plot_title.lower(): + ttype = u"PDR" + elif u"-ndr" in plot_title.lower(): + ttype = u"NDR" else: - logging.warning("Invalid test type: {0}". - format(test["type"])) + logging.warning( + f"Invalid test type: {test[u'type']}" + ) continue + y_tmp_vals[test[u"parent"]][0].append( + test[u"latency"][ttype][u"direction1"][u"min"]) + y_tmp_vals[test[u"parent"]][1].append( + test[u"latency"][ttype][u"direction1"][u"avg"]) + y_tmp_vals[test[u"parent"]][2].append( + test[u"latency"][ttype][u"direction1"][u"max"]) + y_tmp_vals[test[u"parent"]][3].append( + test[u"latency"][ttype][u"direction2"][u"min"]) + y_tmp_vals[test[u"parent"]][4].append( + test[u"latency"][ttype][u"direction2"][u"avg"]) + y_tmp_vals[test[u"parent"]][5].append( + test[u"latency"][ttype][u"direction2"][u"max"]) except (KeyError, TypeError) as err: logging.warning(repr(err)) @@ -329,8 +350,8 @@ def plot_latency_error_bars_name(plot, input_data): y_maxs = list() nr_of_samples = list() for key, val in y_tmp_vals.items(): - name = re.sub(REGEX_NIC, "", key.replace('-ndrpdr', ''). - replace('2n1l-', '')) + name = re.sub(REGEX_NIC, u"", key.replace(u'-ndrpdr', u''). + replace(u'2n1l-', u'')) x_vals.append(name) # dir 1 y_vals.append(mean(val[1]) if val[1] else None) y_mins.append(mean(val[0]) if val[0] else None) @@ -345,22 +366,22 @@ def plot_latency_error_bars_name(plot, input_data): traces = list() annotations = list() - for idx in range(len(x_vals)): + for idx, _ in enumerate(x_vals): if not bool(int(idx % 2)): - direction = "West-East" + direction = u"West-East" else: - direction = "East-West" - hovertext = ("No. of Runs: {nr}<br>" - "Test: {test}<br>" - "Direction: {dir}<br>".format(test=x_vals[idx], - dir=direction, - nr=nr_of_samples[idx])) + direction = u"East-West" + hovertext = ( + f"No. of Runs: {nr_of_samples[idx]}<br>" + f"Test: {x_vals[idx]}<br>" + f"Direction: {direction}<br>" + ) if isinstance(y_maxs[idx], float): - hovertext += "Max: {max:.2f}uSec<br>".format(max=y_maxs[idx]) + hovertext += f"Max: {y_maxs[idx]:.2f}uSec<br>" if isinstance(y_vals[idx], float): - hovertext += "Mean: {avg:.2f}uSec<br>".format(avg=y_vals[idx]) + hovertext += f"Mean: {y_vals[idx]:.2f}uSec<br>" if isinstance(y_mins[idx], float): - hovertext += "Min: {min:.2f}uSec".format(min=y_mins[idx]) + hovertext += f"Min: {y_mins[idx]:.2f}uSec" if isinstance(y_maxs[idx], float) and isinstance(y_vals[idx], float): array = [y_maxs[idx] - y_vals[idx], ] @@ -376,9 +397,9 @@ def plot_latency_error_bars_name(plot, input_data): name=x_vals[idx], legendgroup=x_vals[idx], showlegend=bool(int(idx % 2)), - mode="markers", + mode=u"markers", error_y=dict( - type='data', + type=u"data", symmetric=False, array=array, arrayminus=arrayminus, @@ -389,48 +410,49 @@ def plot_latency_error_bars_name(plot, input_data): color=COLORS[int(idx / 2)], ), text=hovertext, - hoverinfo="text", + hoverinfo=u"text", )) annotations.append(dict( x=idx, y=0, - xref="x", - yref="y", - xanchor="center", - yanchor="top", - text="E-W" if bool(int(idx % 2)) else "W-E", + xref=u"x", + yref=u"y", + xanchor=u"center", + yanchor=u"top", + text=u"E-W" if bool(int(idx % 2)) else u"W-E", font=dict( size=16, ), - align="center", + align=u"center", showarrow=False )) try: # Create plot - file_type = plot.get("output-file-type", ".html") - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], file_type)) - layout = deepcopy(plot["layout"]) - if layout.get("title", None): - layout["title"] = "<b>Latency:</b> {0}".\ - format(layout["title"]) - layout["annotations"] = annotations + file_type = plot.get(u"output-file-type", u".html") + logging.info(f" Writing file {plot[u'output-file']}{file_type}.") + layout = deepcopy(plot[u"layout"]) + if layout.get(u"title", None): + layout[u"title"] = f"<b>Latency:</b> {layout[u'title']}" + layout[u"annotations"] = annotations plpl = plgo.Figure(data=traces, layout=layout) # Export Plot - ploff.plot(plpl, - show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], file_type)) + ploff.plot( + plpl, + show_link=False, auto_open=False, + filename=f"{plot[u'output-file']}{file_type}" + ) except PlotlyError as err: - logging.error(" Finished with error: {}". - format(str(err).replace("\n", " "))) + logging.error( + f" Finished with error: {repr(err)}".replace(u"\n", u" ") + ) return -def plot_throughput_speedup_analysis_name(plot, input_data): +def plot_tsa_name(plot, input_data): """Generate the plot(s) with algorithm: - plot_throughput_speedup_analysis_name + plot_tsa_name specified in the specification file. :param plot: Plot to generate. @@ -440,956 +462,51 @@ def plot_throughput_speedup_analysis_name(plot, input_data): """ # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) + plot_title = plot.get(u"title", u"") + logging.info( + f" Creating data set for the {plot.get(u'type', u'')} {plot_title}." + ) data = input_data.filter_tests_by_name( - plot, params=["throughput", "parent", "tags", "type"]) + plot, params=[u"throughput", u"parent", u"tags", u"type"]) if data is None: - logging.error("No data.") + logging.error(u"No data.") return y_vals = OrderedDict() for job in data: for build in job: for test in build: - if y_vals.get(test["parent"], None) is None: - y_vals[test["parent"]] = {"1": list(), - "2": list(), - "4": list()} - try: - if test["type"] in ("NDRPDR",): - if "-pdr" in plot_title.lower(): - ttype = "PDR" - elif "-ndr" in plot_title.lower(): - ttype = "NDR" - else: - continue - if "1C" in test["tags"]: - y_vals[test["parent"]]["1"]. \ - append(test["throughput"][ttype]["LOWER"]) - elif "2C" in test["tags"]: - y_vals[test["parent"]]["2"]. \ - append(test["throughput"][ttype]["LOWER"]) - elif "4C" in test["tags"]: - y_vals[test["parent"]]["4"]. \ - append(test["throughput"][ttype]["LOWER"]) - except (KeyError, TypeError): - pass - - if not y_vals: - logging.warning("No data for the plot '{}'". - format(plot.get("title", ""))) - return - - y_1c_max = dict() - for test_name, test_vals in y_vals.items(): - for key, test_val in test_vals.items(): - if test_val: - avg_val = sum(test_val) / len(test_val) - y_vals[test_name][key] = (avg_val, len(test_val)) - ideal = avg_val / (int(key) * 1000000.0) - if test_name not in y_1c_max or ideal > y_1c_max[test_name]: - y_1c_max[test_name] = ideal - - vals = OrderedDict() - y_max = list() - nic_limit = 0 - lnk_limit = 0 - pci_limit = plot["limits"]["pci"]["pci-g3-x8"] - for test_name, test_vals in y_vals.items(): - try: - if test_vals["1"][1]: - name = re.sub(REGEX_NIC, "", test_name.replace('-ndrpdr', ''). - replace('2n1l-', '')) - vals[name] = OrderedDict() - y_val_1 = test_vals["1"][0] / 1000000.0 - y_val_2 = test_vals["2"][0] / 1000000.0 if test_vals["2"][0] \ - else None - y_val_4 = test_vals["4"][0] / 1000000.0 if test_vals["4"][0] \ - else None - - vals[name]["val"] = [y_val_1, y_val_2, y_val_4] - vals[name]["rel"] = [1.0, None, None] - vals[name]["ideal"] = [y_1c_max[test_name], - y_1c_max[test_name] * 2, - y_1c_max[test_name] * 4] - vals[name]["diff"] = [(y_val_1 - y_1c_max[test_name]) * 100 / - y_val_1, None, None] - vals[name]["count"] = [test_vals["1"][1], - test_vals["2"][1], - test_vals["4"][1]] - - try: - val_max = max(vals[name]["val"]) - except ValueError as err: - logging.error(repr(err)) - continue - if val_max: - y_max.append(val_max) - - if y_val_2: - vals[name]["rel"][1] = round(y_val_2 / y_val_1, 2) - vals[name]["diff"][1] = \ - (y_val_2 - vals[name]["ideal"][1]) * 100 / y_val_2 - if y_val_4: - vals[name]["rel"][2] = round(y_val_4 / y_val_1, 2) - vals[name]["diff"][2] = \ - (y_val_4 - vals[name]["ideal"][2]) * 100 / y_val_4 - except IndexError as err: - logging.warning("No data for '{0}'".format(test_name)) - logging.warning(repr(err)) - - # Limits: - if "x520" in test_name: - limit = plot["limits"]["nic"]["x520"] - elif "x710" in test_name: - limit = plot["limits"]["nic"]["x710"] - elif "xxv710" in test_name: - limit = plot["limits"]["nic"]["xxv710"] - elif "xl710" in test_name: - limit = plot["limits"]["nic"]["xl710"] - elif "x553" in test_name: - limit = plot["limits"]["nic"]["x553"] - else: - limit = 0 - if limit > nic_limit: - nic_limit = limit - - mul = 2 if "ge2p" in test_name else 1 - if "10ge" in test_name: - limit = plot["limits"]["link"]["10ge"] * mul - elif "25ge" in test_name: - limit = plot["limits"]["link"]["25ge"] * mul - elif "40ge" in test_name: - limit = plot["limits"]["link"]["40ge"] * mul - elif "100ge" in test_name: - limit = plot["limits"]["link"]["100ge"] * mul - else: - limit = 0 - if limit > lnk_limit: - lnk_limit = limit - - traces = list() - annotations = list() - x_vals = [1, 2, 4] - - # Limits: - try: - threshold = 1.1 * max(y_max) # 10% - except ValueError as err: - logging.error(err) - return - nic_limit /= 1000000.0 - traces.append(plgo.Scatter( - x=x_vals, - y=[nic_limit, ] * len(x_vals), - name="NIC: {0:.2f}Mpps".format(nic_limit), - showlegend=False, - mode="lines", - line=dict( - dash="dot", - color=COLORS[-1], - width=1), - hoverinfo="none" - )) - annotations.append(dict( - x=1, - y=nic_limit, - xref="x", - yref="y", - xanchor="left", - yanchor="bottom", - text="NIC: {0:.2f}Mpps".format(nic_limit), - font=dict( - size=14, - color=COLORS[-1], - ), - align="left", - showarrow=False - )) - y_max.append(nic_limit) - - lnk_limit /= 1000000.0 - if lnk_limit < threshold: - traces.append(plgo.Scatter( - x=x_vals, - y=[lnk_limit, ] * len(x_vals), - name="Link: {0:.2f}Mpps".format(lnk_limit), - showlegend=False, - mode="lines", - line=dict( - dash="dot", - color=COLORS[-2], - width=1), - hoverinfo="none" - )) - annotations.append(dict( - x=1, - y=lnk_limit, - xref="x", - yref="y", - xanchor="left", - yanchor="bottom", - text="Link: {0:.2f}Mpps".format(lnk_limit), - font=dict( - size=14, - color=COLORS[-2], - ), - align="left", - showarrow=False - )) - y_max.append(lnk_limit) - - pci_limit /= 1000000.0 - if (pci_limit < threshold and - (pci_limit < lnk_limit * 0.95 or lnk_limit > lnk_limit * 1.05)): - traces.append(plgo.Scatter( - x=x_vals, - y=[pci_limit, ] * len(x_vals), - name="PCIe: {0:.2f}Mpps".format(pci_limit), - showlegend=False, - mode="lines", - line=dict( - dash="dot", - color=COLORS[-3], - width=1), - hoverinfo="none" - )) - annotations.append(dict( - x=1, - y=pci_limit, - xref="x", - yref="y", - xanchor="left", - yanchor="bottom", - text="PCIe: {0:.2f}Mpps".format(pci_limit), - font=dict( - size=14, - color=COLORS[-3], - ), - align="left", - showarrow=False - )) - y_max.append(pci_limit) - - # Perfect and measured: - cidx = 0 - for name, val in vals.iteritems(): - hovertext = list() - try: - for idx in range(len(val["val"])): - htext = "" - if isinstance(val["val"][idx], float): - htext += "No. of Runs: {1}<br>" \ - "Mean: {0:.2f}Mpps<br>".format(val["val"][idx], - val["count"][idx]) - if isinstance(val["diff"][idx], float): - htext += "Diff: {0:.0f}%<br>".format( - round(val["diff"][idx])) - if isinstance(val["rel"][idx], float): - htext += "Speedup: {0:.2f}".format(val["rel"][idx]) - hovertext.append(htext) - traces.append(plgo.Scatter(x=x_vals, - y=val["val"], - name=name, - legendgroup=name, - mode="lines+markers", - line=dict( - color=COLORS[cidx], - width=2), - marker=dict( - symbol="circle", - size=10 - ), - text=hovertext, - hoverinfo="text+name" - )) - traces.append(plgo.Scatter(x=x_vals, - y=val["ideal"], - name="{0} perfect".format(name), - legendgroup=name, - showlegend=False, - mode="lines", - line=dict( - color=COLORS[cidx], - width=2, - dash="dash"), - text=["Perfect: {0:.2f}Mpps".format(y) - for y in val["ideal"]], - hoverinfo="text" - )) - cidx += 1 - except (IndexError, ValueError, KeyError) as err: - logging.warning("No data for '{0}'".format(name)) - logging.warning(repr(err)) - - try: - # Create plot - file_type = plot.get("output-file-type", ".html") - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], file_type)) - layout = deepcopy(plot["layout"]) - if layout.get("title", None): - layout["title"] = "<b>Speedup Multi-core:</b> {0}". \ - format(layout["title"]) - layout["yaxis"]["range"] = [0, int(max(y_max) * 1.1)] - layout["annotations"].extend(annotations) - plpl = plgo.Figure(data=traces, layout=layout) - - # Export Plot - ploff.plot(plpl, - show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], file_type)) - except PlotlyError as err: - logging.error(" Finished with error: {}". - format(repr(err).replace("\n", " "))) - return - - -def plot_performance_box(plot, input_data): - """Generate the plot(s) with algorithm: plot_performance_box - specified in the specification file. - - TODO: Remove when not needed. - - :param plot: Plot to generate. - :param input_data: Data to process. - :type plot: pandas.Series - :type input_data: InputData - """ - - # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) - data = input_data.filter_data(plot) - if data is None: - logging.error("No data.") - return - - # Prepare the data for the plot - y_vals = dict() - y_tags = dict() - for job in data: - for build in job: - for test in build: - if y_vals.get(test["parent"], None) is None: - y_vals[test["parent"]] = list() - y_tags[test["parent"]] = test.get("tags", None) + if y_vals.get(test[u"parent"], None) is None: + y_vals[test[u"parent"]] = { + u"1": list(), + u"2": list(), + u"4": list() + } try: - if test["type"] in ("NDRPDR", ): - if "-pdr" in plot_title.lower(): - y_vals[test["parent"]].\ - append(test["throughput"]["PDR"]["LOWER"]) - elif "-ndr" in plot_title.lower(): - y_vals[test["parent"]]. \ - append(test["throughput"]["NDR"]["LOWER"]) - else: - continue - elif test["type"] in ("SOAK", ): - y_vals[test["parent"]].\ - append(test["throughput"]["LOWER"]) - else: - continue - except (KeyError, TypeError): - y_vals[test["parent"]].append(None) - - # Sort the tests - order = plot.get("sort", None) - if order and y_tags: - y_sorted = OrderedDict() - y_tags_l = {s: [t.lower() for t in ts] for s, ts in y_tags.items()} - for tag in order: - logging.debug(tag) - for suite, tags in y_tags_l.items(): - if "not " in tag: - tag = tag.split(" ")[-1] - if tag.lower() in tags: - continue - else: - if tag.lower() not in tags: + if test[u"type"] not in (u"NDRPDR",): continue - try: - y_sorted[suite] = y_vals.pop(suite) - y_tags_l.pop(suite) - logging.debug(suite) - except KeyError as err: - logging.error("Not found: {0}".format(repr(err))) - finally: - break - else: - y_sorted = y_vals - - # Add None to the lists with missing data - max_len = 0 - nr_of_samples = list() - for val in y_sorted.values(): - if len(val) > max_len: - max_len = len(val) - nr_of_samples.append(len(val)) - for key, val in y_sorted.items(): - if len(val) < max_len: - val.extend([None for _ in range(max_len - len(val))]) - - # Add plot traces - traces = list() - df = pd.DataFrame(y_sorted) - df.head() - y_max = list() - for i, col in enumerate(df.columns): - tst_name = re.sub(REGEX_NIC, "", - col.lower().replace('-ndrpdr', ''). - replace('2n1l-', '')) - name = "{nr}. ({samples:02d} run{plural}) {name}".\ - format(nr=(i + 1), - samples=nr_of_samples[i], - plural='s' if nr_of_samples[i] > 1 else '', - name=tst_name) - - logging.debug(name) - traces.append(plgo.Box(x=[str(i + 1) + '.'] * len(df[col]), - y=[y / 1000000 if y else None for y in df[col]], - name=name, - **plot["traces"])) - try: - val_max = max(df[col]) - except ValueError as err: - logging.error(repr(err)) - continue - if val_max: - y_max.append(int(val_max / 1000000) + 2) - - try: - # Create plot - layout = deepcopy(plot["layout"]) - if layout.get("title", None): - layout["title"] = "<b>Throughput:</b> {0}". \ - format(layout["title"]) - if y_max: - layout["yaxis"]["range"] = [0, max(y_max)] - plpl = plgo.Figure(data=traces, layout=layout) - - # Export Plot - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], plot["output-file-type"])) - ploff.plot(plpl, show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], - plot["output-file-type"])) - except PlotlyError as err: - logging.error(" Finished with error: {}". - format(repr(err).replace("\n", " "))) - return - - -def plot_soak_bars(plot, input_data): - """Generate the plot(s) with algorithm: plot_soak_bars - specified in the specification file. - - :param plot: Plot to generate. - :param input_data: Data to process. - :type plot: pandas.Series - :type input_data: InputData - """ - - # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) - data = input_data.filter_data(plot) - if data is None: - logging.error("No data.") - return - - # Prepare the data for the plot - y_vals = dict() - y_tags = dict() - for job in data: - for build in job: - for test in build: - if y_vals.get(test["parent"], None) is None: - y_tags[test["parent"]] = test.get("tags", None) - try: - if test["type"] in ("SOAK", ): - y_vals[test["parent"]] = test["throughput"] - else: - continue - except (KeyError, TypeError): - y_vals[test["parent"]] = dict() - - # Sort the tests - order = plot.get("sort", None) - if order and y_tags: - y_sorted = OrderedDict() - y_tags_l = {s: [t.lower() for t in ts] for s, ts in y_tags.items()} - for tag in order: - logging.debug(tag) - for suite, tags in y_tags_l.items(): - if "not " in tag: - tag = tag.split(" ")[-1] - if tag.lower() in tags: - continue - else: - if tag.lower() not in tags: - continue - try: - y_sorted[suite] = y_vals.pop(suite) - y_tags_l.pop(suite) - logging.debug(suite) - except KeyError as err: - logging.error("Not found: {0}".format(repr(err))) - finally: - break - else: - y_sorted = y_vals - - idx = 0 - y_max = 0 - traces = list() - for test_name, test_data in y_sorted.items(): - idx += 1 - name = "{nr}. {name}".\ - format(nr=idx, name=test_name.lower().replace('-soak', '')) - if len(name) > 50: - name_lst = name.split('-') - name = "" - split_name = True - for segment in name_lst: - if (len(name) + len(segment) + 1) > 50 and split_name: - name += "<br> " - split_name = False - name += segment + '-' - name = name[:-1] - - y_val = test_data.get("LOWER", None) - if y_val: - y_val /= 1000000 - if y_val > y_max: - y_max = y_val - - time = "No Information" - result = "No Information" - hovertext = ("{name}<br>" - "Packet Throughput: {val:.2f}Mpps<br>" - "Final Duration: {time}<br>" - "Result: {result}".format(name=name, - val=y_val, - time=time, - result=result)) - traces.append(plgo.Bar(x=[str(idx) + '.', ], - y=[y_val, ], - name=name, - text=hovertext, - hoverinfo="text")) - try: - # Create plot - layout = deepcopy(plot["layout"]) - if layout.get("title", None): - layout["title"] = "<b>Packet Throughput:</b> {0}". \ - format(layout["title"]) - if y_max: - layout["yaxis"]["range"] = [0, y_max + 1] - plpl = plgo.Figure(data=traces, layout=layout) - # Export Plot - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], plot["output-file-type"])) - ploff.plot(plpl, show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], - plot["output-file-type"])) - except PlotlyError as err: - logging.error(" Finished with error: {}". - format(repr(err).replace("\n", " "))) - return - - -def plot_soak_boxes(plot, input_data): - """Generate the plot(s) with algorithm: plot_soak_boxes - specified in the specification file. - - :param plot: Plot to generate. - :param input_data: Data to process. - :type plot: pandas.Series - :type input_data: InputData - """ - - # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) - data = input_data.filter_data(plot) - if data is None: - logging.error("No data.") - return - - # Prepare the data for the plot - y_vals = dict() - y_tags = dict() - for job in data: - for build in job: - for test in build: - if y_vals.get(test["parent"], None) is None: - y_tags[test["parent"]] = test.get("tags", None) - try: - if test["type"] in ("SOAK", ): - y_vals[test["parent"]] = test["throughput"] - else: - continue - except (KeyError, TypeError): - y_vals[test["parent"]] = dict() - - # Sort the tests - order = plot.get("sort", None) - if order and y_tags: - y_sorted = OrderedDict() - y_tags_l = {s: [t.lower() for t in ts] for s, ts in y_tags.items()} - for tag in order: - logging.debug(tag) - for suite, tags in y_tags_l.items(): - if "not " in tag: - tag = tag.split(" ")[-1] - if tag.lower() in tags: - continue - else: - if tag.lower() not in tags: - continue - try: - y_sorted[suite] = y_vals.pop(suite) - y_tags_l.pop(suite) - logging.debug(suite) - except KeyError as err: - logging.error("Not found: {0}".format(repr(err))) - finally: - break - else: - y_sorted = y_vals - - idx = 0 - y_max = 0 - traces = list() - for test_name, test_data in y_sorted.items(): - idx += 1 - name = "{nr}. {name}".\ - format(nr=idx, name=test_name.lower().replace('-soak', ''). - replace('2n1l-', '')) - if len(name) > 55: - name_lst = name.split('-') - name = "" - split_name = True - for segment in name_lst: - if (len(name) + len(segment) + 1) > 55 and split_name: - name += "<br> " - split_name = False - name += segment + '-' - name = name[:-1] - - y_val = test_data.get("UPPER", None) - if y_val: - y_val /= 1000000 - if y_val > y_max: - y_max = y_val - - y_base = test_data.get("LOWER", None) - if y_base: - y_base /= 1000000 - - hovertext = ("Upper bound: {upper:.2f}<br>" - "Lower bound: {lower:.2f}".format(upper=y_val, - lower=y_base)) - traces.append(plgo.Bar(x=[str(idx) + '.', ], - # +0.05 to see the value in case lower == upper - y=[y_val - y_base + 0.05, ], - base=y_base, - name=name, - text=hovertext, - hoverinfo="text")) - try: - # Create plot - layout = deepcopy(plot["layout"]) - if layout.get("title", None): - layout["title"] = "<b>Throughput:</b> {0}". \ - format(layout["title"]) - if y_max: - layout["yaxis"]["range"] = [0, y_max + 1] - plpl = plgo.Figure(data=traces, layout=layout) - # Export Plot - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], plot["output-file-type"])) - ploff.plot(plpl, show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], - plot["output-file-type"])) - except PlotlyError as err: - logging.error(" Finished with error: {}". - format(repr(err).replace("\n", " "))) - return - - -def plot_latency_error_bars(plot, input_data): - """Generate the plot(s) with algorithm: plot_latency_error_bars - specified in the specification file. - - TODO: Remove when not needed. - - :param plot: Plot to generate. - :param input_data: Data to process. - :type plot: pandas.Series - :type input_data: InputData - """ - - # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) - data = input_data.filter_data(plot) - if data is None: - logging.error("No data.") - return - # Prepare the data for the plot - y_tmp_vals = dict() - y_tags = dict() - for job in data: - for build in job: - for test in build: - try: - logging.debug("test['latency']: {0}\n". - format(test["latency"])) - except ValueError as err: - logging.warning(repr(err)) - if y_tmp_vals.get(test["parent"], None) is None: - y_tmp_vals[test["parent"]] = [ - list(), # direction1, min - list(), # direction1, avg - list(), # direction1, max - list(), # direction2, min - list(), # direction2, avg - list() # direction2, max - ] - y_tags[test["parent"]] = test.get("tags", None) - try: - if test["type"] in ("NDRPDR", ): - if "-pdr" in plot_title.lower(): - ttype = "PDR" - elif "-ndr" in plot_title.lower(): - ttype = "NDR" - else: - logging.warning("Invalid test type: {0}". - format(test["type"])) - continue - y_tmp_vals[test["parent"]][0].append( - test["latency"][ttype]["direction1"]["min"]) - y_tmp_vals[test["parent"]][1].append( - test["latency"][ttype]["direction1"]["avg"]) - y_tmp_vals[test["parent"]][2].append( - test["latency"][ttype]["direction1"]["max"]) - y_tmp_vals[test["parent"]][3].append( - test["latency"][ttype]["direction2"]["min"]) - y_tmp_vals[test["parent"]][4].append( - test["latency"][ttype]["direction2"]["avg"]) - y_tmp_vals[test["parent"]][5].append( - test["latency"][ttype]["direction2"]["max"]) + if u"-pdr" in plot_title.lower(): + ttype = u"PDR" + elif u"-ndr" in plot_title.lower(): + ttype = u"NDR" else: - logging.warning("Invalid test type: {0}". - format(test["type"])) - continue - except (KeyError, TypeError) as err: - logging.warning(repr(err)) - logging.debug("y_tmp_vals: {0}\n".format(y_tmp_vals)) - - # Sort the tests - order = plot.get("sort", None) - if order and y_tags: - y_sorted = OrderedDict() - y_tags_l = {s: [t.lower() for t in ts] for s, ts in y_tags.items()} - for tag in order: - logging.debug(tag) - for suite, tags in y_tags_l.items(): - if "not " in tag: - tag = tag.split(" ")[-1] - if tag.lower() in tags: continue - else: - if tag.lower() not in tags: - continue - try: - y_sorted[suite] = y_tmp_vals.pop(suite) - y_tags_l.pop(suite) - logging.debug(suite) - except KeyError as err: - logging.error("Not found: {0}".format(repr(err))) - finally: - break - else: - y_sorted = y_tmp_vals - - logging.debug("y_sorted: {0}\n".format(y_sorted)) - x_vals = list() - y_vals = list() - y_mins = list() - y_maxs = list() - nr_of_samples = list() - for key, val in y_sorted.items(): - name = re.sub(REGEX_NIC, "", key.replace('-ndrpdr', ''). - replace('2n1l-', '')) - x_vals.append(name) # dir 1 - y_vals.append(mean(val[1]) if val[1] else None) - y_mins.append(mean(val[0]) if val[0] else None) - y_maxs.append(mean(val[2]) if val[2] else None) - nr_of_samples.append(len(val[1]) if val[1] else 0) - x_vals.append(name) # dir 2 - y_vals.append(mean(val[4]) if val[4] else None) - y_mins.append(mean(val[3]) if val[3] else None) - y_maxs.append(mean(val[5]) if val[5] else None) - nr_of_samples.append(len(val[3]) if val[3] else 0) - logging.debug("x_vals :{0}\n".format(x_vals)) - logging.debug("y_vals :{0}\n".format(y_vals)) - logging.debug("y_mins :{0}\n".format(y_mins)) - logging.debug("y_maxs :{0}\n".format(y_maxs)) - logging.debug("nr_of_samples :{0}\n".format(nr_of_samples)) - traces = list() - annotations = list() - - for idx in range(len(x_vals)): - if not bool(int(idx % 2)): - direction = "West-East" - else: - direction = "East-West" - hovertext = ("No. of Runs: {nr}<br>" - "Test: {test}<br>" - "Direction: {dir}<br>".format(test=x_vals[idx], - dir=direction, - nr=nr_of_samples[idx])) - if isinstance(y_maxs[idx], float): - hovertext += "Max: {max:.2f}uSec<br>".format(max=y_maxs[idx]) - if isinstance(y_vals[idx], float): - hovertext += "Mean: {avg:.2f}uSec<br>".format(avg=y_vals[idx]) - if isinstance(y_mins[idx], float): - hovertext += "Min: {min:.2f}uSec".format(min=y_mins[idx]) - - if isinstance(y_maxs[idx], float) and isinstance(y_vals[idx], float): - array = [y_maxs[idx] - y_vals[idx], ] - else: - array = [None, ] - if isinstance(y_mins[idx], float) and isinstance(y_vals[idx], float): - arrayminus = [y_vals[idx] - y_mins[idx], ] - else: - arrayminus = [None, ] - logging.debug("y_vals[{1}] :{0}\n".format(y_vals[idx], idx)) - logging.debug("array :{0}\n".format(array)) - logging.debug("arrayminus :{0}\n".format(arrayminus)) - traces.append(plgo.Scatter( - x=[idx, ], - y=[y_vals[idx], ], - name=x_vals[idx], - legendgroup=x_vals[idx], - showlegend=bool(int(idx % 2)), - mode="markers", - error_y=dict( - type='data', - symmetric=False, - array=array, - arrayminus=arrayminus, - color=COLORS[int(idx / 2)] - ), - marker=dict( - size=10, - color=COLORS[int(idx / 2)], - ), - text=hovertext, - hoverinfo="text", - )) - annotations.append(dict( - x=idx, - y=0, - xref="x", - yref="y", - xanchor="center", - yanchor="top", - text="E-W" if bool(int(idx % 2)) else "W-E", - font=dict( - size=16, - ), - align="center", - showarrow=False - )) - - try: - # Create plot - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], plot["output-file-type"])) - layout = deepcopy(plot["layout"]) - if layout.get("title", None): - layout["title"] = "<b>Latency:</b> {0}".\ - format(layout["title"]) - layout["annotations"] = annotations - plpl = plgo.Figure(data=traces, layout=layout) - - # Export Plot - ploff.plot(plpl, - show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], - plot["output-file-type"])) - except PlotlyError as err: - logging.error(" Finished with error: {}". - format(str(err).replace("\n", " "))) - return - - -def plot_throughput_speedup_analysis(plot, input_data): - """Generate the plot(s) with algorithm: - plot_throughput_speedup_analysis - specified in the specification file. - - TODO: Remove when not needed. - - :param plot: Plot to generate. - :param input_data: Data to process. - :type plot: pandas.Series - :type input_data: InputData - """ - - # Transform the data - plot_title = plot.get("title", "") - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot_title)) - data = input_data.filter_data(plot) - if data is None: - logging.error("No data.") - return - - y_vals = dict() - y_tags = dict() - for job in data: - for build in job: - for test in build: - if y_vals.get(test["parent"], None) is None: - y_vals[test["parent"]] = {"1": list(), - "2": list(), - "4": list()} - y_tags[test["parent"]] = test.get("tags", None) - try: - if test["type"] in ("NDRPDR",): - if "-pdr" in plot_title.lower(): - ttype = "PDR" - elif "-ndr" in plot_title.lower(): - ttype = "NDR" - else: - continue - if "1C" in test["tags"]: - y_vals[test["parent"]]["1"]. \ - append(test["throughput"][ttype]["LOWER"]) - elif "2C" in test["tags"]: - y_vals[test["parent"]]["2"]. \ - append(test["throughput"][ttype]["LOWER"]) - elif "4C" in test["tags"]: - y_vals[test["parent"]]["4"]. \ - append(test["throughput"][ttype]["LOWER"]) + if u"1C" in test[u"tags"]: + y_vals[test[u"parent"]][u"1"]. \ + append(test[u"throughput"][ttype][u"LOWER"]) + elif u"2C" in test[u"tags"]: + y_vals[test[u"parent"]][u"2"]. \ + append(test[u"throughput"][ttype][u"LOWER"]) + elif u"4C" in test[u"tags"]: + y_vals[test[u"parent"]][u"4"]. \ + append(test[u"throughput"][ttype][u"LOWER"]) except (KeyError, TypeError): pass if not y_vals: - logging.warning("No data for the plot '{}'". - format(plot.get("title", ""))) + logging.warning(f"No data for the plot {plot.get(u'title', u'')}") return y_1c_max = dict() @@ -1397,112 +514,97 @@ def plot_throughput_speedup_analysis(plot, input_data): for key, test_val in test_vals.items(): if test_val: avg_val = sum(test_val) / len(test_val) - y_vals[test_name][key] = (avg_val, len(test_val)) + y_vals[test_name][key] = [avg_val, len(test_val)] ideal = avg_val / (int(key) * 1000000.0) if test_name not in y_1c_max or ideal > y_1c_max[test_name]: y_1c_max[test_name] = ideal - vals = dict() + vals = OrderedDict() y_max = list() nic_limit = 0 lnk_limit = 0 - pci_limit = plot["limits"]["pci"]["pci-g3-x8"] + pci_limit = plot[u"limits"][u"pci"][u"pci-g3-x8"] for test_name, test_vals in y_vals.items(): try: - if test_vals["1"][1]: - name = re.sub(REGEX_NIC, "", test_name.replace('-ndrpdr', ''). - replace('2n1l-', '')) - vals[name] = dict() - y_val_1 = test_vals["1"][0] / 1000000.0 - y_val_2 = test_vals["2"][0] / 1000000.0 if test_vals["2"][0] \ + if test_vals[u"1"][1]: + name = re.sub( + REGEX_NIC, + u"", + test_name.replace(u'-ndrpdr', u'').replace(u'2n1l-', u'') + ) + vals[name] = OrderedDict() + y_val_1 = test_vals[u"1"][0] / 1000000.0 + y_val_2 = test_vals[u"2"][0] / 1000000.0 if test_vals[u"2"][0] \ else None - y_val_4 = test_vals["4"][0] / 1000000.0 if test_vals["4"][0] \ + y_val_4 = test_vals[u"4"][0] / 1000000.0 if test_vals[u"4"][0] \ else None - vals[name]["val"] = [y_val_1, y_val_2, y_val_4] - vals[name]["rel"] = [1.0, None, None] - vals[name]["ideal"] = [y_1c_max[test_name], - y_1c_max[test_name] * 2, - y_1c_max[test_name] * 4] - vals[name]["diff"] = [(y_val_1 - y_1c_max[test_name]) * 100 / - y_val_1, None, None] - vals[name]["count"] = [test_vals["1"][1], - test_vals["2"][1], - test_vals["4"][1]] + vals[name][u"val"] = [y_val_1, y_val_2, y_val_4] + vals[name][u"rel"] = [1.0, None, None] + vals[name][u"ideal"] = [ + y_1c_max[test_name], + y_1c_max[test_name] * 2, + y_1c_max[test_name] * 4 + ] + vals[name][u"diff"] = [ + (y_val_1 - y_1c_max[test_name]) * 100 / y_val_1, None, None + ] + vals[name][u"count"] = [ + test_vals[u"1"][1], + test_vals[u"2"][1], + test_vals[u"4"][1] + ] try: - # val_max = max(max(vals[name]["val"], vals[name]["ideal"])) - val_max = max(vals[name]["val"]) + val_max = max(vals[name][u"val"]) except ValueError as err: - logging.error(err) + logging.error(repr(err)) continue if val_max: - # y_max.append(int((val_max / 10) + 1) * 10) y_max.append(val_max) if y_val_2: - vals[name]["rel"][1] = round(y_val_2 / y_val_1, 2) - vals[name]["diff"][1] = \ - (y_val_2 - vals[name]["ideal"][1]) * 100 / y_val_2 + vals[name][u"rel"][1] = round(y_val_2 / y_val_1, 2) + vals[name][u"diff"][1] = \ + (y_val_2 - vals[name][u"ideal"][1]) * 100 / y_val_2 if y_val_4: - vals[name]["rel"][2] = round(y_val_4 / y_val_1, 2) - vals[name]["diff"][2] = \ - (y_val_4 - vals[name]["ideal"][2]) * 100 / y_val_4 + vals[name][u"rel"][2] = round(y_val_4 / y_val_1, 2) + vals[name][u"diff"][2] = \ + (y_val_4 - vals[name][u"ideal"][2]) * 100 / y_val_4 except IndexError as err: - logging.warning("No data for '{0}'".format(test_name)) + logging.warning(f"No data for {test_name}") logging.warning(repr(err)) # Limits: - if "x520" in test_name: - limit = plot["limits"]["nic"]["x520"] - elif "x710" in test_name: - limit = plot["limits"]["nic"]["x710"] - elif "xxv710" in test_name: - limit = plot["limits"]["nic"]["xxv710"] - elif "xl710" in test_name: - limit = plot["limits"]["nic"]["xl710"] - elif "x553" in test_name: - limit = plot["limits"]["nic"]["x553"] + if u"x520" in test_name: + limit = plot[u"limits"][u"nic"][u"x520"] + elif u"x710" in test_name: + limit = plot[u"limits"][u"nic"][u"x710"] + elif u"xxv710" in test_name: + limit = plot[u"limits"][u"nic"][u"xxv710"] + elif u"xl710" in test_name: + limit = plot[u"limits"][u"nic"][u"xl710"] + elif u"x553" in test_name: + limit = plot[u"limits"][u"nic"][u"x553"] else: limit = 0 if limit > nic_limit: nic_limit = limit - mul = 2 if "ge2p" in test_name else 1 - if "10ge" in test_name: - limit = plot["limits"]["link"]["10ge"] * mul - elif "25ge" in test_name: - limit = plot["limits"]["link"]["25ge"] * mul - elif "40ge" in test_name: - limit = plot["limits"]["link"]["40ge"] * mul - elif "100ge" in test_name: - limit = plot["limits"]["link"]["100ge"] * mul + mul = 2 if u"ge2p" in test_name else 1 + if u"10ge" in test_name: + limit = plot[u"limits"][u"link"][u"10ge"] * mul + elif u"25ge" in test_name: + limit = plot[u"limits"][u"link"][u"25ge"] * mul + elif u"40ge" in test_name: + limit = plot[u"limits"][u"link"][u"40ge"] * mul + elif u"100ge" in test_name: + limit = plot[u"limits"][u"link"][u"100ge"] * mul else: limit = 0 if limit > lnk_limit: lnk_limit = limit - # Sort the tests - order = plot.get("sort", None) - if order and y_tags: - y_sorted = OrderedDict() - y_tags_l = {s: [t.lower() for t in ts] for s, ts in y_tags.items()} - for tag in order: - for test, tags in y_tags_l.items(): - if tag.lower() in tags: - name = re.sub(REGEX_NIC, "", - test.replace('-ndrpdr', ''). - replace('2n1l-', '')) - try: - y_sorted[name] = vals.pop(name) - y_tags_l.pop(test) - except KeyError as err: - logging.error("Not found: {0}".format(err)) - finally: - break - else: - y_sorted = vals - traces = list() annotations = list() x_vals = [1, 2, 4] @@ -1514,35 +616,33 @@ def plot_throughput_speedup_analysis(plot, input_data): logging.error(err) return nic_limit /= 1000000.0 - # if nic_limit < threshold: traces.append(plgo.Scatter( x=x_vals, y=[nic_limit, ] * len(x_vals), - name="NIC: {0:.2f}Mpps".format(nic_limit), + name=f"NIC: {nic_limit:.2f}Mpps", showlegend=False, - mode="lines", + mode=u"lines", line=dict( - dash="dot", + dash=u"dot", color=COLORS[-1], width=1), - hoverinfo="none" + hoverinfo=u"none" )) annotations.append(dict( x=1, y=nic_limit, - xref="x", - yref="y", - xanchor="left", - yanchor="bottom", - text="NIC: {0:.2f}Mpps".format(nic_limit), + xref=u"x", + yref=u"y", + xanchor=u"left", + yanchor=u"bottom", + text=f"NIC: {nic_limit:.2f}Mpps", font=dict( size=14, color=COLORS[-1], ), - align="left", + align=u"left", showarrow=False )) - # y_max.append(int((nic_limit / 10) + 1) * 10) y_max.append(nic_limit) lnk_limit /= 1000000.0 @@ -1550,142 +650,146 @@ def plot_throughput_speedup_analysis(plot, input_data): traces.append(plgo.Scatter( x=x_vals, y=[lnk_limit, ] * len(x_vals), - name="Link: {0:.2f}Mpps".format(lnk_limit), + name=f"Link: {lnk_limit:.2f}Mpps", showlegend=False, - mode="lines", + mode=u"lines", line=dict( - dash="dot", + dash=u"dot", color=COLORS[-2], width=1), - hoverinfo="none" + hoverinfo=u"none" )) annotations.append(dict( x=1, y=lnk_limit, - xref="x", - yref="y", - xanchor="left", - yanchor="bottom", - text="Link: {0:.2f}Mpps".format(lnk_limit), + xref=u"x", + yref=u"y", + xanchor=u"left", + yanchor=u"bottom", + text=f"Link: {lnk_limit:.2f}Mpps", font=dict( size=14, color=COLORS[-2], ), - align="left", + align=u"left", showarrow=False )) - # y_max.append(int((lnk_limit / 10) + 1) * 10) y_max.append(lnk_limit) pci_limit /= 1000000.0 if (pci_limit < threshold and - (pci_limit < lnk_limit * 0.95 or lnk_limit > lnk_limit * 1.05)): + (pci_limit < lnk_limit * 0.95 or lnk_limit > lnk_limit * 1.05)): traces.append(plgo.Scatter( x=x_vals, y=[pci_limit, ] * len(x_vals), - name="PCIe: {0:.2f}Mpps".format(pci_limit), + name=f"PCIe: {pci_limit:.2f}Mpps", showlegend=False, - mode="lines", + mode=u"lines", line=dict( - dash="dot", + dash=u"dot", color=COLORS[-3], width=1), - hoverinfo="none" + hoverinfo=u"none" )) annotations.append(dict( x=1, y=pci_limit, - xref="x", - yref="y", - xanchor="left", - yanchor="bottom", - text="PCIe: {0:.2f}Mpps".format(pci_limit), + xref=u"x", + yref=u"y", + xanchor=u"left", + yanchor=u"bottom", + text=f"PCIe: {pci_limit:.2f}Mpps", font=dict( size=14, color=COLORS[-3], ), - align="left", + align=u"left", showarrow=False )) - # y_max.append(int((pci_limit / 10) + 1) * 10) y_max.append(pci_limit) # Perfect and measured: cidx = 0 - for name, val in y_sorted.iteritems(): + for name, val in vals.items(): hovertext = list() try: - for idx in range(len(val["val"])): + for idx in range(len(val[u"val"])): htext = "" - if isinstance(val["val"][idx], float): - htext += "No. of Runs: {1}<br>" \ - "Mean: {0:.2f}Mpps<br>".format(val["val"][idx], - val["count"][idx]) - if isinstance(val["diff"][idx], float): - htext += "Diff: {0:.0f}%<br>".format(round(val["diff"][idx])) - if isinstance(val["rel"][idx], float): - htext += "Speedup: {0:.2f}".format(val["rel"][idx]) + if isinstance(val[u"val"][idx], float): + htext += ( + f"No. of Runs: {val[u'count'][idx]}<br>" + f"Mean: {val[u'val'][idx]:.2f}Mpps<br>" + ) + if isinstance(val[u"diff"][idx], float): + htext += f"Diff: {round(val[u'diff'][idx]):.0f}%<br>" + if isinstance(val[u"rel"][idx], float): + htext += f"Speedup: {val[u'rel'][idx]:.2f}" hovertext.append(htext) - traces.append(plgo.Scatter(x=x_vals, - y=val["val"], - name=name, - legendgroup=name, - mode="lines+markers", - line=dict( - color=COLORS[cidx], - width=2), - marker=dict( - symbol="circle", - size=10 - ), - text=hovertext, - hoverinfo="text+name" - )) - traces.append(plgo.Scatter(x=x_vals, - y=val["ideal"], - name="{0} perfect".format(name), - legendgroup=name, - showlegend=False, - mode="lines", - line=dict( - color=COLORS[cidx], - width=2, - dash="dash"), - text=["Perfect: {0:.2f}Mpps".format(y) - for y in val["ideal"]], - hoverinfo="text" - )) + traces.append( + plgo.Scatter( + x=x_vals, + y=val[u"val"], + name=name, + legendgroup=name, + mode=u"lines+markers", + line=dict( + color=COLORS[cidx], + width=2), + marker=dict( + symbol=u"circle", + size=10 + ), + text=hovertext, + hoverinfo=u"text+name" + ) + ) + traces.append( + plgo.Scatter( + x=x_vals, + y=val[u"ideal"], + name=f"{name} perfect", + legendgroup=name, + showlegend=False, + mode=u"lines", + line=dict( + color=COLORS[cidx], + width=2, + dash=u"dash"), + text=[f"Perfect: {y:.2f}Mpps" for y in val[u"ideal"]], + hoverinfo=u"text" + ) + ) cidx += 1 except (IndexError, ValueError, KeyError) as err: - logging.warning("No data for '{0}'".format(name)) - logging.warning(repr(err)) + logging.warning(f"No data for {name}\n{repr(err)}") try: # Create plot - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], plot["output-file-type"])) - layout = deepcopy(plot["layout"]) - if layout.get("title", None): - layout["title"] = "<b>Speedup Multi-core:</b> {0}". \ - format(layout["title"]) - # layout["yaxis"]["range"] = [0, int((max(y_max) / 10) + 1) * 10] - layout["yaxis"]["range"] = [0, int(max(y_max) * 1.1)] - layout["annotations"].extend(annotations) + file_type = plot.get(u"output-file-type", u".html") + logging.info(f" Writing file {plot[u'output-file']}{file_type}.") + layout = deepcopy(plot[u"layout"]) + if layout.get(u"title", None): + layout[u"title"] = f"<b>Speedup Multi-core:</b> {layout[u'title']}" + layout[u"yaxis"][u"range"] = [0, int(max(y_max) * 1.1)] + layout[u"annotations"].extend(annotations) plpl = plgo.Figure(data=traces, layout=layout) # Export Plot - ploff.plot(plpl, - show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], - plot["output-file-type"])) + ploff.plot( + plpl, + show_link=False, + auto_open=False, + filename=f"{plot[u'output-file']}{file_type}" + ) except PlotlyError as err: - logging.error(" Finished with error: {}". - format(str(err).replace("\n", " "))) + logging.error( + f" Finished with error: {repr(err)}".replace(u"\n", u" ") + ) return -def plot_http_server_performance_box(plot, input_data): - """Generate the plot(s) with algorithm: plot_http_server_performance_box +def plot_http_server_perf_box(plot, input_data): + """Generate the plot(s) with algorithm: plot_http_server_perf_box specified in the specification file. :param plot: Plot to generate. @@ -1695,11 +799,13 @@ def plot_http_server_performance_box(plot, input_data): """ # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot.get("title", ""))) + logging.info( + f" Creating the data set for the {plot.get(u'type', u'')} " + f"{plot.get(u'title', u'')}." + ) data = input_data.filter_data(plot) if data is None: - logging.error("No data.") + logging.error(u"No data.") return # Prepare the data for the plot @@ -1707,12 +813,12 @@ def plot_http_server_performance_box(plot, input_data): for job in data: for build in job: for test in build: - if y_vals.get(test["name"], None) is None: - y_vals[test["name"]] = list() + if y_vals.get(test[u"name"], None) is None: + y_vals[test[u"name"]] = list() try: - y_vals[test["name"]].append(test["result"]) + y_vals[test[u"name"]].append(test[u"result"]) except (KeyError, TypeError): - y_vals[test["name"]].append(None) + y_vals[test[u"name"]].append(None) # Add None to the lists with missing data max_len = 0 @@ -1721,53 +827,59 @@ def plot_http_server_performance_box(plot, input_data): if len(val) > max_len: max_len = len(val) nr_of_samples.append(len(val)) - for key, val in y_vals.items(): + for val in y_vals.values(): if len(val) < max_len: val.extend([None for _ in range(max_len - len(val))]) # Add plot traces traces = list() - df = pd.DataFrame(y_vals) - df.head() - for i, col in enumerate(df.columns): - name = "{nr}. ({samples:02d} run{plural}) {name}".\ - format(nr=(i + 1), - samples=nr_of_samples[i], - plural='s' if nr_of_samples[i] > 1 else '', - name=col.lower().replace('-ndrpdr', '')) + df_y = pd.DataFrame(y_vals) + df_y.head() + for i, col in enumerate(df_y.columns): + name = \ + f"{i + 1}. " \ + f"({nr_of_samples[i]:02d} " \ + f"run{u's' if nr_of_samples[i] > 1 else u''}) " \ + f"{col.lower().replace(u'-ndrpdr', u'')}" if len(name) > 50: - name_lst = name.split('-') - name = "" + name_lst = name.split(u'-') + name = u"" split_name = True for segment in name_lst: if (len(name) + len(segment) + 1) > 50 and split_name: - name += "<br> " + name += u"<br> " split_name = False - name += segment + '-' + name += segment + u'-' name = name[:-1] - traces.append(plgo.Box(x=[str(i + 1) + '.'] * len(df[col]), - y=df[col], + traces.append(plgo.Box(x=[str(i + 1) + u'.'] * len(df_y[col]), + y=df_y[col], name=name, - **plot["traces"])) + **plot[u"traces"])) try: # Create plot - plpl = plgo.Figure(data=traces, layout=plot["layout"]) + plpl = plgo.Figure(data=traces, layout=plot[u"layout"]) # Export Plot - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], plot["output-file-type"])) - ploff.plot(plpl, show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], - plot["output-file-type"])) + logging.info( + f" Writing file {plot[u'output-file']}" + f"{plot[u'output-file-type']}." + ) + ploff.plot( + plpl, + show_link=False, + auto_open=False, + filename=f"{plot[u'output-file']}{plot[u'output-file-type']}" + ) except PlotlyError as err: - logging.error(" Finished with error: {}". - format(str(err).replace("\n", " "))) + logging.error( + f" Finished with error: {repr(err)}".replace(u"\n", u" ") + ) return -def plot_service_density_heatmap(plot, input_data): - """Generate the plot(s) with algorithm: plot_service_density_heatmap +def plot_nf_heatmap(plot, input_data): + """Generate the plot(s) with algorithm: plot_nf_heatmap specified in the specification file. :param plot: Plot to generate. @@ -1776,729 +888,320 @@ def plot_service_density_heatmap(plot, input_data): :type input_data: InputData """ - REGEX_CN = re.compile(r'^(\d*)R(\d*)C$') - REGEX_TEST_NAME = re.compile(r'^.*-(\d+ch|\d+pl)-' + regex_cn = re.compile(r'^(\d*)R(\d*)C$') + regex_test_name = re.compile(r'^.*-(\d+ch|\d+pl)-' r'(\d+mif|\d+vh)-' r'(\d+vm\d+t|\d+dcr\d+t).*$') - - txt_chains = list() - txt_nodes = list() vals = dict() # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot.get("title", ""))) + logging.info( + f" Creating the data set for the {plot.get(u'type', u'')} " + f"{plot.get(u'title', u'')}." + ) data = input_data.filter_data(plot, continue_on_error=True) if data is None or data.empty: - logging.error("No data.") + logging.error(u"No data.") return for job in data: for build in job: for test in build: - for tag in test['tags']: - groups = re.search(REGEX_CN, tag) + for tag in test[u"tags"]: + groups = re.search(regex_cn, tag) if groups: - c = str(groups.group(1)) - n = str(groups.group(2)) + chain = str(groups.group(1)) + node = str(groups.group(2)) break else: continue - groups = re.search(REGEX_TEST_NAME, test["name"]) + groups = re.search(regex_test_name, test[u"name"]) if groups and len(groups.groups()) == 3: - hover_name = "{chain}-{vhost}-{vm}".format( - chain=str(groups.group(1)), - vhost=str(groups.group(2)), - vm=str(groups.group(3))) + hover_name = ( + f"{str(groups.group(1))}-" + f"{str(groups.group(2))}-" + f"{str(groups.group(3))}" + ) else: - hover_name = "" - if vals.get(c, None) is None: - vals[c] = dict() - if vals[c].get(n, None) is None: - vals[c][n] = dict(name=hover_name, - vals=list(), - nr=None, - mean=None, - stdev=None) + hover_name = u"" + if vals.get(chain, None) is None: + vals[chain] = dict() + if vals[chain].get(node, None) is None: + vals[chain][node] = dict( + name=hover_name, + vals=list(), + nr=None, + mean=None, + stdev=None + ) try: - if plot["include-tests"] == "MRR": - result = test["result"]["receive-rate"] # .avg - elif plot["include-tests"] == "PDR": - result = test["throughput"]["PDR"]["LOWER"] - elif plot["include-tests"] == "NDR": - result = test["throughput"]["NDR"]["LOWER"] + if plot[u"include-tests"] == u"MRR": + result = test[u"result"][u"receive-rate"] + elif plot[u"include-tests"] == u"PDR": + result = test[u"throughput"][u"PDR"][u"LOWER"] + elif plot[u"include-tests"] == u"NDR": + result = test[u"throughput"][u"NDR"][u"LOWER"] else: result = None except TypeError: result = None if result: - vals[c][n]["vals"].append(result) + vals[chain][node][u"vals"].append(result) if not vals: - logging.error("No data.") + logging.error(u"No data.") return - for key_c in vals.keys(): + txt_chains = list() + txt_nodes = list() + for key_c in vals: txt_chains.append(key_c) for key_n in vals[key_c].keys(): txt_nodes.append(key_n) - if vals[key_c][key_n]["vals"]: - vals[key_c][key_n]["nr"] = len(vals[key_c][key_n]["vals"]) - vals[key_c][key_n]["mean"] = \ - round(mean(vals[key_c][key_n]["vals"]) / 1000000, 1) - vals[key_c][key_n]["stdev"] = \ - round(stdev(vals[key_c][key_n]["vals"]) / 1000000, 1) + if vals[key_c][key_n][u"vals"]: + vals[key_c][key_n][u"nr"] = len(vals[key_c][key_n][u"vals"]) + vals[key_c][key_n][u"mean"] = \ + round(mean(vals[key_c][key_n][u"vals"]) / 1000000, 1) + vals[key_c][key_n][u"stdev"] = \ + round(stdev(vals[key_c][key_n][u"vals"]) / 1000000, 1) txt_nodes = list(set(txt_nodes)) - txt_chains = sorted(txt_chains, key=lambda chain: int(chain)) - txt_nodes = sorted(txt_nodes, key=lambda node: int(node)) + def sort_by_int(value): + """Makes possible to sort a list of strings which represent integers. + + :param value: Integer as a string. + :type value: str + :returns: Integer representation of input parameter 'value'. + :rtype: int + """ + return int(value) + + txt_chains = sorted(txt_chains, key=sort_by_int) + txt_nodes = sorted(txt_nodes, key=sort_by_int) chains = [i + 1 for i in range(len(txt_chains))] nodes = [i + 1 for i in range(len(txt_nodes))] data = [list() for _ in range(len(chains))] - for c in chains: - for n in nodes: + for chain in chains: + for node in nodes: try: - val = vals[txt_chains[c - 1]][txt_nodes[n - 1]]["mean"] + val = vals[txt_chains[chain - 1]][txt_nodes[node - 1]][u"mean"] except (KeyError, IndexError): val = None - data[c - 1].append(val) + data[chain - 1].append(val) - # Colorscales: - my_green = [[0.0, 'rgb(235, 249, 242)'], - [1.0, 'rgb(45, 134, 89)']] + # Color scales: + my_green = [[0.0, u"rgb(235, 249, 242)"], + [1.0, u"rgb(45, 134, 89)"]] - my_blue = [[0.0, 'rgb(236, 242, 248)'], - [1.0, 'rgb(57, 115, 172)']] + my_blue = [[0.0, u"rgb(236, 242, 248)"], + [1.0, u"rgb(57, 115, 172)"]] - my_grey = [[0.0, 'rgb(230, 230, 230)'], - [1.0, 'rgb(102, 102, 102)']] + my_grey = [[0.0, u"rgb(230, 230, 230)"], + [1.0, u"rgb(102, 102, 102)"]] hovertext = list() annotations = list() - text = ("Test: {name}<br>" - "Runs: {nr}<br>" - "Thput: {val}<br>" - "StDev: {stdev}") + text = (u"Test: {name}<br>" + u"Runs: {nr}<br>" + u"Thput: {val}<br>" + u"StDev: {stdev}") - for c in range(len(txt_chains)): + for chain, _ in enumerate(txt_chains): hover_line = list() - for n in range(len(txt_nodes)): - if data[c][n] is not None: - annotations.append(dict( - x=n+1, - y=c+1, - xref="x", - yref="y", - xanchor="center", - yanchor="middle", - text=str(data[c][n]), - font=dict( - size=14, - ), - align="center", - showarrow=False - )) + for node, _ in enumerate(txt_nodes): + if data[chain][node] is not None: + annotations.append( + dict( + x=node+1, + y=chain+1, + xref=u"x", + yref=u"y", + xanchor=u"center", + yanchor=u"middle", + text=str(data[chain][node]), + font=dict( + size=14, + ), + align=u"center", + showarrow=False + ) + ) hover_line.append(text.format( - name=vals[txt_chains[c]][txt_nodes[n]]["name"], - nr=vals[txt_chains[c]][txt_nodes[n]]["nr"], - val=data[c][n], - stdev=vals[txt_chains[c]][txt_nodes[n]]["stdev"])) + name=vals[txt_chains[chain]][txt_nodes[node]][u"name"], + nr=vals[txt_chains[chain]][txt_nodes[node]][u"nr"], + val=data[chain][node], + stdev=vals[txt_chains[chain]][txt_nodes[node]][u"stdev"])) hovertext.append(hover_line) traces = [ - plgo.Heatmap(x=nodes, - y=chains, - z=data, - colorbar=dict( - title=plot.get("z-axis", ""), - titleside="right", - titlefont=dict( - size=16 - ), - tickfont=dict( - size=16, - ), - tickformat=".1f", - yanchor="bottom", - y=-0.02, - len=0.925, - ), - showscale=True, - colorscale=my_green, - text=hovertext, - hoverinfo="text") + plgo.Heatmap( + x=nodes, + y=chains, + z=data, + colorbar=dict( + title=plot.get(u"z-axis", u""), + titleside=u"right", + titlefont=dict( + size=16 + ), + tickfont=dict( + size=16, + ), + tickformat=u".1f", + yanchor=u"bottom", + y=-0.02, + len=0.925, + ), + showscale=True, + colorscale=my_green, + text=hovertext, + hoverinfo=u"text" + ) ] for idx, item in enumerate(txt_nodes): # X-axis, numbers: - annotations.append(dict( - x=idx+1, - y=0.05, - xref="x", - yref="y", - xanchor="center", - yanchor="top", - text=item, - font=dict( - size=16, - ), - align="center", - showarrow=False - )) - for idx, item in enumerate(txt_chains): - # Y-axis, numbers: - annotations.append(dict( - x=0.35, - y=idx+1, - xref="x", - yref="y", - xanchor="right", - yanchor="middle", - text=item, - font=dict( - size=16, - ), - align="center", - showarrow=False - )) - # X-axis, title: - annotations.append(dict( - x=0.55, - y=-0.15, - xref="paper", - yref="y", - xanchor="center", - yanchor="bottom", - text=plot.get("x-axis", ""), - font=dict( - size=16, - ), - align="center", - showarrow=False - )) - # Y-axis, title: - annotations.append(dict( - x=-0.1, - y=0.5, - xref="x", - yref="paper", - xanchor="center", - yanchor="middle", - text=plot.get("y-axis", ""), - font=dict( - size=16, - ), - align="center", - textangle=270, - showarrow=False - )) - updatemenus = list([ - dict( - x=1.0, - y=0.0, - xanchor='right', - yanchor='bottom', - direction='up', - buttons=list([ - dict( - args=[{"colorscale": [my_green, ], "reversescale": False}], - label="Green", - method="update" - ), - dict( - args=[{"colorscale": [my_blue, ], "reversescale": False}], - label="Blue", - method="update" + annotations.append( + dict( + x=idx+1, + y=0.05, + xref=u"x", + yref=u"y", + xanchor=u"center", + yanchor=u"top", + text=item, + font=dict( + size=16, ), - dict( - args=[{"colorscale": [my_grey, ], "reversescale": False}], - label="Grey", - method="update" - ) - ]) + align=u"center", + showarrow=False + ) ) - ]) - - try: - layout = deepcopy(plot["layout"]) - except KeyError as err: - logging.error("Finished with error: No layout defined") - logging.error(repr(err)) - return - - layout["annotations"] = annotations - layout['updatemenus'] = updatemenus - - try: - # Create plot - plpl = plgo.Figure(data=traces, layout=layout) - - # Export Plot - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], plot["output-file-type"])) - ploff.plot(plpl, show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], - plot["output-file-type"])) - except PlotlyError as err: - logging.error(" Finished with error: {}". - format(str(err).replace("\n", " "))) - return - - -def plot_service_density_heatmap_compare(plot, input_data): - """Generate the plot(s) with algorithm: plot_service_density_heatmap_compare - specified in the specification file. - - :param plot: Plot to generate. - :param input_data: Data to process. - :type plot: pandas.Series - :type input_data: InputData - """ - - REGEX_CN = re.compile(r'^(\d*)R(\d*)C$') - REGEX_TEST_NAME = re.compile(r'^.*-(\d+ch|\d+pl)-' - r'(\d+mif|\d+vh)-' - r'(\d+vm\d+t|\d+dcr\d+t).*$') - REGEX_THREADS = re.compile(r'^(\d+)(VM|DCR)(\d+)T$') - - txt_chains = list() - txt_nodes = list() - vals = dict() - - # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(plot.get("type", ""), plot.get("title", ""))) - data = input_data.filter_data(plot, continue_on_error=True) - if data is None or data.empty: - logging.error("No data.") - return - - for job in data: - for build in job: - for test in build: - for tag in test['tags']: - groups = re.search(REGEX_CN, tag) - if groups: - c = str(groups.group(1)) - n = str(groups.group(2)) - break - else: - continue - groups = re.search(REGEX_TEST_NAME, test["name"]) - if groups and len(groups.groups()) == 3: - hover_name = "{chain}-{vhost}-{vm}".format( - chain=str(groups.group(1)), - vhost=str(groups.group(2)), - vm=str(groups.group(3))) - else: - hover_name = "" - if vals.get(c, None) is None: - vals[c] = dict() - if vals[c].get(n, None) is None: - vals[c][n] = dict(name=hover_name, - vals_r=list(), - vals_c=list(), - nr_r=None, - nr_c=None, - mean_r=None, - mean_c=None, - stdev_r=None, - stdev_c=None) - try: - if plot["include-tests"] == "MRR": - result = test["result"]["receive-rate"] # .avg - elif plot["include-tests"] == "PDR": - result = test["throughput"]["PDR"]["LOWER"] - elif plot["include-tests"] == "NDR": - result = test["throughput"]["NDR"]["LOWER"] - else: - result = None - except TypeError: - result = None - - if result: - for tag in test['tags']: - groups = re.search(REGEX_THREADS, tag) - if groups and len(groups.groups()) == 3: - if str(groups.group(3)) == \ - plot["reference"]["include"]: - vals[c][n]["vals_r"].append(result) - elif str(groups.group(3)) == \ - plot["compare"]["include"]: - vals[c][n]["vals_c"].append(result) - break - if not vals: - logging.error("No data.") - return - - for key_c in vals.keys(): - txt_chains.append(key_c) - for key_n in vals[key_c].keys(): - txt_nodes.append(key_n) - if vals[key_c][key_n]["vals_r"]: - vals[key_c][key_n]["nr_r"] = len(vals[key_c][key_n]["vals_r"]) - vals[key_c][key_n]["mean_r"] = \ - mean(vals[key_c][key_n]["vals_r"]) - vals[key_c][key_n]["stdev_r"] = \ - round(stdev(vals[key_c][key_n]["vals_r"]) / 1000000, 1) - if vals[key_c][key_n]["vals_c"]: - vals[key_c][key_n]["nr_c"] = len(vals[key_c][key_n]["vals_c"]) - vals[key_c][key_n]["mean_c"] = \ - mean(vals[key_c][key_n]["vals_c"]) - vals[key_c][key_n]["stdev_c"] = \ - round(stdev(vals[key_c][key_n]["vals_c"]) / 1000000, 1) - - txt_nodes = list(set(txt_nodes)) - - txt_chains = sorted(txt_chains, key=lambda chain: int(chain)) - txt_nodes = sorted(txt_nodes, key=lambda node: int(node)) - - chains = [i + 1 for i in range(len(txt_chains))] - nodes = [i + 1 for i in range(len(txt_nodes))] - - data_r = [list() for _ in range(len(chains))] - data_c = [list() for _ in range(len(chains))] - diff = [list() for _ in range(len(chains))] - for c in chains: - for n in nodes: - try: - val_r = vals[txt_chains[c - 1]][txt_nodes[n - 1]]["mean_r"] - except (KeyError, IndexError): - val_r = None - try: - val_c = vals[txt_chains[c - 1]][txt_nodes[n - 1]]["mean_c"] - except (KeyError, IndexError): - val_c = None - if val_c is not None and val_r: - val_d = (val_c - val_r) * 100 / val_r - else: - val_d = None - - if val_r is not None: - val_r = round(val_r / 1000000, 1) - data_r[c - 1].append(val_r) - if val_c is not None: - val_c = round(val_c / 1000000, 1) - data_c[c - 1].append(val_c) - if val_d is not None: - val_d = int(round(val_d, 0)) - diff[c - 1].append(val_d) - - # Colorscales: - my_green = [[0.0, 'rgb(235, 249, 242)'], - [1.0, 'rgb(45, 134, 89)']] - - my_blue = [[0.0, 'rgb(236, 242, 248)'], - [1.0, 'rgb(57, 115, 172)']] - - my_grey = [[0.0, 'rgb(230, 230, 230)'], - [1.0, 'rgb(102, 102, 102)']] - - hovertext = list() - - annotations = list() - annotations_r = list() - annotations_c = list() - annotations_diff = list() - - text = ("Test: {name}" - "<br>{title_r}: {text_r}" - "<br>{title_c}: {text_c}{text_diff}") - text_r = "Thput: {val_r}; StDev: {stdev_r}; Runs: {nr_r}" - text_c = "Thput: {val_c}; StDev: {stdev_c}; Runs: {nr_c}" - text_diff = "<br>Relative Difference {title_c} vs. {title_r}: {diff}%" - - for c in range(len(txt_chains)): - hover_line = list() - for n in range(len(txt_nodes)): - point = dict( - x=n + 1, - y=c + 1, - xref="x", - yref="y", - xanchor="center", - yanchor="middle", - text="", + for idx, item in enumerate(txt_chains): + # Y-axis, numbers: + annotations.append( + dict( + x=0.35, + y=idx+1, + xref=u"x", + yref=u"y", + xanchor=u"right", + yanchor=u"middle", + text=item, font=dict( - size=14, + size=16, ), - align="center", + align=u"center", showarrow=False ) - - point_text_r = "Not present" - point_text_c = "Not present" - point_text_diff = "" - try: - point_r = data_r[c][n] - if point_r is not None: - point_text_r = text_r.format( - val_r=point_r, - stdev_r=vals[txt_chains[c]][txt_nodes[n]]["stdev_r"], - nr_r=vals[txt_chains[c]][txt_nodes[n]]["nr_r"]) - except KeyError: - point_r = None - point["text"] = "" if point_r is None else point_r - annotations_r.append(deepcopy(point)) - - try: - point_c = data_c[c][n] - if point_c is not None: - point_text_c = text_c.format( - val_c=point_c, - stdev_c=vals[txt_chains[c]][txt_nodes[n]]["stdev_c"], - nr_c=vals[txt_chains[c]][txt_nodes[n]]["nr_c"]) - except KeyError: - point_c = None - point["text"] = "" if point_c is None else point_c - annotations_c.append(deepcopy(point)) - - try: - point_d = diff[c][n] - if point_d is not None: - point_text_diff = text_diff.format( - title_r=plot["reference"]["name"], - title_c=plot["compare"]["name"], - diff=point_d) - except KeyError: - point_d = None - point["text"] = "" if point_d is None else point_d - annotations_diff.append(deepcopy(point)) - - try: - name = vals[txt_chains[c]][txt_nodes[n]]["name"] - except KeyError: - continue - - hover_line.append(text.format( - name=name, - title_r=plot["reference"]["name"], - text_r=point_text_r, - title_c=plot["compare"]["name"], - text_c=point_text_c, - text_diff=point_text_diff - )) - - hovertext.append(hover_line) - - traces = [ - plgo.Heatmap(x=nodes, - y=chains, - z=data_r, - visible=True, - colorbar=dict( - title=plot.get("z-axis", ""), - titleside="right", - titlefont=dict( - size=16 - ), - tickfont=dict( - size=16, - ), - tickformat=".1f", - yanchor="bottom", - y=-0.02, - len=0.925, - ), - showscale=True, - colorscale=my_green, - reversescale=False, - text=hovertext, - hoverinfo="text"), - plgo.Heatmap(x=nodes, - y=chains, - z=data_c, - visible=False, - colorbar=dict( - title=plot.get("z-axis", ""), - titleside="right", - titlefont=dict( - size=16 - ), - tickfont=dict( - size=16, - ), - tickformat=".1f", - yanchor="bottom", - y=-0.02, - len=0.925, - ), - showscale=True, - colorscale=my_blue, - reversescale=False, - text=hovertext, - hoverinfo="text"), - plgo.Heatmap(x=nodes, - y=chains, - z=diff, - name="Diff", - visible=False, - colorbar=dict( - title="Relative Difference {name_c} vs. {name_r} [%]". - format(name_c=plot["compare"]["name"], - name_r=plot["reference"]["name"]), - titleside="right", - titlefont=dict( - size=16 - ), - tickfont=dict( - size=16, - ), - tickformat=".1f", - yanchor="bottom", - y=-0.02, - len=0.925, - ), - showscale=True, - colorscale=my_grey, - reversescale=False, - text=hovertext, - hoverinfo="text") - ] - - for idx, item in enumerate(txt_nodes): - # X-axis, numbers: - annotations.append(dict( - x=idx+1, - y=0.05, - xref="x", - yref="y", - xanchor="center", - yanchor="top", - text=item, + ) + # X-axis, title: + annotations.append( + dict( + x=0.55, + y=-0.15, + xref=u"paper", + yref=u"y", + xanchor=u"center", + yanchor=u"bottom", + text=plot.get(u"x-axis", u""), font=dict( size=16, ), - align="center", + align=u"center", showarrow=False - )) - for idx, item in enumerate(txt_chains): - # Y-axis, numbers: - annotations.append(dict( - x=0.35, - y=idx+1, - xref="x", - yref="y", - xanchor="right", - yanchor="middle", - text=item, + ) + ) + # Y-axis, title: + annotations.append( + dict( + x=-0.1, + y=0.5, + xref=u"x", + yref=u"paper", + xanchor=u"center", + yanchor=u"middle", + text=plot.get(u"y-axis", u""), font=dict( size=16, ), - align="center", + align=u"center", + textangle=270, showarrow=False - )) - # X-axis, title: - annotations.append(dict( - x=0.55, - y=-0.15, - xref="paper", - yref="y", - xanchor="center", - yanchor="bottom", - text=plot.get("x-axis", ""), - font=dict( - size=16, - ), - align="center", - showarrow=False - )) - # Y-axis, title: - annotations.append(dict( - x=-0.1, - y=0.5, - xref="x", - yref="paper", - xanchor="center", - yanchor="middle", - text=plot.get("y-axis", ""), - font=dict( - size=16, - ), - align="center", - textangle=270, - showarrow=False - )) + ) + ) updatemenus = list([ dict( - active=0, x=1.0, y=0.0, - xanchor='right', - yanchor='bottom', - direction='up', + xanchor=u"right", + yanchor=u"bottom", + direction=u"up", buttons=list([ dict( - label=plot["reference"]["name"], - method="update", args=[ { - "visible": [True, False, False] - }, - { - "colorscale": [my_green, ], - "reversescale": False, - "annotations": annotations + annotations_r, - }, - ] + u"colorscale": [my_green, ], + u"reversescale": False + } + ], + label=u"Green", + method=u"update" ), dict( - label=plot["compare"]["name"], - method="update", args=[ { - "visible": [False, True, False] - }, - { - "colorscale": [my_blue, ], - "reversescale": False, - "annotations": annotations + annotations_c, - }, - ] + u"colorscale": [my_blue, ], + u"reversescale": False + } + ], + label=u"Blue", + method=u"update" ), dict( - label="Diff", - method="update", args=[ { - "visible": [False, False, True] - }, - { - "colorscale": [my_grey, ], - "reversescale": False, - "annotations": annotations + annotations_diff, - }, - ] - ), + u"colorscale": [my_grey, ], + u"reversescale": False + } + ], + label=u"Grey", + method=u"update" + ) ]) ) ]) try: - layout = deepcopy(plot["layout"]) + layout = deepcopy(plot[u"layout"]) except KeyError as err: - logging.error("Finished with error: No layout defined") - logging.error(repr(err)) + logging.error(f"Finished with error: No layout defined\n{repr(err)}") return - layout["annotations"] = annotations + annotations_r - layout['updatemenus'] = updatemenus + layout[u"annotations"] = annotations + layout[u'updatemenus'] = updatemenus try: # Create plot plpl = plgo.Figure(data=traces, layout=layout) # Export Plot - logging.info(" Writing file '{0}{1}'.". - format(plot["output-file"], plot["output-file-type"])) - ploff.plot(plpl, show_link=False, auto_open=False, - filename='{0}{1}'.format(plot["output-file"], - plot["output-file-type"])) + logging.info( + f" Writing file {plot[u'output-file']}" + f"{plot[u'output-file-type']}." + ) + ploff.plot( + plpl, + show_link=False, + auto_open=False, + filename=f"{plot[u'output-file']}{plot[u'output-file-type']}" + ) except PlotlyError as err: - logging.error(" Finished with error: {}". - format(str(err).replace("\n", " "))) + logging.error( + f" Finished with error: {repr(err)}".replace(u"\n", u" ") + ) return diff --git a/resources/tools/presentation/generator_report.py b/resources/tools/presentation/generator_report.py index 13ca75c1b9..f2bf8a1cb2 100644 --- a/resources/tools/presentation/generator_report.py +++ b/resources/tools/presentation/generator_report.py @@ -19,11 +19,11 @@ import datetime from shutil import make_archive -from utils import get_files, execute_command, archive_input_data +from pal_utils import get_files, execute_command, archive_input_data # .css file for the html format of the report -THEME_OVERRIDES = """/* override table width restrictions */ +THEME_OVERRIDES = u"""/* override table width restrictions */ @media screen and (min-width: 767px) { .wy-table-responsive table td, .wy-table-responsive table th { white-space: normal !important; @@ -87,22 +87,22 @@ THEME_OVERRIDES = """/* override table width restrictions */ """ # Command to build the html format of the report -HTML_BUILDER = 'sphinx-build -v -c . -a ' \ - '-b html -E ' \ - '-t html ' \ - '-D release={release} ' \ - '-D version="Test Report {date}" ' \ - '{working_dir} ' \ - '{build_dir}/' +HTML_BUILDER = u'sphinx-build -v -c . -a ' \ + u'-b html -E ' \ + u'-t html ' \ + u'-D release={release} ' \ + u'-D version="Test Report {date}" ' \ + u'{working_dir} ' \ + u'{build_dir}/' # Command to build the pdf format of the report -PDF_BUILDER = 'sphinx-build -v -c . -a ' \ - '-b latex -E ' \ - '-t latex ' \ - '-D release={release} ' \ - '-D version="Test Report {date}" ' \ - '{working_dir} ' \ - '{build_dir}' +PDF_BUILDER = u'sphinx-build -v -c . -a ' \ + u'-b latex -E ' \ + u'-t latex ' \ + u'-D release={release} ' \ + u'-D version="Test Report {date}" ' \ + u'{working_dir} ' \ + u'{build_dir}' def generate_report(release, spec, report_week): @@ -116,126 +116,115 @@ def generate_report(release, spec, report_week): :type report_week: str """ - logging.info("Generating the report ...") + logging.info(u"Generating the report ...") report = { - "html": generate_html_report, - "pdf": generate_pdf_report + u"html": generate_html_report, + u"pdf": generate_pdf_report } - for report_format, versions in spec.output["format"].items(): - report[report_format](release, spec, versions, report_week) + for report_format in spec.output[u"format"]: + report[report_format](release, spec, report_week) archive_input_data(spec) - logging.info("Done.") + logging.info(u"Done.") -def generate_html_report(release, spec, versions, report_version): +def generate_html_report(release, spec, report_version): """Generate html format of the report. :param release: Release string of the product. :param spec: Specification read from the specification file. - :param versions: List of versions to generate. :param report_version: Version of the report. :type release: str :type spec: Specification - :type versions: list :type report_version: str """ - logging.info(" Generating the html report, give me a few minutes, please " - "...") + _ = report_version - working_dir = spec.environment["paths"]["DIR[WORKING,SRC]"] + logging.info(u" Generating the html report, give me a few minutes, please " + u"...") - cmd = 'cd {working_dir} && mv -f index.html.template index.rst'.\ - format(working_dir=working_dir) - execute_command(cmd) + working_dir = spec.environment[u"paths"][u"DIR[WORKING,SRC]"] + + execute_command(f"cd {working_dir} && mv -f index.html.template index.rst") cmd = HTML_BUILDER.format( release=release, - date=datetime.datetime.utcnow().strftime('%Y-%m-%d %H:%M UTC'), + date=datetime.datetime.utcnow().strftime(u'%Y-%m-%d %H:%M UTC'), working_dir=working_dir, - build_dir=spec.environment["paths"]["DIR[BUILD,HTML]"]) + build_dir=spec.environment[u"paths"][u"DIR[BUILD,HTML]"]) execute_command(cmd) - with open(spec.environment["paths"]["DIR[CSS_PATCH_FILE]"], "w") as \ + with open(spec.environment[u"paths"][u"DIR[CSS_PATCH_FILE]"], u"w") as \ css_file: css_file.write(THEME_OVERRIDES) - with open(spec.environment["paths"]["DIR[CSS_PATCH_FILE2]"], "w") as \ + with open(spec.environment[u"paths"][u"DIR[CSS_PATCH_FILE2]"], u"w") as \ css_file: css_file.write(THEME_OVERRIDES) - logging.info(" Done.") + logging.info(u" Done.") -def generate_pdf_report(release, spec, versions, report_week): +def generate_pdf_report(release, spec, report_week): """Generate html format of the report. :param release: Release string of the product. :param spec: Specification read from the specification file. - :param versions: List of versions to generate. Not implemented yet. :param report_week: Calendar week when the report is published. :type release: str :type spec: Specification - :type versions: list :type report_week: str """ - logging.info(" Generating the pdf report, give me a few minutes, please " - "...") + logging.info(u" Generating the pdf report, give me a few minutes, please " + u"...") - working_dir = spec.environment["paths"]["DIR[WORKING,SRC]"] + working_dir = spec.environment[u"paths"][u"DIR[WORKING,SRC]"] - cmd = 'cd {working_dir} && mv -f index.pdf.template index.rst'.\ - format(working_dir=working_dir) - execute_command(cmd) + execute_command(f"cd {working_dir} && mv -f index.pdf.template index.rst") - _convert_all_svg_to_pdf(spec.environment["paths"]["DIR[WORKING,SRC]"]) + _convert_all_svg_to_pdf(spec.environment[u"paths"][u"DIR[WORKING,SRC]"]) # Convert PyPLOT graphs in HTML format to PDF. - convert_plots = "xvfb-run -a wkhtmltopdf {html} {pdf}" - plots = get_files(spec.environment["paths"]["DIR[STATIC,VPP]"], "html") - plots.extend(get_files(spec.environment["paths"]["DIR[STATIC,DPDK]"], - "html")) + convert_plots = u"xvfb-run -a wkhtmltopdf {html} {pdf}" + plots = get_files(spec.environment[u"paths"][u"DIR[STATIC,VPP]"], u"html") + plots.extend( + get_files(spec.environment[u"paths"][u"DIR[STATIC,DPDK]"], u"html") + ) for plot in plots: - file_name = "{0}.pdf".format(plot.rsplit(".", 1)[0]) - logging.info("Converting '{0}' to '{1}'".format(plot, file_name)) + file_name = f"{plot.rsplit(u'.', 1)[0]}.pdf" + logging.info(f"Converting {plot} to {file_name}") execute_command(convert_plots.format(html=plot, pdf=file_name)) # Generate the LaTeX documentation - build_dir = spec.environment["paths"]["DIR[BUILD,LATEX]"] + build_dir = spec.environment[u"paths"][u"DIR[BUILD,LATEX]"] cmd = PDF_BUILDER.format( release=release, - date=datetime.datetime.utcnow().strftime('%Y-%m-%d %H:%M UTC'), + date=datetime.datetime.utcnow().strftime(u'%Y-%m-%d %H:%M UTC'), working_dir=working_dir, build_dir=build_dir) execute_command(cmd) # Build pdf documentation - archive_dir = spec.environment["paths"]["DIR[STATIC,ARCH]"] + archive_dir = spec.environment[u"paths"][u"DIR[STATIC,ARCH]"] cmds = [ - 'cd {build_dir} && ' - 'pdflatex -shell-escape -interaction nonstopmode csit.tex || true'. - format(build_dir=build_dir), - 'cd {build_dir} && ' - 'pdflatex -interaction nonstopmode csit.tex || true'. - format(build_dir=build_dir), - 'cd {build_dir} && ' - 'cp csit.pdf ../{archive_dir}/csit_{release}.{week}.pdf &&' - 'cp csit.pdf ../{archive_dir}/csit_{release}.pdf'. - format(build_dir=build_dir, - archive_dir=archive_dir, - release=release, - week=report_week) + f'cd {build_dir} && ' + f'pdflatex -shell-escape -interaction nonstopmode csit.tex || true', + f'cd {build_dir} && ' + f'pdflatex -interaction nonstopmode csit.tex || true', + f'cd {build_dir} && ' + f'cp csit.pdf ../{archive_dir}/csit_{release}.{report_week}.pdf &&' + f'cp csit.pdf ../{archive_dir}/csit_{release}.pdf' ] for cmd in cmds: execute_command(cmd) - logging.info(" Done.") + logging.info(u" Done.") def archive_report(spec): @@ -245,13 +234,15 @@ def archive_report(spec): :type spec: Specification """ - logging.info(" Archiving the report ...") + logging.info(u" Archiving the report ...") - make_archive("csit.report", - "gztar", - base_dir=spec.environment["paths"]["DIR[BUILD,HTML]"]) + make_archive( + u"csit.report", + u"gztar", + base_dir=spec.environment[u"paths"][u"DIR[BUILD,HTML]"] + ) - logging.info(" Done.") + logging.info(u" Done.") def _convert_all_svg_to_pdf(path): @@ -261,10 +252,10 @@ def _convert_all_svg_to_pdf(path): :type path: str """ - cmd = "inkscape -D -z --file={svg} --export-pdf={pdf}" - - svg_files = get_files(path, "svg", full_path=True) + svg_files = get_files(path, u"svg", full_path=True) for svg_file in svg_files: - pdf_file = "{0}.pdf".format(svg_file.rsplit('.', 1)[0]) - logging.info("Converting '{0}' to '{1}'".format(svg_file, pdf_file)) - execute_command(cmd.format(svg=svg_file, pdf=pdf_file)) + pdf_file = f"{svg_file.rsplit(u'.', 1)[0]}.pdf" + logging.info(f"Converting {svg_file} to {pdf_file}") + execute_command( + f"inkscape -D -z --file={svg_file} --export-pdf={pdf_file}" + ) diff --git a/resources/tools/presentation/generator_tables.py b/resources/tools/presentation/generator_tables.py index 4a1ac0ef71..b931344b4b 100644 --- a/resources/tools/presentation/generator_tables.py +++ b/resources/tools/presentation/generator_tables.py @@ -19,18 +19,18 @@ import logging import csv import re +from collections import OrderedDict +from xml.etree import ElementTree as ET +from datetime import datetime as dt +from datetime import timedelta + import plotly.graph_objects as go import plotly.offline as ploff import pandas as pd -from string import replace -from collections import OrderedDict from numpy import nan, isnan -from xml.etree import ElementTree as ET -from datetime import datetime as dt -from datetime import timedelta -from utils import mean, stdev, relative_change, classify_anomalies, \ +from pal_utils import mean, stdev, relative_change, classify_anomalies, \ convert_csv_to_pretty_txt, relative_change_stdev @@ -46,14 +46,30 @@ def generate_tables(spec, data): :type data: InputData """ - logging.info("Generating the tables ...") + generator = { + u"table_details": table_details, + u"table_merged_details": table_merged_details, + u"table_perf_comparison": table_perf_comparison, + u"table_perf_comparison_nic": table_perf_comparison_nic, + u"table_nics_comparison": table_nics_comparison, + u"table_soak_vs_ndr": table_soak_vs_ndr, + u"table_perf_trending_dash": table_perf_trending_dash, + u"table_perf_trending_dash_html": table_perf_trending_dash_html, + u"table_last_failed_tests": table_last_failed_tests, + u"table_failed_tests": table_failed_tests, + u"table_failed_tests_html": table_failed_tests_html + } + + logging.info(u"Generating the tables ...") for table in spec.tables: try: - eval(table["algorithm"])(table, data) + generator[table[u"algorithm"]](table, data) except NameError as err: - logging.error("Probably algorithm '{alg}' is not defined: {err}". - format(alg=table["algorithm"], err=repr(err))) - logging.info("Done.") + logging.error( + f"Probably algorithm {table[u'algorithm']} is not defined: " + f"{repr(err)}" + ) + logging.info(u"Done.") def table_details(table, input_data): @@ -66,62 +82,68 @@ def table_details(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) data = input_data.filter_data(table) # Prepare the header of the tables header = list() - for column in table["columns"]: - header.append('"{0}"'.format(str(column["title"]).replace('"', '""'))) + for column in table[u"columns"]: + header.append( + u'"{0}"'.format(str(column[u"title"]).replace(u'"', u'""')) + ) # Generate the data for the table according to the model in the table # specification - job = table["data"].keys()[0] - build = str(table["data"][job][0]) + job = list(table[u"data"].keys())[0] + build = str(table[u"data"][job][0]) try: suites = input_data.suites(job, build) except KeyError: - logging.error(" No data available. The table will not be generated.") + logging.error( + u" No data available. The table will not be generated." + ) return - for suite_longname, suite in suites.iteritems(): + for suite in suites.values: # Generate data - suite_name = suite["name"] + suite_name = suite[u"name"] table_lst = list() for test in data[job][build].keys(): - if data[job][build][test]["parent"] in suite_name: - row_lst = list() - for column in table["columns"]: - try: - col_data = str(data[job][build][test][column["data"]. - split(" ")[1]]).replace('"', '""') - if column["data"].split(" ")[1] in ("conf-history", - "show-run"): - col_data = replace(col_data, " |br| ", "", - maxreplace=1) - col_data = " |prein| {0} |preout| ".\ - format(col_data[:-5]) - row_lst.append('"{0}"'.format(col_data)) - except KeyError: - row_lst.append("No data") - table_lst.append(row_lst) + if data[job][build][test][u"parent"] not in suite_name: + continue + row_lst = list() + for column in table[u"columns"]: + try: + col_data = str(data[job][build][test][column[ + u"data"].split(" ")[1]]).replace(u'"', u'""') + if column[u"data"].split(u" ")[1] in \ + (u"conf-history", u"show-run"): + col_data = col_data.replace(u" |br| ", u"", ) + col_data = f" |prein| {col_data[:-5]} |preout| " + row_lst.append(f'"{col_data}"') + except KeyError: + row_lst.append(u"No data") + table_lst.append(row_lst) # Write the data to file if table_lst: - file_name = "{0}_{1}{2}".format(table["output-file"], suite_name, - table["output-file-ext"]) - logging.info(" Writing file: '{}'".format(file_name)) - with open(file_name, "w") as file_handler: - file_handler.write(",".join(header) + "\n") + file_name = ( + f"{table[u'output-file']}_{suite_name}" + f"{table[u'output-file-ext']}" + ) + logging.info(f" Writing file: {file_name}") + with open(file_name, u"w") as file_handler: + file_handler.write(u",".join(header) + u"\n") for item in table_lst: - file_handler.write(",".join(item) + "\n") + file_handler.write(u",".join(item) + u"\n") - logging.info(" Done.") + logging.info(u" Done.") def table_merged_details(table, input_data): @@ -134,107 +156,155 @@ def table_merged_details(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) - + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) data = input_data.filter_data(table, continue_on_error=True) data = input_data.merge_data(data) data.sort_index(inplace=True) - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) suites = input_data.filter_data( - table, continue_on_error=True, data_set="suites") + table, continue_on_error=True, data_set=u"suites") suites = input_data.merge_data(suites) # Prepare the header of the tables header = list() - for column in table["columns"]: - header.append('"{0}"'.format(str(column["title"]).replace('"', '""'))) + for column in table[u"columns"]: + header.append( + u'"{0}"'.format(str(column[u"title"]).replace(u'"', u'""')) + ) - for _, suite in suites.iteritems(): + for suite in suites.values: # Generate data - suite_name = suite["name"] + suite_name = suite[u"name"] table_lst = list() for test in data.keys(): - if data[test]["parent"] in suite_name: - row_lst = list() - for column in table["columns"]: - try: - col_data = str(data[test][column["data"]. - split(" ")[1]]).replace('"', '""') - col_data = replace(col_data, "No Data", - "Not Captured ") - if column["data"].split(" ")[1] in ("conf-history", - "show-run"): - col_data = replace(col_data, " |br| ", "", - maxreplace=1) - col_data = " |prein| {0} |preout| ".\ - format(col_data[:-5]) - row_lst.append('"{0}"'.format(col_data)) - except KeyError: - row_lst.append('"Not captured"') - table_lst.append(row_lst) + if data[test][u"parent"] not in suite_name: + continue + row_lst = list() + for column in table[u"columns"]: + try: + col_data = str(data[test][column[ + u"data"].split(u" ")[1]]).replace(u'"', u'""') + col_data = col_data.replace( + u"No Data", u"Not Captured " + ) + if column[u"data"].split(u" ")[1] in \ + (u"conf-history", u"show-run"): + col_data = col_data.replace(u" |br| ", u"", 1) + col_data = f" |prein| {col_data[:-5]} |preout| " + row_lst.append(f'"{col_data}"') + except KeyError: + row_lst.append(u'"Not captured"') + table_lst.append(row_lst) # Write the data to file if table_lst: - file_name = "{0}_{1}{2}".format(table["output-file"], suite_name, - table["output-file-ext"]) - logging.info(" Writing file: '{}'".format(file_name)) - with open(file_name, "w") as file_handler: - file_handler.write(",".join(header) + "\n") + file_name = ( + f"{table[u'output-file']}_{suite_name}" + f"{table[u'output-file-ext']}" + ) + logging.info(f" Writing file: {file_name}") + with open(file_name, u"w") as file_handler: + file_handler.write(u",".join(header) + "u\n") for item in table_lst: - file_handler.write(",".join(item) + "\n") + file_handler.write(u",".join(item) + u"\n") - logging.info(" Done.") + logging.info(u" Done.") def _tpc_modify_test_name(test_name): - test_name_mod = test_name.replace("-ndrpdrdisc", ""). \ - replace("-ndrpdr", "").replace("-pdrdisc", ""). \ - replace("-ndrdisc", "").replace("-pdr", ""). \ - replace("-ndr", ""). \ - replace("1t1c", "1c").replace("2t1c", "1c"). \ - replace("2t2c", "2c").replace("4t2c", "2c"). \ - replace("4t4c", "4c").replace("8t4c", "4c") - test_name_mod = re.sub(REGEX_NIC, "", test_name_mod) - return test_name_mod + """Modify a test name by replacing its parts. + + :param test_name: Test name to be modified. + :type test_name: str + :returns: Modified test name. + :rtype: str + """ + test_name_mod = test_name.\ + replace(u"-ndrpdrdisc", u""). \ + replace(u"-ndrpdr", u"").\ + replace(u"-pdrdisc", u""). \ + replace(u"-ndrdisc", u"").\ + replace(u"-pdr", u""). \ + replace(u"-ndr", u""). \ + replace(u"1t1c", u"1c").\ + replace(u"2t1c", u"1c"). \ + replace(u"2t2c", u"2c").\ + replace(u"4t2c", u"2c"). \ + replace(u"4t4c", u"4c").\ + replace(u"8t4c", u"4c") + + return re.sub(REGEX_NIC, u"", test_name_mod) def _tpc_modify_displayed_test_name(test_name): - return test_name.replace("1t1c", "1c").replace("2t1c", "1c"). \ - replace("2t2c", "2c").replace("4t2c", "2c"). \ - replace("4t4c", "4c").replace("8t4c", "4c") + """Modify a test name which is displayed in a table by replacing its parts. + + :param test_name: Test name to be modified. + :type test_name: str + :returns: Modified test name. + :rtype: str + """ + return test_name.\ + replace(u"1t1c", u"1c").\ + replace(u"2t1c", u"1c"). \ + replace(u"2t2c", u"2c").\ + replace(u"4t2c", u"2c"). \ + replace(u"4t4c", u"4c").\ + replace(u"8t4c", u"4c") def _tpc_insert_data(target, src, include_tests): + """Insert src data to the target structure. + + :param target: Target structure where the data is placed. + :param src: Source data to be placed into the target stucture. + :param include_tests: Which results will be included (MRR, NDR, PDR). + :type target: list + :type src: dict + :type include_tests: str + """ try: - if include_tests == "MRR": - target.append(src["result"]["receive-rate"]) # .avg) - elif include_tests == "PDR": - target.append(src["throughput"]["PDR"]["LOWER"]) - elif include_tests == "NDR": - target.append(src["throughput"]["NDR"]["LOWER"]) + if include_tests == u"MRR": + target.append(src[u"result"][u"receive-rate"]) + elif include_tests == u"PDR": + target.append(src[u"throughput"][u"PDR"][u"LOWER"]) + elif include_tests == u"NDR": + target.append(src[u"throughput"][u"NDR"][u"LOWER"]) except (KeyError, TypeError): pass def _tpc_sort_table(table): - # Sort the table: - # 1. New in CSIT-XXXX - # 2. See footnote - # 3. Delta + """Sort the table this way: + + 1. Put "New in CSIT-XXXX" at the first place. + 2. Put "See footnote" at the second place. + 3. Sort the rest by "Delta". + + :param table: Table to sort. + :type table: list + :returns: Sorted table. + :rtype: list + """ + + tbl_new = list() tbl_see = list() tbl_delta = list() for item in table: if isinstance(item[-1], str): - if "New in CSIT" in item[-1]: + if u"New in CSIT" in item[-1]: tbl_new.append(item) - elif "See footnote" in item[-1]: + elif u"See footnote" in item[-1]: tbl_see.append(item) else: tbl_delta.append(item) @@ -269,22 +339,22 @@ def _tpc_generate_html_table(header, data, output_file_name): :type output_file_name: str """ - df = pd.DataFrame(data, columns=header) + df_data = pd.DataFrame(data, columns=header) - df_sorted = [df.sort_values( + df_sorted = [df_data.sort_values( by=[key, header[0]], ascending=[True, True] if key != header[0] else [False, True]) for key in header] - df_sorted_rev = [df.sort_values( + df_sorted_rev = [df_data.sort_values( by=[key, header[0]], ascending=[False, True] if key != header[0] else [True, True]) for key in header] df_sorted.extend(df_sorted_rev) - fill_color = [["#d4e4f7" if idx % 2 else "#e9f1fb" - for idx in range(len(df))]] + fill_color = [[u"#d4e4f7" if idx % 2 else u"#e9f1fb" + for idx in range(len(df_data))]] table_header = dict( - values=["<b>{item}</b>".format(item=item) for item in header], - fill_color="#7eade7", - align=["left", "center"] + values=[f"<b>{item}</b>" for item in header], + fill_color=u"#7eade7", + align=[u"left", u"center"] ) fig = go.Figure() @@ -298,47 +368,47 @@ def _tpc_generate_html_table(header, data, output_file_name): cells=dict( values=columns, fill_color=fill_color, - align=["left", "right"] + align=[u"left", u"right"] ) ) ) buttons = list() - menu_items = ["<b>{0}</b> (ascending)".format(itm) for itm in header] - menu_items_rev = ["<b>{0}</b> (descending)".format(itm) for itm in header] + menu_items = [f"<b>{itm}</b> (ascending)" for itm in header] + menu_items_rev = [f"<b>{itm}</b> (descending)" for itm in header] menu_items.extend(menu_items_rev) for idx, hdr in enumerate(menu_items): visible = [False, ] * len(menu_items) visible[idx] = True buttons.append( dict( - label=hdr.replace(" [Mpps]", ""), - method="update", - args=[{"visible": visible}], + label=hdr.replace(u" [Mpps]", u""), + method=u"update", + args=[{u"visible": visible}], ) ) fig.update_layout( updatemenus=[ go.layout.Updatemenu( - type="dropdown", - direction="down", + type=u"dropdown", + direction=u"down", x=0.03, - xanchor="left", + xanchor=u"left", y=1.045, - yanchor="top", + yanchor=u"top", active=len(menu_items) - 1, buttons=list(buttons) ) ], annotations=[ go.layout.Annotation( - text="<b>Sort by:</b>", + text=u"<b>Sort by:</b>", x=0, - xref="paper", + xref=u"paper", y=1.035, - yref="paper", - align="left", + yref=u"paper", + align=u"left", showarrow=False ) ] @@ -347,8 +417,8 @@ def _tpc_generate_html_table(header, data, output_file_name): ploff.plot(fig, show_link=False, auto_open=False, filename=output_file_name) -def table_performance_comparison(table, input_data): - """Generate the table(s) with algorithm: table_performance_comparison +def table_perf_comparison(table, input_data): + """Generate the table(s) with algorithm: table_perf_comparison specified in the specification file. :param table: Table to generate. @@ -357,231 +427,229 @@ def table_performance_comparison(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) data = input_data.filter_data(table, continue_on_error=True) # Prepare the header of the tables try: - header = ["Test case", ] + header = [u"Test case", ] - if table["include-tests"] == "MRR": - hdr_param = "Rec Rate" + if table[u"include-tests"] == u"MRR": + hdr_param = u"Rec Rate" else: - hdr_param = "Thput" + hdr_param = u"Thput" - history = table.get("history", None) - if history: - for item in history: - header.extend( - ["{0} {1} [Mpps]".format(item["title"], hdr_param), - "{0} Stdev [Mpps]".format(item["title"])]) + history = table.get(u"history", list()) + for item in history: + header.extend( + [ + f"{item[u'title']} {hdr_param} [Mpps]", + f"{item[u'title']} Stdev [Mpps]" + ] + ) header.extend( - ["{0} {1} [Mpps]".format(table["reference"]["title"], hdr_param), - "{0} Stdev [Mpps]".format(table["reference"]["title"]), - "{0} {1} [Mpps]".format(table["compare"]["title"], hdr_param), - "{0} Stdev [Mpps]".format(table["compare"]["title"]), - "Delta [%]"]) - header_str = ",".join(header) + "\n" + [ + f"{table[u'reference'][u'title']} {hdr_param} [Mpps]", + f"{table[u'reference'][u'title']} Stdev [Mpps]", + f"{table[u'compare'][u'title']} {hdr_param} [Mpps]", + f"{table[u'compare'][u'title']} Stdev [Mpps]", + u"Delta [%]" + ] + ) + header_str = u",".join(header) + u"\n" except (AttributeError, KeyError) as err: - logging.error("The model is invalid, missing parameter: {0}". - format(err)) + logging.error(f"The model is invalid, missing parameter: {repr(err)}") return # Prepare data to the table: tbl_dict = dict() topo = "" - for job, builds in table["reference"]["data"].items(): - topo = "2n-skx" if "2n-skx" in job else "" + for job, builds in table[u"reference"][u"data"].items(): + topo = u"2n-skx" if u"2n-skx" in job else u"" for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): + for tst_name, tst_data in data[job][str(build)].items(): tst_name_mod = _tpc_modify_test_name(tst_name) - if "across topologies" in table["title"].lower(): - tst_name_mod = tst_name_mod.replace("2n1l-", "") + if u"across topologies" in table[u"title"].lower(): + tst_name_mod = tst_name_mod.replace(u"2n1l-", u"") if tbl_dict.get(tst_name_mod, None) is None: - groups = re.search(REGEX_NIC, tst_data["parent"]) - nic = groups.group(0) if groups else "" - name = "{0}-{1}".format(nic, "-".join(tst_data["name"]. - split("-")[:-1])) - if "across testbeds" in table["title"].lower() or \ - "across topologies" in table["title"].lower(): + groups = re.search(REGEX_NIC, tst_data[u"parent"]) + nic = groups.group(0) if groups else u"" + name = \ + f"{nic}-{u'-'.join(tst_data[u'name'].split(u'-')[:-1])}" + if u"across testbeds" in table[u"title"].lower() or \ + u"across topologies" in table[u"title"].lower(): name = _tpc_modify_displayed_test_name(name) - tbl_dict[tst_name_mod] = {"name": name, - "ref-data": list(), - "cmp-data": list()} - _tpc_insert_data(target=tbl_dict[tst_name_mod]["ref-data"], + tbl_dict[tst_name_mod] = { + u"name": name, + u"ref-data": list(), + u"cmp-data": list() + } + _tpc_insert_data(target=tbl_dict[tst_name_mod][u"ref-data"], src=tst_data, - include_tests=table["include-tests"]) + include_tests=table[u"include-tests"]) - for job, builds in table["compare"]["data"].items(): + for job, builds in table[u"compare"][u"data"].items(): for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): + for tst_name, tst_data in data[job][str(build)].items(): tst_name_mod = _tpc_modify_test_name(tst_name) - if "across topologies" in table["title"].lower(): - tst_name_mod = tst_name_mod.replace("2n1l-", "") + if u"across topologies" in table[u"title"].lower(): + tst_name_mod = tst_name_mod.replace(u"2n1l-", u"") if tbl_dict.get(tst_name_mod, None) is None: - groups = re.search(REGEX_NIC, tst_data["parent"]) - nic = groups.group(0) if groups else "" - name = "{0}-{1}".format(nic, "-".join(tst_data["name"]. - split("-")[:-1])) - if "across testbeds" in table["title"].lower() or \ - "across topologies" in table["title"].lower(): + groups = re.search(REGEX_NIC, tst_data[u"parent"]) + nic = groups.group(0) if groups else u"" + name = \ + f"{nic}-{u'-'.join(tst_data[u'name'].split(u'-')[:-1])}" + if u"across testbeds" in table[u"title"].lower() or \ + u"across topologies" in table[u"title"].lower(): name = _tpc_modify_displayed_test_name(name) - tbl_dict[tst_name_mod] = {"name": name, - "ref-data": list(), - "cmp-data": list()} - _tpc_insert_data(target=tbl_dict[tst_name_mod]["cmp-data"], - src=tst_data, - include_tests=table["include-tests"]) + tbl_dict[tst_name_mod] = { + u"name": name, + u"ref-data": list(), + u"cmp-data": list() + } + _tpc_insert_data( + target=tbl_dict[tst_name_mod][u"cmp-data"], + src=tst_data, + include_tests=table[u"include-tests"] + ) - replacement = table["compare"].get("data-replacement", None) + replacement = table[u"compare"].get(u"data-replacement", None) if replacement: create_new_list = True rpl_data = input_data.filter_data( table, data=replacement, continue_on_error=True) for job, builds in replacement.items(): for build in builds: - for tst_name, tst_data in rpl_data[job][str(build)].iteritems(): + for tst_name, tst_data in rpl_data[job][str(build)].items(): tst_name_mod = _tpc_modify_test_name(tst_name) - if "across topologies" in table["title"].lower(): - tst_name_mod = tst_name_mod.replace("2n1l-", "") + if u"across topologies" in table[u"title"].lower(): + tst_name_mod = tst_name_mod.replace(u"2n1l-", u"") if tbl_dict.get(tst_name_mod, None) is None: - name = "{0}".format("-".join(tst_data["name"]. - split("-")[:-1])) - if "across testbeds" in table["title"].lower() or \ - "across topologies" in table["title"].lower(): + name = \ + f"{u'-'.join(tst_data[u'name'].split(u'-')[:-1])}" + if u"across testbeds" in table[u"title"].lower() or \ + u"across topologies" in table[u"title"].lower(): name = _tpc_modify_displayed_test_name(name) - tbl_dict[tst_name_mod] = {"name": name, - "ref-data": list(), - "cmp-data": list()} + tbl_dict[tst_name_mod] = { + u"name": name, + u"ref-data": list(), + u"cmp-data": list() + } if create_new_list: create_new_list = False - tbl_dict[tst_name_mod]["cmp-data"] = list() + tbl_dict[tst_name_mod][u"cmp-data"] = list() - _tpc_insert_data(target=tbl_dict[tst_name_mod]["cmp-data"], - src=tst_data, - include_tests=table["include-tests"]) + _tpc_insert_data( + target=tbl_dict[tst_name_mod][u"cmp-data"], + src=tst_data, + include_tests=table[u"include-tests"] + ) - if history: - for item in history: - for job, builds in item["data"].items(): - for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): - tst_name_mod = _tpc_modify_test_name(tst_name) - if "across topologies" in table["title"].lower(): - tst_name_mod = tst_name_mod.replace("2n1l-", "") - if tbl_dict.get(tst_name_mod, None) is None: + for item in history: + for job, builds in item[u"data"].items(): + for build in builds: + for tst_name, tst_data in data[job][str(build)].items(): + tst_name_mod = _tpc_modify_test_name(tst_name) + if u"across topologies" in table[u"title"].lower(): + tst_name_mod = tst_name_mod.replace(u"2n1l-", u"") + if tbl_dict.get(tst_name_mod, None) is None: + continue + if tbl_dict[tst_name_mod].get(u"history", None) is None: + tbl_dict[tst_name_mod][u"history"] = OrderedDict() + if tbl_dict[tst_name_mod][u"history"].\ + get(item[u"title"], None) is None: + tbl_dict[tst_name_mod][u"history"][item[ + u"title"]] = list() + try: + if table[u"include-tests"] == u"MRR": + res = tst_data[u"result"][u"receive-rate"] + elif table[u"include-tests"] == u"PDR": + res = tst_data[u"throughput"][u"PDR"][u"LOWER"] + elif table[u"include-tests"] == u"NDR": + res = tst_data[u"throughput"][u"NDR"][u"LOWER"] + else: continue - if tbl_dict[tst_name_mod].get("history", None) is None: - tbl_dict[tst_name_mod]["history"] = OrderedDict() - if tbl_dict[tst_name_mod]["history"].\ - get(item["title"], None) is None: - tbl_dict[tst_name_mod]["history"][item["title"]] = \ - list() - try: - # TODO: Re-work when NDRPDRDISC tests are not used - if table["include-tests"] == "MRR": - tbl_dict[tst_name_mod]["history"][item[ - "title"]].append(tst_data["result"][ - "receive-rate"].avg) - elif table["include-tests"] == "PDR": - if tst_data["type"] == "PDR": - tbl_dict[tst_name_mod]["history"][ - item["title"]].\ - append(tst_data["throughput"]["value"]) - elif tst_data["type"] == "NDRPDR": - tbl_dict[tst_name_mod]["history"][item[ - "title"]].append(tst_data["throughput"][ - "PDR"]["LOWER"]) - elif table["include-tests"] == "NDR": - if tst_data["type"] == "NDR": - tbl_dict[tst_name_mod]["history"][ - item["title"]].\ - append(tst_data["throughput"]["value"]) - elif tst_data["type"] == "NDRPDR": - tbl_dict[tst_name_mod]["history"][item[ - "title"]].append(tst_data["throughput"][ - "NDR"]["LOWER"]) - else: - continue - except (TypeError, KeyError): - pass + tbl_dict[tst_name_mod][u"history"][item[u"title"]].\ + append(res) + except (TypeError, KeyError): + pass tbl_lst = list() footnote = False - for tst_name in tbl_dict.keys(): - item = [tbl_dict[tst_name]["name"], ] + for tst_name in tbl_dict: + item = [tbl_dict[tst_name][u"name"], ] if history: - if tbl_dict[tst_name].get("history", None) is not None: - for hist_data in tbl_dict[tst_name]["history"].values(): + if tbl_dict[tst_name].get(u"history", None) is not None: + for hist_data in tbl_dict[tst_name][u"history"].values(): if hist_data: item.append(round(mean(hist_data) / 1000000, 2)) item.append(round(stdev(hist_data) / 1000000, 2)) else: - item.extend(["Not tested", "Not tested"]) + item.extend([u"Not tested", u"Not tested"]) else: - item.extend(["Not tested", "Not tested"]) - data_t = tbl_dict[tst_name]["ref-data"] + item.extend([u"Not tested", u"Not tested"]) + data_t = tbl_dict[tst_name][u"ref-data"] if data_t: item.append(round(mean(data_t) / 1000000, 2)) item.append(round(stdev(data_t) / 1000000, 2)) else: - item.extend(["Not tested", "Not tested"]) - data_t = tbl_dict[tst_name]["cmp-data"] + item.extend([u"Not tested", u"Not tested"]) + data_t = tbl_dict[tst_name][u"cmp-data"] if data_t: item.append(round(mean(data_t) / 1000000, 2)) item.append(round(stdev(data_t) / 1000000, 2)) else: - item.extend(["Not tested", "Not tested"]) - if item[-2] == "Not tested": + item.extend([u"Not tested", u"Not tested"]) + if item[-2] == u"Not tested": pass - elif item[-4] == "Not tested": - item.append("New in CSIT-1908") - elif topo == "2n-skx" and "dot1q" in tbl_dict[tst_name]["name"]: - item.append("See footnote [1]") + elif item[-4] == u"Not tested": + item.append(u"New in CSIT-1908") + elif topo == u"2n-skx" and u"dot1q" in tbl_dict[tst_name][u"name"]: + item.append(u"See footnote [1]") footnote = True elif item[-4] != 0: item.append(int(relative_change(float(item[-4]), float(item[-2])))) - if (len(item) == len(header)) and (item[-3] != "Not tested"): + if (len(item) == len(header)) and (item[-3] != u"Not tested"): tbl_lst.append(item) tbl_lst = _tpc_sort_table(tbl_lst) # Generate csv tables: - csv_file = "{0}.csv".format(table["output-file"]) - with open(csv_file, "w") as file_handler: + csv_file = f"{table[u'output-file']}.csv" + with open(csv_file, u"w") as file_handler: file_handler.write(header_str) for test in tbl_lst: - file_handler.write(",".join([str(item) for item in test]) + "\n") + file_handler.write(u",".join([str(item) for item in test]) + u"\n") - txt_file_name = "{0}.txt".format(table["output-file"]) + txt_file_name = f"{table[u'output-file']}.txt" convert_csv_to_pretty_txt(csv_file, txt_file_name) if footnote: - with open(txt_file_name, 'a') as txt_file: + with open(txt_file_name, u'a') as txt_file: txt_file.writelines([ - "\nFootnotes:\n", - "[1] CSIT-1908 changed test methodology of dot1q tests in " - "2-node testbeds, dot1q encapsulation is now used on both " - "links of SUT.\n", - " Previously dot1q was used only on a single link with the " - "other link carrying untagged Ethernet frames. This changes " - "results\n", - " in slightly lower throughput in CSIT-1908 for these " - "tests. See release notes." + u"\nFootnotes:\n", + u"[1] CSIT-1908 changed test methodology of dot1q tests in " + u"2-node testbeds, dot1q encapsulation is now used on both " + u"links of SUT.\n", + u" Previously dot1q was used only on a single link with the " + u"other link carrying untagged Ethernet frames. This changes " + u"results\n", + u" in slightly lower throughput in CSIT-1908 for these " + u"tests. See release notes." ]) # Generate html table: - _tpc_generate_html_table(header, tbl_lst, - "{0}.html".format(table["output-file"])) + _tpc_generate_html_table(header, tbl_lst, f"{table[u'output-file']}.html") -def table_performance_comparison_nic(table, input_data): - """Generate the table(s) with algorithm: table_performance_comparison +def table_perf_comparison_nic(table, input_data): + """Generate the table(s) with algorithm: table_perf_comparison specified in the specification file. :param table: Table to generate. @@ -590,231 +658,229 @@ def table_performance_comparison_nic(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) data = input_data.filter_data(table, continue_on_error=True) # Prepare the header of the tables try: - header = ["Test case", ] + header = [u"Test case", ] - if table["include-tests"] == "MRR": - hdr_param = "Rec Rate" + if table[u"include-tests"] == u"MRR": + hdr_param = u"Rec Rate" else: - hdr_param = "Thput" + hdr_param = u"Thput" - history = table.get("history", None) - if history: - for item in history: - header.extend( - ["{0} {1} [Mpps]".format(item["title"], hdr_param), - "{0} Stdev [Mpps]".format(item["title"])]) + history = table.get(u"history", list()) + for item in history: + header.extend( + [ + f"{item[u'title']} {hdr_param} [Mpps]", + f"{item[u'title']} Stdev [Mpps]" + ] + ) header.extend( - ["{0} {1} [Mpps]".format(table["reference"]["title"], hdr_param), - "{0} Stdev [Mpps]".format(table["reference"]["title"]), - "{0} {1} [Mpps]".format(table["compare"]["title"], hdr_param), - "{0} Stdev [Mpps]".format(table["compare"]["title"]), - "Delta [%]"]) - header_str = ",".join(header) + "\n" + [ + f"{table[u'reference'][u'title']} {hdr_param} [Mpps]", + f"{table[u'reference'][u'title']} Stdev [Mpps]", + f"{table[u'compare'][u'title']} {hdr_param} [Mpps]", + f"{table[u'compare'][u'title']} Stdev [Mpps]", + u"Delta [%]" + ] + ) + header_str = u",".join(header) + u"\n" except (AttributeError, KeyError) as err: - logging.error("The model is invalid, missing parameter: {0}". - format(err)) + logging.error(f"The model is invalid, missing parameter: {repr(err)}") return # Prepare data to the table: tbl_dict = dict() - topo = "" - for job, builds in table["reference"]["data"].items(): - topo = "2n-skx" if "2n-skx" in job else "" + topo = u"" + for job, builds in table[u"reference"][u"data"].items(): + topo = u"2n-skx" if u"2n-skx" in job else u"" for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): - if table["reference"]["nic"] not in tst_data["tags"]: + for tst_name, tst_data in data[job][str(build)].items(): + if table[u"reference"][u"nic"] not in tst_data[u"tags"]: continue tst_name_mod = _tpc_modify_test_name(tst_name) - if "across topologies" in table["title"].lower(): - tst_name_mod = tst_name_mod.replace("2n1l-", "") + if u"across topologies" in table[u"title"].lower(): + tst_name_mod = tst_name_mod.replace(u"2n1l-", u"") if tbl_dict.get(tst_name_mod, None) is None: - name = "{0}".format("-".join(tst_data["name"]. - split("-")[:-1])) - if "across testbeds" in table["title"].lower() or \ - "across topologies" in table["title"].lower(): + name = f"{u'-'.join(tst_data[u'name'].split(u'-')[:-1])}" + if u"across testbeds" in table[u"title"].lower() or \ + u"across topologies" in table[u"title"].lower(): name = _tpc_modify_displayed_test_name(name) - tbl_dict[tst_name_mod] = {"name": name, - "ref-data": list(), - "cmp-data": list()} - _tpc_insert_data(target=tbl_dict[tst_name_mod]["ref-data"], - src=tst_data, - include_tests=table["include-tests"]) + tbl_dict[tst_name_mod] = { + u"name": name, + u"ref-data": list(), + u"cmp-data": list() + } + _tpc_insert_data( + target=tbl_dict[tst_name_mod][u"ref-data"], + src=tst_data, + include_tests=table[u"include-tests"] + ) - for job, builds in table["compare"]["data"].items(): + for job, builds in table[u"compare"][u"data"].items(): for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): - if table["compare"]["nic"] not in tst_data["tags"]: + for tst_name, tst_data in data[job][str(build)].items(): + if table[u"compare"][u"nic"] not in tst_data[u"tags"]: continue tst_name_mod = _tpc_modify_test_name(tst_name) - if "across topologies" in table["title"].lower(): - tst_name_mod = tst_name_mod.replace("2n1l-", "") + if u"across topologies" in table[u"title"].lower(): + tst_name_mod = tst_name_mod.replace(u"2n1l-", u"") if tbl_dict.get(tst_name_mod, None) is None: - name = "{0}".format("-".join(tst_data["name"]. - split("-")[:-1])) - if "across testbeds" in table["title"].lower() or \ - "across topologies" in table["title"].lower(): + name = f"{u'-'.join(tst_data[u'name'].split(u'-')[:-1])}" + if u"across testbeds" in table[u"title"].lower() or \ + u"across topologies" in table[u"title"].lower(): name = _tpc_modify_displayed_test_name(name) - tbl_dict[tst_name_mod] = {"name": name, - "ref-data": list(), - "cmp-data": list()} - _tpc_insert_data(target=tbl_dict[tst_name_mod]["cmp-data"], - src=tst_data, - include_tests=table["include-tests"]) + tbl_dict[tst_name_mod] = { + u"name": name, + u"ref-data": list(), + u"cmp-data": list() + } + _tpc_insert_data( + target=tbl_dict[tst_name_mod][u"cmp-data"], + src=tst_data, + include_tests=table[u"include-tests"] + ) - replacement = table["compare"].get("data-replacement", None) + replacement = table[u"compare"].get(u"data-replacement", None) if replacement: create_new_list = True rpl_data = input_data.filter_data( table, data=replacement, continue_on_error=True) for job, builds in replacement.items(): for build in builds: - for tst_name, tst_data in rpl_data[job][str(build)].iteritems(): - if table["compare"]["nic"] not in tst_data["tags"]: + for tst_name, tst_data in rpl_data[job][str(build)].items(): + if table[u"compare"][u"nic"] not in tst_data[u"tags"]: continue tst_name_mod = _tpc_modify_test_name(tst_name) - if "across topologies" in table["title"].lower(): - tst_name_mod = tst_name_mod.replace("2n1l-", "") + if u"across topologies" in table[u"title"].lower(): + tst_name_mod = tst_name_mod.replace(u"2n1l-", u"") if tbl_dict.get(tst_name_mod, None) is None: - name = "{0}".format("-".join(tst_data["name"]. - split("-")[:-1])) - if "across testbeds" in table["title"].lower() or \ - "across topologies" in table["title"].lower(): + name = \ + f"{u'-'.join(tst_data[u'name'].split(u'-')[:-1])}" + if u"across testbeds" in table[u"title"].lower() or \ + u"across topologies" in table[u"title"].lower(): name = _tpc_modify_displayed_test_name(name) - tbl_dict[tst_name_mod] = {"name": name, - "ref-data": list(), - "cmp-data": list()} + tbl_dict[tst_name_mod] = { + u"name": name, + u"ref-data": list(), + u"cmp-data": list() + } if create_new_list: create_new_list = False - tbl_dict[tst_name_mod]["cmp-data"] = list() + tbl_dict[tst_name_mod][u"cmp-data"] = list() - _tpc_insert_data(target=tbl_dict[tst_name_mod]["cmp-data"], - src=tst_data, - include_tests=table["include-tests"]) + _tpc_insert_data( + target=tbl_dict[tst_name_mod][u"cmp-data"], + src=tst_data, + include_tests=table[u"include-tests"] + ) - if history: - for item in history: - for job, builds in item["data"].items(): - for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): - if item["nic"] not in tst_data["tags"]: - continue - tst_name_mod = _tpc_modify_test_name(tst_name) - if "across topologies" in table["title"].lower(): - tst_name_mod = tst_name_mod.replace("2n1l-", "") - if tbl_dict.get(tst_name_mod, None) is None: + for item in history: + for job, builds in item[u"data"].items(): + for build in builds: + for tst_name, tst_data in data[job][str(build)].items(): + if item[u"nic"] not in tst_data[u"tags"]: + continue + tst_name_mod = _tpc_modify_test_name(tst_name) + if u"across topologies" in table[u"title"].lower(): + tst_name_mod = tst_name_mod.replace(u"2n1l-", u"") + if tbl_dict.get(tst_name_mod, None) is None: + continue + if tbl_dict[tst_name_mod].get(u"history", None) is None: + tbl_dict[tst_name_mod][u"history"] = OrderedDict() + if tbl_dict[tst_name_mod][u"history"].\ + get(item[u"title"], None) is None: + tbl_dict[tst_name_mod][u"history"][item[ + u"title"]] = list() + try: + if table[u"include-tests"] == u"MRR": + res = tst_data[u"result"][u"receive-rate"] + elif table[u"include-tests"] == u"PDR": + res = tst_data[u"throughput"][u"PDR"][u"LOWER"] + elif table[u"include-tests"] == u"NDR": + res = tst_data[u"throughput"][u"NDR"][u"LOWER"] + else: continue - if tbl_dict[tst_name_mod].get("history", None) is None: - tbl_dict[tst_name_mod]["history"] = OrderedDict() - if tbl_dict[tst_name_mod]["history"].\ - get(item["title"], None) is None: - tbl_dict[tst_name_mod]["history"][item["title"]] = \ - list() - try: - # TODO: Re-work when NDRPDRDISC tests are not used - if table["include-tests"] == "MRR": - tbl_dict[tst_name_mod]["history"][item[ - "title"]].append(tst_data["result"][ - "receive-rate"].avg) - elif table["include-tests"] == "PDR": - if tst_data["type"] == "PDR": - tbl_dict[tst_name_mod]["history"][ - item["title"]].\ - append(tst_data["throughput"]["value"]) - elif tst_data["type"] == "NDRPDR": - tbl_dict[tst_name_mod]["history"][item[ - "title"]].append(tst_data["throughput"][ - "PDR"]["LOWER"]) - elif table["include-tests"] == "NDR": - if tst_data["type"] == "NDR": - tbl_dict[tst_name_mod]["history"][ - item["title"]].\ - append(tst_data["throughput"]["value"]) - elif tst_data["type"] == "NDRPDR": - tbl_dict[tst_name_mod]["history"][item[ - "title"]].append(tst_data["throughput"][ - "NDR"]["LOWER"]) - else: - continue - except (TypeError, KeyError): - pass + tbl_dict[tst_name_mod][u"history"][item[u"title"]].\ + append(res) + except (TypeError, KeyError): + pass tbl_lst = list() footnote = False - for tst_name in tbl_dict.keys(): - item = [tbl_dict[tst_name]["name"], ] + for tst_name in tbl_dict: + item = [tbl_dict[tst_name][u"name"], ] if history: - if tbl_dict[tst_name].get("history", None) is not None: - for hist_data in tbl_dict[tst_name]["history"].values(): + if tbl_dict[tst_name].get(u"history", None) is not None: + for hist_data in tbl_dict[tst_name][u"history"].values(): if hist_data: item.append(round(mean(hist_data) / 1000000, 2)) item.append(round(stdev(hist_data) / 1000000, 2)) else: - item.extend(["Not tested", "Not tested"]) + item.extend([u"Not tested", u"Not tested"]) else: - item.extend(["Not tested", "Not tested"]) - data_t = tbl_dict[tst_name]["ref-data"] + item.extend([u"Not tested", u"Not tested"]) + data_t = tbl_dict[tst_name][u"ref-data"] if data_t: item.append(round(mean(data_t) / 1000000, 2)) item.append(round(stdev(data_t) / 1000000, 2)) else: - item.extend(["Not tested", "Not tested"]) - data_t = tbl_dict[tst_name]["cmp-data"] + item.extend([u"Not tested", u"Not tested"]) + data_t = tbl_dict[tst_name][u"cmp-data"] if data_t: item.append(round(mean(data_t) / 1000000, 2)) item.append(round(stdev(data_t) / 1000000, 2)) else: - item.extend(["Not tested", "Not tested"]) - if item[-2] == "Not tested": + item.extend([u"Not tested", u"Not tested"]) + if item[-2] == u"Not tested": pass - elif item[-4] == "Not tested": - item.append("New in CSIT-1908") - elif topo == "2n-skx" and "dot1q" in tbl_dict[tst_name]["name"]: - item.append("See footnote [1]") + elif item[-4] == u"Not tested": + item.append(u"New in CSIT-1908") + elif topo == u"2n-skx" and u"dot1q" in tbl_dict[tst_name][u"name"]: + item.append(u"See footnote [1]") footnote = True elif item[-4] != 0: item.append(int(relative_change(float(item[-4]), float(item[-2])))) - if (len(item) == len(header)) and (item[-3] != "Not tested"): + if (len(item) == len(header)) and (item[-3] != u"Not tested"): tbl_lst.append(item) tbl_lst = _tpc_sort_table(tbl_lst) # Generate csv tables: - csv_file = "{0}.csv".format(table["output-file"]) - with open(csv_file, "w") as file_handler: + csv_file = f"{table[u'output-file']}.csv" + with open(csv_file, u"w") as file_handler: file_handler.write(header_str) for test in tbl_lst: - file_handler.write(",".join([str(item) for item in test]) + "\n") + file_handler.write(u",".join([str(item) for item in test]) + u"\n") - txt_file_name = "{0}.txt".format(table["output-file"]) + txt_file_name = f"{table[u'output-file']}.txt" convert_csv_to_pretty_txt(csv_file, txt_file_name) if footnote: - with open(txt_file_name, 'a') as txt_file: + with open(txt_file_name, u'a') as txt_file: txt_file.writelines([ - "\nFootnotes:\n", - "[1] CSIT-1908 changed test methodology of dot1q tests in " - "2-node testbeds, dot1q encapsulation is now used on both " - "links of SUT.\n", - " Previously dot1q was used only on a single link with the " - "other link carrying untagged Ethernet frames. This changes " - "results\n", - " in slightly lower throughput in CSIT-1908 for these " - "tests. See release notes." + u"\nFootnotes:\n", + u"[1] CSIT-1908 changed test methodology of dot1q tests in " + u"2-node testbeds, dot1q encapsulation is now used on both " + u"links of SUT.\n", + u" Previously dot1q was used only on a single link with the " + u"other link carrying untagged Ethernet frames. This changes " + u"results\n", + u" in slightly lower throughput in CSIT-1908 for these " + u"tests. See release notes." ]) # Generate html table: - _tpc_generate_html_table(header, tbl_lst, - "{0}.html".format(table["output-file"])) + _tpc_generate_html_table(header, tbl_lst, f"{table[u'output-file']}.html") def table_nics_comparison(table, input_data): @@ -827,83 +893,82 @@ def table_nics_comparison(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) data = input_data.filter_data(table, continue_on_error=True) # Prepare the header of the tables try: - header = ["Test case", ] + header = [u"Test case", ] - if table["include-tests"] == "MRR": - hdr_param = "Rec Rate" + if table[u"include-tests"] == u"MRR": + hdr_param = u"Rec Rate" else: - hdr_param = "Thput" + hdr_param = u"Thput" header.extend( - ["{0} {1} [Mpps]".format(table["reference"]["title"], hdr_param), - "{0} Stdev [Mpps]".format(table["reference"]["title"]), - "{0} {1} [Mpps]".format(table["compare"]["title"], hdr_param), - "{0} Stdev [Mpps]".format(table["compare"]["title"]), - "Delta [%]"]) - header_str = ",".join(header) + "\n" + [ + f"{table[u'reference'][u'title']} {hdr_param} [Mpps]", + f"{table[u'reference'][u'title']} Stdev [Mpps]", + f"{table[u'compare'][u'title']} {hdr_param} [Mpps]", + f"{table[u'compare'][u'title']} Stdev [Mpps]", + u"Delta [%]" + ] + ) + except (AttributeError, KeyError) as err: - logging.error("The model is invalid, missing parameter: {0}". - format(err)) + logging.error(f"The model is invalid, missing parameter: {repr(err)}") return # Prepare data to the table: tbl_dict = dict() - for job, builds in table["data"].items(): + for job, builds in table[u"data"].items(): for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): - tst_name_mod = tst_name.replace("-ndrpdrdisc", "").\ - replace("-ndrpdr", "").replace("-pdrdisc", "").\ - replace("-ndrdisc", "").replace("-pdr", "").\ - replace("-ndr", "").\ - replace("1t1c", "1c").replace("2t1c", "1c").\ - replace("2t2c", "2c").replace("4t2c", "2c").\ - replace("4t4c", "4c").replace("8t4c", "4c") - tst_name_mod = re.sub(REGEX_NIC, "", tst_name_mod) + for tst_name, tst_data in data[job][str(build)].items(): + tst_name_mod = _tpc_modify_test_name(tst_name) if tbl_dict.get(tst_name_mod, None) is None: - name = "-".join(tst_data["name"].split("-")[:-1]) - tbl_dict[tst_name_mod] = {"name": name, - "ref-data": list(), - "cmp-data": list()} + name = u"-".join(tst_data[u"name"].split(u"-")[:-1]) + tbl_dict[tst_name_mod] = { + u"name": name, + u"ref-data": list(), + u"cmp-data": list() + } try: - if table["include-tests"] == "MRR": - result = tst_data["result"]["receive-rate"] # .avg - elif table["include-tests"] == "PDR": - result = tst_data["throughput"]["PDR"]["LOWER"] - elif table["include-tests"] == "NDR": - result = tst_data["throughput"]["NDR"]["LOWER"] + result = None + if table[u"include-tests"] == u"MRR": + result = tst_data[u"result"][u"receive-rate"] + elif table[u"include-tests"] == u"PDR": + result = tst_data[u"throughput"][u"PDR"][u"LOWER"] + elif table[u"include-tests"] == u"NDR": + result = tst_data[u"throughput"][u"NDR"][u"LOWER"] else: - result = None + continue - if result: - if table["reference"]["nic"] in tst_data["tags"]: - tbl_dict[tst_name_mod]["ref-data"].append(result) - elif table["compare"]["nic"] in tst_data["tags"]: - tbl_dict[tst_name_mod]["cmp-data"].append(result) + if result and \ + table[u"reference"][u"nic"] in tst_data[u"tags"]: + tbl_dict[tst_name_mod][u"ref-data"].append(result) + elif result and \ + table[u"compare"][u"nic"] in tst_data[u"tags"]: + tbl_dict[tst_name_mod][u"cmp-data"].append(result) except (TypeError, KeyError) as err: - logging.debug("No data for {0}".format(tst_name)) - logging.debug(repr(err)) + logging.debug(f"No data for {tst_name}\n{repr(err)}") # No data in output.xml for this test tbl_lst = list() - for tst_name in tbl_dict.keys(): - item = [tbl_dict[tst_name]["name"], ] - data_t = tbl_dict[tst_name]["ref-data"] + for tst_name in tbl_dict: + item = [tbl_dict[tst_name][u"name"], ] + data_t = tbl_dict[tst_name][u"ref-data"] if data_t: item.append(round(mean(data_t) / 1000000, 2)) item.append(round(stdev(data_t) / 1000000, 2)) else: item.extend([None, None]) - data_t = tbl_dict[tst_name]["cmp-data"] + data_t = tbl_dict[tst_name][u"cmp-data"] if data_t: item.append(round(mean(data_t) / 1000000, 2)) item.append(round(stdev(data_t) / 1000000, 2)) @@ -918,17 +983,16 @@ def table_nics_comparison(table, input_data): tbl_lst.sort(key=lambda rel: rel[-1], reverse=True) # Generate csv tables: - csv_file = "{0}.csv".format(table["output-file"]) - with open(csv_file, "w") as file_handler: - file_handler.write(header_str) + with open(f"{table[u'output-file']}.csv", u"w") as file_handler: + file_handler.write(u",".join(header) + u"\n") for test in tbl_lst: - file_handler.write(",".join([str(item) for item in test]) + "\n") + file_handler.write(u",".join([str(item) for item in test]) + u"\n") - convert_csv_to_pretty_txt(csv_file, "{0}.txt".format(table["output-file"])) + convert_csv_to_pretty_txt(f"{table[u'output-file']}.csv", + f"{table[u'output-file']}.txt") # Generate html table: - _tpc_generate_html_table(header, tbl_lst, - "{0}.html".format(table["output-file"])) + _tpc_generate_html_table(header, tbl_lst, f"{table[u'output-file']}.html") def table_soak_vs_ndr(table, input_data): @@ -941,80 +1005,87 @@ def table_soak_vs_ndr(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) data = input_data.filter_data(table, continue_on_error=True) # Prepare the header of the table try: header = [ - "Test case", - "{0} Thput [Mpps]".format(table["reference"]["title"]), - "{0} Stdev [Mpps]".format(table["reference"]["title"]), - "{0} Thput [Mpps]".format(table["compare"]["title"]), - "{0} Stdev [Mpps]".format(table["compare"]["title"]), - "Delta [%]", "Stdev of delta [%]"] - header_str = ",".join(header) + "\n" + u"Test case", + f"{table[u'reference'][u'title']} Thput [Mpps]", + f"{table[u'reference'][u'title']} Stdev [Mpps]", + f"{table[u'compare'][u'title']} Thput [Mpps]", + f"{table[u'compare'][u'title']} Stdev [Mpps]", + u"Delta [%]", u"Stdev of delta [%]" + ] + header_str = u",".join(header) + u"\n" except (AttributeError, KeyError) as err: - logging.error("The model is invalid, missing parameter: {0}". - format(err)) + logging.error(f"The model is invalid, missing parameter: {repr(err)}") return # Create a list of available SOAK test results: tbl_dict = dict() - for job, builds in table["compare"]["data"].items(): + for job, builds in table[u"compare"][u"data"].items(): for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): - if tst_data["type"] == "SOAK": - tst_name_mod = tst_name.replace("-soak", "") + for tst_name, tst_data in data[job][str(build)].items(): + if tst_data[u"type"] == u"SOAK": + tst_name_mod = tst_name.replace(u"-soak", u"") if tbl_dict.get(tst_name_mod, None) is None: - groups = re.search(REGEX_NIC, tst_data["parent"]) - nic = groups.group(0) if groups else "" - name = "{0}-{1}".format(nic, "-".join(tst_data["name"]. - split("-")[:-1])) + groups = re.search(REGEX_NIC, tst_data[u"parent"]) + nic = groups.group(0) if groups else u"" + name = ( + f"{nic}-" + f"{u'-'.join(tst_data[u'name'].split(u'-')[:-1])}" + ) tbl_dict[tst_name_mod] = { - "name": name, - "ref-data": list(), - "cmp-data": list() + u"name": name, + u"ref-data": list(), + u"cmp-data": list() } try: - tbl_dict[tst_name_mod]["cmp-data"].append( - tst_data["throughput"]["LOWER"]) + tbl_dict[tst_name_mod][u"cmp-data"].append( + tst_data[u"throughput"][u"LOWER"]) except (KeyError, TypeError): pass tests_lst = tbl_dict.keys() # Add corresponding NDR test results: - for job, builds in table["reference"]["data"].items(): + for job, builds in table[u"reference"][u"data"].items(): for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): - tst_name_mod = tst_name.replace("-ndrpdr", "").\ - replace("-mrr", "") - if tst_name_mod in tests_lst: - try: - if tst_data["type"] in ("NDRPDR", "MRR", "BMRR"): - if table["include-tests"] == "MRR": - result = tst_data["result"]["receive-rate"] - elif table["include-tests"] == "PDR": - result = tst_data["throughput"]["PDR"]["LOWER"] - elif table["include-tests"] == "NDR": - result = tst_data["throughput"]["NDR"]["LOWER"] - else: - result = None - if result is not None: - tbl_dict[tst_name_mod]["ref-data"].append( - result) - except (KeyError, TypeError): + for tst_name, tst_data in data[job][str(build)].items(): + tst_name_mod = tst_name.replace(u"-ndrpdr", u"").\ + replace(u"-mrr", u"") + if tst_name_mod not in tests_lst: + continue + try: + if tst_data[u"type"] not in (u"NDRPDR", u"MRR", u"BMRR"): continue + if table[u"include-tests"] == u"MRR": + result = tst_data[u"result"][u"receive-rate"] + elif table[u"include-tests"] == u"PDR": + result = \ + tst_data[u"throughput"][u"PDR"][u"LOWER"] + elif table[u"include-tests"] == u"NDR": + result = \ + tst_data[u"throughput"][u"NDR"][u"LOWER"] + else: + result = None + if result is not None: + tbl_dict[tst_name_mod][u"ref-data"].append( + result) + except (KeyError, TypeError): + continue tbl_lst = list() - for tst_name in tbl_dict.keys(): - item = [tbl_dict[tst_name]["name"], ] - data_r = tbl_dict[tst_name]["ref-data"] + for tst_name in tbl_dict: + item = [tbl_dict[tst_name][u"name"], ] + data_r = tbl_dict[tst_name][u"ref-data"] if data_r: data_r_mean = mean(data_r) item.append(round(data_r_mean / 1000000, 2)) @@ -1024,7 +1095,7 @@ def table_soak_vs_ndr(table, input_data): data_r_mean = None data_r_stdev = None item.extend([None, None]) - data_c = tbl_dict[tst_name]["cmp-data"] + data_c = tbl_dict[tst_name][u"cmp-data"] if data_c: data_c_mean = mean(data_c) item.append(round(data_c_mean / 1000000, 2)) @@ -1045,22 +1116,21 @@ def table_soak_vs_ndr(table, input_data): tbl_lst.sort(key=lambda rel: rel[-1], reverse=True) # Generate csv tables: - csv_file = "{0}.csv".format(table["output-file"]) - with open(csv_file, "w") as file_handler: + csv_file = f"{table[u'output-file']}.csv" + with open(csv_file, u"w") as file_handler: file_handler.write(header_str) for test in tbl_lst: - file_handler.write(",".join([str(item) for item in test]) + "\n") + file_handler.write(u",".join([str(item) for item in test]) + u"\n") - convert_csv_to_pretty_txt(csv_file, "{0}.txt".format(table["output-file"])) + convert_csv_to_pretty_txt(csv_file, f"{table[u'output-file']}.txt") # Generate html table: - _tpc_generate_html_table(header, tbl_lst, - "{0}.html".format(table["output-file"])) + _tpc_generate_html_table(header, tbl_lst, f"{table[u'output-file']}.html") -def table_performance_trending_dashboard(table, input_data): +def table_perf_trending_dash(table, input_data): """Generate the table(s) with algorithm: - table_performance_trending_dashboard + table_perf_trending_dash specified in the specification file. :param table: Table to generate. @@ -1069,55 +1139,58 @@ def table_performance_trending_dashboard(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) data = input_data.filter_data(table, continue_on_error=True) # Prepare the header of the tables - header = ["Test Case", - "Trend [Mpps]", - "Short-Term Change [%]", - "Long-Term Change [%]", - "Regressions [#]", - "Progressions [#]" - ] - header_str = ",".join(header) + "\n" + header = [ + u"Test Case", + u"Trend [Mpps]", + u"Short-Term Change [%]", + u"Long-Term Change [%]", + u"Regressions [#]", + u"Progressions [#]" + ] + header_str = u",".join(header) + u"\n" # Prepare data to the table: tbl_dict = dict() - for job, builds in table["data"].items(): + for job, builds in table[u"data"].items(): for build in builds: - for tst_name, tst_data in data[job][str(build)].iteritems(): - if tst_name.lower() in table.get("ignore-list", list()): + for tst_name, tst_data in data[job][str(build)].items(): + if tst_name.lower() in table.get(u"ignore-list", list()): continue if tbl_dict.get(tst_name, None) is None: - groups = re.search(REGEX_NIC, tst_data["parent"]) + groups = re.search(REGEX_NIC, tst_data[u"parent"]) if not groups: continue nic = groups.group(0) tbl_dict[tst_name] = { - "name": "{0}-{1}".format(nic, tst_data["name"]), - "data": OrderedDict()} + u"name": f"{nic}-{tst_data[u'name']}", + u"data": OrderedDict() + } try: - tbl_dict[tst_name]["data"][str(build)] = \ - tst_data["result"]["receive-rate"] + tbl_dict[tst_name][u"data"][str(build)] = \ + tst_data[u"result"][u"receive-rate"] except (TypeError, KeyError): pass # No data in output.xml for this test tbl_lst = list() - for tst_name in tbl_dict.keys(): - data_t = tbl_dict[tst_name]["data"] + for tst_name in tbl_dict: + data_t = tbl_dict[tst_name][u"data"] if len(data_t) < 2: continue classification_lst, avgs = classify_anomalies(data_t) - win_size = min(len(data_t), table["window"]) - long_win_size = min(len(data_t), table["long-trend-window"]) + win_size = min(len(data_t), table[u"window"]) + long_win_size = min(len(data_t), table[u"long-trend-window"]) try: max_long_avg = max( @@ -1147,34 +1220,33 @@ def table_performance_trending_dashboard(table, input_data): isnan(rel_change_long): continue tbl_lst.append( - [tbl_dict[tst_name]["name"], + [tbl_dict[tst_name][u"name"], round(last_avg / 1000000, 2), rel_change_last, rel_change_long, - classification_lst[-win_size:].count("regression"), - classification_lst[-win_size:].count("progression")]) + classification_lst[-win_size:].count(u"regression"), + classification_lst[-win_size:].count(u"progression")]) tbl_lst.sort(key=lambda rel: rel[0]) tbl_sorted = list() - for nrr in range(table["window"], -1, -1): + for nrr in range(table[u"window"], -1, -1): tbl_reg = [item for item in tbl_lst if item[4] == nrr] - for nrp in range(table["window"], -1, -1): + for nrp in range(table[u"window"], -1, -1): tbl_out = [item for item in tbl_reg if item[5] == nrp] tbl_out.sort(key=lambda rel: rel[2]) tbl_sorted.extend(tbl_out) - file_name = "{0}{1}".format(table["output-file"], table["output-file-ext"]) + file_name = f"{table[u'output-file']}{table[u'output-file-ext']}" - logging.info(" Writing file: '{0}'".format(file_name)) - with open(file_name, "w") as file_handler: + logging.info(f" Writing file: {file_name}") + with open(file_name, u"w") as file_handler: file_handler.write(header_str) for test in tbl_sorted: - file_handler.write(",".join([str(item) for item in test]) + '\n') + file_handler.write(u",".join([str(item) for item in test]) + u'\n') - txt_file_name = "{0}.txt".format(table["output-file"]) - logging.info(" Writing file: '{0}'".format(txt_file_name)) - convert_csv_to_pretty_txt(file_name, txt_file_name) + logging.info(f" Writing file: {table[u'output-file']}.txt") + convert_csv_to_pretty_txt(file_name, f"{table[u'output-file']}.txt") def _generate_url(base, testbed, test_name): @@ -1192,145 +1264,144 @@ def _generate_url(base, testbed, test_name): """ url = base - file_name = "" - anchor = ".html#" - feature = "" - - if "lbdpdk" in test_name or "lbvpp" in test_name: - file_name = "link_bonding" - - elif "114b" in test_name and "vhost" in test_name: - file_name = "vts" - - elif "testpmd" in test_name or "l3fwd" in test_name: - file_name = "dpdk" - - elif "memif" in test_name: - file_name = "container_memif" - feature = "-base" - - elif "srv6" in test_name: - file_name = "srv6" - - elif "vhost" in test_name: - if "l2xcbase" in test_name or "l2bdbasemaclrn" in test_name: - file_name = "vm_vhost_l2" - if "114b" in test_name: - feature = "" - elif "l2xcbase" in test_name and "x520" in test_name: - feature = "-base-l2xc" - elif "l2bdbasemaclrn" in test_name and "x520" in test_name: - feature = "-base-l2bd" + file_name = u"" + anchor = u".html#" + feature = u"" + + if u"lbdpdk" in test_name or u"lbvpp" in test_name: + file_name = u"link_bonding" + + elif u"114b" in test_name and u"vhost" in test_name: + file_name = u"vts" + + elif u"testpmd" in test_name or u"l3fwd" in test_name: + file_name = u"dpdk" + + elif u"memif" in test_name: + file_name = u"container_memif" + feature = u"-base" + + elif u"srv6" in test_name: + file_name = u"srv6" + + elif u"vhost" in test_name: + if u"l2xcbase" in test_name or u"l2bdbasemaclrn" in test_name: + file_name = u"vm_vhost_l2" + if u"114b" in test_name: + feature = u"" + elif u"l2xcbase" in test_name and u"x520" in test_name: + feature = u"-base-l2xc" + elif u"l2bdbasemaclrn" in test_name and u"x520" in test_name: + feature = u"-base-l2bd" else: - feature = "-base" - elif "ip4base" in test_name: - file_name = "vm_vhost_ip4" - feature = "-base" - - elif "ipsecbasetnlsw" in test_name: - file_name = "ipsecsw" - feature = "-base-scale" - - elif "ipsec" in test_name: - file_name = "ipsec" - feature = "-base-scale" - if "hw-" in test_name: - file_name = "ipsechw" - elif "sw-" in test_name: - file_name = "ipsecsw" - if "-int-" in test_name: - feature = "-base-scale-int" - elif "tnl" in test_name: - feature = "-base-scale-tnl" - - elif "ethip4lispip" in test_name or "ethip4vxlan" in test_name: - file_name = "ip4_tunnels" - feature = "-base" - - elif "ip4base" in test_name or "ip4scale" in test_name: - file_name = "ip4" - if "xl710" in test_name: - feature = "-base-scale-features" - elif "iacl" in test_name: - feature = "-features-iacl" - elif "oacl" in test_name: - feature = "-features-oacl" - elif "snat" in test_name or "cop" in test_name: - feature = "-features" + feature = u"-base" + elif u"ip4base" in test_name: + file_name = u"vm_vhost_ip4" + feature = u"-base" + + elif u"ipsecbasetnlsw" in test_name: + file_name = u"ipsecsw" + feature = u"-base-scale" + + elif u"ipsec" in test_name: + file_name = u"ipsec" + feature = u"-base-scale" + if u"hw-" in test_name: + file_name = u"ipsechw" + elif u"sw-" in test_name: + file_name = u"ipsecsw" + if u"-int-" in test_name: + feature = u"-base-scale-int" + elif u"tnl" in test_name: + feature = u"-base-scale-tnl" + + elif u"ethip4lispip" in test_name or u"ethip4vxlan" in test_name: + file_name = u"ip4_tunnels" + feature = u"-base" + + elif u"ip4base" in test_name or u"ip4scale" in test_name: + file_name = u"ip4" + if u"xl710" in test_name: + feature = u"-base-scale-features" + elif u"iacl" in test_name: + feature = u"-features-iacl" + elif u"oacl" in test_name: + feature = u"-features-oacl" + elif u"snat" in test_name or u"cop" in test_name: + feature = u"-features" else: - feature = "-base-scale" - - elif "ip6base" in test_name or "ip6scale" in test_name: - file_name = "ip6" - feature = "-base-scale" - - elif "l2xcbase" in test_name or "l2xcscale" in test_name \ - or "l2bdbasemaclrn" in test_name or "l2bdscale" in test_name \ - or "l2dbbasemaclrn" in test_name or "l2dbscale" in test_name: - file_name = "l2" - if "macip" in test_name: - feature = "-features-macip" - elif "iacl" in test_name: - feature = "-features-iacl" - elif "oacl" in test_name: - feature = "-features-oacl" + feature = u"-base-scale" + + elif u"ip6base" in test_name or u"ip6scale" in test_name: + file_name = u"ip6" + feature = u"-base-scale" + + elif u"l2xcbase" in test_name or u"l2xcscale" in test_name \ + or u"l2bdbasemaclrn" in test_name or u"l2bdscale" in test_name: + file_name = u"l2" + if u"macip" in test_name: + feature = u"-features-macip" + elif u"iacl" in test_name: + feature = u"-features-iacl" + elif u"oacl" in test_name: + feature = u"-features-oacl" else: - feature = "-base-scale" - - if "x520" in test_name: - nic = "x520-" - elif "x710" in test_name: - nic = "x710-" - elif "xl710" in test_name: - nic = "xl710-" - elif "xxv710" in test_name: - nic = "xxv710-" - elif "vic1227" in test_name: - nic = "vic1227-" - elif "vic1385" in test_name: - nic = "vic1385-" - elif "x553" in test_name: - nic = "x553-" + feature = u"-base-scale" + + if u"x520" in test_name: + nic = u"x520-" + elif u"x710" in test_name: + nic = u"x710-" + elif u"xl710" in test_name: + nic = u"xl710-" + elif u"xxv710" in test_name: + nic = u"xxv710-" + elif u"vic1227" in test_name: + nic = u"vic1227-" + elif u"vic1385" in test_name: + nic = u"vic1385-" + elif u"x553" in test_name: + nic = u"x553-" else: - nic = "" + nic = u"" anchor += nic - if "64b" in test_name: - framesize = "64b" - elif "78b" in test_name: - framesize = "78b" - elif "imix" in test_name: - framesize = "imix" - elif "9000b" in test_name: - framesize = "9000b" - elif "1518b" in test_name: - framesize = "1518b" - elif "114b" in test_name: - framesize = "114b" + if u"64b" in test_name: + framesize = u"64b" + elif u"78b" in test_name: + framesize = u"78b" + elif u"imix" in test_name: + framesize = u"imix" + elif u"9000b" in test_name: + framesize = u"9000b" + elif u"1518b" in test_name: + framesize = u"1518b" + elif u"114b" in test_name: + framesize = u"114b" else: - framesize = "" - anchor += framesize + '-' - - if "1t1c" in test_name: - anchor += "1t1c" - elif "2t2c" in test_name: - anchor += "2t2c" - elif "4t4c" in test_name: - anchor += "4t4c" - elif "2t1c" in test_name: - anchor += "2t1c" - elif "4t2c" in test_name: - anchor += "4t2c" - elif "8t4c" in test_name: - anchor += "8t4c" - - return url + file_name + '-' + testbed + '-' + nic + framesize + \ - feature.replace("-int", "").replace("-tnl", "") + anchor + feature - - -def table_performance_trending_dashboard_html(table, input_data): + framesize = u"" + anchor += framesize + u"-" + + if u"1t1c" in test_name: + anchor += u"1t1c" + elif u"2t2c" in test_name: + anchor += u"2t2c" + elif u"4t4c" in test_name: + anchor += u"4t4c" + elif u"2t1c" in test_name: + anchor += u"2t1c" + elif u"4t2c" in test_name: + anchor += u"4t2c" + elif u"8t4c" in test_name: + anchor += u"8t4c" + + return url + file_name + u"-" + testbed + u"-" + nic + framesize + \ + feature.replace("-int", u"").replace("-tnl", u"") + anchor + feature + + +def table_perf_trending_dash_html(table, input_data): """Generate the table(s) with algorithm: - table_performance_trending_dashboard_html specified in the specification + table_perf_trending_dash_html specified in the specification file. :param table: Table to generate. @@ -1339,70 +1410,97 @@ def table_performance_trending_dashboard_html(table, input_data): :type input_data: InputData """ - testbed = table.get("testbed", None) - if testbed is None: - logging.error("The testbed is not defined for the table '{0}'.". - format(table.get("title", ""))) + _ = input_data + + if not table.get(u"testbed", None): + logging.error( + f"The testbed is not defined for the table " + f"{table.get(u'title', u'')}." + ) return - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") try: - with open(table["input-file"], 'rb') as csv_file: - csv_content = csv.reader(csv_file, delimiter=',', quotechar='"') - csv_lst = [item for item in csv_content] + with open(table[u"input-file"], u'rt') as csv_file: + csv_lst = list(csv.reader(csv_file, delimiter=u',', quotechar=u'"')) except KeyError: - logging.warning("The input file is not defined.") + logging.warning(u"The input file is not defined.") return except csv.Error as err: - logging.warning("Not possible to process the file '{0}'.\n{1}". - format(table["input-file"], err)) + logging.warning( + f"Not possible to process the file {table[u'input-file']}.\n" + f"{repr(err)}" + ) return # Table: - dashboard = ET.Element("table", attrib=dict(width="100%", border='0')) + dashboard = ET.Element(u"table", attrib=dict(width=u"100%", border=u'0')) # Table header: - tr = ET.SubElement(dashboard, "tr", attrib=dict(bgcolor="#7eade7")) + trow = ET.SubElement(dashboard, u"tr", attrib=dict(bgcolor=u"#7eade7")) for idx, item in enumerate(csv_lst[0]): - alignment = "left" if idx == 0 else "center" - th = ET.SubElement(tr, "th", attrib=dict(align=alignment)) - th.text = item + alignment = u"left" if idx == 0 else u"center" + thead = ET.SubElement(trow, u"th", attrib=dict(align=alignment)) + thead.text = item # Rows: - colors = {"regression": ("#ffcccc", "#ff9999"), - "progression": ("#c6ecc6", "#9fdf9f"), - "normal": ("#e9f1fb", "#d4e4f7")} + colors = { + u"regression": ( + u"#ffcccc", + u"#ff9999" + ), + u"progression": ( + u"#c6ecc6", + u"#9fdf9f" + ), + u"normal": ( + u"#e9f1fb", + u"#d4e4f7" + ) + } for r_idx, row in enumerate(csv_lst[1:]): if int(row[4]): - color = "regression" + color = u"regression" elif int(row[5]): - color = "progression" + color = u"progression" else: - color = "normal" - background = colors[color][r_idx % 2] - tr = ET.SubElement(dashboard, "tr", attrib=dict(bgcolor=background)) + color = u"normal" + trow = ET.SubElement( + dashboard, u"tr", attrib=dict(bgcolor=colors[color][r_idx % 2]) + ) # Columns: for c_idx, item in enumerate(row): - alignment = "left" if c_idx == 0 else "center" - td = ET.SubElement(tr, "td", attrib=dict(align=alignment)) + tdata = ET.SubElement( + trow, + u"td", + attrib=dict(align=u"left" if c_idx == 0 else u"center") + ) # Name: if c_idx == 0: - url = _generate_url("../trending/", testbed, item) - ref = ET.SubElement(td, "a", attrib=dict(href=url)) + ref = ET.SubElement( + tdata, + u"a", + attrib=dict( + href=_generate_url( + u"../trending/", + table.get(u"testbed", None), + item + ) + ) + ) ref.text = item else: - td.text = item + tdata.text = item try: - with open(table["output-file"], 'w') as html_file: - logging.info(" Writing file: '{0}'".format(table["output-file"])) - html_file.write(".. raw:: html\n\n\t") - html_file.write(ET.tostring(dashboard)) - html_file.write("\n\t<p><br><br></p>\n") + with open(table[u"output-file"], u'w') as html_file: + logging.info(f" Writing file: {table[u'output-file']}") + html_file.write(u".. raw:: html\n\n\t") + html_file.write(str(ET.tostring(dashboard, encoding=u"unicode"))) + html_file.write(u"\n\t<p><br><br></p>\n") except KeyError: - logging.warning("The output file is not defined.") + logging.warning(u"The output file is not defined.") return @@ -1416,53 +1514,56 @@ def table_last_failed_tests(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) + data = input_data.filter_data(table, continue_on_error=True) if data is None or data.empty: - logging.warn(" No data for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.warning( + f" No data for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) return tbl_list = list() - for job, builds in table["data"].items(): + for job, builds in table[u"data"].items(): for build in builds: build = str(build) try: - version = input_data.metadata(job, build).get("version", "") + version = input_data.metadata(job, build).get(u"version", u"") except KeyError: - logging.error("Data for {job}: {build} is not present.". - format(job=job, build=build)) + logging.error(f"Data for {job}: {build} is not present.") return tbl_list.append(build) tbl_list.append(version) failed_tests = list() passed = 0 failed = 0 - for tst_name, tst_data in data[job][build].iteritems(): - if tst_data["status"] != "FAIL": + for tst_data in data[job][build].values: + if tst_data[u"status"] != u"FAIL": passed += 1 continue failed += 1 - groups = re.search(REGEX_NIC, tst_data["parent"]) + groups = re.search(REGEX_NIC, tst_data[u"parent"]) if not groups: continue nic = groups.group(0) - failed_tests.append("{0}-{1}".format(nic, tst_data["name"])) + failed_tests.append(f"{nic}-{tst_data[u'name']}") tbl_list.append(str(passed)) tbl_list.append(str(failed)) tbl_list.extend(failed_tests) - file_name = "{0}{1}".format(table["output-file"], table["output-file-ext"]) - logging.info(" Writing file: '{0}'".format(file_name)) - with open(file_name, "w") as file_handler: + file_name = f"{table[u'output-file']}{table[u'output-file-ext']}" + logging.info(f" Writing file: {file_name}") + with open(file_name, u"w") as file_handler: for test in tbl_list: - file_handler.write(test + '\n') + file_handler.write(test + u'\n') def table_failed_tests(table, input_data): @@ -1475,95 +1576,103 @@ def table_failed_tests(table, input_data): :type input_data: InputData """ - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") # Transform the data - logging.info(" Creating the data set for the {0} '{1}'.". - format(table.get("type", ""), table.get("title", ""))) + logging.info( + f" Creating the data set for the {table.get(u'type', u'')} " + f"{table.get(u'title', u'')}." + ) data = input_data.filter_data(table, continue_on_error=True) # Prepare the header of the tables - header = ["Test Case", - "Failures [#]", - "Last Failure [Time]", - "Last Failure [VPP-Build-Id]", - "Last Failure [CSIT-Job-Build-Id]"] + header = [ + u"Test Case", + u"Failures [#]", + u"Last Failure [Time]", + u"Last Failure [VPP-Build-Id]", + u"Last Failure [CSIT-Job-Build-Id]" + ] # Generate the data for the table according to the model in the table # specification now = dt.utcnow() - timeperiod = timedelta(int(table.get("window", 7))) + timeperiod = timedelta(int(table.get(u"window", 7))) tbl_dict = dict() - for job, builds in table["data"].items(): + for job, builds in table[u"data"].items(): for build in builds: build = str(build) - for tst_name, tst_data in data[job][build].iteritems(): - if tst_name.lower() in table.get("ignore-list", list()): + for tst_name, tst_data in data[job][build].items(): + if tst_name.lower() in table.get(u"ignore-list", list()): continue if tbl_dict.get(tst_name, None) is None: - groups = re.search(REGEX_NIC, tst_data["parent"]) + groups = re.search(REGEX_NIC, tst_data[u"parent"]) if not groups: continue nic = groups.group(0) tbl_dict[tst_name] = { - "name": "{0}-{1}".format(nic, tst_data["name"]), - "data": OrderedDict()} + u"name": f"{nic}-{tst_data[u'name']}", + u"data": OrderedDict() + } try: generated = input_data.metadata(job, build).\ - get("generated", "") + get(u"generated", u"") if not generated: continue - then = dt.strptime(generated, "%Y%m%d %H:%M") + then = dt.strptime(generated, u"%Y%m%d %H:%M") if (now - then) <= timeperiod: - tbl_dict[tst_name]["data"][build] = ( - tst_data["status"], + tbl_dict[tst_name][u"data"][build] = ( + tst_data[u"status"], generated, - input_data.metadata(job, build).get("version", ""), - build) + input_data.metadata(job, build).get(u"version", + u""), + build + ) except (TypeError, KeyError) as err: - logging.warning("tst_name: {} - err: {}". - format(tst_name, repr(err))) + logging.warning(f"tst_name: {tst_name} - err: {repr(err)}") max_fails = 0 tbl_lst = list() for tst_data in tbl_dict.values(): fails_nr = 0 - fails_last_date = "" - fails_last_vpp = "" - fails_last_csit = "" - for val in tst_data["data"].values(): - if val[0] == "FAIL": + fails_last_date = u"" + fails_last_vpp = u"" + fails_last_csit = u"" + for val in tst_data[u"data"].values(): + if val[0] == u"FAIL": fails_nr += 1 fails_last_date = val[1] fails_last_vpp = val[2] fails_last_csit = val[3] if fails_nr: max_fails = fails_nr if fails_nr > max_fails else max_fails - tbl_lst.append([tst_data["name"], - fails_nr, - fails_last_date, - fails_last_vpp, - "mrr-daily-build-{0}".format(fails_last_csit)]) + tbl_lst.append( + [ + tst_data[u"name"], + fails_nr, + fails_last_date, + fails_last_vpp, + f"mrr-daily-build-{fails_last_csit}" + ] + ) tbl_lst.sort(key=lambda rel: rel[2], reverse=True) tbl_sorted = list() for nrf in range(max_fails, -1, -1): tbl_fails = [item for item in tbl_lst if item[1] == nrf] tbl_sorted.extend(tbl_fails) - file_name = "{0}{1}".format(table["output-file"], table["output-file-ext"]) - logging.info(" Writing file: '{0}'".format(file_name)) - with open(file_name, "w") as file_handler: - file_handler.write(",".join(header) + "\n") + file_name = f"{table[u'output-file']}{table[u'output-file-ext']}" + logging.info(f" Writing file: {file_name}") + with open(file_name, u"w") as file_handler: + file_handler.write(u",".join(header) + u"\n") for test in tbl_sorted: - file_handler.write(",".join([str(item) for item in test]) + '\n') + file_handler.write(u",".join([str(item) for item in test]) + u'\n') - txt_file_name = "{0}.txt".format(table["output-file"]) - logging.info(" Writing file: '{0}'".format(txt_file_name)) - convert_csv_to_pretty_txt(file_name, txt_file_name) + logging.info(f" Writing file: {table[u'output-file']}.txt") + convert_csv_to_pretty_txt(file_name, f"{table[u'output-file']}.txt") def table_failed_tests_html(table, input_data): @@ -1576,60 +1685,77 @@ def table_failed_tests_html(table, input_data): :type input_data: InputData """ - testbed = table.get("testbed", None) - if testbed is None: - logging.error("The testbed is not defined for the table '{0}'.". - format(table.get("title", ""))) + _ = input_data + + if not table.get(u"testbed", None): + logging.error( + f"The testbed is not defined for the table " + f"{table.get(u'title', u'')}." + ) return - logging.info(" Generating the table {0} ...". - format(table.get("title", ""))) + logging.info(f" Generating the table {table.get(u'title', u'')} ...") try: - with open(table["input-file"], 'rb') as csv_file: - csv_content = csv.reader(csv_file, delimiter=',', quotechar='"') - csv_lst = [item for item in csv_content] + with open(table[u"input-file"], u'rt') as csv_file: + csv_lst = list(csv.reader(csv_file, delimiter=u',', quotechar=u'"')) except KeyError: - logging.warning("The input file is not defined.") + logging.warning(u"The input file is not defined.") return except csv.Error as err: - logging.warning("Not possible to process the file '{0}'.\n{1}". - format(table["input-file"], err)) + logging.warning( + f"Not possible to process the file {table[u'input-file']}.\n" + f"{repr(err)}" + ) return # Table: - failed_tests = ET.Element("table", attrib=dict(width="100%", border='0')) + failed_tests = ET.Element(u"table", attrib=dict(width=u"100%", border=u'0')) # Table header: - tr = ET.SubElement(failed_tests, "tr", attrib=dict(bgcolor="#7eade7")) + trow = ET.SubElement(failed_tests, u"tr", attrib=dict(bgcolor=u"#7eade7")) for idx, item in enumerate(csv_lst[0]): - alignment = "left" if idx == 0 else "center" - th = ET.SubElement(tr, "th", attrib=dict(align=alignment)) - th.text = item + alignment = u"left" if idx == 0 else u"center" + thead = ET.SubElement(trow, u"th", attrib=dict(align=alignment)) + thead.text = item # Rows: - colors = ("#e9f1fb", "#d4e4f7") + colors = (u"#e9f1fb", u"#d4e4f7") for r_idx, row in enumerate(csv_lst[1:]): background = colors[r_idx % 2] - tr = ET.SubElement(failed_tests, "tr", attrib=dict(bgcolor=background)) + trow = ET.SubElement( + failed_tests, u"tr", attrib=dict(bgcolor=background) + ) # Columns: for c_idx, item in enumerate(row): - alignment = "left" if c_idx == 0 else "center" - td = ET.SubElement(tr, "td", attrib=dict(align=alignment)) + tdata = ET.SubElement( + trow, + u"td", + attrib=dict(align=u"left" if c_idx == 0 else u"center") + ) # Name: if c_idx == 0: - url = _generate_url("../trending/", testbed, item) - ref = ET.SubElement(td, "a", attrib=dict(href=url)) + ref = ET.SubElement( + tdata, + u"a", + attrib=dict( + href=_generate_url( + u"../trending/", + table.get(u"testbed", None), + item + ) + ) + ) ref.text = item else: - td.text = item + tdata.text = item try: - with open(table["output-file"], 'w') as html_file: - logging.info(" Writing file: '{0}'".format(table["output-file"])) - html_file.write(".. raw:: html\n\n\t") - html_file.write(ET.tostring(failed_tests)) - html_file.write("\n\t<p><br><br></p>\n") + with open(table[u"output-file"], u'w') as html_file: + logging.info(f" Writing file: {table[u'output-file']}") + html_file.write(u".. raw:: html\n\n\t") + html_file.write(str(ET.tostring(failed_tests, encoding=u"unicode"))) + html_file.write(u"\n\t<p><br><br></p>\n") except KeyError: - logging.warning("The output file is not defined.") + logging.warning(u"The output file is not defined.") return diff --git a/resources/tools/presentation/input_data_files.py b/resources/tools/presentation/input_data_files.py index 0a723cedc7..b1bb8cd20f 100644 --- a/resources/tools/presentation/input_data_files.py +++ b/resources/tools/presentation/input_data_files.py @@ -16,26 +16,27 @@ Download all data. """ import re -import requests import logging from os import rename, mkdir from os.path import join +from http.client import responses from zipfile import ZipFile, is_zipfile, BadZipfile -from httplib import responses -from requests.adapters import HTTPAdapter -from requests.packages.urllib3.util.retry import Retry -from requests import codes, RequestException, Timeout, TooManyRedirects, \ - HTTPError, ConnectionError -from errors import PresentationError +import requests + +from requests.adapters import HTTPAdapter, Retry +from requests.exceptions import RequestException +from requests import codes + +from pal_errors import PresentationError # Chunk size used for file download CHUNK_SIZE = 512 # Separator used in file names -SEPARATOR = "__" +SEPARATOR = u"__" REGEX_RELEASE = re.compile(r'(\D*)(\d{4}|master)(\D*)') @@ -81,79 +82,70 @@ def _download_file(url, file_name, log, arch=False): ) adapter = HTTPAdapter(max_retries=retry) session = requests.Session() - session.mount('http://', adapter) - session.mount('https://', adapter) + session.mount(u"http://", adapter) + session.mount(u"https://", adapter) return session success = False session = None try: - log.append(("INFO", " Connecting to '{0}' ...".format(url))) + log.append((u"INFO", f" Connecting to {url} ...")) session = requests_retry_session() response = session.get(url, stream=True) code = response.status_code - log.append(("INFO", " {0}: {1}".format(code, responses[code]))) + log.append((u"INFO", f" {code}: {responses[code]}")) - if code != codes["OK"]: + if code != codes[u"OK"]: if session: session.close() - url = url.replace("_info", "") - log.append(("INFO", " Connecting to '{0}' ...".format(url))) + url = url.replace(u"_info", u"") + log.append((u"INFO", f" Connecting to {url} ...")) session = requests_retry_session() response = session.get(url, stream=True) code = response.status_code - log.append(("INFO", " {0}: {1}".format(code, responses[code]))) - if code != codes["OK"]: + log.append((u"INFO", f" {code}: {responses[code]}")) + if code != codes[u"OK"]: return False, file_name - file_name = file_name.replace("_info", "") + file_name = file_name.replace(u"_info", u"") - dst_file_name = file_name.replace(".gz", "") - log.append(("INFO", " Downloading the file '{0}' to '{1}' ...". - format(url, dst_file_name))) - with open(dst_file_name, "wb") as file_handle: + dst_file_name = file_name.replace(u".gz", u"") + log.append( + (u"INFO", f" Downloading the file {url} to {dst_file_name} ...") + ) + with open(dst_file_name, u"wb") as file_handle: for chunk in response.iter_content(chunk_size=CHUNK_SIZE): if chunk: file_handle.write(chunk) - if arch and ".gz" in file_name: + if arch and u".gz" in file_name: if session: session.close() - log.append(("INFO", " Downloading the file '{0}' to '{1}' ...". - format(url, file_name))) + log.append( + (u"INFO", f" Downloading the file {url} to {file_name} ...") + ) session = requests_retry_session() response = session.get(url, stream=True) - if response.status_code == codes["OK"]: - with open(file_name, "wb") as file_handle: + if response.status_code == codes[u"OK"]: + with open(file_name, u"wb") as file_handle: file_handle.write(response.raw.read()) else: - log.append(("ERROR", "Not possible to download the file '{0}' " - "to '{1}' ...".format(url, file_name))) + log.append( + (u"ERROR", f"Not possible to download the file {url} to " + f"{file_name} ...") + ) success = True - except ConnectionError as err: - log.append(("ERROR", "Not possible to connect to '{0}'.".format(url))) - log.append(("DEBUG", repr(err))) - except HTTPError as err: - log.append(("ERROR", "Invalid HTTP response from '{0}'.".format(url))) - log.append(("DEBUG", repr(err))) - except TooManyRedirects as err: - log.append(("ERROR", "Request exceeded the configured number " - "of maximum re-directions.")) - log.append(("DEBUG", repr(err))) - except Timeout as err: - log.append(("ERROR", "Request timed out.")) - log.append(("DEBUG", repr(err))) except RequestException as err: - log.append(("ERROR", "Unexpected HTTP request exception.")) - log.append(("DEBUG", repr(err))) + log.append( + (u"ERROR", f"HTTP Request exception:\n{repr(err)}") + ) except (IOError, ValueError, KeyError) as err: - log.append(("ERROR", "Download failed.")) - log.append(("DEBUG", repr(err))) + log.append((u"ERROR", f"Download failed.\n{repr(err)}")) finally: if session: session.close() - log.append(("INFO", " Download finished.")) + log.append((u"INFO", u" Download finished.")) return success, file_name @@ -170,39 +162,41 @@ def _unzip_file(spec, build, pid, log): :rtype: bool """ - file_name = build["file-name"] - if ".zip" in file_name: - data_file = spec.input["zip-extract"] + file_name = build[u"file-name"] + if u".zip" in file_name: + data_file = spec.input[u"zip-extract"] else: - data_file = spec.input["extract"] + data_file = spec.input[u"extract"] - directory = spec.environment["paths"]["DIR[WORKING,DATA]"] + directory = spec.environment[u"paths"][u"DIR[WORKING,DATA]"] tmp_dir = join(directory, str(pid)) try: mkdir(tmp_dir) except OSError: pass - new_name = "{0}{1}{2}".format(file_name.rsplit('.')[-2], - SEPARATOR, - data_file.split("/")[-1]) + new_name = \ + f"{file_name.rsplit(u'.')[-2]}{SEPARATOR}{data_file.split(u'/')[-1]}" - log.append(("INFO", " Unzipping: '{0}' from '{1}'.". - format(data_file, file_name))) + log.append((u"INFO", f" Unzipping: {data_file} from {file_name}.")) try: - with ZipFile(file_name, 'r') as zip_file: + with ZipFile(file_name, u'r') as zip_file: zip_file.extract(data_file, tmp_dir) - log.append(("INFO", " Renaming the file '{0}' to '{1}'". - format(join(tmp_dir, data_file), new_name))) + log.append( + (u"INFO", f" Renaming the file {join(tmp_dir, data_file)} to " + f"{new_name}") + ) rename(join(tmp_dir, data_file), new_name) - build["file-name"] = new_name + build[u"file-name"] = new_name return True except (BadZipfile, RuntimeError) as err: - log.append(("ERROR", "Failed to unzip the file '{0}': {1}.". - format(file_name, str(err)))) + log.append( + (u"ERROR", f"Failed to unzip the file {file_name}: {repr(err)}.") + ) return False except OSError as err: - log.append(("ERROR", "Failed to rename the file '{0}': {1}.". - format(data_file, str(err)))) + log.append( + (u"ERROR", f"Failed to rename the file {data_file}: {repr(err)}.") + ) return False @@ -225,67 +219,69 @@ def download_and_unzip_data_file(spec, job, build, pid, log): # Try to download .gz from logs.fd.io - file_name = spec.input["file-name"] - url = "{0}/{1}".format( - spec.environment["urls"]["URL[NEXUS,LOG]"], - spec.input["download-path"].format( - job=job, build=build["build"], filename=file_name)) - new_name = join(spec.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}".format( - job=job, sep=SEPARATOR, build=build["build"], - name=file_name)) + file_name = spec.input[u"file-name"] + url = u"{0}/{1}".format( + spec.environment[u'urls'][u'URL[NEXUS,LOG]'], + spec.input[u'download-path'].format( + job=job, build=build[u'build'], filename=file_name + ) + ) + new_name = join( + spec.environment[u"paths"][u"DIR[WORKING,DATA]"], + f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}" + ) - logging.info("Trying to download {0}".format(url)) + logging.info(f"Trying to download {url}") - arch = True if spec.configuration.get("archive-inputs", True) else False + arch = bool(spec.configuration.get(u"archive-inputs", True)) success, downloaded_name = _download_file(url, new_name, log, arch=arch) if not success: # Try to download .gz from docs.fd.io - file_name = spec.input["file-name"] - url = "{0}/{1}".format( - spec.environment["urls"]["URL[NEXUS,DOC]"], - spec.input["download-path"].format( - job=job, build=build["build"], filename=file_name)) - new_name = join(spec.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}".format( - job=job, sep=SEPARATOR, build=build["build"], - name=file_name)) + file_name = spec.input[u"file-name"] + url = u"{0}/{1}".format( + spec.environment[u"urls"][u"URL[NEXUS,DOC]"], + spec.input[u"download-path"].format( + job=job, build=build[u"build"], filename=file_name + ) + ) + new_name = join( + spec.environment[u"paths"][u"DIR[WORKING,DATA]"], + f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}" + ) - logging.info("Downloading {0}".format(url)) + logging.info(f"Downloading {url}") - if spec.configuration.get("archive-inputs", True): - arch = True success, downloaded_name = _download_file(url, new_name, log, arch=arch) if not success: # Try to download .zip from docs.fd.io - file_name = spec.input["zip-file-name"] - new_name = join(spec.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}".format( - job=job, sep=SEPARATOR, build=build["build"], - name=file_name)) + file_name = spec.input[u"zip-file-name"] + new_name = join( + spec.environment[u"paths"][u"DIR[WORKING,DATA]"], + f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}" + ) release = re.search(REGEX_RELEASE, job).group(2) - for rls in (release, "master"): - nexus_file_name = "{job}{sep}{build}{sep}{name}". \ - format(job=job, sep=SEPARATOR, build=build["build"], - name=file_name) + for rls in (release, u"master"): + nexus_file_name = \ + f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}" try: - rls = "rls{0}".format(int(rls)) + rls = f"rls{int(rls)}" except ValueError: - # It is 'master' + # It is master pass - url = "{url}/{release}/{dir}/{file}". \ - format(url=spec.environment["urls"]["URL[NEXUS,DOC]"], - release=rls, - dir=spec.environment["urls"]["DIR[NEXUS,DOC]"], - file=nexus_file_name) + url = ( + f"{spec.environment[u'urls'][u'URL[NEXUS,DOC]']}/" + f"{rls}/" + f"{spec.environment[u'urls'][u'DIR[NEXUS,DOC]']}/" + f"{nexus_file_name}" + ) - logging.info("Downloading {0}".format(url)) + logging.info(f"Downloading {url}") success, downloaded_name = _download_file(url, new_name, log) if success: @@ -295,41 +291,40 @@ def download_and_unzip_data_file(spec, job, build, pid, log): # Try to download .zip from jenkins.fd.io - file_name = spec.input["zip-file-name"] - download_path = spec.input["zip-download-path"] - if job.startswith("csit-"): - url = spec.environment["urls"]["URL[JENKINS,CSIT]"] - elif job.startswith("hc2vpp-"): - url = spec.environment["urls"]["URL[JENKINS,HC]"] + file_name = spec.input[u"zip-file-name"] + download_path = spec.input[u"zip-download-path"] + if job.startswith(u"csit-"): + url = spec.environment[u"urls"][u"URL[JENKINS,CSIT]"] + elif job.startswith(u"hc2vpp-"): + url = spec.environment[u"urls"][u"URL[JENKINS,HC]"] else: - raise PresentationError( - "No url defined for the job '{}'.".format(job)) + raise PresentationError(f"No url defined for the job {job}.") full_name = download_path.format( - job=job, build=build["build"], filename=file_name) - url = "{0}/{1}".format(url, full_name) - new_name = join(spec.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}". - format(job=job, sep=SEPARATOR, build=build["build"], - name=file_name)) + job=job, build=build[u"build"], filename=file_name + ) + url = u"{0}/{1}".format(url, full_name) + new_name = join( + spec.environment[u"paths"][u"DIR[WORKING,DATA]"], + f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}" + ) - logging.info("Downloading {0}".format(url)) + logging.info(f"Downloading {url}") success, downloaded_name = _download_file(url, new_name, log) - if success and downloaded_name.endswith(".zip"): + if success and downloaded_name.endswith(u".zip"): if not is_zipfile(downloaded_name): - log.append(("ERROR", - "Zip file '{0}' is corrupted.".format(new_name))) + log.append((u"ERROR", f"Zip file {new_name} is corrupted.")) success = False if success: - build["file-name"] = downloaded_name + build[u"file-name"] = downloaded_name - if file_name.endswith(".gz"): - build["file-name"] = downloaded_name[:-3] + if file_name.endswith(u".gz"): + build[u"file-name"] = downloaded_name[:-3] - if downloaded_name.endswith(".zip"): + if downloaded_name.endswith(u".zip"): success = _unzip_file(spec, build, pid, log) return success diff --git a/resources/tools/presentation/input_data_parser.py b/resources/tools/presentation/input_data_parser.py index e48b271489..1177defe35 100644 --- a/resources/tools/presentation/input_data_parser.py +++ b/resources/tools/presentation/input_data_parser.py @@ -19,28 +19,29 @@ - filter the data using tags, """ -import copy import re +import copy import resource -import pandas as pd import logging -import prettytable -from robot.api import ExecutionResult, ResultVisitor -from robot import errors from collections import OrderedDict -from string import replace from os import remove from datetime import datetime as dt from datetime import timedelta from json import loads +import prettytable +import pandas as pd + +from robot.api import ExecutionResult, ResultVisitor +from robot import errors + from resources.libraries.python import jumpavg from input_data_files import download_and_unzip_data_file # Separator used in file names -SEPARATOR = "__" +SEPARATOR = u"__" class ExecutionChecker(ResultVisitor): @@ -200,9 +201,6 @@ class ExecutionChecker(ResultVisitor): .. note:: ID is the lowercase full path to the test. """ - # TODO: Remove when definitely no NDRPDRDISC tests are used: - REGEX_RATE = re.compile(r'^[\D\d]*FINAL_RATE:\s(\d+\.\d+)\s(\w+)') - REGEX_PLR_RATE = re.compile(r'PLRsearch lower bound::?\s(\d+.\d+).*\n' r'PLRsearch upper bound::?\s(\d+.\d+)') @@ -274,10 +272,10 @@ class ExecutionChecker(ResultVisitor): # Ignore list self._ignore = ignore - # Number of VAT History messages found: + # Number of PAPI History messages found: # 0 - no message - # 1 - VAT History of DUT1 - # 2 - VAT History of DUT2 + # 1 - PAPI History of DUT1 + # 2 - PAPI History of DUT2 self._lookup_kw_nr = 0 self._conf_history_lookup_nr = 0 @@ -288,29 +286,30 @@ class ExecutionChecker(ResultVisitor): # Test ID of currently processed test- the lowercase full path to the # test - self._test_ID = None + self._test_id = None # The main data structure self._data = { - "metadata": OrderedDict(), - "suites": OrderedDict(), - "tests": OrderedDict() + u"metadata": OrderedDict(), + u"suites": OrderedDict(), + u"tests": OrderedDict() } # Save the provided metadata for key, val in metadata.items(): - self._data["metadata"][key] = val + self._data[u"metadata"][key] = val # Dictionary defining the methods used to parse different types of # messages self.parse_msg = { - "timestamp": self._get_timestamp, - "vpp-version": self._get_vpp_version, - "dpdk-version": self._get_dpdk_version, - "teardown-vat-history": self._get_vat_history, - "teardown-papi-history": self._get_papi_history, - "test-show-runtime": self._get_show_run, - "testbed": self._get_testbed + u"timestamp": self._get_timestamp, + u"vpp-version": self._get_vpp_version, + u"dpdk-version": self._get_dpdk_version, + # TODO: Remove when not needed: + u"teardown-vat-history": self._get_vat_history, + u"teardown-papi-history": self._get_papi_history, + u"test-show-runtime": self._get_show_run, + u"testbed": self._get_testbed } @property @@ -331,15 +330,16 @@ class ExecutionChecker(ResultVisitor): :returns: Nothing. """ - if msg.message.count("Setup of TG node"): + if msg.message.count(u"Setup of TG node") or \ + msg.message.count(u"Setup of node TG host"): reg_tg_ip = re.compile( - r'Setup of TG node (\d{1,3}.\d{1,3}.\d{1,3}.\d{1,3}) done') + r'.*TG .* (\d{1,3}.\d{1,3}.\d{1,3}.\d{1,3}).*') try: self._testbed = str(re.search(reg_tg_ip, msg.message).group(1)) except (KeyError, ValueError, IndexError, AttributeError): pass finally: - self._data["metadata"]["testbed"] = self._testbed + self._data[u"metadata"][u"testbed"] = self._testbed self._msg_type = None def _get_vpp_version(self, msg): @@ -350,12 +350,12 @@ class ExecutionChecker(ResultVisitor): :returns: Nothing. """ - if msg.message.count("return STDOUT Version:") or \ - msg.message.count("VPP Version:") or \ - msg.message.count("VPP version:"): + if msg.message.count(u"return STDOUT Version:") or \ + msg.message.count(u"VPP Version:") or \ + msg.message.count(u"VPP version:"): self._version = str(re.search(self.REGEX_VERSION_VPP, msg.message). group(2)) - self._data["metadata"]["version"] = self._version + self._data[u"metadata"][u"version"] = self._version self._msg_type = None def _get_dpdk_version(self, msg): @@ -366,11 +366,11 @@ class ExecutionChecker(ResultVisitor): :returns: Nothing. """ - if msg.message.count("DPDK Version:"): + if msg.message.count(u"DPDK Version:"): try: self._version = str(re.search( - self.REGEX_VERSION_DPDK, msg.message). group(2)) - self._data["metadata"]["version"] = self._version + self.REGEX_VERSION_DPDK, msg.message).group(2)) + self._data[u"metadata"][u"version"] = self._version except IndexError: pass finally: @@ -385,30 +385,32 @@ class ExecutionChecker(ResultVisitor): """ self._timestamp = msg.timestamp[:14] - self._data["metadata"]["generated"] = self._timestamp + self._data[u"metadata"][u"generated"] = self._timestamp self._msg_type = None def _get_vat_history(self, msg): """Called when extraction of VAT command history is required. + TODO: Remove when not needed. + :param msg: Message to process. :type msg: Message :returns: Nothing. """ - if msg.message.count("VAT command history:"): + if msg.message.count(u"VAT command history:"): self._conf_history_lookup_nr += 1 if self._conf_history_lookup_nr == 1: - self._data["tests"][self._test_ID]["conf-history"] = str() + self._data[u"tests"][self._test_id][u"conf-history"] = str() else: self._msg_type = None - text = re.sub("\d{1,3}.\d{1,3}.\d{1,3}.\d{1,3} " - "VAT command history:", "", msg.message, count=1). \ - replace("\n\n", "\n").replace('\n', ' |br| ').\ - replace('\r', '').replace('"', "'") + text = re.sub(r"\d{1,3}.\d{1,3}.\d{1,3}.\d{1,3} " + r"VAT command history:", u"", msg.message, count=1). \ + replace(u"\n\n", u"\n").replace(u'\n', u' |br| ').\ + replace(u'\r', u'').replace(u'"', u"'") - self._data["tests"][self._test_ID]["conf-history"] += " |br| " - self._data["tests"][self._test_ID]["conf-history"] += \ - "**DUT" + str(self._conf_history_lookup_nr) + ":** " + text + self._data[u"tests"][self._test_id][u"conf-history"] += ( + f" |br| **DUT{str(self._conf_history_lookup_nr)}:** {text}" + ) def _get_papi_history(self, msg): """Called when extraction of PAPI command history is required. @@ -417,20 +419,21 @@ class ExecutionChecker(ResultVisitor): :type msg: Message :returns: Nothing. """ - if msg.message.count("PAPI command history:"): + if msg.message.count(u"PAPI command history:"): self._conf_history_lookup_nr += 1 if self._conf_history_lookup_nr == 1: - self._data["tests"][self._test_ID]["conf-history"] = str() + self._data[u"tests"][self._test_id][u"conf-history"] = str() else: self._msg_type = None - text = re.sub("\d{1,3}.\d{1,3}.\d{1,3}.\d{1,3} " - "PAPI command history:", "", msg.message, count=1). \ - replace("\n\n", "\n").replace('\n', ' |br| ').\ - replace('\r', '').replace('"', "'") + text = re.sub(r"\d{1,3}.\d{1,3}.\d{1,3}.\d{1,3} " + r"PAPI command history:", u"", + msg.message, count=1). \ + replace(u"\n\n", u"\n").replace(u'\n', u' |br| ').\ + replace(u'\r', u'').replace(u'"', u"'") - self._data["tests"][self._test_ID]["conf-history"] += " |br| " - self._data["tests"][self._test_ID]["conf-history"] += \ - "**DUT" + str(self._conf_history_lookup_nr) + ":** " + text + self._data[u"tests"][self._test_id][u"conf-history"] += ( + f" |br| **DUT{str(self._conf_history_lookup_nr)}:** {text}" + ) def _get_show_run(self, msg): """Called when extraction of VPP operational data (output of CLI command @@ -440,56 +443,66 @@ class ExecutionChecker(ResultVisitor): :type msg: Message :returns: Nothing. """ - if not "show-run" in self._data["tests"][self._test_ID].keys(): - self._data["tests"][self._test_ID]["show-run"] = str() + if u"show-run" not in self._data[u"tests"][self._test_id].keys(): + self._data[u"tests"][self._test_id][u"show-run"] = str() - if msg.message.count("stats runtime"): - host = str(re.search(self.REGEX_TC_PAPI_CLI, msg.message).\ + if msg.message.count(u"stats runtime"): + host = str(re.search(self.REGEX_TC_PAPI_CLI, msg.message). group(1)) - socket = str(re.search(self.REGEX_TC_PAPI_CLI, msg.message).\ + socket = str(re.search(self.REGEX_TC_PAPI_CLI, msg.message). group(2)) - message = str(msg.message).replace(' ', '').replace('\n', '').\ - replace("'", '"').replace('b"', '"').replace('u"', '"').\ - split(":",1)[1] - runtime = loads(message) + runtime = loads( + str(msg.message). + replace(u' ', u''). + replace(u'\n', u''). + replace(u"'", u'"'). + replace(u'b"', u'"'). + replace(u'u"', u'"'). + split(u":", 1)[1] + ) try: - threads_nr = len(runtime[0]["clocks"]) + threads_nr = len(runtime[0][u"clocks"]) except (IndexError, KeyError): return - tbl_hdr = ["Name", "Calls", "Vectors", "Suspends", "Clocks", - "Vectors/Calls"] + tbl_hdr = [ + u"Name", + u"Calls", + u"Vectors", + u"Suspends", + u"Clocks", + u"Vectors/Calls" + ] table = [[tbl_hdr, ] for _ in range(threads_nr)] for item in runtime: for idx in range(threads_nr): - name = format(item["name"]) - calls = format(item["calls"][idx]) - vectors = format(item["vectors"][idx]) - suspends = format(item["suspends"][idx]) - if item["vectors"][idx] > 0: + name = format(item[u"name"]) + calls = format(item[u"calls"][idx]) + vectors = format(item[u"vectors"][idx]) + suspends = format(item[u"suspends"][idx]) + if item[u"vectors"][idx] > 0: clocks = format( - item["clocks"][idx]/item["vectors"][idx], ".2e") - elif item["calls"][idx] > 0: + item[u"clocks"][idx]/item[u"vectors"][idx], u".2e") + elif item[u"calls"][idx] > 0: clocks = format( - item["clocks"][idx]/item["calls"][idx], ".2e") - elif item["suspends"][idx] > 0: + item[u"clocks"][idx]/item[u"calls"][idx], u".2e") + elif item[u"suspends"][idx] > 0: clocks = format( - item["clocks"][idx]/item["suspends"][idx], ".2e") + item[u"clocks"][idx]/item[u"suspends"][idx], u".2e") else: clocks = 0 - if item["calls"][idx] > 0: + if item[u"calls"][idx] > 0: vectors_call = format( - item["vectors"][idx]/item["calls"][idx], ".2f") + item[u"vectors"][idx]/item[u"calls"][idx], u".2f") else: - vectors_call = format(0, ".2f") + vectors_call = format(0, u".2f") if int(calls) + int(vectors) + int(suspends): table[idx].append([ name, calls, vectors, suspends, clocks, vectors_call ]) text = "" for idx in range(threads_nr): - text += "Thread {idx} ".format(idx=idx) - text += "vpp_main\n" if idx == 0 else \ - "vpp_wk_{idx}\n".format(idx=idx-1) + text += f"Thread {idx} " + text += u"vpp_main\n" if idx == 0 else f"vpp_wk_{idx-1}\n" txt_table = None for row in table[idx]: if txt_table is None: @@ -498,19 +511,19 @@ class ExecutionChecker(ResultVisitor): if any(row[1:]): txt_table.add_row(row) txt_table.set_style(prettytable.MSWORD_FRIENDLY) - txt_table.align["Name"] = "l" - txt_table.align["Calls"] = "r" - txt_table.align["Vectors"] = "r" - txt_table.align["Suspends"] = "r" - txt_table.align["Clocks"] = "r" - txt_table.align["Vectors/Calls"] = "r" - - text += txt_table.get_string(sortby="Name") + '\n' - text = (" \n **DUT: {host}/{socket}** \n {text}". - format(host=host, socket=socket, text=text)) - text = text.replace('\n', ' |br| ').replace('\r', '').\ - replace('"', "'") - self._data["tests"][self._test_ID]["show-run"] += text + txt_table.align[u"Name"] = u"l" + txt_table.align[u"Calls"] = u"r" + txt_table.align[u"Vectors"] = u"r" + txt_table.align[u"Suspends"] = u"r" + txt_table.align[u"Clocks"] = u"r" + txt_table.align[u"Vectors/Calls"] = u"r" + + text += txt_table.get_string(sortby=u"Name") + u'\n' + text = f" \n **DUT: {host}/{socket}** \n {text}".\ + replace(u'\n', u' |br| ').\ + replace(u'\r', u'').\ + replace(u'"', u"'") + self._data[u"tests"][self._test_id][u"show-run"] += text def _get_ndrpdr_throughput(self, msg): """Get NDR_LOWER, NDR_UPPER, PDR_LOWER and PDR_UPPER from the test @@ -523,19 +536,19 @@ class ExecutionChecker(ResultVisitor): """ throughput = { - "NDR": {"LOWER": -1.0, "UPPER": -1.0}, - "PDR": {"LOWER": -1.0, "UPPER": -1.0} + u"NDR": {u"LOWER": -1.0, u"UPPER": -1.0}, + u"PDR": {u"LOWER": -1.0, u"UPPER": -1.0} } - status = "FAIL" + status = u"FAIL" groups = re.search(self.REGEX_NDRPDR_RATE, msg) if groups is not None: try: - throughput["NDR"]["LOWER"] = float(groups.group(1)) - throughput["NDR"]["UPPER"] = float(groups.group(2)) - throughput["PDR"]["LOWER"] = float(groups.group(3)) - throughput["PDR"]["UPPER"] = float(groups.group(4)) - status = "PASS" + throughput[u"NDR"][u"LOWER"] = float(groups.group(1)) + throughput[u"NDR"][u"UPPER"] = float(groups.group(2)) + throughput[u"PDR"][u"LOWER"] = float(groups.group(3)) + throughput[u"PDR"][u"UPPER"] = float(groups.group(4)) + status = u"PASS" except (IndexError, ValueError): pass @@ -552,17 +565,17 @@ class ExecutionChecker(ResultVisitor): """ throughput = { - "LOWER": -1.0, - "UPPER": -1.0 + u"LOWER": -1.0, + u"UPPER": -1.0 } - status = "FAIL" + status = u"FAIL" groups = re.search(self.REGEX_PLR_RATE, msg) if groups is not None: try: - throughput["LOWER"] = float(groups.group(1)) - throughput["UPPER"] = float(groups.group(2)) - status = "PASS" + throughput[u"LOWER"] = float(groups.group(1)) + throughput[u"UPPER"] = float(groups.group(2)) + status = u"PASS" except (IndexError, ValueError): pass @@ -576,18 +589,23 @@ class ExecutionChecker(ResultVisitor): :returns: Parsed data as a dict and the status (PASS/FAIL). :rtype: tuple(dict, str) """ - latency_default = {"min": -1.0, "avg": -1.0, "max": -1.0, "hdrh": ""} + latency_default = { + u"min": -1.0, + u"avg": -1.0, + u"max": -1.0, + u"hdrh": u"" + } latency = { - "NDR": { - "direction1": copy.copy(latency_default), - "direction2": copy.copy(latency_default) + u"NDR": { + u"direction1": copy.copy(latency_default), + u"direction2": copy.copy(latency_default) }, - "PDR": { - "direction1": copy.copy(latency_default), - "direction2": copy.copy(latency_default) + u"PDR": { + u"direction1": copy.copy(latency_default), + u"direction2": copy.copy(latency_default) } } - status = "FAIL" + status = u"FAIL" groups = re.search(self.REGEX_NDRPDR_LAT, msg) def process_latency(in_str): @@ -605,24 +623,28 @@ class ExecutionChecker(ResultVisitor): in_list = in_str.split('/') rval = { - "min": float(in_list[0]), - "avg": float(in_list[1]), - "max": float(in_list[2]), - "hdrh": "" + u"min": float(in_list[0]), + u"avg": float(in_list[1]), + u"max": float(in_list[2]), + u"hdrh": u"" } if len(in_list) == 4: - rval["hdrh"] = str(in_list[3]) + rval[u"hdrh"] = str(in_list[3]) return rval if groups is not None: try: - latency["NDR"]["direction1"] = process_latency(groups.group(1)) - latency["NDR"]["direction2"] = process_latency(groups.group(2)) - latency["PDR"]["direction1"] = process_latency(groups.group(3)) - latency["PDR"]["direction2"] = process_latency(groups.group(4)) - status = "PASS" + latency[u"NDR"][u"direction1"] = \ + process_latency(groups.group(1)) + latency[u"NDR"][u"direction2"] = \ + process_latency(groups.group(2)) + latency[u"PDR"][u"direction1"] = \ + process_latency(groups.group(3)) + latency[u"PDR"][u"direction2"] = \ + process_latency(groups.group(4)) + status = u"PASS" except (IndexError, ValueError): pass @@ -653,17 +675,22 @@ class ExecutionChecker(ResultVisitor): except AttributeError: return - doc_str = suite.doc.replace('"', "'").replace('\n', ' ').\ - replace('\r', '').replace('*[', ' |br| *[').replace("*", "**") - doc_str = replace(doc_str, ' |br| *[', '*[', maxreplace=1) - - self._data["suites"][suite.longname.lower().replace('"', "'"). - replace(" ", "_")] = { - "name": suite.name.lower(), - "doc": doc_str, - "parent": parent_name, - "level": len(suite.longname.split(".")) - } + doc_str = suite.doc.\ + replace(u'"', u"'").\ + replace(u'\n', u' ').\ + replace(u'\r', u'').\ + replace(u'*[', u' |br| *[').\ + replace(u"*", u"**").\ + replace(u' |br| *[', u'*[', 1) + + self._data[u"suites"][suite.longname.lower(). + replace(u'"', u"'"). + replace(u" ", u"_")] = { + u"name": suite.name.lower(), + u"doc": doc_str, + u"parent": parent_name, + u"level": len(suite.longname.split(u".")) + } suite.keywords.visit(self) @@ -674,7 +701,6 @@ class ExecutionChecker(ResultVisitor): :type suite: Suite :returns: Nothing. """ - pass def visit_test(self, test): """Implements traversing through the test. @@ -707,128 +733,90 @@ class ExecutionChecker(ResultVisitor): # Change the TC long name and name if defined in the mapping table longname = self._mapping.get(longname_orig, None) if longname is not None: - name = longname.split('.')[-1] - logging.debug("{0}\n{1}\n{2}\n{3}".format( - self._data["metadata"], longname_orig, longname, name)) + name = longname.split(u'.')[-1] + logging.debug( + f"{self._data[u'metadata']}\n{longname_orig}\n{longname}\n" + f"{name}" + ) else: longname = longname_orig name = test.name.lower() # Remove TC number from the TC long name (backward compatibility): - self._test_ID = re.sub(self.REGEX_TC_NUMBER, "", longname) + self._test_id = re.sub(self.REGEX_TC_NUMBER, u"", longname) # Remove TC number from the TC name (not needed): - test_result["name"] = re.sub(self.REGEX_TC_NUMBER, "", name) - - test_result["parent"] = test.parent.name.lower() - test_result["tags"] = tags - doc_str = test.doc.replace('"', "'").replace('\n', ' '). \ - replace('\r', '').replace('[', ' |br| [') - test_result["doc"] = replace(doc_str, ' |br| [', '[', maxreplace=1) - test_result["msg"] = test.message.replace('\n', ' |br| '). \ - replace('\r', '').replace('"', "'") - test_result["type"] = "FUNC" - test_result["status"] = test.status - - if "PERFTEST" in tags: + test_result[u"name"] = re.sub(self.REGEX_TC_NUMBER, "", name) + + test_result[u"parent"] = test.parent.name.lower() + test_result[u"tags"] = tags + test_result["doc"] = test.doc.\ + replace(u'"', u"'").\ + replace(u'\n', u' ').\ + replace(u'\r', u'').\ + replace(u'[', u' |br| [').\ + replace(u' |br| [', u'[', 1) + test_result[u"msg"] = test.message.\ + replace(u'\n', u' |br| ').\ + replace(u'\r', u'').\ + replace(u'"', u"'") + test_result[u"type"] = u"FUNC" + test_result[u"status"] = test.status + + if u"PERFTEST" in tags: # Replace info about cores (e.g. -1c-) with the info about threads # and cores (e.g. -1t1c-) in the long test case names and in the # test case names if necessary. - groups = re.search(self.REGEX_TC_NAME_OLD, self._test_ID) + groups = re.search(self.REGEX_TC_NAME_OLD, self._test_id) if not groups: tag_count = 0 tag_tc = str() - for tag in test_result["tags"]: + for tag in test_result[u"tags"]: groups = re.search(self.REGEX_TC_TAG, tag) if groups: tag_count += 1 tag_tc = tag if tag_count == 1: - self._test_ID = re.sub(self.REGEX_TC_NAME_NEW, - "-{0}-".format(tag_tc.lower()), - self._test_ID, + self._test_id = re.sub(self.REGEX_TC_NAME_NEW, + f"-{tag_tc.lower()}-", + self._test_id, count=1) - test_result["name"] = re.sub(self.REGEX_TC_NAME_NEW, - "-{0}-".format(tag_tc.lower()), - test_result["name"], - count=1) + test_result[u"name"] = re.sub(self.REGEX_TC_NAME_NEW, + f"-{tag_tc.lower()}-", + test_result["name"], + count=1) else: - test_result["status"] = "FAIL" - self._data["tests"][self._test_ID] = test_result - logging.debug("The test '{0}' has no or more than one " - "multi-threading tags.".format(self._test_ID)) - logging.debug("Tags: {0}".format(test_result["tags"])) + test_result[u"status"] = u"FAIL" + self._data[u"tests"][self._test_id] = test_result + logging.debug( + f"The test {self._test_id} has no or more than one " + f"multi-threading tags.\n" + f"Tags: {test_result[u'tags']}" + ) return - if test.status == "PASS" and ("NDRPDRDISC" in tags or - "NDRPDR" in tags or - "SOAK" in tags or - "TCP" in tags or - "MRR" in tags or - "BMRR" in tags or - "RECONF" in tags): - # TODO: Remove when definitely no NDRPDRDISC tests are used: - if "NDRDISC" in tags: - test_result["type"] = "NDR" - # TODO: Remove when definitely no NDRPDRDISC tests are used: - elif "PDRDISC" in tags: - test_result["type"] = "PDR" - elif "NDRPDR" in tags: - test_result["type"] = "NDRPDR" - elif "SOAK" in tags: - test_result["type"] = "SOAK" - elif "TCP" in tags: - test_result["type"] = "TCP" - elif "MRR" in tags: - test_result["type"] = "MRR" - elif "FRMOBL" in tags or "BMRR" in tags: - test_result["type"] = "BMRR" - elif "RECONF" in tags: - test_result["type"] = "RECONF" - else: - test_result["status"] = "FAIL" - self._data["tests"][self._test_ID] = test_result - return - - # TODO: Remove when definitely no NDRPDRDISC tests are used: - if test_result["type"] in ("NDR", "PDR"): - try: - rate_value = str(re.search( - self.REGEX_RATE, test.message).group(1)) - except AttributeError: - rate_value = "-1" - try: - rate_unit = str(re.search( - self.REGEX_RATE, test.message).group(2)) - except AttributeError: - rate_unit = "-1" - - test_result["throughput"] = dict() - test_result["throughput"]["value"] = \ - int(rate_value.split('.')[0]) - test_result["throughput"]["unit"] = rate_unit - test_result["latency"] = \ - self._get_latency(test.message, test_result["type"]) - if test_result["type"] == "PDR": - test_result["lossTolerance"] = str(re.search( - self.REGEX_TOLERANCE, test.message).group(1)) - - elif test_result["type"] in ("NDRPDR", ): - test_result["throughput"], test_result["status"] = \ + if test.status == u"PASS": + if u"NDRPDR" in tags: + test_result[u"type"] = u"NDRPDR" + test_result[u"throughput"], test_result[u"status"] = \ self._get_ndrpdr_throughput(test.message) - test_result["latency"], test_result["status"] = \ + test_result[u"latency"], test_result[u"status"] = \ self._get_ndrpdr_latency(test.message) - - elif test_result["type"] in ("SOAK", ): - test_result["throughput"], test_result["status"] = \ + elif u"SOAK" in tags: + test_result[u"type"] = u"SOAK" + test_result[u"throughput"], test_result[u"status"] = \ self._get_plr_throughput(test.message) - - elif test_result["type"] in ("TCP", ): + elif u"TCP" in tags: + test_result[u"type"] = u"TCP" groups = re.search(self.REGEX_TCP, test.message) - test_result["result"] = int(groups.group(2)) + test_result[u"result"] = int(groups.group(2)) + elif u"MRR" in tags or u"FRMOBL" in tags or u"BMRR" in tags: + if u"MRR" in tags: + test_result[u"type"] = u"MRR" + else: + test_result[u"type"] = u"BMRR" - elif test_result["type"] in ("MRR", "BMRR"): - test_result["result"] = dict() + test_result[u"result"] = dict() groups = re.search(self.REGEX_BMRR, test.message) if groups is not None: items_str = groups.group(1) @@ -836,25 +824,29 @@ class ExecutionChecker(ResultVisitor): in items_str.split(",")] # Use whole list in CSIT-1180. stats = jumpavg.AvgStdevStats.for_runs(items_float) - test_result["result"]["receive-rate"] = stats.avg + test_result[u"result"][u"receive-rate"] = stats.avg else: groups = re.search(self.REGEX_MRR, test.message) - test_result["result"]["receive-rate"] = \ + test_result[u"result"][u"receive-rate"] = \ float(groups.group(3)) / float(groups.group(1)) - - elif test_result["type"] == "RECONF": - test_result["result"] = None + elif u"RECONF" in tags: + test_result[u"type"] = u"RECONF" + test_result[u"result"] = None try: grps_loss = re.search(self.REGEX_RECONF_LOSS, test.message) grps_time = re.search(self.REGEX_RECONF_TIME, test.message) - test_result["result"] = { - "loss": int(grps_loss.group(1)), - "time": float(grps_time.group(1)) + test_result[u"result"] = { + u"loss": int(grps_loss.group(1)), + u"time": float(grps_time.group(1)) } except (AttributeError, IndexError, ValueError, TypeError): - test_result["status"] = "FAIL" + test_result[u"status"] = u"FAIL" + else: + test_result[u"status"] = u"FAIL" + self._data[u"tests"][self._test_id] = test_result + return - self._data["tests"][self._test_ID] = test_result + self._data[u"tests"][self._test_id] = test_result def end_test(self, test): """Called when test ends. @@ -863,7 +855,6 @@ class ExecutionChecker(ResultVisitor): :type test: Test :returns: Nothing. """ - pass def visit_keyword(self, keyword): """Implements traversing through the keyword and its child keywords. @@ -883,9 +874,9 @@ class ExecutionChecker(ResultVisitor): :returns: Nothing. """ try: - if keyword.type == "setup": + if keyword.type == u"setup": self.visit_setup_kw(keyword) - elif keyword.type == "teardown": + elif keyword.type == u"teardown": self._lookup_kw_nr = 0 self.visit_teardown_kw(keyword) else: @@ -901,7 +892,6 @@ class ExecutionChecker(ResultVisitor): :type keyword: Keyword :returns: Nothing. """ - pass def visit_test_kw(self, test_kw): """Implements traversing through the test keyword and its child @@ -924,12 +914,12 @@ class ExecutionChecker(ResultVisitor): :type test_kw: Keyword :returns: Nothing. """ - if test_kw.name.count("Show Runtime Counters On All Duts"): + if test_kw.name.count(u"Show Runtime Counters On All Duts"): self._lookup_kw_nr += 1 self._show_run_lookup_nr = 0 - self._msg_type = "test-show-runtime" - elif test_kw.name.count("Install Dpdk Test") and not self._version: - self._msg_type = "dpdk-version" + self._msg_type = u"test-show-runtime" + elif test_kw.name.count(u"Install Dpdk Test") and not self._version: + self._msg_type = u"dpdk-version" else: return test_kw.messages.visit(self) @@ -941,7 +931,6 @@ class ExecutionChecker(ResultVisitor): :type test_kw: Keyword :returns: Nothing. """ - pass def visit_setup_kw(self, setup_kw): """Implements traversing through the teardown keyword and its child @@ -964,14 +953,14 @@ class ExecutionChecker(ResultVisitor): :type setup_kw: Keyword :returns: Nothing. """ - if setup_kw.name.count("Show Vpp Version On All Duts") \ + if setup_kw.name.count(u"Show Vpp Version On All Duts") \ and not self._version: - self._msg_type = "vpp-version" - elif setup_kw.name.count("Set Global Variable") \ + self._msg_type = u"vpp-version" + elif setup_kw.name.count(u"Set Global Variable") \ and not self._timestamp: - self._msg_type = "timestamp" - elif setup_kw.name.count("Setup Framework") and not self._testbed: - self._msg_type = "testbed" + self._msg_type = u"timestamp" + elif setup_kw.name.count(u"Setup Framework") and not self._testbed: + self._msg_type = u"testbed" else: return setup_kw.messages.visit(self) @@ -983,7 +972,6 @@ class ExecutionChecker(ResultVisitor): :type setup_kw: Keyword :returns: Nothing. """ - pass def visit_teardown_kw(self, teardown_kw): """Implements traversing through the teardown keyword and its child @@ -999,21 +987,21 @@ class ExecutionChecker(ResultVisitor): self.end_teardown_kw(keyword) def start_teardown_kw(self, teardown_kw): - """Called when teardown keyword starts. Default implementation does - nothing. + """Called when teardown keyword starts :param teardown_kw: Keyword to process. :type teardown_kw: Keyword :returns: Nothing. """ - if teardown_kw.name.count("Show Vat History On All Duts"): + if teardown_kw.name.count(u"Show Vat History On All Duts"): + # TODO: Remove when not needed: self._conf_history_lookup_nr = 0 - self._msg_type = "teardown-vat-history" + self._msg_type = u"teardown-vat-history" teardown_kw.messages.visit(self) - elif teardown_kw.name.count("Show Papi History On All Duts"): + elif teardown_kw.name.count(u"Show Papi History On All Duts"): self._conf_history_lookup_nr = 0 - self._msg_type = "teardown-papi-history" + self._msg_type = u"teardown-papi-history" teardown_kw.messages.visit(self) def end_teardown_kw(self, teardown_kw): @@ -1023,7 +1011,6 @@ class ExecutionChecker(ResultVisitor): :type teardown_kw: Keyword :returns: Nothing. """ - pass def visit_message(self, msg): """Implements visiting the message. @@ -1054,7 +1041,6 @@ class ExecutionChecker(ResultVisitor): :type msg: Message :returns: Nothing. """ - pass class InputData: @@ -1107,7 +1093,7 @@ class InputData: :rtype: pandas.Series """ - return self.data[job][build]["metadata"] + return self.data[job][build][u"metadata"] def suites(self, job, build): """Getter - suites @@ -1120,7 +1106,7 @@ class InputData: :rtype: pandas.Series """ - return self.data[job][str(build)]["suites"] + return self.data[job][str(build)][u"suites"] def tests(self, job, build): """Getter - tests @@ -1133,7 +1119,7 @@ class InputData: :rtype: pandas.Series """ - return self.data[job][build]["tests"] + return self.data[job][build][u"tests"] def _parse_tests(self, job, build, log): """Process data from robot output.xml file and return JSON structured @@ -1150,16 +1136,18 @@ class InputData: """ metadata = { - "job": job, - "build": build + u"job": job, + u"build": build } - with open(build["file-name"], 'r') as data_file: + with open(build[u"file-name"], u'r') as data_file: try: result = ExecutionResult(data_file) except errors.DataError as err: - log.append(("ERROR", "Error occurred while parsing output.xml: " - "{0}".format(err))) + log.append( + (u"ERROR", f"Error occurred while parsing output.xml: " + f"{repr(err)}") + ) return None checker = ExecutionChecker(metadata, self._cfg.mapping, self._cfg.ignore) @@ -1185,10 +1173,11 @@ class InputData: logs = list() - logs.append(("INFO", " Processing the job/build: {0}: {1}". - format(job, build["build"]))) + logs.append( + (u"INFO", f" Processing the job/build: {job}: {build[u'build']}") + ) - state = "failed" + state = u"failed" success = False data = None do_repeat = repeat @@ -1199,61 +1188,70 @@ class InputData: break do_repeat -= 1 if not success: - logs.append(("ERROR", "It is not possible to download the input " - "data file from the job '{job}', build " - "'{build}', or it is damaged. Skipped.". - format(job=job, build=build["build"]))) + logs.append( + (u"ERROR", + f"It is not possible to download the input data file from the " + f"job {job}, build {build[u'build']}, or it is damaged. " + f"Skipped.") + ) if success: - logs.append(("INFO", " Processing data from the build '{0}' ...". - format(build["build"]))) + logs.append( + (u"INFO", + f" Processing data from the build {build[u'build']} ...") + ) data = self._parse_tests(job, build, logs) if data is None: - logs.append(("ERROR", "Input data file from the job '{job}', " - "build '{build}' is damaged. Skipped.". - format(job=job, build=build["build"]))) + logs.append( + (u"ERROR", + f"Input data file from the job {job}, build " + f"{build[u'build']} is damaged. Skipped.") + ) else: - state = "processed" + state = u"processed" try: - remove(build["file-name"]) + remove(build[u"file-name"]) except OSError as err: - logs.append(("ERROR", "Cannot remove the file '{0}': {1}". - format(build["file-name"], repr(err)))) + logs.append( + ("ERROR", f"Cannot remove the file {build[u'file-name']}: " + f"{repr(err)}") + ) # If the time-period is defined in the specification file, remove all # files which are outside the time period. - timeperiod = self._cfg.input.get("time-period", None) + timeperiod = self._cfg.input.get(u"time-period", None) if timeperiod and data: now = dt.utcnow() timeperiod = timedelta(int(timeperiod)) - metadata = data.get("metadata", None) + metadata = data.get(u"metadata", None) if metadata: - generated = metadata.get("generated", None) + generated = metadata.get(u"generated", None) if generated: - generated = dt.strptime(generated, "%Y%m%d %H:%M") + generated = dt.strptime(generated, u"%Y%m%d %H:%M") if (now - generated) > timeperiod: # Remove the data and the file: - state = "removed" + state = u"removed" data = None logs.append( - ("INFO", - " The build {job}/{build} is outdated, will be " - "removed".format(job=job, build=build["build"]))) - logs.append(("INFO", " Done.")) + (u"INFO", + f" The build {job}/{build[u'build']} is " + f"outdated, will be removed.") + ) + logs.append((u"INFO", u" Done.")) for level, line in logs: - if level == "INFO": + if level == u"INFO": logging.info(line) - elif level == "ERROR": + elif level == u"ERROR": logging.error(line) - elif level == "DEBUG": + elif level == u"DEBUG": logging.debug(line) - elif level == "CRITICAL": + elif level == u"CRITICAL": logging.critical(line) - elif level == "WARNING": + elif level == u"WARNING": logging.warning(line) - return {"data": data, "state": state, "job": job, "build": build} + return {u"data": data, u"state": state, u"job": job, u"build": build} def download_and_parse_data(self, repeat=1): """Download the input data files, parse input data from input files and @@ -1264,41 +1262,48 @@ class InputData: :type repeat: int """ - logging.info("Downloading and parsing input files ...") + logging.info(u"Downloading and parsing input files ...") for job, builds in self._cfg.builds.items(): for build in builds: result = self._download_and_parse_build(job, build, repeat) - build_nr = result["build"]["build"] + build_nr = result[u"build"][u"build"] - if result["data"]: - data = result["data"] + if result[u"data"]: + data = result[u"data"] build_data = pd.Series({ - "metadata": pd.Series( - data["metadata"].values(), - index=data["metadata"].keys()), - "suites": pd.Series(data["suites"].values(), - index=data["suites"].keys()), - "tests": pd.Series(data["tests"].values(), - index=data["tests"].keys())}) + u"metadata": pd.Series( + list(data[u"metadata"].values()), + index=list(data[u"metadata"].keys()) + ), + u"suites": pd.Series( + list(data[u"suites"].values()), + index=list(data[u"suites"].keys()) + ), + u"tests": pd.Series( + list(data[u"tests"].values()), + index=list(data[u"tests"].keys()) + ) + }) if self._input_data.get(job, None) is None: self._input_data[job] = pd.Series() self._input_data[job][str(build_nr)] = build_data self._cfg.set_input_file_name( - job, build_nr, result["build"]["file-name"]) + job, build_nr, result[u"build"][u"file-name"]) - self._cfg.set_input_state(job, build_nr, result["state"]) + self._cfg.set_input_state(job, build_nr, result[u"state"]) - logging.info("Memory allocation: {0:,d}MB".format( - resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1000)) + mem_alloc = \ + resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1000 + logging.info(f"Memory allocation: {mem_alloc:.0f}MB") - logging.info("Done.") + logging.info(u"Done.") @staticmethod - def _end_of_tag(tag_filter, start=0, closer="'"): + def _end_of_tag(tag_filter, start=0, closer=u"'"): """Return the index of character in the string which is the end of tag. :param tag_filter: The string where the end of tag is being searched. @@ -1333,9 +1338,9 @@ class InputData: if index is None: return tag_filter index += 1 - tag_filter = tag_filter[:index] + " in tags" + tag_filter[index:] + tag_filter = tag_filter[:index] + u" in tags" + tag_filter[index:] - def filter_data(self, element, params=None, data=None, data_set="tests", + def filter_data(self, element, params=None, data=None, data_set=u"tests", continue_on_error=False): """Filter required data from the given jobs and builds. @@ -1375,21 +1380,21 @@ class InputData: """ try: - if element["filter"] in ("all", "template"): - cond = "True" + if element[u"filter"] in (u"all", u"template"): + cond = u"True" else: - cond = InputData._condition(element["filter"]) - logging.debug(" Filter: {0}".format(cond)) + cond = InputData._condition(element[u"filter"]) + logging.debug(f" Filter: {cond}") except KeyError: - logging.error(" No filter defined.") + logging.error(u" No filter defined.") return None if params is None: - params = element.get("parameters", None) + params = element.get(u"parameters", None) if params: - params.append("type") + params.append(u"type") - data_to_filter = data if data else element["data"] + data_to_filter = data if data else element[u"data"] data = pd.Series() try: for job, builds in data_to_filter.items(): @@ -1397,41 +1402,46 @@ class InputData: for build in builds: data[job][str(build)] = pd.Series() try: - data_iter = self.data[job][str(build)][data_set].\ - iteritems() + data_dict = dict( + self.data[job][str(build)][data_set].items()) except KeyError: if continue_on_error: continue - else: - return None - for test_ID, test_data in data_iter: - if eval(cond, {"tags": test_data.get("tags", "")}): - data[job][str(build)][test_ID] = pd.Series() + return None + + for test_id, test_data in data_dict.items(): + if eval(cond, {u"tags": test_data.get(u"tags", u"")}): + data[job][str(build)][test_id] = pd.Series() if params is None: for param, val in test_data.items(): - data[job][str(build)][test_ID][param] = val + data[job][str(build)][test_id][param] = val else: for param in params: try: - data[job][str(build)][test_ID][param] =\ + data[job][str(build)][test_id][param] =\ test_data[param] except KeyError: - data[job][str(build)][test_ID][param] =\ - "No Data" + data[job][str(build)][test_id][param] =\ + u"No Data" return data except (KeyError, IndexError, ValueError) as err: - logging.error(" Missing mandatory parameter in the element " - "specification: {0}".format(err)) + logging.error( + f"Missing mandatory parameter in the element specification: " + f"{repr(err)}" + ) return None - except AttributeError: + except AttributeError as err: + logging.error(repr(err)) return None - except SyntaxError: - logging.error(" The filter '{0}' is not correct. Check if all " - "tags are enclosed by apostrophes.".format(cond)) + except SyntaxError as err: + logging.error( + f"The filter {cond} is not correct. Check if all tags are " + f"enclosed by apostrophes.\n{repr(err)}" + ) return None - def filter_tests_by_name(self, element, params=None, data_set="tests", + def filter_tests_by_name(self, element, params=None, data_set=u"tests", continue_on_error=False): """Filter required data from the given jobs and builds. @@ -1467,60 +1477,61 @@ class InputData: :rtype pandas.Series """ - include = element.get("include", None) + include = element.get(u"include", None) if not include: - logging.warning("No tests to include, skipping the element.") + logging.warning(u"No tests to include, skipping the element.") return None if params is None: - params = element.get("parameters", None) + params = element.get(u"parameters", None) if params: - params.append("type") + params.append(u"type") data = pd.Series() try: - for job, builds in element["data"].items(): + for job, builds in element[u"data"].items(): data[job] = pd.Series() for build in builds: data[job][str(build)] = pd.Series() for test in include: try: reg_ex = re.compile(str(test).lower()) - for test_ID in self.data[job][str(build)]\ - [data_set].keys(): - if re.match(reg_ex, str(test_ID).lower()): - test_data = self.data[job][str(build)]\ - [data_set][test_ID] - data[job][str(build)][test_ID] = pd.Series() + for test_id in self.data[job][ + str(build)][data_set].keys(): + if re.match(reg_ex, str(test_id).lower()): + test_data = self.data[job][ + str(build)][data_set][test_id] + data[job][str(build)][test_id] = pd.Series() if params is None: for param, val in test_data.items(): - data[job][str(build)][test_ID]\ + data[job][str(build)][test_id]\ [param] = val else: for param in params: try: - data[job][str(build)][test_ID]\ - [param] = test_data[param] + data[job][str(build)][ + test_id][param] = \ + test_data[param] except KeyError: - data[job][str(build)][test_ID]\ - [param] = "No Data" + data[job][str(build)][ + test_id][param] = u"No Data" except KeyError as err: - logging.error("{err!r}".format(err=err)) + logging.error(repr(err)) if continue_on_error: continue - else: - return None + return None return data except (KeyError, IndexError, ValueError) as err: - logging.error("Missing mandatory parameter in the element " - "specification: {err!r}".format(err=err)) + logging.error( + f"Missing mandatory parameter in the element " + f"specification: {repr(err)}" + ) return None except AttributeError as err: - logging.error("{err!r}".format(err=err)) + logging.error(repr(err)) return None - @staticmethod def merge_data(data): """Merge data from more jobs and builds to a simple data structure. @@ -1542,12 +1553,12 @@ class InputData: :rtype: pandas.Series """ - logging.info(" Merging data ...") + logging.info(u" Merging data ...") merged_data = pd.Series() - for _, builds in data.iteritems(): - for _, item in builds.iteritems(): - for ID, item_data in item.iteritems(): - merged_data[ID] = item_data + for builds in data.values: + for item in builds.values: + for item_id, item_data in item.items(): + merged_data[item_id] = item_data return merged_data diff --git a/resources/tools/presentation/pal.py b/resources/tools/presentation/pal.py index 94742b07da..01eb8f64ff 100644 --- a/resources/tools/presentation/pal.py +++ b/resources/tools/presentation/pal.py @@ -18,7 +18,7 @@ import sys import argparse import logging -from errors import PresentationError +from pal_errors import PresentationError from environment import Environment, clean_environment from specification_parser import Specification from input_data_parser import InputData @@ -27,7 +27,7 @@ from generator_plots import generate_plots from generator_files import generate_files from static_content import prepare_static_content from generator_report import generate_report -from generator_CPTA import generate_cpta +from generator_cpta import generate_cpta from generator_alerts import Alerting, AlertingError @@ -41,26 +41,26 @@ def parse_args(): parser = argparse.ArgumentParser(description=__doc__, formatter_class=argparse. RawDescriptionHelpFormatter) - parser.add_argument("-s", "--specification", + parser.add_argument(u"-s", u"--specification", required=True, - type=argparse.FileType('r'), - help="Specification YAML file.") - parser.add_argument("-r", "--release", - default="master", + type=argparse.FileType(u'r'), + help=u"Specification YAML file.") + parser.add_argument(u"-r", u"--release", + default=u"master", type=str, - help="Release string of the product.") - parser.add_argument("-w", "--week", - default="1", + help=u"Release string of the product.") + parser.add_argument(u"-w", u"--week", + default=u"1", type=str, - help="Calendar week when the report is published.") - parser.add_argument("-l", "--logging", - choices=["DEBUG", "INFO", "WARNING", - "ERROR", "CRITICAL"], - default="ERROR", - help="Logging level.") - parser.add_argument("-f", "--force", - action='store_true', - help="Force removing the old build(s) if present.") + help=u"Calendar week when the report is published.") + parser.add_argument(u"-l", u"--logging", + choices=[u"DEBUG", u"INFO", u"WARNING", + u"ERROR", u"CRITICAL"], + default=u"ERROR", + help=u"Logging level.") + parser.add_argument(u"-f", u"--force", + action=u"store_true", + help=u"Force removing the old build(s) if present.") return parser.parse_args() @@ -68,75 +68,70 @@ def parse_args(): def main(): """Main function.""" - log_levels = {"NOTSET": logging.NOTSET, - "DEBUG": logging.DEBUG, - "INFO": logging.INFO, - "WARNING": logging.WARNING, - "ERROR": logging.ERROR, - "CRITICAL": logging.CRITICAL} + log_levels = {u"NOTSET": logging.NOTSET, + u"DEBUG": logging.DEBUG, + u"INFO": logging.INFO, + u"WARNING": logging.WARNING, + u"ERROR": logging.ERROR, + u"CRITICAL": logging.CRITICAL} args = parse_args() - logging.basicConfig(format='%(asctime)s: %(levelname)s: %(message)s', - datefmt='%Y/%m/%d %H:%M:%S', + logging.basicConfig(format=u"%(asctime)s: %(levelname)s: %(message)s", + datefmt=u"%Y/%m/%d %H:%M:%S", level=log_levels[args.logging]) - logging.info("Application started.") + logging.info(u"Application started.") try: spec = Specification(args.specification) spec.read_specification() except PresentationError: - logging.critical("Finished with error.") + logging.critical(u"Finished with error.") return 1 - if spec.output["output"] not in ("report", "CPTA"): - logging.critical("The output '{0}' is not supported.". - format(spec.output["output"])) + if spec.output[u"output"] not in (u"report", u"CPTA"): + logging.critical( + f"The output {spec.output[u'output']} is not supported." + ) return 1 - # ret_code = 1 - # try: - env = Environment(spec.environment, args.force) - env.set_environment() - - prepare_static_content(spec) - - data = InputData(spec) - data.download_and_parse_data(repeat=1) - - generate_tables(spec, data) - generate_plots(spec, data) - generate_files(spec, data) - - if spec.output["output"] == "report": - generate_report(args.release, spec, args.week) - logging.info("Successfully finished.") - elif spec.output["output"] == "CPTA": - sys.stdout.write(generate_cpta(spec, data)) - try: - alert = Alerting(spec) - alert.generate_alerts() - except AlertingError as err: - logging.warning(repr(err)) - logging.info("Successfully finished.") - ret_code = 0 - - # except AlertingError as err: - # logging.critical("Finished with an alerting error.") - # logging.critical(repr(err)) - # except PresentationError as err: - # logging.critical("Finished with an PAL error.") - # logging.critical(repr(err)) - # except (KeyError, ValueError) as err: - # logging.critical("Finished with an error.") - # logging.critical(repr(err)) - # except Exception as err: - # logging.critical("Finished with an unexpected error.") - # logging.critical(repr(err)) - # finally: - # if spec is not None: - # clean_environment(spec.environment) - # return ret_code - - -if __name__ == '__main__': + ret_code = 1 + try: + env = Environment(spec.environment, args.force) + env.set_environment() + + prepare_static_content(spec) + + data = InputData(spec) + data.download_and_parse_data(repeat=1) + + generate_tables(spec, data) + generate_plots(spec, data) + generate_files(spec, data) + + if spec.output[u"output"] == u"report": + generate_report(args.release, spec, args.week) + elif spec.output[u"output"] == u"CPTA": + sys.stdout.write(generate_cpta(spec, data)) + try: + alert = Alerting(spec) + alert.generate_alerts() + except AlertingError as err: + logging.warning(repr(err)) + + logging.info(u"Successfully finished.") + ret_code = 0 + + except AlertingError as err: + logging.critical(f"Finished with an alerting error.\n{repr(err)}") + except PresentationError as err: + logging.critical(f"Finished with an PAL error.\n{repr(err)}") + except (KeyError, ValueError) as err: + logging.critical(f"Finished with an error.\n{repr(err)}") + finally: + if spec is not None: + clean_environment(spec.environment) + return ret_code + + +if __name__ == u"__main__": sys.exit(main()) diff --git a/resources/tools/presentation/errors.py b/resources/tools/presentation/pal_errors.py index 64084f6a40..649c86cde8 100644 --- a/resources/tools/presentation/errors.py +++ b/resources/tools/presentation/pal_errors.py @@ -1,4 +1,4 @@ -# Copyright (c) 2017 Cisco and/or its affiliates. +# Copyright (c) 2019 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -29,21 +29,21 @@ class PresentationError(Exception): - relevant data if there are any collected (optional parameter details). """ - log_exception = {"DEBUG": logging.debug, - "INFO": logging.info, - "WARNING": logging.warning, - "ERROR": logging.error, - "CRITICAL": logging.critical} + log_exception = {u"DEBUG": logging.debug, + u"INFO": logging.info, + u"WARNING": logging.warning, + u"ERROR": logging.error, + u"CRITICAL": logging.critical} - def __init__(self, msg, details='', level="CRITICAL"): + def __init__(self, msg, details=u'', level=u"CRITICAL"): """Sets the exception message and the level. :param msg: Short description of the encountered problem. :param details: Relevant messages if there are any collected, e.g., - from caught exception (optional parameter details), or relevant data if - there are any collected (optional parameter details). + from caught exception (optional parameter details), or relevant data + if there are any collected (optional parameter details). :param level: Level of the error, possible choices are: "DEBUG", "INFO", - "WARNING", "ERROR" and "CRITICAL". + "WARNING", "ERROR" and "CRITICAL". :type msg: str :type details: str :type level: str @@ -59,13 +59,14 @@ class PresentationError(Exception): if self._details: self.log_exception[self._level](self._details) except KeyError: - print("Wrong log level.") + print(u"Wrong log level.") sys.exit(1) def __repr__(self): return ( - "PresentationError(msg={msg!r},details={dets!r},level={level!r})". - format(msg=self._msg, dets=self._details, level=self._level)) + f"PresentationError(msg={self._msg!r},details={self._details!r}," + f"level={self._level!r})" + ) def __str__(self): return str(self._msg) diff --git a/resources/tools/presentation/utils.py b/resources/tools/presentation/pal_utils.py index 3bd5a71e00..5949432340 100644 --- a/resources/tools/presentation/utils.py +++ b/resources/tools/presentation/pal_utils.py @@ -14,22 +14,24 @@ """General purpose utilities. """ -import multiprocessing import subprocess import math -import numpy as np import logging import csv -import prettytable from os import walk, makedirs, environ from os.path import join, isdir from shutil import move, Error from datetime import datetime + +import numpy as np +import prettytable + from pandas import Series from resources.libraries.python import jumpavg -from errors import PresentationError + +from pal_errors import PresentationError def mean(items): @@ -132,11 +134,10 @@ def get_rst_title_char(level): :returns: Character used for the given title level in rst files. :rtype: str """ - chars = ('=', '-', '`', "'", '.', '~', '*', '+', '^') + chars = (u'=', u'-', u'`', u"'", u'.', u'~', u'*', u'+', u'^') if level < len(chars): return chars[level] - else: - return chars[-1] + return chars[-1] def execute_command(cmd): @@ -159,16 +160,16 @@ def execute_command(cmd): stdout, stderr = proc.communicate() if stdout: - logging.info(stdout) + logging.info(stdout.decode()) if stderr: - logging.info(stderr) + logging.info(stderr.decode()) if proc.returncode != 0: - logging.error(" Command execution failed.") - return proc.returncode, stdout, stderr + logging.error(u" Command execution failed.") + return proc.returncode, stdout.decode(), stderr.decode() -def get_last_successful_build_number(jenkins_url, job_name): +def get_last_successful_build_nr(jenkins_url, job_name): """Get the number of the last successful build of the given job. :param jenkins_url: Jenkins URL. @@ -178,11 +179,9 @@ def get_last_successful_build_number(jenkins_url, job_name): :returns: The build number as a string. :rtype: str """ - - url = "{}/{}/lastSuccessfulBuild/buildNumber".format(jenkins_url, job_name) - cmd = "wget -qO- {url}".format(url=url) - - return execute_command(cmd) + return execute_command( + f"wget -qO- {jenkins_url}/{job_name}/lastSuccessfulBuild/buildNumber" + ) def get_last_completed_build_number(jenkins_url, job_name): @@ -195,11 +194,9 @@ def get_last_completed_build_number(jenkins_url, job_name): :returns: The build number as a string. :rtype: str """ - - url = "{}/{}/lastCompletedBuild/buildNumber".format(jenkins_url, job_name) - cmd = "wget -qO- {url}".format(url=url) - - return execute_command(cmd) + return execute_command( + f"wget -qO- {jenkins_url}/{job_name}/lastCompletedBuild/buildNumber" + ) def get_build_timestamp(jenkins_url, job_name, build_nr): @@ -214,14 +211,9 @@ def get_build_timestamp(jenkins_url, job_name, build_nr): :returns: The timestamp. :rtype: datetime.datetime """ - - url = "{jenkins_url}/{job_name}/{build_nr}".format(jenkins_url=jenkins_url, - job_name=job_name, - build_nr=build_nr) - cmd = "wget -qO- {url}".format(url=url) - - timestamp = execute_command(cmd) - + timestamp = execute_command( + f"wget -qO- {jenkins_url}/{job_name}/{build_nr}" + ) return datetime.fromtimestamp(timestamp/1000) @@ -233,29 +225,31 @@ def archive_input_data(spec): :raises PresentationError: If it is not possible to archive the input data. """ - logging.info(" Archiving the input data files ...") + logging.info(u" Archiving the input data files ...") - extension = spec.input["arch-file-format"] + extension = spec.input[u"arch-file-format"] data_files = list() for ext in extension: data_files.extend(get_files( - spec.environment["paths"]["DIR[WORKING,DATA]"], extension=ext)) - dst = spec.environment["paths"]["DIR[STATIC,ARCH]"] - logging.info(" Destination: {0}".format(dst)) + spec.environment[u"paths"][u"DIR[WORKING,DATA]"], extension=ext)) + dst = spec.environment[u"paths"][u"DIR[STATIC,ARCH]"] + logging.info(f" Destination: {dst}") try: if not isdir(dst): makedirs(dst) for data_file in data_files: - logging.info(" Moving the file: {0} ...".format(data_file)) + logging.info(f" Moving the file: {data_file} ...") move(data_file, dst) except (Error, OSError) as err: - raise PresentationError("Not possible to archive the input data.", - str(err)) + raise PresentationError( + u"Not possible to archive the input data.", + repr(err) + ) - logging.info(" Done.") + logging.info(u" Done.") def classify_anomalies(data): @@ -273,7 +267,7 @@ def classify_anomalies(data): # Nan means something went wrong. # Use 0.0 to cause that being reported as a severe regression. bare_data = [0.0 if np.isnan(sample) else sample - for sample in data.itervalues()] + for sample in data.values()] # TODO: Make BitCountingGroupList a subclass of list again? group_list = jumpavg.classify(bare_data).group_list group_list.reverse() # Just to use .pop() for FIFO. @@ -282,9 +276,9 @@ def classify_anomalies(data): active_group = None values_left = 0 avg = 0.0 - for sample in data.itervalues(): + for sample in data.values(): if np.isnan(sample): - classification.append("outlier") + classification.append(u"outlier") avgs.append(sample) continue if values_left < 1 or active_group is None: @@ -297,70 +291,30 @@ def classify_anomalies(data): avgs.append(avg) values_left -= 1 continue - classification.append("normal") + classification.append(u"normal") avgs.append(avg) values_left -= 1 return classification, avgs -def convert_csv_to_pretty_txt(csv_file, txt_file): +def convert_csv_to_pretty_txt(csv_file_name, txt_file_name): """Convert the given csv table to pretty text table. - :param csv_file: The path to the input csv file. - :param txt_file: The path to the output pretty text file. - :type csv_file: str - :type txt_file: str + :param csv_file_name: The path to the input csv file. + :param txt_file_name: The path to the output pretty text file. + :type csv_file_name: str + :type txt_file_name: str """ txt_table = None - with open(csv_file, 'rb') as csv_file: - csv_content = csv.reader(csv_file, delimiter=',', quotechar='"') + with open(csv_file_name, u"rt") as csv_file: + csv_content = csv.reader(csv_file, delimiter=u',', quotechar=u'"') for row in csv_content: if txt_table is None: txt_table = prettytable.PrettyTable(row) else: txt_table.add_row(row) - txt_table.align["Test case"] = "l" + txt_table.align[u"Test case"] = u"l" if txt_table: - with open(txt_file, "w") as txt_file: + with open(txt_file_name, u"w") as txt_file: txt_file.write(str(txt_table)) - - -class Worker(multiprocessing.Process): - """Worker class used to process tasks in separate parallel processes. - """ - - def __init__(self, work_queue, data_queue, func): - """Initialization. - - :param work_queue: Queue with items to process. - :param data_queue: Shared memory between processes. Queue which keeps - the result data. This data is then read by the main process and used - in further processing. - :param func: Function which is executed by the worker. - :type work_queue: multiprocessing.JoinableQueue - :type data_queue: multiprocessing.Manager().Queue() - :type func: Callable object - """ - super(Worker, self).__init__() - self._work_queue = work_queue - self._data_queue = data_queue - self._func = func - - def run(self): - """Method representing the process's activity. - """ - - while True: - try: - self.process(self._work_queue.get()) - finally: - self._work_queue.task_done() - - def process(self, item_to_process): - """Method executed by the runner. - - :param item_to_process: Data to be processed by the function. - :type item_to_process: tuple - """ - self._func(self.pid, self._data_queue, *item_to_process) diff --git a/resources/tools/presentation/requirements.txt b/resources/tools/presentation/requirements.txt index 1676983658..537bbf2fda 100644 --- a/resources/tools/presentation/requirements.txt +++ b/resources/tools/presentation/requirements.txt @@ -1,9 +1,9 @@ -Sphinx==1.7.6 +Sphinx==2.2.1 sphinx-rtd-theme==0.4.0 -robotframework==2.9.2 -sphinxcontrib-programoutput -PyYAML==5.1.1 -numpy==1.16.4 -pandas==0.24.2 +sphinxcontrib-programoutput==0.15 +robotframework==3.1.2 +PyYAML==5.1 +numpy==1.17.3 +pandas==0.25.3 plotly==4.1.1 PTable==0.9.2 diff --git a/resources/tools/presentation/run_cpta.sh b/resources/tools/presentation/run_cpta.sh index a60c1cc7a9..1aa5703162 100755 --- a/resources/tools/presentation/run_cpta.sh +++ b/resources/tools/presentation/run_cpta.sh @@ -22,7 +22,7 @@ source ${DIR[WORKING]}/env/bin/activate # Install python dependencies: pip3 install -r requirements.txt -export PYTHONPATH=`pwd` +export PYTHONPATH=`pwd`:`pwd`/../../../:`pwd`/../../libraries/python STATUS=$(python pal.py \ --specification specification_CPTA.yaml \ diff --git a/resources/tools/presentation/run_report.sh b/resources/tools/presentation/run_report.sh index 56c671a8bb..9f6290e4b2 100755 --- a/resources/tools/presentation/run_report.sh +++ b/resources/tools/presentation/run_report.sh @@ -33,7 +33,7 @@ source ${DIR[WORKING]}/env/bin/activate # Install python dependencies: pip3 install -r requirements.txt -export PYTHONPATH=`pwd` +export PYTHONPATH=`pwd`:`pwd`/../../../:`pwd`/../../libraries/python python pal.py \ --specification specification.yaml \ diff --git a/resources/tools/presentation/specification.yaml b/resources/tools/presentation/specification.yaml index f4ad61686d..f24e6f355d 100644 --- a/resources/tools/presentation/specification.yaml +++ b/resources/tools/presentation/specification.yaml @@ -141,52 +141,52 @@ - 12 # NDRPDR sel - 13 # NDRPDR sel - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 21 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel +# - 21 # NDRPDR sel plot-vpp-throughput-lat-tsa-3n-hsw-vhost: csit-vpp-perf-verify-1908-3n-hsw: - 48 # NDRPDR sel vhost - 49 # NDRPDR sel vhost - 52 # NDRPDR sel vhost - - 53 # NDRPDR sel vhost - - 54 # NDRPDR sel vhost - - 55 # NDRPDR sel vhost - - 56 # NDRPDR sel vhost - - 57 # NDRPDR sel vhost - - 58 # NDRPDR sel vhost - - 59 # NDRPDR sel vhost +# - 53 # NDRPDR sel vhost +# - 54 # NDRPDR sel vhost +# - 55 # NDRPDR sel vhost +# - 56 # NDRPDR sel vhost +# - 57 # NDRPDR sel vhost +# - 58 # NDRPDR sel vhost +# - 59 # NDRPDR sel vhost plot-vpp-http-server-performance: csit-vpp-perf-verify-1908-2n-skx: - 29 # TCP - 30 # TCP - 58 # TCP - - 59 # TCP - - 60 # TCP - - 61 # TCP - - 62 # TCP - - 63 # TCP - - 64 # TCP - - 65 # TCP +# - 59 # TCP +# - 60 # TCP +# - 61 # TCP +# - 62 # TCP +# - 63 # TCP +# - 64 # TCP +# - 65 # TCP plot-dpdk-throughput-latency-3n-hsw: csit-dpdk-perf-verify-1908-3n-hsw: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel +# - 5 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel # 3n-skx @@ -195,39 +195,39 @@ - 14 # NDRPDR sel - 16 # NDRPDR sel - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 22 # NDRPDR sel - - 23 # NDRPDR sel - - 24 # NDRPDR sel - - 25 # NDRPDR sel - - 26 # NDRPDR sel - - 29 # NDRPDR sel +# - 20 # NDRPDR sel +# - 22 # NDRPDR sel +# - 23 # NDRPDR sel +# - 24 # NDRPDR sel +# - 25 # NDRPDR sel +# - 26 # NDRPDR sel +# - 29 # NDRPDR sel plot-vpp-throughput-lat-tsa-3n-skx-vhost: csit-vpp-perf-verify-1908-3n-skx: - 52 # NDRPDR sel vhost - 53 # NDRPDR sel vhost - 55 # NDRPDR sel vhost - - 61 # NDRPDR sel vhost - - 62 # NDRPDR sel vhost - - 63 # NDRPDR sel vhost - - 64 # NDRPDR sel vhost - - 65 # NDRPDR sel vhost - - 66 # NDRPDR sel vhost - - 67 # NDRPDR sel vhost +# - 61 # NDRPDR sel vhost +# - 62 # NDRPDR sel vhost +# - 63 # NDRPDR sel vhost +# - 64 # NDRPDR sel vhost +# - 65 # NDRPDR sel vhost +# - 66 # NDRPDR sel vhost +# - 67 # NDRPDR sel vhost plot-dpdk-throughput-latency-3n-skx: csit-dpdk-perf-verify-1908-3n-skx: - 3 # NDRPDR sel - 4 # NDRPDR sel - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel # 2n-skx @@ -236,26 +236,26 @@ - 26 # NDRPDR sel - 27 # NDRPDR sel - 34 # NDRPDR sel - - 35 # NDRPDR sel - - 36 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel - - 44 # NDRPDR sel - - 45 # NDRPDR sel - - 46 # NDRPDR sel +# - 35 # NDRPDR sel +# - 36 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel +# - 44 # NDRPDR sel +# - 45 # NDRPDR sel +# - 46 # NDRPDR sel plot-vpp-throughput-lat-tsa-2n-skx-vhost: csit-vpp-perf-verify-1908-2n-skx: - 88 # NDRPDR sel vhost - 89 # NDRPDR sel vhost - 91 # NDRPDR sel vhost - - 92 # NDRPDR sel vhost - - 93 # NDRPDR sel vhost - - 94 # NDRPDR sel vhost - - 95 # NDRPDR sel vhost - - 96 # NDRPDR sel vhost - - 97 # NDRPDR sel vhost - - 98 # NDRPDR sel vhost +# - 92 # NDRPDR sel vhost +# - 93 # NDRPDR sel vhost +# - 94 # NDRPDR sel vhost +# - 95 # NDRPDR sel vhost +# - 96 # NDRPDR sel vhost +# - 97 # NDRPDR sel vhost +# - 98 # NDRPDR sel vhost plot-vpp-soak-2n-skx: csit-vpp-perf-verify-1908-2n-skx: @@ -267,13 +267,13 @@ - 22 # NFV - 31 # NFV - 32 # NFV - - 37 # NFV - - 38 # NFV - - 39 # NFV - - 53 # NFV - - 67 # NFV - - 68 # NFV - - 70 # NFV +# - 37 # NFV +# - 38 # NFV +# - 39 # NFV +# - 53 # NFV +# - 67 # NFV +# - 68 # NFV +# - 70 # NFV plot-vpp-nfv-2n-skx-mrr: csit-vpp-perf-verify-1908-2n-skx: @@ -284,26 +284,26 @@ - 113 # NFV reconf sel - 114 # NFV reconf sel - 115 # NFV reconf sel - - 116 # NFV reconf sel - - 118 # NFV reconf sel - - 119 # NFV reconf sel - - 120 # NFV reconf sel - - 121 # NFV reconf sel - - 122 # NFV reconf sel - - 123 # NFV reconf sel +# - 116 # NFV reconf sel +# - 118 # NFV reconf sel +# - 119 # NFV reconf sel +# - 120 # NFV reconf sel +# - 121 # NFV reconf sel +# - 122 # NFV reconf sel +# - 123 # NFV reconf sel plot-dpdk-throughput-latency-2n-skx: csit-dpdk-perf-verify-1908-2n-skx: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 5 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel # 3n-tsh @@ -312,13 +312,13 @@ - 7 # NDRPDR sel - 8 # NDRPDR sel - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel - - 13 # NDRPDR sel - - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel +# - 13 # NDRPDR sel +# - 14 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel # 3n-dnv @@ -327,13 +327,13 @@ - 6 # NDRPDR sel - 7 # NDRPDR sel - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 12 # NDRPDR sel - - 13 # NDRPDR sel - - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 12 # NDRPDR sel +# - 13 # NDRPDR sel +# - 14 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel # 2n-dnv @@ -342,13 +342,13 @@ - 1 # NDRPDR sel - 2 # NDRPDR sel - 3 # NDRPDR sel - - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel +# - 4 # NDRPDR sel +# - 5 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel ########################################################################### # Simple tables (only one data set): @@ -360,26 +360,26 @@ - 14 # NDRPDR sel - 16 # NDRPDR sel - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 22 # NDRPDR sel - - 23 # NDRPDR sel - - 24 # NDRPDR sel - - 25 # NDRPDR sel - - 26 # NDRPDR sel - - 29 # NDRPDR sel +# - 20 # NDRPDR sel +# - 22 # NDRPDR sel +# - 23 # NDRPDR sel +# - 24 # NDRPDR sel +# - 25 # NDRPDR sel +# - 26 # NDRPDR sel +# - 29 # NDRPDR sel vpp-nic-comparison-2n-skx: csit-vpp-perf-verify-1908-2n-skx: - 26 # NDRPDR sel - 27 # NDRPDR sel - 34 # NDRPDR sel - - 35 # NDRPDR sel - - 36 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel - - 44 # NDRPDR sel - - 45 # NDRPDR sel - - 46 # NDRPDR sel +# - 35 # NDRPDR sel +# - 36 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel +# - 44 # NDRPDR sel +# - 45 # NDRPDR sel +# - 46 # NDRPDR sel vpp-nic-comparison-3n-skx-mrr: csit-vpp-perf-verify-1908-3n-skx: @@ -396,45 +396,45 @@ - 24 # NDRPDR full - 25 # NDRPDR full - 26 # NDRPDR full - - 27 # NDRPDR full - - 28 # NDRPDR full - - 29 # NDRPDR full - - 32 # NDRPDR full - - 33 # NDRPDR full - - 36 # NDRPDR full - - 38 # NDRPDR full - - 50 # NDRPDR full vhost +# - 27 # NDRPDR full +# - 28 # NDRPDR full +# - 29 # NDRPDR full +# - 32 # NDRPDR full +# - 33 # NDRPDR full +# - 36 # NDRPDR full +# - 38 # NDRPDR full +# - 50 # NDRPDR full vhost vpp-perf-results-3n-skx: csit-vpp-perf-verify-1908-3n-skx: - 27 # NDRPDR full - 28 # NDRPDR full - 30 # NDRPDR full - - 31 # NDRPDR full - - 32 # NDRPDR full - - 33 # NDRPDR full - - 34 # NDRPDR full - - 35 # NDRPDR full - - 37 # NDRPDR full - - 39 # NDRPDR full - - 40 # NDRPDR full - - 41 # NDRPDR full - - 54 # NDRPDR full +# - 31 # NDRPDR full +# - 32 # NDRPDR full +# - 33 # NDRPDR full +# - 34 # NDRPDR full +# - 35 # NDRPDR full +# - 37 # NDRPDR full +# - 39 # NDRPDR full +# - 40 # NDRPDR full +# - 41 # NDRPDR full +# - 54 # NDRPDR full vpp-perf-results-2n-skx: csit-vpp-perf-verify-1908-2n-skx: - 48 # NDRPDR full - 49 # NDRPDR full - 50 # NDRPDR full - - 51 # NDRPDR full - - 54 # NDRPDR full - - 55 # NDRPDR full - - 90 # NDRPDR full vhost - - 110 # NFV full - - 111 # NFV full - - 126 # NFV full - - 127 # NFV full - - 112 # NFV reconf full +# - 51 # NDRPDR full +# - 54 # NDRPDR full +# - 55 # NDRPDR full +# - 90 # NDRPDR full vhost +# - 110 # NFV full +# - 111 # NFV full +# - 126 # NFV full +# - 127 # NFV full +# - 112 # NFV reconf full vpp-perf-results-3n-tsh: csit-vpp-perf-verify-1908-3n-tsh: @@ -506,47 +506,47 @@ - 10 # NDRPDR sel - 19 # NDRPDR sel - 31 # NDRPDR sel - - 36 # NDRPDR sel - - 40 # NDRPDR sel - - 42 # NDRPDR sel - - 44 # NDRPDR sel - - 46 # NDRPDR sel - - 47 # NDRPDR sel - - 48 # NDRPDR sel - - 61 # NDRPDR LB - - 62 # NDRPDR LB - - 63 # NDRPDR LB - - 64 # NDRPDR LB - - 65 # NDRPDR LB - - 66 # NDRPDR LB - - 67 # NDRPDR LB - - 69 # NDRPDR LB - - 70 # NDRPDR LB - - 71 # NDRPDR LB +# - 36 # NDRPDR sel +# - 40 # NDRPDR sel +# - 42 # NDRPDR sel +# - 44 # NDRPDR sel +# - 46 # NDRPDR sel +# - 47 # NDRPDR sel +# - 48 # NDRPDR sel +# - 61 # NDRPDR LB +# - 62 # NDRPDR LB +# - 63 # NDRPDR LB +# - 64 # NDRPDR LB +# - 65 # NDRPDR LB +# - 66 # NDRPDR LB +# - 67 # NDRPDR LB +# - 69 # NDRPDR LB +# - 70 # NDRPDR LB +# - 71 # NDRPDR LB vpp-performance-changes-3n-hsw-ref: csit-vpp-perf-verify-1904-3n-hsw: - 11 # NDRPDR sel - 12 # NDRPDR sel - 20 # NDRPDR sel - - 21 # NDRPDR sel - - 23 # NDRPDR sel - - 24 # NDRPDR sel - - 25 # NDRPDR sel - - 26 # NDRPDR sel - - 29 # NDRPDR sel - - 30 # NDRPDR sel +# - 21 # NDRPDR sel +# - 23 # NDRPDR sel +# - 24 # NDRPDR sel +# - 25 # NDRPDR sel +# - 26 # NDRPDR sel +# - 29 # NDRPDR sel +# - 30 # NDRPDR sel vpp-performance-changes-3n-hsw-cmp: csit-vpp-perf-verify-1908-3n-hsw: - 12 # NDRPDR sel - 13 # NDRPDR sel - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 21 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel +# - 21 # NDRPDR sel vpp-performance-changes-3n-hsw: - "vpp-performance-changes-3n-hsw-h1" - "vpp-performance-changes-3n-hsw-ref" @@ -557,47 +557,47 @@ - 18 # NDRPDR sel - 19 # NDRPDR sel - 20 # NDRPDR sel - - 28 # NDRPDR sel - - 31 # NDRPDR sel - - 33 # NDRPDR sel - - 36 # NDRPDR sel - - 38 # NDRPDR sel - - 47 # NDRPDR sel - - 48 # NDRPDR sel - - 54 # NDRPDR LB - - 57 # NDRPDR LB - - 58 # NDRPDR LB - - 59 # NDRPDR LB - - 60 # NDRPDR LB - - 63 # NDRPDR LB - - 64 # NDRPDR LB - - 65 # NDRPDR LB - - 66 # NDRPDR LB - - 67 # NDRPDR LB +# - 28 # NDRPDR sel +# - 31 # NDRPDR sel +# - 33 # NDRPDR sel +# - 36 # NDRPDR sel +# - 38 # NDRPDR sel +# - 47 # NDRPDR sel +# - 48 # NDRPDR sel +# - 54 # NDRPDR LB +# - 57 # NDRPDR LB +# - 58 # NDRPDR LB +# - 59 # NDRPDR LB +# - 60 # NDRPDR LB +# - 63 # NDRPDR LB +# - 64 # NDRPDR LB +# - 65 # NDRPDR LB +# - 66 # NDRPDR LB +# - 67 # NDRPDR LB vpp-performance-changes-3n-skx-ref: csit-vpp-perf-verify-1904-3n-skx: - 8 # NDRPDR sel - 9 # NDRPDR sel - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel - - 14 # NDRPDR sel - - 16 # NDRPDR sel - - 21 # NDRPDR sel - - 23 # NDRPDR sel - - 25 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel +# - 14 # NDRPDR sel +# - 16 # NDRPDR sel +# - 21 # NDRPDR sel +# - 23 # NDRPDR sel +# - 25 # NDRPDR sel vpp-performance-changes-3n-skx-cmp: csit-vpp-perf-verify-1908-3n-skx: - 14 # NDRPDR sel - 16 # NDRPDR sel - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 22 # NDRPDR sel - - 23 # NDRPDR sel - - 24 # NDRPDR sel - - 25 # NDRPDR sel - - 26 # NDRPDR sel - - 29 # NDRPDR sel +# - 20 # NDRPDR sel +# - 22 # NDRPDR sel +# - 23 # NDRPDR sel +# - 24 # NDRPDR sel +# - 25 # NDRPDR sel +# - 26 # NDRPDR sel +# - 29 # NDRPDR sel vpp-performance-changes-3n-skx: - "vpp-performance-changes-3n-skx-h1" - "vpp-performance-changes-3n-skx-ref" @@ -608,37 +608,37 @@ - 5 # NDRPDR sel - 12 # NDRPDR sel - 13 # NDRPDR sel - - 21 # NDRPDR sel - - 23 # NDRPDR sel - - 31 # NDRPDR sel - - 34 # NDRPDR sel - - 40 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel +# - 21 # NDRPDR sel +# - 23 # NDRPDR sel +# - 31 # NDRPDR sel +# - 34 # NDRPDR sel +# - 40 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel vpp-performance-changes-2n-skx-ref: csit-vpp-perf-verify-1904-2n-skx: - 12 # NDRPDR sel - 14 # NDRPDR sel - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 21 # NDRPDR sel - - 27 # NDRPDR sel - - 30 # NDRPDR sel - - 31 # NDRPDR sel - - 32 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 21 # NDRPDR sel +# - 27 # NDRPDR sel +# - 30 # NDRPDR sel +# - 31 # NDRPDR sel +# - 32 # NDRPDR sel vpp-performance-changes-2n-skx-cmp: csit-vpp-perf-verify-1908-2n-skx: - 26 # NDRPDR sel - 27 # NDRPDR sel - 34 # NDRPDR sel - - 35 # NDRPDR sel - - 36 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel - - 44 # NDRPDR sel - - 45 # NDRPDR sel - - 46 # NDRPDR sel +# - 35 # NDRPDR sel +# - 36 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel +# - 44 # NDRPDR sel +# - 45 # NDRPDR sel +# - 46 # NDRPDR sel vpp-performance-changes-2n-skx: - "vpp-performance-changes-2n-skx-h1" - "vpp-performance-changes-2n-skx-ref" @@ -649,13 +649,13 @@ - 26 # NDRPDR sel - 27 # NDRPDR sel - 34 # NDRPDR sel - - 35 # NDRPDR sel - - 36 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel - - 44 # NDRPDR sel - - 45 # NDRPDR sel - - 46 # NDRPDR sel +# - 35 # NDRPDR sel +# - 36 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel +# - 44 # NDRPDR sel +# - 45 # NDRPDR sel +# - 46 # NDRPDR sel vpp-soak-vs-ndr-2n-skx-cmp: csit-vpp-perf-verify-1908-2n-skx: - 83 # SOAK @@ -665,13 +665,13 @@ - 26 # NDRPDR sel - 27 # NDRPDR sel - 34 # NDRPDR sel - - 35 # NDRPDR sel - - 36 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel - - 44 # NDRPDR sel - - 45 # NDRPDR sel - - 46 # NDRPDR sel +# - 35 # NDRPDR sel +# - 36 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel +# - 44 # NDRPDR sel +# - 45 # NDRPDR sel +# - 46 # NDRPDR sel - 83 # SOAK - 86 # SOAK @@ -722,37 +722,37 @@ - 13 # NDRPDR sel - 14 # NDRPDR sel - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 21 # NDRPDR sel - - 22 # NDRPDR sel - - 23 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel +# - 21 # NDRPDR sel +# - 22 # NDRPDR sel +# - 23 # NDRPDR sel dpdk-performance-changes-3n-hsw-ref: csit-dpdk-perf-verify-1904-3n-hsw: - 2 # NDRPDR sel - 3 # NDRPDR sel - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 13 # NDRPDR sel - - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel +# - 6 # NDRPDR sel +# - 13 # NDRPDR sel +# - 14 # NDRPDR sel +# - 15 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel dpdk-performance-changes-3n-hsw-cmp: csit-dpdk-perf-verify-1908-3n-hsw: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel +# - 5 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel dpdk-performance-changes-3n-hsw: - "dpdk-performance-changes-3n-hsw-h1" - "dpdk-performance-changes-3n-hsw-ref" @@ -763,37 +763,37 @@ - 3 # NDRPDR sel - 5 # NDRPDR sel - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel - - 13 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel +# - 13 # NDRPDR sel dpdk-performance-changes-3n-skx-ref: csit-dpdk-perf-verify-1904-3n-skx: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel - - 13 # NDRPDR sel - - 14 # NDRPDR sel +# - 5 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel +# - 13 # NDRPDR sel +# - 14 # NDRPDR sel dpdk-performance-changes-3n-skx-cmp: csit-dpdk-perf-verify-1908-3n-skx: - 3 # NDRPDR sel - 4 # NDRPDR sel - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel dpdk-performance-changes-3n-skx: - "dpdk-performance-changes-3n-skx-h1" - "dpdk-performance-changes-3n-skx-ref" @@ -804,37 +804,37 @@ - 12 # NDRPDR sel - 13 # NDRPDR sel - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 21 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel +# - 21 # NDRPDR sel dpdk-performance-changes-2n-skx-ref: csit-dpdk-perf-verify-1904-2n-skx: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 5 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel dpdk-performance-changes-2n-skx-cmp: csit-dpdk-perf-verify-1908-2n-skx: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 5 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel dpdk-performance-changes-2n-skx: - "dpdk-performance-changes-2n-skx-h1" - "dpdk-performance-changes-2n-skx-ref" @@ -847,25 +847,25 @@ - 12 # NDRPDR sel - 13 # NDRPDR sel - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 21 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel +# - 21 # NDRPDR sel vpp-compare-testbeds-cmp: csit-vpp-perf-verify-1908-3n-skx: - 14 # NDRPDR sel - 16 # NDRPDR sel - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 22 # NDRPDR sel - - 23 # NDRPDR sel - - 24 # NDRPDR sel - - 25 # NDRPDR sel - - 26 # NDRPDR sel - - 29 # NDRPDR sel +# - 20 # NDRPDR sel +# - 22 # NDRPDR sel +# - 23 # NDRPDR sel +# - 24 # NDRPDR sel +# - 25 # NDRPDR sel +# - 26 # NDRPDR sel +# - 29 # NDRPDR sel vpp-compare-testbeds: - "vpp-compare-testbeds-ref" - "vpp-compare-testbeds-cmp" @@ -885,25 +885,25 @@ - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel +# - 5 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel dpdk-compare-testbeds-cmp: csit-dpdk-perf-verify-1908-3n-skx: - 3 # NDRPDR sel - 4 # NDRPDR sel - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel dpdk-compare-testbeds: - "dpdk-compare-testbeds-ref" - "dpdk-compare-testbeds-cmp" @@ -915,25 +915,25 @@ - 14 # NDRPDR sel - 16 # NDRPDR sel - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 22 # NDRPDR sel - - 23 # NDRPDR sel - - 24 # NDRPDR sel - - 25 # NDRPDR sel - - 26 # NDRPDR sel - - 29 # NDRPDR sel +# - 20 # NDRPDR sel +# - 22 # NDRPDR sel +# - 23 # NDRPDR sel +# - 24 # NDRPDR sel +# - 25 # NDRPDR sel +# - 26 # NDRPDR sel +# - 29 # NDRPDR sel vpp-compare-topologies-cmp: csit-vpp-perf-verify-1908-2n-skx: - 26 # NDRPDR sel - 27 # NDRPDR sel - 34 # NDRPDR sel - - 35 # NDRPDR sel - - 36 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel - - 44 # NDRPDR sel - - 45 # NDRPDR sel - - 46 # NDRPDR sel +# - 35 # NDRPDR sel +# - 36 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel +# - 44 # NDRPDR sel +# - 45 # NDRPDR sel +# - 46 # NDRPDR sel vpp-compare-topologies: - "vpp-compare-topologies-ref" - "vpp-compare-topologies-cmp" @@ -953,25 +953,25 @@ - 3 # NDRPDR sel - 4 # NDRPDR sel - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel dpdk-compare-topologies-cmp: csit-dpdk-perf-verify-1908-2n-skx: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 5 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel dpdk-compare-topologies: - "dpdk-compare-topologies-ref" - "dpdk-compare-topologies-cmp" @@ -1412,67 +1412,67 @@ - 10 # NDRPDR sel - 19 # NDRPDR sel - 31 # NDRPDR sel - - 36 # NDRPDR sel - - 40 # NDRPDR sel - - 42 # NDRPDR sel - - 44 # NDRPDR sel - - 46 # NDRPDR sel - - 47 # NDRPDR sel - - 48 # NDRPDR sel - - 61 # NDRPDR LB - - 62 # NDRPDR LB - - 63 # NDRPDR LB - - 64 # NDRPDR LB - - 65 # NDRPDR LB - - 66 # NDRPDR LB - - 67 # NDRPDR LB - - 69 # NDRPDR LB - - 70 # NDRPDR LB - - 71 # NDRPDR LB +# - 36 # NDRPDR sel +# - 40 # NDRPDR sel +# - 42 # NDRPDR sel +# - 44 # NDRPDR sel +# - 46 # NDRPDR sel +# - 47 # NDRPDR sel +# - 48 # NDRPDR sel +# - 61 # NDRPDR LB +# - 62 # NDRPDR LB +# - 63 # NDRPDR LB +# - 64 # NDRPDR LB +# - 65 # NDRPDR LB +# - 66 # NDRPDR LB +# - 67 # NDRPDR LB +# - 69 # NDRPDR LB +# - 70 # NDRPDR LB +# - 71 # NDRPDR LB csit-vpp-perf-verify-1904-3n-hsw: - 9 # MRR sel - 11 # NDRPDR sel - 12 # NDRPDR sel - 20 # NDRPDR sel - - 21 # NDRPDR sel - - 23 # NDRPDR sel - - 24 # NDRPDR sel - - 25 # NDRPDR sel - - 26 # NDRPDR sel - - 29 # NDRPDR sel - - 30 # NDRPDR sel +# - 21 # NDRPDR sel +# - 23 # NDRPDR sel +# - 24 # NDRPDR sel +# - 25 # NDRPDR sel +# - 26 # NDRPDR sel +# - 29 # NDRPDR sel +# - 30 # NDRPDR sel csit-vpp-perf-verify-1908-3n-hsw: - 12 # NDRPDR sel - 13 # NDRPDR sel - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 21 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel +# - 21 # NDRPDR sel - 48 # NDRPDR sel vhost - 49 # NDRPDR sel vhost - 52 # NDRPDR sel vhost - - 53 # NDRPDR sel vhost - - 54 # NDRPDR sel vhost - - 55 # NDRPDR sel vhost - - 56 # NDRPDR sel vhost - - 57 # NDRPDR sel vhost - - 58 # NDRPDR sel vhost - - 59 # NDRPDR sel vhost +# - 53 # NDRPDR sel vhost +# - 54 # NDRPDR sel vhost +# - 55 # NDRPDR sel vhost +# - 56 # NDRPDR sel vhost +# - 57 # NDRPDR sel vhost +# - 58 # NDRPDR sel vhost +# - 59 # NDRPDR sel vhost - 24 # NDRPDR full - 25 # NDRPDR full - 26 # NDRPDR full - - 27 # NDRPDR full - - 28 # NDRPDR full - - 29 # NDRPDR full - - 32 # NDRPDR full - - 33 # NDRPDR full - - 36 # NDRPDR full - - 38 # NDRPDR full - - 50 # NDRPDR full vhost +# - 27 # NDRPDR full +# - 28 # NDRPDR full +# - 29 # NDRPDR full +# - 32 # NDRPDR full +# - 33 # NDRPDR full +# - 36 # NDRPDR full +# - 38 # NDRPDR full +# - 50 # NDRPDR full vhost - 39 # MRR sel csit-vpp-perf-verify-1901-3n-skx: @@ -1480,69 +1480,69 @@ - 18 # NDRPDR sel - 19 # NDRPDR sel - 20 # NDRPDR sel - - 28 # NDRPDR sel - - 31 # NDRPDR sel - - 33 # NDRPDR sel - - 36 # NDRPDR sel - - 38 # NDRPDR sel - - 47 # NDRPDR sel - - 48 # NDRPDR sel - - 54 # NDRPDR LB - - 57 # NDRPDR LB - - 58 # NDRPDR LB - - 59 # NDRPDR LB - - 60 # NDRPDR LB - - 63 # NDRPDR LB - - 64 # NDRPDR LB - - 65 # NDRPDR LB - - 66 # NDRPDR LB - - 67 # NDRPDR LB +# - 28 # NDRPDR sel +# - 31 # NDRPDR sel +# - 33 # NDRPDR sel +# - 36 # NDRPDR sel +# - 38 # NDRPDR sel +# - 47 # NDRPDR sel +# - 48 # NDRPDR sel +# - 54 # NDRPDR LB +# - 57 # NDRPDR LB +# - 58 # NDRPDR LB +# - 59 # NDRPDR LB +# - 60 # NDRPDR LB +# - 63 # NDRPDR LB +# - 64 # NDRPDR LB +# - 65 # NDRPDR LB +# - 66 # NDRPDR LB +# - 67 # NDRPDR LB csit-vpp-perf-verify-1904-3n-skx: - 7 # MRR sel - 8 # NDRPDR sel - 9 # NDRPDR sel - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel - - 14 # NDRPDR sel - - 16 # NDRPDR sel - - 21 # NDRPDR sel - - 23 # NDRPDR sel - - 25 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel +# - 14 # NDRPDR sel +# - 16 # NDRPDR sel +# - 21 # NDRPDR sel +# - 23 # NDRPDR sel +# - 25 # NDRPDR sel csit-vpp-perf-verify-1908-3n-skx: - 14 # NDRPDR sel - 16 # NDRPDR sel - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 22 # NDRPDR sel - - 23 # NDRPDR sel - - 24 # NDRPDR sel - - 25 # NDRPDR sel - - 26 # NDRPDR sel - - 29 # NDRPDR sel - - 52 # NDRPDR sel vhost - - 53 # NDRPDR sel vhost - - 55 # NDRPDR sel vhost - - 61 # NDRPDR sel vhost - - 62 # NDRPDR sel vhost - - 63 # NDRPDR sel vhost - - 64 # NDRPDR sel vhost - - 65 # NDRPDR sel vhost - - 66 # NDRPDR sel vhost - - 67 # NDRPDR sel vhost +# - 20 # NDRPDR sel +# - 22 # NDRPDR sel +# - 23 # NDRPDR sel +# - 24 # NDRPDR sel +# - 25 # NDRPDR sel +# - 26 # NDRPDR sel +# - 29 # NDRPDR sel +# - 52 # NDRPDR sel vhost +# - 53 # NDRPDR sel vhost +# - 55 # NDRPDR sel vhost +# - 61 # NDRPDR sel vhost +# - 62 # NDRPDR sel vhost +# - 63 # NDRPDR sel vhost +# - 64 # NDRPDR sel vhost +# - 65 # NDRPDR sel vhost +# - 66 # NDRPDR sel vhost +# - 67 # NDRPDR sel vhost - 27 # NDRPDR full - 28 # NDRPDR full - 30 # NDRPDR full - - 31 # NDRPDR full - - 32 # NDRPDR full - - 33 # NDRPDR full - - 34 # NDRPDR full - - 35 # NDRPDR full - - 37 # NDRPDR full - - 39 # NDRPDR full - - 40 # NDRPDR full - - 41 # NDRPDR full - - 54 # NDRPDR full +# - 31 # NDRPDR full +# - 32 # NDRPDR full +# - 33 # NDRPDR full +# - 34 # NDRPDR full +# - 35 # NDRPDR full +# - 37 # NDRPDR full +# - 39 # NDRPDR full +# - 40 # NDRPDR full +# - 41 # NDRPDR full +# - 54 # NDRPDR full - 36 # MRR sel csit-vpp-perf-verify-1901-2n-skx: @@ -1550,90 +1550,90 @@ - 5 # NDRPDR sel - 12 # NDRPDR sel - 13 # NDRPDR sel - - 21 # NDRPDR sel - - 23 # NDRPDR sel - - 31 # NDRPDR sel - - 34 # NDRPDR sel - - 40 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel +# - 21 # NDRPDR sel +# - 23 # NDRPDR sel +# - 31 # NDRPDR sel +# - 34 # NDRPDR sel +# - 40 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel csit-vpp-perf-verify-1904-2n-skx: - 11 # MRR sel - 12 # NDRPDR sel - 14 # NDRPDR sel - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 21 # NDRPDR sel - - 27 # NDRPDR sel - - 30 # NDRPDR sel - - 31 # NDRPDR sel - - 32 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 21 # NDRPDR sel +# - 27 # NDRPDR sel +# - 30 # NDRPDR sel +# - 31 # NDRPDR sel +# - 32 # NDRPDR sel csit-vpp-perf-verify-1908-2n-skx: - 26 # NDRPDR sel - 27 # NDRPDR sel - 34 # NDRPDR sel - - 35 # NDRPDR sel - - 36 # NDRPDR sel - - 42 # NDRPDR sel - - 43 # NDRPDR sel - - 44 # NDRPDR sel - - 45 # NDRPDR sel - - 46 # NDRPDR sel +# - 35 # NDRPDR sel +# - 36 # NDRPDR sel +# - 42 # NDRPDR sel +# - 43 # NDRPDR sel +# - 44 # NDRPDR sel +# - 45 # NDRPDR sel +# - 46 # NDRPDR sel - 88 # NDRPDR sel vhost - 89 # NDRPDR sel vhost - 91 # NDRPDR sel vhost - - 92 # NDRPDR sel vhost - - 93 # NDRPDR sel vhost - - 94 # NDRPDR sel vhost - - 95 # NDRPDR sel vhost - - 96 # NDRPDR sel vhost - - 97 # NDRPDR sel vhost - - 98 # NDRPDR sel vhost +# - 92 # NDRPDR sel vhost +# - 93 # NDRPDR sel vhost +# - 94 # NDRPDR sel vhost +# - 95 # NDRPDR sel vhost +# - 96 # NDRPDR sel vhost +# - 97 # NDRPDR sel vhost +# - 98 # NDRPDR sel vhost - 48 # NDRPDR full - 49 # NDRPDR full - 50 # NDRPDR full - - 51 # NDRPDR full - - 54 # NDRPDR full - - 55 # NDRPDR full - - 90 # NDRPDR full vhost +# - 51 # NDRPDR full +# - 54 # NDRPDR full +# - 55 # NDRPDR full +# - 90 # NDRPDR full vhost - 69 # MRR sel - 29 # TCP - 30 # TCP - 58 # TCP - - 59 # TCP - - 60 # TCP - - 61 # TCP - - 62 # TCP - - 63 # TCP - - 64 # TCP - - 65 # TCP +# - 59 # TCP +# - 60 # TCP +# - 61 # TCP +# - 62 # TCP +# - 63 # TCP +# - 64 # TCP +# - 65 # TCP - 22 # NFV - 31 # NFV - 32 # NFV - - 37 # NFV - - 38 # NFV - - 39 # NFV - - 53 # NFV - - 67 # NFV - - 68 # NFV - - 70 # NFV - - 110 # NFV full - - 111 # NFV full - - 126 # NFV full - - 127 # NFV full +# - 37 # NFV +# - 38 # NFV +# - 39 # NFV +# - 53 # NFV +# - 67 # NFV +# - 68 # NFV +# - 70 # NFV +# - 110 # NFV full +# - 111 # NFV full +# - 126 # NFV full +# - 127 # NFV full - 71 # NFV MRR - - 112 # NFV reconf full +# - 112 # NFV reconf full - 113 # NFV reconf sel - 114 # NFV reconf sel - 115 # NFV reconf sel - - 116 # NFV reconf sel - - 118 # NFV reconf sel - - 119 # NFV reconf sel - - 120 # NFV reconf sel - - 121 # NFV reconf sel - - 122 # NFV reconf sel - - 123 # NFV reconf sel +# - 116 # NFV reconf sel +# - 118 # NFV reconf sel +# - 119 # NFV reconf sel +# - 120 # NFV reconf sel +# - 121 # NFV reconf sel +# - 122 # NFV reconf sel +# - 123 # NFV reconf sel - 83 # SOAK - 86 # SOAK @@ -1641,142 +1641,142 @@ - 7 # NDRPDR sel - 8 # NDRPDR sel - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel - - 13 # NDRPDR sel - - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel +# - 13 # NDRPDR sel +# - 14 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel - 17 # MRR sel csit-vpp-perf-verify-1908-3n-dnv: - 6 # NDRPDR sel - 7 # NDRPDR sel - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 12 # NDRPDR sel - - 13 # NDRPDR sel - - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 12 # NDRPDR sel +# - 13 # NDRPDR sel +# - 14 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel - 17 # MRR sel csit-vpp-perf-verify-1908-2n-dnv: - 1 # NDRPDR sel - 2 # NDRPDR sel - 3 # NDRPDR sel - - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel +# - 4 # NDRPDR sel +# - 5 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel - 11 # MRR sel csit-dpdk-perf-verify-1901-3n-hsw: - 13 # NDRPDR sel - 14 # NDRPDR sel - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 21 # NDRPDR sel - - 22 # NDRPDR sel - - 23 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel +# - 21 # NDRPDR sel +# - 22 # NDRPDR sel +# - 23 # NDRPDR sel csit-dpdk-perf-verify-1904-3n-hsw: - 2 # NDRPDR sel - 3 # NDRPDR sel - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 13 # NDRPDR sel - - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel +# - 6 # NDRPDR sel +# - 13 # NDRPDR sel +# - 14 # NDRPDR sel +# - 15 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel csit-dpdk-perf-verify-1908-3n-hsw: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel +# - 5 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel csit-dpdk-perf-verify-1901-3n-skx: - 3 # NDRPDR sel - 5 # NDRPDR sel - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel - - 13 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel +# - 13 # NDRPDR sel csit-dpdk-perf-verify-1904-3n-skx: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel - - 13 # NDRPDR sel - - 14 # NDRPDR sel +# - 5 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel +# - 13 # NDRPDR sel +# - 14 # NDRPDR sel csit-dpdk-perf-verify-1908-3n-skx: - 3 # NDRPDR sel - 4 # NDRPDR sel - 5 # NDRPDR sel - - 6 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 6 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel csit-dpdk-perf-verify-1901-2n-skx: - 12 # NDRPDR sel - 13 # NDRPDR sel - 14 # NDRPDR sel - - 15 # NDRPDR sel - - 16 # NDRPDR sel - - 17 # NDRPDR sel - - 18 # NDRPDR sel - - 19 # NDRPDR sel - - 20 # NDRPDR sel - - 21 # NDRPDR sel +# - 15 # NDRPDR sel +# - 16 # NDRPDR sel +# - 17 # NDRPDR sel +# - 18 # NDRPDR sel +# - 19 # NDRPDR sel +# - 20 # NDRPDR sel +# - 21 # NDRPDR sel csit-dpdk-perf-verify-1904-2n-skx: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 5 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel csit-dpdk-perf-verify-1908-2n-skx: - 2 # NDRPDR sel - 3 # NDRPDR sel - 4 # NDRPDR sel - - 5 # NDRPDR sel - - 7 # NDRPDR sel - - 8 # NDRPDR sel - - 9 # NDRPDR sel - - 10 # NDRPDR sel - - 11 # NDRPDR sel - - 12 # NDRPDR sel +# - 5 # NDRPDR sel +# - 7 # NDRPDR sel +# - 8 # NDRPDR sel +# - 9 # NDRPDR sel +# - 10 # NDRPDR sel +# - 11 # NDRPDR sel +# - 12 # NDRPDR sel csit-vpp-functional-1908-ubuntu1604-virl: - 64 @@ -2104,7 +2104,7 @@ # VPP Performance Changes 3n-hsw 1t1c pdr - type: "table" title: "VPP Performance Changes 3n-hsw 1t1c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-hsw-1t1c-pdr" history: - title: "rls1901" @@ -2127,7 +2127,7 @@ # VPP Performance Changes 3n-hsw 2t2c pdr - type: "table" title: "VPP Performance Changes 3n-hsw 2t2c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-hsw-2t2c-pdr" history: - title: "rls1901" @@ -2150,7 +2150,7 @@ # VPP Performance Changes 3n-hsw 1t1c ndr - type: "table" title: "VPP Performance Changes 3n-hsw 1t1c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-hsw-1t1c-ndr" history: - title: "rls1901" @@ -2173,7 +2173,7 @@ # VPP Performance Changes 3n-hsw 2t2c ndr - type: "table" title: "VPP Performance Changes 3n-hsw 2t2c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-hsw-2t2c-ndr" history: - title: "rls1901" @@ -2196,7 +2196,7 @@ # VPP Performance Changes 3n-skx 2t1c pdr - type: "table" title: "VPP Performance Changes 3n-skx 2t1c pdr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-skx-2t1c-pdr" history: - title: "rls1901 x710" @@ -2223,7 +2223,7 @@ # VPP Performance Changes 3n-skx 4t2c pdr - type: "table" title: "VPP Performance Changes 3n-skx 4t2c pdr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-skx-4t2c-pdr" history: - title: "rls1901 x710" @@ -2250,7 +2250,7 @@ # VPP Performance Changes 3n-skx 2t1c ndr - type: "table" title: "VPP Performance Changes 3n-skx 2t1c ndr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-skx-2t1c-ndr" history: - title: "rls1901 x710" @@ -2277,7 +2277,7 @@ # VPP Performance Changes 3n-skx 4t2c ndr - type: "table" title: "VPP Performance Changes 3n-skx 4t2c ndr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-skx-4t2c-ndr" history: - title: "rls1901 x710" @@ -2304,7 +2304,7 @@ # VPP Performance Changes 2n-skx 2t1c pdr - type: "table" title: "VPP Performance Changes 2n-skx 2t1c pdr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-changes-2n-skx-2t1c-pdr" history: - title: "rls1901 x710" @@ -2331,7 +2331,7 @@ # VPP Performance Changes 2n-skx 4t2c pdr - type: "table" title: "VPP Performance Changes 2n-skx 4t2c pdr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-changes-2n-skx-4t2c-pdr" history: - title: "rls1901 x710" @@ -2358,7 +2358,7 @@ # VPP Performance Changes 2n-skx 2t1c ndr - type: "table" title: "VPP Performance Changes 2n-skx 2t1c ndr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-changes-2n-skx-2t1c-ndr" history: - title: "rls1901 x710" @@ -2385,7 +2385,7 @@ # VPP Performance Changes 2n-skx 4t2c ndr - type: "table" title: "VPP Performance Changes 2n-skx 4t2c ndr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-changes-2n-skx-4t2c-ndr" history: - title: "rls1901 x710" @@ -2412,7 +2412,7 @@ # VPP Comparison Across Testbeds 3n-hsw to 3n-skx 1c ndr - type: "table" title: "VPP Comparison Across Testbeds 3n-hsw to 3n-skx 1c ndr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-compare-testbeds-3n-hsw-3n-skx-1c-ndr" reference: title: "3n-hsw xl710" @@ -2434,7 +2434,7 @@ # VPP Comparison Across Testbeds 3n-hsw to 3n-skx 2c ndr - type: "table" title: "VPP Comparison Across Testbeds 3n-hsw to 3n-skx 2c ndr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-compare-testbeds-3n-hsw-3n-skx-2c-ndr" reference: title: "3n-hsw xl710" @@ -2456,7 +2456,7 @@ # VPP Comparison Across Testbeds 3n-hsw to 3n-skx 1c pdr - type: "table" title: "VPP Comparison Across Testbeds 3n-hsw to 3n-skx 1c pdr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-compare-testbeds-3n-hsw-3n-skx-1c-pdr" reference: title: "3n-hsw xl710" @@ -2478,7 +2478,7 @@ # VPP Comparison Across Testbeds 3n-hsw to 3n-skx 2c pdr - type: "table" title: "VPP Comparison Across Testbeds 3n-hsw to 3n-skx 2c pdr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-compare-testbeds-3n-hsw-3n-skx-2c-pdr" reference: title: "3n-hsw xl710" @@ -2500,7 +2500,7 @@ # VPP Comparison Across Topologies 3n-skx to 2n-skx 1c ndr - type: "table" title: "VPP Comparison Across Topologies 3n-skx to 2n-skx 1c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-compare-topologies-3n-skx-2n-skx-1c-ndr" reference: title: "3-Node Skx" @@ -2519,7 +2519,7 @@ # VPP Comparison Across Topologies 3n-skx to 2n-skx 2c ndr - type: "table" title: "VPP Comparison Across Topologies 3n-skx to 2n-skx 2c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-compare-topologies-3n-skx-2n-skx-2c-ndr" reference: title: "3-Node Skx" @@ -2538,7 +2538,7 @@ # VPP Comparison Across Topologies 3n-skx to 2n-skx 1c pdr - type: "table" title: "VPP Comparison Across Topologies 3n-skx to 2n-skx 1c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-compare-topologies-3n-skx-2n-skx-1c-pdr" reference: title: "3-Node Skx" @@ -2557,7 +2557,7 @@ # VPP Comparison Across Topologies 3n-skx to 2n-skx 2c pdr - type: "table" title: "VPP Comparison Across Topologies 3n-skx to 2n-skx 2c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-compare-topologies-3n-skx-2n-skx-2c-pdr" reference: title: "3-Node Skx" @@ -2576,7 +2576,7 @@ # VPP Performance Changes 3n-hsw 1t1c MRR - type: "table" title: "VPP Performance Changes 3n-hsw 1t1c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-hsw-1t1c-mrr" history: - title: "rls1901" @@ -2598,7 +2598,7 @@ # VPP Performance Changes 3n-hsw 2t2c MRR - type: "table" title: "VPP Performance Changes 3n-hsw 2t2c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-hsw-2t2c-mrr" history: - title: "rls1901" @@ -2620,7 +2620,7 @@ # VPP Performance Changes 3n-hsw 4t4c MRR - type: "table" title: "VPP Performance Changes 3n-hsw 4t4c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-hsw-4t4c-mrr" history: - title: "rls1901" @@ -2642,7 +2642,7 @@ # VPP Performance Changes 3n-skx 2t1c MRR - type: "table" title: "VPP Performance Changes 3n-skx 2t1c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-skx-2t1c-mrr" history: - title: "rls1901" @@ -2664,7 +2664,7 @@ # VPP Performance Changes 3n-skx 4t2c MRR - type: "table" title: "VPP Performance Changes 3n-skx 4t2c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-skx-4t2c-mrr" history: - title: "rls1901" @@ -2686,7 +2686,7 @@ # VPP Performance Changes 3n-skx 8t4c MRR - type: "table" title: "VPP Performance Changes 3n-skx 8t4c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-3n-skx-8t4c-mrr" history: - title: "rls1901" @@ -2708,7 +2708,7 @@ # VPP Performance Changes 2n-skx 2t1c MRR - type: "table" title: "VPP Performance Changes 2n-skx 2t1c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-2n-skx-2t1c-mrr" history: - title: "rls1901" @@ -2730,7 +2730,7 @@ # VPP Performance Changes 2n-skx 4t2c MRR - type: "table" title: "VPP Performance Changes 2n-skx 4t2c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-2n-skx-4t2c-mrr" history: - title: "rls1901" @@ -2752,7 +2752,7 @@ # VPP Performance Changes 2n-skx 8t4c MRR - type: "table" title: "VPP Performance Changes 2n-skx 8t4c MRR" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-changes-2n-skx-8t4c-mrr" history: - title: "rls1901" @@ -2774,7 +2774,7 @@ # VPP Comparison Across Testbeds 3n-hsw to 3n-skx 1c mrr - type: "table" title: "VPP Comparison Across Testbeds 3n-hsw to 3n-skx 1c mrr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-compare-testbeds-3n-hsw-3n-skx-1c-mrr" reference: title: "3n-hsw xl710" @@ -2796,7 +2796,7 @@ # VPP Comparison Across Testbeds 3n-hsw to 3n-skx 2c mrr - type: "table" title: "VPP Comparison Across Testbeds 3n-hsw to 3n-skx 2c mrr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-compare-testbeds-3n-hsw-3n-skx-2c-mrr" reference: title: "3n-hsw xl710" @@ -2818,7 +2818,7 @@ # VPP Comparison Across Testbeds 3n-hsw to 3n-skx 4c mrr - type: "table" title: "VPP Comparison Across Testbeds 3n-hsw to 3n-skx 4c mrr" - algorithm: "table_performance_comparison_nic" + algorithm: "table_perf_comparison_nic" output-file: "{DIR[STATIC,VPP]}/performance-compare-testbeds-3n-hsw-3n-skx-4c-mrr" reference: title: "3n-hsw xl710" @@ -2840,7 +2840,7 @@ # VPP Comparison Across Topologies 3n-skx to 2n-skx 1c mrr - type: "table" title: "VPP Comparison Across Topologies 3n-skx to 2n-skx 1c mrr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-compare-topologies-3n-skx-2n-skx-1c-mrr" reference: title: "3-Node Skx" @@ -2859,7 +2859,7 @@ # VPP Comparison Across Topologies 3n-skx to 2n-skx 2c mrr - type: "table" title: "VPP Comparison Across Topologies 3n-skx to 2n-skx 2c mrr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-compare-topologies-3n-skx-2n-skx-2c-mrr" reference: title: "3-Node Skx" @@ -2878,7 +2878,7 @@ # VPP Comparison Across Topologies 3n-skx to 2n-skx 4c mrr - type: "table" title: "VPP Comparison Across Topologies 3n-skx to 2n-skx 4c mrr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,VPP]}/performance-compare-topologies-3n-skx-2n-skx-4c-mrr" reference: title: "3-Node Skx" @@ -3613,7 +3613,7 @@ # DPDK Performance Changes 3n-hsw 1t1c pdr - type: "table" title: "DPDK Performance Changes 3n-hsw 1t1c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-3n-hsw-1t1c-pdr" history: - title: "rls1901" @@ -3635,7 +3635,7 @@ # DPDK Performance Changes 3n-hsw 2t2c pdr - type: "table" title: "DPDK Performance Changes 3n-hsw 2t2c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-3n-hsw-2t2c-pdr" history: - title: "rls1901" @@ -3657,7 +3657,7 @@ # DPDK Performance Changes 3n-hsw 1t1c ndr - type: "table" title: "DPDK Performance Changes 3n-hsw 1t1c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-3n-hsw-1t1c-ndr" history: - title: "rls1901" @@ -3679,7 +3679,7 @@ # DPDK Performance Changes 3n-hsw 2t2c ndr - type: "table" title: "DPDK Performance Changes 3n-hsw 2t2c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-3n-hsw-2t2c-ndr" history: - title: "rls1901" @@ -3701,7 +3701,7 @@ # DPDK Performance Changes 3n-skx 2t1c pdr - type: "table" title: "DPDK Performance Changes 3n-skx 2t1c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-3n-skx-2t1c-pdr" history: - title: "rls1901" @@ -3723,7 +3723,7 @@ # DPDK Performance Changes 3n-skx 4t2c pdr - type: "table" title: "DPDK Performance Changes 3n-skx 4t2c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-3n-skx-4t2c-pdr" history: - title: "rls1901" @@ -3745,7 +3745,7 @@ # DPDK Performance Changes 3n-skx 2t1c ndr - type: "table" title: "DPDK Performance Changes 3n-skx 2t1c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-3n-skx-2t1c-ndr" history: - title: "rls1901" @@ -3767,7 +3767,7 @@ # DPDK Performance Changes 3n-skx 4t2c ndr - type: "table" title: "DPDK Performance Changes 3n-skx 4t2c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-3n-skx-4t2c-ndr" history: - title: "rls1901" @@ -3789,7 +3789,7 @@ # DPDK Comparison Across Testbeds 3n-hsw to 3n-skx ndr - type: "table" title: "DPDK Comparison Across Testbeds 3n-hsw to 3n-skx ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-compare-testbeds-3n-hsw-3n-skx-ndr" reference: title: "3-Node Hsw" @@ -3808,7 +3808,7 @@ # DPDK Comparison Across Testbeds 3n-hsw to 3n-skx pdr - type: "table" title: "DPDK Comparison Across Testbeds 3n-hsw to 3n-skx pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-compare-testbeds-3n-hsw-3n-skx-pdr" reference: title: "3-Node Hsw" @@ -3827,7 +3827,7 @@ # DPDK Comparison Across Topologies 3n-skx to 2n-skx ndr - type: "table" title: "DPDK Comparison Across Topologies 3n-skx to 2n-skx ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-compare-topologies-3n-skx-2n-skx-ndr" reference: title: "3-Node Skx" @@ -3846,7 +3846,7 @@ # DPDK Comparison Across Topologies 3n-skx to 2n-skx pdr - type: "table" title: "DPDK Comparison Across Topologies 3n-skx to 2n-skx pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-compare-topologies-3n-skx-2n-skx-pdr" reference: title: "3-Node Skx" @@ -3865,7 +3865,7 @@ # DPDK Performance Changes 2n-skx 2t1c pdr - type: "table" title: "DPDK Performance Changes 2n-skx 2t1c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-2n-skx-2t1c-pdr" history: - title: "rls1901" @@ -3887,7 +3887,7 @@ # DPDK Performance Changes 2n-skx 4t2c pdr - type: "table" title: "DPDK Performance Changes 2n-skx 4t2c pdr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-2n-skx-4t2c-pdr" history: - title: "rls1901" @@ -3909,7 +3909,7 @@ # DPDK Performance Changes 2n-skx 2t1c ndr - type: "table" title: "DPDK Performance Changes 2n-skx 2t1c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-2n-skx-2t1c-ndr" history: - title: "rls1901" @@ -3931,7 +3931,7 @@ # DPDK Performance Changes 2n-skx 4t2c ndr - type: "table" title: "DPDK Performance Changes 2n-skx 4t2c ndr" - algorithm: "table_performance_comparison" + algorithm: "table_perf_comparison" output-file: "{DIR[STATIC,DPDK]}/performance-changes-2n-skx-4t2c-ndr" history: - title: "rls1901" @@ -4683,7 +4683,7 @@ # Plots VPP HTTP Server Performance - type: "plot" title: "VPP HTTP Server Performance" - algorithm: "plot_http_server_performance_box" + algorithm: "plot_http_server_perf_box" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/http-server-performance-cps" data: @@ -4705,7 +4705,7 @@ - type: "plot" title: "VPP HTTP Server Performance" - algorithm: "plot_http_server_performance_box" + algorithm: "plot_http_server_perf_box" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/http-server-performance-rps" data: @@ -4728,26 +4728,17 @@ # Soak Test (PLRsearch), boxes - type: "plot" title: "VPP Critical rate: 30 Minutes Soak Test (PLRsearch)" - algorithm: "plot_performance_box" + algorithm: "plot_perf_box_name" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/soak-test-1" data: "plot-vpp-soak-2n-skx" - filter: "('L2BDMACLRN' or 'L2PATCH' or 'L2XCBASE') and not 'VHOST' and not 'MEMIF'" - parameters: - - "throughput" - - "parent" - - "tags" - sort: - - "DRV_AVF" - - "DRV_AVF" - - "DRV_AVF" - - "L2BDMACLRN" - - "L2XCBASE" - - "L2PATCH" - traces: - hoverinfo: "y+name" - boxpoints: "all" - whiskerwidth: 0 + include: + - "Tests.Vpp.Perf.L2.2N1L-25Ge2P1Xxv710-Avf-Eth-L2Patch-Soak.64B-2t1c-avf-eth-l2patch-soak" + - "Tests.Vpp.Perf.L2.2N1L-25Ge2P1Xxv710-Avf-Eth-L2Xcbase-Soak.64B-2t1c-avf-eth-l2xcbase-soak" + - "Tests.Vpp.Perf.L2.2N1L-25Ge2P1Xxv710-Avf-Eth-L2Bdbasemaclrn-Soak.64B-2t1c-avf-eth-l2bdbasemaclrn-soak" + - "Tests.Vpp.Perf.L2.2N1L-25Ge2P1Xxv710-Eth-L2Patch-Soak.64B-2t1c-eth-l2patch-soak" + - "Tests.Vpp.Perf.L2.2N1L-25Ge2P1Xxv710-Eth-L2Xcbase-Soak.64B-2t1c-eth-l2xcbase-soak" + - "Tests.Vpp.Perf.L2.2N1L-25Ge2P1Xxv710-Eth-L2Bdbasemaclrn-Soak.64B-2t1c-eth-l2bdbasemaclrn-soak" layout: title: "30 Minutes Soak Test (PLRsearch)" layout: "plot-soak-throughput" @@ -4755,26 +4746,17 @@ # Soak Test (PLRsearch), boxes - type: "plot" title: "VPP Critical rate: 30 Minutes Soak Test (PLRsearch)" - algorithm: "plot_performance_box" + algorithm: "plot_perf_box_name" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/soak-test-2" data: "plot-vpp-soak-2n-skx" - filter: "'L2BDMACLRN' and ('VHOST' or 'MEMIF') or 'IP4BASE' or 'IP6BASE'" - parameters: - - "throughput" - - "parent" - - "tags" - sort: - - "VHOST" - - "VHOST" - - "MEMIF" - - "IP4FWD" - - "IP4FWD" - - "IP6FWD" - traces: - hoverinfo: "y+name" - boxpoints: "all" - whiskerwidth: 0 + include: + - "Tests.Vpp.Perf.Vm Vhost.2N1L-25Ge2P1Xxv710-Eth-L2Bdbasemaclrn-Eth-2Vhostvr1024-1Vm-Vppl2Xc-Soak.64B-2t1c-eth-l2bdbasemaclrn-eth-2vhostvr1024-1vm-vppl2xc-soak" + - "Tests.Vpp.Perf.Vm Vhost.2N1L-25Ge2P1Xxv710-Eth-L2Bdbasemaclrn-Eth-2Vhostvr1024-1Vm-Soak.64B-2t1c-eth-l2bdbasemaclrn-eth-2vhostvr1024-1vm-soak" + - "Tests.Vpp.Perf.Container Memif.2N1L-25Ge2P1Xxv710-Eth-L2Bdbasemaclrn-Eth-2Memif-1Dcr-Soak.64B-2t1c-eth-l2bdbasemaclrn-eth-2memif-1dcr-soak" + - "Tests.Vpp.Perf.Ip4.2N1L-25Ge2P1Xxv710-Avf-Ethip4-Ip4Base-Soak.64B-2t1c-avf-ethip4-ip4base-soak" + - "Tests.Vpp.Perf.Ip4.2N1L-25Ge2P1Xxv710-Ethip4-Ip4Base-Soak.64B-2t1c-ethip4-ip4base-soak" + - "Tests.Vpp.Perf.Ip6.2N1L-25Ge2P1Xxv710-Ethip6-Ip6Base-Soak.78B-2t1c-ethip6-ip6base-soak" layout: title: "30 Minutes Soak Test (PLRsearch)" layout: "plot-soak-throughput" @@ -4783,7 +4765,7 @@ - type: "plot" title: "NFV Implied time lost: 2n-skx-xxv710-imix-2t1c-dot1qip4vxlan-l2bd-{Y}ch-1ach-{2XY}vh-{XY}vm-reconf" - algorithm: "plot_service_density_reconf_box_name" + algorithm: "plot_nf_reconf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-imix-2t1c-dot1qip4vxlan-l2bd-reconf" data: "plot-vpp-nfv-reconf-2n-skx" include: @@ -4799,7 +4781,7 @@ - type: "plot" title: "NFV Implied time lost: 2n-skx-xxv710-imix-4t2c-dot1qip4vxlan-l2bd-{Y}ch-1ach-{2XY}vh-{XY}vm-reconf" - algorithm: "plot_service_density_reconf_box_name" + algorithm: "plot_nf_reconf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-imix-4t2c-dot1qip4vxlan-l2bd-reconf" data: "plot-vpp-nfv-reconf-2n-skx" include: @@ -4815,7 +4797,7 @@ - type: "plot" title: "NFV Implied time lost: 2n-skx-xxv710-imix-8t4c-dot1qip4vxlan-l2bd-{Y}ch-1ach-{2XY}vh-{XY}vm-reconf" - algorithm: "plot_service_density_reconf_box_name" + algorithm: "plot_nf_reconf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-imix-8t4c-dot1qip4vxlan-l2bd-reconf" data: "plot-vpp-nfv-reconf-2n-skx" include: @@ -4832,7 +4814,7 @@ # Plots - NF Density - VSC - MRR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-vsc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -4864,7 +4846,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-vsc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -4896,7 +4878,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-vsc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -4929,7 +4911,7 @@ # Plots - NF Density - VSC - MRR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-vsc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -4961,7 +4943,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-vsc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -4993,7 +4975,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-vsc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -5026,7 +5008,7 @@ # Plots - NF Density - VSC - NDR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-vsc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5058,7 +5040,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-vsc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5090,7 +5072,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-vsc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5123,7 +5105,7 @@ # Plots - NF Density - VSC - NDR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-vsc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5155,7 +5137,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-vsc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5187,7 +5169,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-vsc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5220,7 +5202,7 @@ # Plots - NF Density - VSC - PDR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-vsc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5252,7 +5234,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-vsc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5284,7 +5266,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-vsc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5317,7 +5299,7 @@ # Plots - NF Density - VSC - PDR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-vsc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5349,7 +5331,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-vsc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5381,7 +5363,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}vh-{XY}vm-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-vsc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5415,7 +5397,7 @@ # Plots - NF Density - CSC - MRR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-csc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -5447,7 +5429,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-csc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -5479,7 +5461,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-csc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -5512,7 +5494,7 @@ # Plots - NF Density - CSC - MRR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-csc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -5544,7 +5526,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-csc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -5576,7 +5558,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-csc-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -5609,7 +5591,7 @@ # Plots - NF Density - CSC - NDR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-csc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5641,7 +5623,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-csc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5673,7 +5655,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-csc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5706,7 +5688,7 @@ # Plots - NF Density - CSC - NDR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-csc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5738,7 +5720,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-csc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5770,7 +5752,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-csc-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5803,7 +5785,7 @@ # Plots - NF Density - CSC - PDR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-csc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5835,7 +5817,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-csc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5867,7 +5849,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-csc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5900,7 +5882,7 @@ # Plots - NF Density - CSC - PDR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-csc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5932,7 +5914,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-csc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5964,7 +5946,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}ch-{2XY}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-csc-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -5998,7 +5980,7 @@ # Plots - NF Density - CSP - MRR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-csp-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -6030,7 +6012,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-csp-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -6062,7 +6044,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-csp-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -6095,7 +6077,7 @@ # Plots - NF Density - CSP - MRR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-csp-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -6127,7 +6109,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-csp-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -6159,7 +6141,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-mrr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-csp-mrr" data: "plot-vpp-nfv-2n-skx-mrr" @@ -6192,7 +6174,7 @@ # Plots - NF Density - CSP - NDR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-csp-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6224,7 +6206,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-csp-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6256,7 +6238,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-csp-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6289,7 +6271,7 @@ # Plots - NF Density - CSP - NDR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-csp-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6321,7 +6303,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-csp-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6353,7 +6335,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-ndr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-csp-ndr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6386,7 +6368,7 @@ # Plots - NF Density - CSP - PDR 64b - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-2t1c-base-csp-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6418,7 +6400,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-4t2c-base-csp-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6450,7 +6432,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-64b-8t4c-base-csp-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6483,7 +6465,7 @@ # Plots - NF Density - CSP - PDR IMIX - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-2t1c-base-csp-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6515,7 +6497,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-4t2c-base-csp-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6547,7 +6529,7 @@ - type: "plot" title: "Packet Throughput: eth-l2bd-{Y}pl-{2Y}mif-{XY}dcr-vppip4-pdr" - algorithm: "plot_service_density_heatmap" + algorithm: "plot_nf_heatmap" output-file-type: ".html" output-file: "{DIR[STATIC,VPP]}/l2bd-2n-skx-xxv710-imix-8t4c-base-csp-pdr" data: "plot-vpp-nfv-2n-skx-ndrpdr" @@ -6583,7 +6565,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-memif-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-memif-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6597,7 +6579,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-memif-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-memif-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6611,7 +6593,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-memif-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-memif-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6626,7 +6608,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-memif-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-memif-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6641,7 +6623,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-memif-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-memif-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -6656,7 +6638,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-memif-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-memif-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -6671,7 +6653,7 @@ - type: "plot" title: "Throughput: 2n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6686,7 +6668,7 @@ - type: "plot" title: "Throughput: 2n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6701,7 +6683,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6715,7 +6697,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6729,7 +6711,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6744,7 +6726,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6759,7 +6741,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6775,7 +6757,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -6791,7 +6773,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-ip4routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-ip4routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -6804,7 +6786,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-ip4routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-ip4routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -6817,7 +6799,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6832,7 +6814,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6847,7 +6829,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6861,7 +6843,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6875,7 +6857,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6890,7 +6872,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6905,7 +6887,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6921,7 +6903,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -6937,7 +6919,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-ip4routing-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-ip4routing-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -6952,7 +6934,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-ip4routing-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-ip4routing-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -6967,7 +6949,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-features-ip4routing-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-features-ip4routing-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -6983,7 +6965,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-features-ip4routing-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-features-ip4routing-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -6999,7 +6981,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-64b-1t1c-ip4routing-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-1t1c-ip4routing-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -7014,7 +6996,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-64b-1t1c-ip4routing-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-1t1c-ip4routing-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -7029,7 +7011,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-64b-1t1c-features-ip4routing-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-1t1c-features-ip4routing-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -7045,7 +7027,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-64b-1t1c-features-ip4routing-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-1t1c-features-ip4routing-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -7061,7 +7043,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-ip4routing-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-ip4routing-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7076,7 +7058,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-ip4routing-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-ip4routing-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7091,7 +7073,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-features-ip4routing-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-features-ip4routing-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7107,7 +7089,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-features-ip4routing-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-features-ip4routing-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7123,7 +7105,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-ip4tunnel-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-ip4tunnel-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7135,7 +7117,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-ip4tunnel-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-ip4tunnel-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7147,7 +7129,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-ip4tunnel-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4tunnel-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7161,7 +7143,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-ip4tunnel-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4tunnel-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7175,7 +7157,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-ip4tunnel-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-ip4tunnel-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7189,7 +7171,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-ip4tunnel-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-ip4tunnel-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7203,7 +7185,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-ip4tunnel-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-ip4tunnel-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7217,7 +7199,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-ip4tunnel-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-ip4tunnel-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7231,7 +7213,7 @@ - type: "plot" title: "Throughput: 2n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7244,7 +7226,7 @@ - type: "plot" title: "Throughput: 2n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7257,7 +7239,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7272,7 +7254,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7287,7 +7269,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-78b-1t1c-ip6routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-1t1c-ip6routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7300,7 +7282,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-78b-1t1c-ip6routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-1t1c-ip6routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7313,7 +7295,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7326,7 +7308,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7339,7 +7321,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7354,7 +7336,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7369,7 +7351,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-78b-1t1c-ip6routing-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-1t1c-ip6routing-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7384,7 +7366,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-78b-1t1c-ip6routing-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-1t1c-ip6routing-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7399,7 +7381,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-78b-1t1c-ip6routing-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-78b-1t1c-ip6routing-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -7414,7 +7396,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-78b-1t1c-ip6routing-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-78b-1t1c-ip6routing-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -7429,7 +7411,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-78b-1t1c-ip6routing-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-78b-1t1c-ip6routing-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7444,7 +7426,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-78b-1t1c-ip6routing-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-78b-1t1c-ip6routing-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7459,7 +7441,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-sw-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-sw-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7475,7 +7457,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-sw-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-sw-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7491,7 +7473,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-hw-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-hw-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7505,7 +7487,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-hw-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-hw-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7519,7 +7501,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-imix-2t1c-ipsec-ip4routing-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-imix-2t1c-ipsec-ip4routing-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7535,7 +7517,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-imix-2t1c-ipsec-ip4routing-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-imix-2t1c-ipsec-ip4routing-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7551,7 +7533,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7567,7 +7549,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7583,7 +7565,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7599,7 +7581,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -7615,7 +7597,7 @@ - type: "plot" title: "Throughput: 2n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7631,7 +7613,7 @@ - type: "plot" title: "Throughput: 2n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7647,7 +7629,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-l2switching-base-avf-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-avf-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7661,7 +7643,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-l2switching-base-avf-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-avf-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7675,7 +7657,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-l2switching-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7690,7 +7672,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-l2switching-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7705,7 +7687,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7719,7 +7701,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -7733,7 +7715,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-l2switching-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-l2switching-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7749,7 +7731,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-l2switching-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-l2switching-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -7765,7 +7747,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7780,7 +7762,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7795,7 +7777,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-l2switching-base-avf-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-avf-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7808,7 +7790,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-l2switching-base-avf-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-avf-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7821,7 +7803,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-l2switching-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7835,7 +7817,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-l2switching-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7849,7 +7831,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7865,7 +7847,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7881,7 +7863,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-features-l2switching-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-features-l2switching-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7897,7 +7879,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-features-l2switching-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-features-l2switching-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -7913,7 +7895,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-l2switching-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-l2switching-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7927,7 +7909,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-l2switching-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-l2switching-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7941,7 +7923,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-l2switching-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-l2switching-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7957,7 +7939,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-l2switching-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-l2switching-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7973,7 +7955,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-features-l2switching-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-features-l2switching-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -7989,7 +7971,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-features-l2switching-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-features-l2switching-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8005,7 +7987,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-64b-1t1c-l2switching-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-1t1c-l2switching-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -8019,7 +8001,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-64b-1t1c-l2switching-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-1t1c-l2switching-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -8033,7 +8015,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-64b-1t1c-l2switching-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-1t1c-l2switching-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -8049,7 +8031,7 @@ - type: "plot" title: "Throughput: 2n-dnv-x553-64b-1t1c-l2switching-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-1t1c-l2switching-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -8065,7 +8047,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-l2switching-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-l2switching-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -8079,7 +8061,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-l2switching-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-l2switching-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -8093,7 +8075,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-l2switching-base-scale-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-l2switching-base-scale-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -8109,7 +8091,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-l2switching-base-scale-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-l2switching-base-scale-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -8125,7 +8107,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-features-l2switching-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-features-l2switching-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -8141,7 +8123,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-64b-1t1c-features-l2switching-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-1t1c-features-l2switching-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -8157,7 +8139,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-78b-1t1c-srv6-ip6routing-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-1t1c-srv6-ip6routing-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -8173,7 +8155,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-78b-1t1c-srv6-ip6routing-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-1t1c-srv6-ip6routing-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -8189,7 +8171,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-78b-2t1c-srv6-ip6routing-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-2t1c-srv6-ip6routing-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8205,7 +8187,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-78b-2t1c-srv6-ip6routing-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-2t1c-srv6-ip6routing-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8221,7 +8203,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-78b-1t1c-srv6-ip6routing-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-1t1c-srv6-ip6routing-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8237,7 +8219,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-78b-1t1c-srv6-ip6routing-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-1t1c-srv6-ip6routing-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8253,7 +8235,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-78b-1t1c-srv6-ip6routing-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-78b-1t1c-srv6-ip6routing-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -8266,7 +8248,7 @@ - type: "plot" title: "Throughput: 3n-dnv-x553-78b-1t1c-srv6-ip6routing-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-78b-1t1c-srv6-ip6routing-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -8279,7 +8261,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-vhost-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-vhost-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -8293,7 +8275,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-vhost-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-vhost-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -8307,7 +8289,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-ndr" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -8321,7 +8303,7 @@ - type: "plot" title: "Throughput: 2n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-pdr" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -8335,7 +8317,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-vhost-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-vhost-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -8349,7 +8331,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-vhost-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-vhost-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -8363,7 +8345,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-vhost-base-i40e-vpp-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -8377,7 +8359,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-64b-1t1c-vhost-base-i40e-vpp-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-vhost-base-i40e-vpp-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -8391,7 +8373,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8405,7 +8387,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8419,7 +8401,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8433,7 +8415,7 @@ - type: "plot" title: "Throughput: 3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8447,7 +8429,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-vhost-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-vhost-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8462,7 +8444,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-vhost-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-vhost-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8477,7 +8459,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8492,7 +8474,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8507,7 +8489,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8523,7 +8505,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8539,7 +8521,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8555,7 +8537,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -8571,7 +8553,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-vhost-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-vhost-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8586,7 +8568,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-vhost-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-vhost-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8601,7 +8583,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-link-bonding-vhost-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-link-bonding-vhost-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8617,7 +8599,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-64b-1t1c-link-bonding-vhost-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-link-bonding-vhost-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8633,7 +8615,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-114b-1t1c-vts-l2switching-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-114b-1t1c-vts-l2switching-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -8646,7 +8628,7 @@ - type: "plot" title: "Throughput: 3n-hsw-xl710-114b-1t1c-vts-l2switching-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-114b-1t1c-vts-l2switching-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -8659,7 +8641,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-114b-2t1c-vts-l2switching-base-i40e-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-114b-2t1c-vts-l2switching-base-i40e-ndr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8672,7 +8654,7 @@ - type: "plot" title: "Throughput: 3n-skx-xxv710-114b-2t1c-vts-l2switching-base-i40e-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-114b-2t1c-vts-l2switching-base-i40e-pdr" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8685,7 +8667,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-114b-1t1c-vts-l2switching-base-ixgbe-ndr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-114b-1t1c-vts-l2switching-base-ixgbe-ndr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8698,7 +8680,7 @@ - type: "plot" title: "Throughput: 3n-tsh-x520-114b-1t1c-vts-l2switching-base-ixgbe-pdr" - algorithm: "plot_performance_box_name" + algorithm: "plot_perf_box_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-114b-1t1c-vts-l2switching-base-ixgbe-pdr" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -8715,7 +8697,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-memif-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-memif-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -8729,7 +8711,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-memif-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-memif-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8744,7 +8726,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -8759,7 +8741,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -8773,7 +8755,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -8788,7 +8770,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -8804,7 +8786,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-1t1c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -8817,7 +8799,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-ip4routing-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8832,7 +8814,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8846,7 +8828,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8861,7 +8843,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-features-ip4routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8877,7 +8859,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-1t1c-ip4tunnel-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-ip4tunnel-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -8889,7 +8871,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-ip4tunnel-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-ip4tunnel-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8903,7 +8885,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -8916,7 +8898,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -8931,7 +8913,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-78b-1t1c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-1t1c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -8944,7 +8926,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-78b-2t1c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8957,7 +8939,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-2t1c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -8972,7 +8954,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-sw-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-sw-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -8988,7 +8970,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-hw-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-1t1c-ipsec-ip4routing-base-scale-hw-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -9002,7 +8984,7 @@ #- type: "plot" # title: "Latency: 3n-skx-xxv710-imix-2t1c-ipsec-ip4routing-base-scale-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-imix-2t1c-ipsec-ip4routing-base-scale-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-skx" # include: @@ -9018,7 +9000,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9034,7 +9016,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-l2switching-base-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9048,7 +9030,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9063,7 +9045,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9077,7 +9059,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-1t1c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -9093,7 +9075,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-l2switching-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9108,7 +9090,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-l2switching-base-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9121,7 +9103,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9135,7 +9117,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9151,7 +9133,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-features-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-features-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9167,7 +9149,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-78b-1t1c-srv6-ip6routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-1t1c-srv6-ip6routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -9183,7 +9165,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-78b-2t1c-srv6-ip6routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-2t1c-srv6-ip6routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9199,7 +9181,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -9213,7 +9195,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -9227,7 +9209,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-1t1c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -9241,7 +9223,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-1t1c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-1t1c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -9255,7 +9237,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -9269,7 +9251,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -9283,7 +9265,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -9298,7 +9280,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -9313,7 +9295,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -9329,7 +9311,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-2t1c-link-bonding-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -9345,7 +9327,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-114b-1t1c-vts-l2switching-base-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-114b-1t1c-vts-l2switching-base-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -9358,7 +9340,7 @@ #- type: "plot" # title: "Latency: 3n-skx-xxv710-114b-2t1c-vts-l2switching-base-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-114b-2t1c-vts-l2switching-base-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-skx" # include: @@ -9371,7 +9353,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-memif-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-memif-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9386,7 +9368,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-ip4routing-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-ip4routing-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9401,7 +9383,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-features-ip4routing-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-features-ip4routing-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9417,7 +9399,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-ip4tunnel-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-ip4tunnel-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9431,7 +9413,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-78b-1t1c-ip6routing-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-1t1c-ip6routing-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9446,7 +9428,7 @@ #- type: "plot" # title: "Latency: 3n-tsh-x520-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-imix-1t1c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-tsh" # include: @@ -9462,7 +9444,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-l2switching-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-l2switching-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9476,7 +9458,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-l2switching-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-l2switching-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9492,7 +9474,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-features-l2switching-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-features-l2switching-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9508,7 +9490,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-78b-1t1c-srv6-ip6routing-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-1t1c-srv6-ip6routing-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9524,7 +9506,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-vhost-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-vhost-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9539,7 +9521,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-1t1c-link-bonding-vhost-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-1t1c-link-bonding-vhost-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -9555,7 +9537,7 @@ #- type: "plot" # title: "Latency: 3n-tsh-x520-114b-1t1c-vts-l2switching-base-ixgbe-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-114b-1t1c-vts-l2switching-base-ixgbe-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-tsh" # include: @@ -9572,7 +9554,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-memif-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-memif-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9586,7 +9568,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-memif-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-memif-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9601,7 +9583,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-64b-4t2c-ip4routing-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-4t2c-ip4routing-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9616,7 +9598,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-ip4routing-base-scale-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-ip4routing-base-scale-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9630,7 +9612,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9645,7 +9627,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-features-ip4routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-features-ip4routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9661,7 +9643,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-2t2c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-2t2c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -9674,7 +9656,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-4t2c-ip4routing-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-4t2c-ip4routing-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9689,7 +9671,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-ip4routing-base-scale-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-ip4routing-base-scale-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9703,7 +9685,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9718,7 +9700,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-features-ip4routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-features-ip4routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9734,7 +9716,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-2t2c-ip4tunnel-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-2t2c-ip4tunnel-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -9746,7 +9728,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-ip4tunnel-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-ip4tunnel-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9760,7 +9742,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-78b-4t2c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-78b-4t2c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9773,7 +9755,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-78b-4t2c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-78b-4t2c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9788,7 +9770,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-78b-2t2c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-2t2c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -9801,7 +9783,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-78b-4t2c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-78b-4t2c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9814,7 +9796,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-78b-4t2c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-4t2c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9829,7 +9811,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-imix-2t2c-ipsec-ip4routing-base-scale-sw-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-2t2c-ipsec-ip4routing-base-scale-sw-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -9845,7 +9827,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-imix-2t2c-ipsec-ip4routing-base-scale-hw-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-2t2c-ipsec-ip4routing-base-scale-hw-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -9859,7 +9841,7 @@ #- type: "plot" # title: "Latency: 3n-skx-xxv710-imix-4t2c-ipsec-ip4routing-base-scale-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-imix-4t2c-ipsec-ip4routing-base-scale-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-skx" # include: @@ -9875,7 +9857,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-64b-4t2c-l2switching-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-4t2c-l2switching-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9891,7 +9873,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-l2switching-base-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-l2switching-base-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9905,7 +9887,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9920,7 +9902,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -9934,7 +9916,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-2t2c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-2t2c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -9950,7 +9932,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-4t2c-l2switching-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-4t2c-l2switching-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9965,7 +9947,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-l2switching-base-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-l2switching-base-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9978,7 +9960,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -9992,7 +9974,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10008,7 +9990,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-features-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-features-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10024,7 +10006,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-78b-2t2c-srv6-ip6routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-2t2c-srv6-ip6routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -10040,7 +10022,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-78b-4t2c-srv6-ip6routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-4t2c-srv6-ip6routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10056,7 +10038,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -10070,7 +10052,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-4t2c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-4t2c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -10084,7 +10066,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-2t2c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-2t2c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -10098,7 +10080,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-2t2c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-2t2c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -10112,7 +10094,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-4t2c-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-4t2c-link-bonding-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -10126,7 +10108,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-4t2c-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-4t2c-link-bonding-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -10140,7 +10122,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -10155,7 +10137,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -10170,7 +10152,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-link-bonding-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -10186,7 +10168,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-4t2c-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-4t2c-link-bonding-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -10202,7 +10184,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-114b-2t2c-vts-l2switching-base-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-114b-2t2c-vts-l2switching-base-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -10215,7 +10197,7 @@ #- type: "plot" # title: "Latency: 3n-skx-xxv710-114b-4t2c-vts-l2switching-base-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-114b-4t2c-vts-l2switching-base-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-skx" # include: @@ -10228,7 +10210,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-memif-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-memif-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10243,7 +10225,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-ip4routing-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-ip4routing-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10258,7 +10240,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-features-ip4routing-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-features-ip4routing-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10274,7 +10256,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-ip4tunnel-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-ip4tunnel-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10288,7 +10270,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-78b-2t2c-ip6routing-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-2t2c-ip6routing-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10303,7 +10285,7 @@ #- type: "plot" # title: "Latency: 3n-tsh-x520-imix-2t2c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-imix-2t2c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-tsh" # include: @@ -10319,7 +10301,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-l2switching-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-l2switching-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10333,7 +10315,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-l2switching-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-l2switching-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10349,7 +10331,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-features-l2switching-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-features-l2switching-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10365,7 +10347,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-78b-2t2c-srv6-ip6routing-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-2t2c-srv6-ip6routing-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10381,7 +10363,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-vhost-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-vhost-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10396,7 +10378,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-2t2c-link-bonding-vhost-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-2t2c-link-bonding-vhost-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -10412,7 +10394,7 @@ #- type: "plot" # title: "Latency: 3n-tsh-x520-114b-2t2c-vts-l2switching-base-ixgbe-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-114b-2t2c-vts-l2switching-base-ixgbe-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-tsh" # include: @@ -10429,7 +10411,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-memif-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-memif-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10443,7 +10425,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-memif-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-memif-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10458,7 +10440,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-64b-8t4c-ip4routing-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-8t4c-ip4routing-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10473,7 +10455,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-ip4routing-base-scale-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-ip4routing-base-scale-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10487,7 +10469,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10502,7 +10484,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-features-ip4routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-features-ip4routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10518,7 +10500,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-4t4c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-4t4c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -10531,7 +10513,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-8t4c-ip4routing-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-8t4c-ip4routing-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10546,7 +10528,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-ip4routing-base-scale-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-ip4routing-base-scale-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10560,7 +10542,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-ip4routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-ip4routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10575,7 +10557,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-features-ip4routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-features-ip4routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10591,7 +10573,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-4t4c-ip4tunnel-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-4t4c-ip4tunnel-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -10603,7 +10585,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-ip4tunnel-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-ip4tunnel-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10617,7 +10599,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-78b-8t4c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-78b-8t4c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10630,7 +10612,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-78b-8t4c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-78b-8t4c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10645,7 +10627,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-78b-4t4c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-4t4c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -10658,7 +10640,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-78b-8t4c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-78b-8t4c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10671,7 +10653,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-78b-8t4c-ip6routing-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-8t4c-ip6routing-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10686,7 +10668,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-imix-4t4c-ipsec-ip4routing-base-scale-sw-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-4t4c-ipsec-ip4routing-base-scale-sw-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -10702,7 +10684,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-imix-4t4c-ipsec-ip4routing-base-scale-hw-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-4t4c-ipsec-ip4routing-base-scale-hw-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -10716,7 +10698,7 @@ #- type: "plot" # title: "Latency: 3n-skx-xxv710-imix-8t4c-ipsec-ip4routing-base-scale-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-imix-8t4c-ipsec-ip4routing-base-scale-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-skx" # include: @@ -10732,7 +10714,7 @@ - type: "plot" title: "Latency: 2n-skx-x710-64b-8t4c-l2switching-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-8t4c-l2switching-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10748,7 +10730,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-l2switching-base-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-l2switching-base-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10762,7 +10744,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10777,7 +10759,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -10791,7 +10773,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-4t4c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-4t4c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -10807,7 +10789,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-8t4c-l2switching-base-scale-avf-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-8t4c-l2switching-base-scale-avf-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10822,7 +10804,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-l2switching-base-avf-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-l2switching-base-avf-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10835,7 +10817,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10849,7 +10831,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-l2switching-base-scale-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-l2switching-base-scale-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10865,7 +10847,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-features-l2switching-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-features-l2switching-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10881,7 +10863,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-78b-4t4c-srv6-ip6routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-4t4c-srv6-ip6routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -10897,7 +10879,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-78b-8t4c-srv6-ip6routing-base-i40e-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-8t4c-srv6-ip6routing-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -10913,7 +10895,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -10927,7 +10909,7 @@ - type: "plot" title: "Latency: 2n-skx-xxv710-64b-8t4c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-8t4c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -10941,7 +10923,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-4t4c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-4t4c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -10955,7 +10937,7 @@ - type: "plot" title: "Latency: 3n-hsw-xl710-64b-4t4c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-4t4c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -10969,7 +10951,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-8t4c-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-8t4c-link-bonding-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -10983,7 +10965,7 @@ - type: "plot" title: "Latency: 3n-skx-x710-64b-8t4c-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-8t4c-link-bonding-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -10997,7 +10979,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -11012,7 +10994,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -11027,7 +11009,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-link-bonding-vhost-base-i40e-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -11043,7 +11025,7 @@ - type: "plot" title: "Latency: 3n-skx-xxv710-64b-8t4c-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-8t4c-link-bonding-vhost-base-i40e-vpp-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -11059,7 +11041,7 @@ #- type: "plot" # title: "Latency: 3n-hsw-xl710-114b-4t4c-vts-l2switching-base-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-114b-4t4c-vts-l2switching-base-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-hsw" # include: @@ -11072,7 +11054,7 @@ #- type: "plot" # title: "Latency: 3n-skx-xxv710-114b-8t4c-vts-l2switching-base-i40e-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-114b-8t4c-vts-l2switching-base-i40e-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-skx" # include: @@ -11085,7 +11067,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-memif-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-memif-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11100,7 +11082,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-ip4routing-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-ip4routing-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11115,7 +11097,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-features-ip4routing-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-features-ip4routing-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11131,7 +11113,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-ip4tunnel-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-ip4tunnel-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11145,7 +11127,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-78b-4t4c-ip6routing-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-4t4c-ip6routing-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11160,7 +11142,7 @@ #- type: "plot" # title: "Latency: 3n-tsh-x520-imix-4t4c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-imix-4t4c-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-tsh" # include: @@ -11176,7 +11158,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-l2switching-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-l2switching-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11190,7 +11172,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-l2switching-base-scale-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-l2switching-base-scale-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11206,7 +11188,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-features-l2switching-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-features-l2switching-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11222,7 +11204,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-78b-4t4c-srv6-ip6routing-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-4t4c-srv6-ip6routing-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11238,7 +11220,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-vhost-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-vhost-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11253,7 +11235,7 @@ - type: "plot" title: "Latency: 3n-tsh-x520-64b-4t4c-link-bonding-vhost-base-ixgbe-ndr-lat" - algorithm: "plot_latency_error_bars_name" + algorithm: "plot_lat_err_bars_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-4t4c-link-bonding-vhost-base-ixgbe-ndr-lat" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -11269,7 +11251,7 @@ #- type: "plot" # title: "Latency: 3n-tsh-x520-114b-4t4c-vts-l2switching-base-ixgbe-ndr-lat" -# algorithm: "plot_latency_error_bars_name" +# algorithm: "plot_lat_err_bars_name" # output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-114b-4t4c-vts-l2switching-base-ixgbe-ndr-lat" # data: "plot-vpp-throughput-lat-tsa-3n-tsh" # include: @@ -11286,7 +11268,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-memif-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-memif-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11300,7 +11282,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-memif-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-memif-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11314,7 +11296,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-memif-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-memif-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11329,7 +11311,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-memif-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-memif-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11344,7 +11326,7 @@ - type: "plot" title: "Speedup: 2n-skx-x710-64b-ip4routing-base-scale-avf-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-ip4routing-base-scale-avf-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11359,7 +11341,7 @@ - type: "plot" title: "Speedup: 2n-skx-x710-64b-ip4routing-base-scale-avf-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-ip4routing-base-scale-avf-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11374,7 +11356,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-ip4routing-base-scale-avf-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-ip4routing-base-scale-avf-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11388,7 +11370,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-ip4routing-base-scale-avf-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-ip4routing-base-scale-avf-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11402,7 +11384,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-ip4routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-ip4routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11417,7 +11399,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-ip4routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-ip4routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11432,7 +11414,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-features-ip4routing-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-features-ip4routing-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11448,7 +11430,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-features-ip4routing-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-features-ip4routing-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11464,7 +11446,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-ip4routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-ip4routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11477,7 +11459,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-ip4routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-ip4routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11490,7 +11472,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-64b-ip4routing-base-scale-avf-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-ip4routing-base-scale-avf-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11505,7 +11487,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-64b-ip4routing-base-scale-avf-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-ip4routing-base-scale-avf-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11520,7 +11502,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-ip4routing-base-scale-avf-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-ip4routing-base-scale-avf-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11534,7 +11516,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-ip4routing-base-scale-avf-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-ip4routing-base-scale-avf-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11548,7 +11530,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-ip4routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-ip4routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11563,7 +11545,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-ip4routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-ip4routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11578,7 +11560,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-features-ip4routing-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-features-ip4routing-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11594,7 +11576,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-features-ip4routing-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-features-ip4routing-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11610,7 +11592,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-ip4tunnel-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-ip4tunnel-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11622,7 +11604,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-ip4tunnel-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-ip4tunnel-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11634,7 +11616,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-ip4tunnel-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-ip4tunnel-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11648,7 +11630,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-ip4tunnel-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-ip4tunnel-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11662,7 +11644,7 @@ - type: "plot" title: "Speedup: 2n-skx-x710-78b-ip6routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-78b-ip6routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11675,7 +11657,7 @@ - type: "plot" title: "Speedup: 2n-skx-x710-78b-ip6routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-78b-ip6routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11688,7 +11670,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-78b-ip6routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-78b-ip6routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11703,7 +11685,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-78b-ip6routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-78b-ip6routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11718,7 +11700,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-78b-ip6routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-ip6routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11731,7 +11713,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-78b-ip6routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-ip6routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11744,7 +11726,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-78b-ip6routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-78b-ip6routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11757,7 +11739,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-78b-ip6routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-78b-ip6routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11770,7 +11752,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-78b-ip6routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-ip6routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11785,7 +11767,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-78b-ip6routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-ip6routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11800,7 +11782,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-imix-ipsec-ip4routing-base-scale-sw-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-ipsec-ip4routing-base-scale-sw-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11816,7 +11798,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-imix-ipsec-ip4routing-base-scale-sw-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-ipsec-ip4routing-base-scale-sw-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11832,7 +11814,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-imix-ipsec-ip4routing-base-scale-hw-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-ipsec-ip4routing-base-scale-hw-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11846,7 +11828,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-imix-ipsec-ip4routing-base-scale-hw-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-imix-ipsec-ip4routing-base-scale-hw-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -11860,7 +11842,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-imix-ipsec-ip4routing-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-imix-ipsec-ip4routing-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11876,7 +11858,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-imix-ipsec-ip4routing-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-imix-ipsec-ip4routing-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -11892,7 +11874,7 @@ - type: "plot" title: "Speedup: 2n-skx-x710-64b-l2switching-base-scale-avf-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-l2switching-base-scale-avf-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11908,7 +11890,7 @@ - type: "plot" title: "Speedup: 2n-skx-x710-64b-l2switching-base-scale-avf-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-x710-64b-l2switching-base-scale-avf-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11924,7 +11906,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-l2switching-base-avf-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-l2switching-base-avf-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11938,7 +11920,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-l2switching-base-avf-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-l2switching-base-avf-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11952,7 +11934,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-l2switching-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-l2switching-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11967,7 +11949,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-l2switching-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-l2switching-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11982,7 +11964,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-l2switching-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-l2switching-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -11996,7 +11978,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-l2switching-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-l2switching-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx" include: @@ -12010,7 +11992,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-l2switching-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-l2switching-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -12026,7 +12008,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-l2switching-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-l2switching-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -12042,7 +12024,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-64b-l2switching-base-scale-avf-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-l2switching-base-scale-avf-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12057,7 +12039,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-64b-l2switching-base-scale-avf-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-l2switching-base-scale-avf-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12072,7 +12054,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-l2switching-base-avf-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-l2switching-base-avf-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12085,7 +12067,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-l2switching-base-avf-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-l2switching-base-avf-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12098,7 +12080,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-l2switching-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-l2switching-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12112,7 +12094,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-l2switching-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-l2switching-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12126,7 +12108,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-l2switching-base-scale-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-l2switching-base-scale-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12142,7 +12124,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-l2switching-base-scale-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-l2switching-base-scale-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12158,7 +12140,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-features-l2switching-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-features-l2switching-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12174,7 +12156,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-features-l2switching-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-features-l2switching-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12190,7 +12172,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-78b-srv6-ip6routing-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-srv6-ip6routing-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -12206,7 +12188,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-78b-srv6-ip6routing-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-78b-srv6-ip6routing-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -12222,7 +12204,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-78b-srv6-ip6routing-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-srv6-ip6routing-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12238,7 +12220,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-78b-srv6-ip6routing-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-78b-srv6-ip6routing-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12254,7 +12236,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-vhost-base-i40e-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-vhost-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -12268,7 +12250,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-vhost-base-i40e-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-vhost-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -12282,7 +12264,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-vhost-base-i40e-vpp-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-vhost-base-i40e-vpp-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -12296,7 +12278,7 @@ - type: "plot" title: "Speedup: 2n-skx-xxv710-64b-vhost-base-i40e-vpp-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-skx-xxv710-64b-vhost-base-i40e-vpp-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-skx-vhost" include: @@ -12310,7 +12292,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-vhost-base-i40e-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-vhost-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -12324,7 +12306,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-vhost-base-i40e-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-vhost-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -12338,7 +12320,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-vhost-base-i40e-vpp-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-vhost-base-i40e-vpp-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -12352,7 +12334,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-64b-vhost-base-i40e-vpp-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-64b-vhost-base-i40e-vpp-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw-vhost" include: @@ -12366,7 +12348,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-64b-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-link-bonding-vhost-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12380,7 +12362,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-64b-link-bonding-vhost-base-i40e-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-link-bonding-vhost-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12394,7 +12376,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-64b-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-link-bonding-vhost-base-i40e-vpp-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12408,7 +12390,7 @@ - type: "plot" title: "Speedup: 3n-skx-x710-64b-link-bonding-vhost-base-i40e-vpp-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-x710-64b-link-bonding-vhost-base-i40e-vpp-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12422,7 +12404,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-vhost-base-i40e-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-vhost-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12437,7 +12419,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-vhost-base-i40e-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-vhost-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12452,7 +12434,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-vhost-base-i40e-vpp-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-vhost-base-i40e-vpp-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12467,7 +12449,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-vhost-base-i40e-vpp-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-vhost-base-i40e-vpp-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12482,7 +12464,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-link-bonding-vhost-base-i40e-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-link-bonding-vhost-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12498,7 +12480,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-link-bonding-vhost-base-i40e-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-link-bonding-vhost-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12514,7 +12496,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-link-bonding-vhost-base-i40e-vpp-ndr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-link-bonding-vhost-base-i40e-vpp-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12530,7 +12512,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-64b-link-bonding-vhost-base-i40e-vpp-pdr" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-64b-link-bonding-vhost-base-i40e-vpp-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx-vhost" include: @@ -12546,7 +12528,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-114b-vts-l2switching-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-114b-vts-l2switching-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -12559,7 +12541,7 @@ - type: "plot" title: "Speedup: 3n-hsw-xl710-114b-vts-l2switching-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-hsw-xl710-114b-vts-l2switching-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-hsw" include: @@ -12572,7 +12554,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-114b-vts-l2switching-base-i40e-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-114b-vts-l2switching-base-i40e-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12585,7 +12567,7 @@ - type: "plot" title: "Speedup: 3n-skx-xxv710-114b-vts-l2switching-base-i40e-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-skx-xxv710-114b-vts-l2switching-base-i40e-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-skx" include: @@ -12598,7 +12580,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-memif-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-memif-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12613,7 +12595,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-memif-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-memif-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12628,7 +12610,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-ip4routing-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-ip4routing-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12643,7 +12625,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-ip4routing-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-ip4routing-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12658,7 +12640,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-features-ip4routing-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-features-ip4routing-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12674,7 +12656,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-features-ip4routing-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-features-ip4routing-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12690,7 +12672,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-ip4tunnel-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-ip4tunnel-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12704,7 +12686,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-ip4tunnel-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-ip4tunnel-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12718,7 +12700,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-78b-ip6routing-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-ip6routing-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12733,7 +12715,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-78b-ip6routing-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-ip6routing-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12748,7 +12730,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-imix-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-imix-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12764,7 +12746,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-imix-ipsec-ip4routing-base-scale-sw-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-imix-ipsec-ip4routing-base-scale-sw-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12780,7 +12762,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-l2switching-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-l2switching-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12794,7 +12776,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-l2switching-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-l2switching-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12808,7 +12790,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-l2switching-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-l2switching-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12824,7 +12806,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-l2switching-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-l2switching-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12840,7 +12822,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-features-l2switching-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-features-l2switching-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12856,7 +12838,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-features-l2switching-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-features-l2switching-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12872,7 +12854,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-78b-srv6-ip6routing-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-srv6-ip6routing-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12888,7 +12870,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-78b-srv6-ip6routing-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-78b-srv6-ip6routing-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12904,7 +12886,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-vhost-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-vhost-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12919,7 +12901,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-vhost-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-vhost-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12934,7 +12916,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-link-bonding-vhost-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-link-bonding-vhost-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12950,7 +12932,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-64b-link-bonding-vhost-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-64b-link-bonding-vhost-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12966,7 +12948,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-114b-vts-l2switching-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-114b-vts-l2switching-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12979,7 +12961,7 @@ - type: "plot" title: "Speedup: 3n-tsh-x520-114b-vts-l2switching-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-tsh-x520-114b-vts-l2switching-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-tsh" include: @@ -12992,7 +12974,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-64b-ip4routing-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-ip4routing-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13007,7 +12989,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-64b-ip4routing-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-ip4routing-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13022,7 +13004,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-64b-features-ip4routing-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-features-ip4routing-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13038,7 +13020,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-64b-features-ip4routing-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-features-ip4routing-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13054,7 +13036,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-78b-ip6routing-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-78b-ip6routing-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13069,7 +13051,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-78b-ip6routing-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-78b-ip6routing-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13084,7 +13066,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-64b-l2switching-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-l2switching-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13098,7 +13080,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-64b-l2switching-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-l2switching-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13112,7 +13094,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-64b-l2switching-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-l2switching-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13128,7 +13110,7 @@ - type: "plot" title: "Speedup: 2n-dnv-x553-64b-l2switching-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/2n-dnv-x553-64b-l2switching-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-2n-dnv" include: @@ -13144,7 +13126,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-ip4routing-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-ip4routing-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13159,7 +13141,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-ip4routing-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-ip4routing-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13174,7 +13156,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-features-ip4routing-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-features-ip4routing-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13190,7 +13172,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-features-ip4routing-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-features-ip4routing-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13206,7 +13188,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-ip4tunnel-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-ip4tunnel-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13220,7 +13202,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-ip4tunnel-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-ip4tunnel-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13234,7 +13216,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-78b-ip6routing-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-78b-ip6routing-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13249,7 +13231,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-78b-ip6routing-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-78b-ip6routing-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13264,7 +13246,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-imix-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-imix-ipsec-ip4routing-base-scale-sw-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13280,7 +13262,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-imix-ipsec-ip4routing-base-scale-sw-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-imix-ipsec-ip4routing-base-scale-sw-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13296,7 +13278,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-l2switching-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-l2switching-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13310,7 +13292,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-l2switching-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-l2switching-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13324,7 +13306,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-l2switching-base-scale-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-l2switching-base-scale-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13340,7 +13322,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-l2switching-base-scale-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-l2switching-base-scale-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13356,7 +13338,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-features-l2switching-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-features-l2switching-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13372,7 +13354,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-64b-features-l2switching-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-64b-features-l2switching-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13388,7 +13370,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-78b-srv6-ip6routing-base-ixgbe-ndr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-78b-srv6-ip6routing-base-ixgbe-ndr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13401,7 +13383,7 @@ - type: "plot" title: "Speedup: 3n-dnv-x553-78b-srv6-ip6routing-base-ixgbe-pdr-tsa" - algorithm: "plot_throughput_speedup_analysis_name" + algorithm: "plot_tsa_name" output-file: "{DIR[STATIC,VPP]}/3n-dnv-x553-78b-srv6-ip6routing-base-ixgbe-pdr-tsa" data: "plot-vpp-throughput-lat-tsa-3n-dnv" include: @@ -13414,1940 +13396,536 @@ ################################################################################ -# Packet Throughput - DPDK l3fwd 3n-hsw-x520 -- type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-x520-64b-1t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x520-64b-1t1c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-3n-hsw-x520-64b-1t1c-base-ndr" - layout: "plot-throughput" +# Packet Throughput - DPDK -- type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-x520-64b-2t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x520-64b-2t2c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-3n-hsw-x520-64b-2t2c-base-ndr" - layout: "plot-throughput" +# 2n-skx-xxv710 - type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-x520-64b-1t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x520-64b-1t1c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 2n-skx-xxv710-64b-2t1c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-xxv710-64b-2t1c-base-ndr" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-x520-64b-1t1c-base-pdr" + title: "2n-skx-xxv710-64b-2t1c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-x520-64b-2t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x520-64b-2t2c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 2n-skx-xxv710-64b-4t2c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-xxv710-64b-4t2c-base-ndr" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-x520-64b-2t2c-base-pdr" + title: "2n-skx-xxv710-64b-4t2c-base-ndr" layout: "plot-throughput" -# Packet Throughput - DPDK l3fwd 3n-hsw-x710 - type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-x710-64b-1t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x710-64b-1t1c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 2n-skx-xxv710-64b-2t1c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-xxv710-64b-2t1c-base-pdr" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-x710-64b-1t1c-base-ndr" + title: "2n-skx-xxv710-64b-2t1c-base-pdr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-x710-64b-2t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x710-64b-2t2c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 2n-skx-xxv710-64b-4t2c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-xxv710-64b-4t2c-base-pdr" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-x710-64b-2t2c-base-ndr" + title: "2n-skx-xxv710-64b-4t2c-base-pdr" layout: "plot-throughput" -- type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-x710-64b-1t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x710-64b-1t1c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-3n-hsw-x710-64b-1t1c-base-pdr" - layout: "plot-throughput" +# 2n-skx-x710 - type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-x710-64b-2t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x710-64b-2t2c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 2n-skx-x710-64b-2t1c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-x710-64b-2t1c-base-ndr" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-x710-64b-2t2c-base-pdr" + title: "2n-skx-x710-64b-2t1c-base-ndr" layout: "plot-throughput" -# Packet Throughput - DPDK l3fwd 3n-hsw-xl710 - type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-xl710-64b-1t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-xl710-64b-1t1c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 2n-skx-x710-64b-4t2c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-x710-64b-4t2c-base-ndr" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-xl710-64b-1t1c-base-ndr" + title: "2n-skx-x710-64b-4t2c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-xl710-64b-2t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-xl710-64b-2t2c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 2n-skx-x710-64b-2t1c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-x710-64b-2t1c-base-pdr" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-xl710-64b-2t2c-base-ndr" + title: "2n-skx-x710-64b-2t1c-base-pdr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-xl710-64b-1t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-xl710-64b-1t1c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 2n-skx-x710-64b-4t2c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-x710-64b-4t2c-base-pdr" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-xl710-64b-1t1c-base-pdr" + title: "2n-skx-x710-64b-4t2c-base-pdr" layout: "plot-throughput" -- type: "plot" - title: "DPDK Throughput: l3fwd-3n-hsw-xl710-64b-2t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-xl710-64b-2t2c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-3n-hsw-xl710-64b-2t2c-base-pdr" - layout: "plot-throughput" +# 3n-skx-xxv710 -# Packet Throughput - DPDK l3fwd 3n-skx-x710 - type: "plot" - title: "DPDK Throughput: l3fwd-3n-skx-x710-64b-2t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-x710-64b-2t1c-base-ndr" + title: "DPDK Throughput: 3n-skx-xxv710-64b-2t1c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-xxv710-64b-2t1c-base-ndr" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-x710-64b-2t1c-base-ndr" + title: "3n-skx-xxv710-64b-2t1c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-skx-x710-64b-4t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-x710-64b-4t2c-base-ndr" + title: "DPDK Throughput: 3n-skx-xxv710-64b-4t2c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-xxv710-64b-4t2c-base-ndr" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-x710-64b-4t2c-base-ndr" + title: "3n-skx-xxv710-64b-4t2c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-skx-x710-64b-2t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-x710-64b-2t1c-base-pdr" + title: "DPDK Throughput: 3n-skx-xxv710-64b-2t1c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-xxv710-64b-2t1c-base-pdr" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-x710-64b-2t1c-base-pdr" + title: "3n-skx-xxv710-64b-2t1c-base-pdr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-skx-x710-64b-4t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-x710-64b-4t2c-base-pdr" + title: "DPDK Throughput: 3n-skx-xxv710-64b-4t2c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-xxv710-64b-4t2c-base-pdr" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-x710-64b-4t2c-base-pdr" + title: "3n-skx-xxv710-64b-4t2c-base-pdr" layout: "plot-throughput" -# Packet Throughput - DPDK l3fwd 3n-skx-xxv710 -- type: "plot" - title: "DPDK Throughput: l3fwd-3n-skx-xxv710-64b-2t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-xxv710-64b-2t1c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-3n-skx-xxv710-64b-2t1c-base-ndr" - layout: "plot-throughput" +# 3n-skx-x710 - type: "plot" - title: "DPDK Throughput: l3fwd-3n-skx-xxv710-64b-4t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-xxv710-64b-4t2c-base-ndr" + title: "DPDK Throughput: 3n-skx-x710-64b-2t1c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-x710-64b-2t1c-base-ndr" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-xxv710-64b-4t2c-base-ndr" + title: "3n-skx-x710-64b-2t1c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-skx-xxv710-64b-2t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-xxv710-64b-2t1c-base-pdr" + title: "DPDK Throughput: 3n-skx-x710-64b-4t2c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-x710-64b-4t2c-base-ndr" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-xxv710-64b-2t1c-base-pdr" + title: "3n-skx-x710-64b-4t2c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: l3fwd-3n-skx-xxv710-64b-4t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-xxv710-64b-4t2c-base-pdr" + title: "DPDK Throughput: 3n-skx-x710-64b-2t1c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-x710-64b-2t1c-base-pdr" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-3n-skx-xxv710-64b-4t2c-base-pdr" - layout: "plot-throughput" - -# Packet Throughput - DPDK l3fwd 2n-skx-x710 -- type: "plot" - title: "DPDK Throughput: l3fwd-2n-skx-x710-64b-2t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-x710-64b-2t1c-base-ndr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-2n-skx-x710-64b-2t1c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: l3fwd-2n-skx-x710-64b-4t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-x710-64b-4t2c-base-ndr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-2n-skx-x710-64b-4t2c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: l3fwd-2n-skx-x710-64b-2t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-x710-64b-2t1c-base-pdr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-2n-skx-x710-64b-2t1c-base-pdr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: l3fwd-2n-skx-x710-64b-4t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-x710-64b-4t2c-base-pdr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-2n-skx-x710-64b-4t2c-base-pdr" - layout: "plot-throughput" - -# Packet Throughput - DPDK l3fwd 2n-skx-xxv710 -- type: "plot" - title: "DPDK Throughput: l3fwd-2n-skx-xxv710-64b-2t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-xxv710-64b-2t1c-base-ndr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-2n-skx-xxv710-64b-2t1c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: l3fwd-2n-skx-xxv710-64b-4t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-xxv710-64b-4t2c-base-ndr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-2n-skx-xxv710-64b-4t2c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: l3fwd-2n-skx-xxv710-64b-2t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-xxv710-64b-2t1c-base-pdr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "l3fwd-2n-skx-xxv710-64b-2t1c-base-pdr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: l3fwd-2n-skx-xxv710-64b-4t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-xxv710-64b-4t2c-base-pdr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-2n-skx-xxv710-64b-4t2c-base-pdr" + title: "3n-skx-x710-64b-2t1c-base-pdr" layout: "plot-throughput" -################################################################################ - -# Packet Throughput - DPDK testpmd 3n-hsw-x520 - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-x520-64b-1t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x520-64b-1t1c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + title: "DPDK Throughput: 3n-skx-x710-64b-4t2c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-x710-64b-4t2c-base-pdr" + data: "plot-dpdk-throughput-latency-3n-skx" + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-x520-64b-1t1c-base-ndr" + title: "3n-skx-x710-64b-4t2c-base-pdr" layout: "plot-throughput" -- type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-x520-64b-2t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x520-64b-2t2c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-hsw-x520-64b-2t2c-base-ndr" - layout: "plot-throughput" +# 3n-hsw-xl710 - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-x520-64b-1t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x520-64b-1t1c-base-pdr" + title: "DPDK Throughput: 3n-hsw-xl710-64b-1t1c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-xl710-64b-1t1c-base-ndr" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-hsw-x520-64b-1t1c-base-pdr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-x520-64b-2t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x520-64b-2t2c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.40ge2p1xl710-eth-l2xcbase-testpmd-ndrpdr.64b-1t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.40ge2p1xl710-ethip4-ip4base-l3fwd-ndrpdr.64b-1t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-x520-64b-2t2c-base-pdr" + title: "3n-hsw-xl710-64b-1t1c-base-ndr" layout: "plot-throughput" -# Packet Throughput - DPDK testpmd 3n-hsw-x710 - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-x710-64b-1t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x710-64b-1t1c-base-ndr" + title: "DPDK Throughput: 3n-hsw-xl710-64b-2t2c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-xl710-64b-2t2c-base-ndr" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.40ge2p1xl710-eth-l2xcbase-testpmd-ndrpdr.64b-2t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.40ge2p1xl710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-x710-64b-1t1c-base-ndr" + title: "3n-hsw-xl710-64b-2t2c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-x710-64b-2t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x710-64b-2t2c-base-ndr" + title: "DPDK Throughput: 3n-hsw-xl710-64b-1t1c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-xl710-64b-1t1c-base-pdr" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.40ge2p1xl710-eth-l2xcbase-testpmd-ndrpdr.64b-1t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.40ge2p1xl710-ethip4-ip4base-l3fwd-ndrpdr.64b-1t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-x710-64b-2t2c-base-ndr" + title: "3n-hsw-xl710-64b-1t1c-base-pdr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-x710-64b-1t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x710-64b-1t1c-base-pdr" + title: "DPDK Throughput: 3n-hsw-xl710-64b-2t2c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-xl710-64b-2t2c-base-pdr" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.40ge2p1xl710-eth-l2xcbase-testpmd-ndrpdr.64b-2t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.40ge2p1xl710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-x710-64b-1t1c-base-pdr" + title: "3n-hsw-xl710-64b-2t2c-base-pdr" layout: "plot-throughput" -- type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-x710-64b-2t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x710-64b-2t2c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-hsw-x710-64b-2t2c-base-pdr" - layout: "plot-throughput" +# 3n-hsw-x710 -# Packet Throughput - DPDK testpmd 3n-hsw-xl710 - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-xl710-64b-1t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-xl710-64b-1t1c-base-ndr" + title: "DPDK Throughput: 3n-hsw-x710-64b-1t1c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-x710-64b-1t1c-base-ndr" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-1t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-1t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-xl710-64b-1t1c-base-ndr" + title: "3n-hsw-x710-64b-1t1c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-xl710-64b-2t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-xl710-64b-2t2c-base-ndr" + title: "DPDK Throughput: 3n-hsw-x710-64b-2t2c-base-ndr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-x710-64b-2t2c-base-ndr" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-xl710-64b-2t2c-base-ndr" + title: "3n-hsw-x710-64b-2t2c-base-ndr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-xl710-64b-1t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-xl710-64b-1t1c-base-pdr" + title: "DPDK Throughput: 3n-hsw-x710-64b-1t1c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-x710-64b-1t1c-base-pdr" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-1t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-1t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-xl710-64b-1t1c-base-pdr" + title: "3n-hsw-x710-64b-1t1c-base-pdr" layout: "plot-throughput" - type: "plot" - title: "DPDK Throughput: testpmd-3n-hsw-xl710-64b-2t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-xl710-64b-2t2c-base-pdr" + title: "DPDK Throughput: 3n-hsw-x710-64b-2t2c-base-pdr" + algorithm: "plot_perf_box_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-x710-64b-2t2c-base-pdr" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-hsw-xl710-64b-2t2c-base-pdr" - layout: "plot-throughput" - -# Packet Throughput - DPDK testpmd 3n-skx-x710 -- type: "plot" - title: "DPDK Throughput: testpmd-3n-skx-x710-64b-2t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-x710-64b-2t1c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-skx-x710-64b-2t1c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-3n-skx-x710-64b-4t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-x710-64b-4t2c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-skx-x710-64b-4t2c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-3n-skx-x710-64b-2t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-x710-64b-2t1c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-skx-x710-64b-2t1c-base-pdr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-3n-skx-x710-64b-4t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-x710-64b-4t2c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-skx-x710-64b-4t2c-base-pdr" - layout: "plot-throughput" - -# Packet Throughput - DPDK testpmd 3n-skx-xxv710 -- type: "plot" - title: "DPDK Throughput: testpmd-3n-skx-xxv710-64b-2t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-xxv710-64b-2t1c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-skx-xxv710-64b-2t1c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-3n-skx-xxv710-64b-4t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-xxv710-64b-4t2c-base-ndr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-skx-xxv710-64b-4t2c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-3n-skx-xxv710-64b-2t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-xxv710-64b-2t1c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-skx-xxv710-64b-2t1c-base-pdr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-3n-skx-xxv710-64b-4t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-xxv710-64b-4t2c-base-pdr" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-3n-skx-xxv710-64b-4t2c-base-pdr" - layout: "plot-throughput" - -# Packet Throughput - DPDK testpmd 2n-skx-x710 -- type: "plot" - title: "DPDK Throughput: testpmd-2n-skx-x710-64b-2t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-x710-64b-2t1c-base-ndr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-2n-skx-x710-64b-2t1c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-2n-skx-x710-64b-4t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-x710-64b-4t2c-base-ndr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-2n-skx-x710-64b-4t2c-base-ndr" - layout: "plot-throughput" - -- type: "plot" - title: "DPDK Throughput: testpmd-2n-skx-x710-64b-2t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-x710-64b-2t1c-base-pdr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-2n-skx-x710-64b-2t1c-base-pdr" + title: "3n-hsw-x710-64b-2t2c-base-pdr" layout: "plot-throughput" -- type: "plot" - title: "DPDK Throughput: testpmd-2n-skx-x710-64b-4t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-x710-64b-4t2c-base-pdr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-2n-skx-x710-64b-4t2c-base-pdr" - layout: "plot-throughput" +################################################################################ -# Packet Throughput - DPDK testpmd 2n-skx-xxv710 -- type: "plot" - title: "DPDK Throughput: testpmd-2n-skx-xxv710-64b-2t1c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-xxv710-64b-2t1c-base-ndr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-2n-skx-xxv710-64b-2t1c-base-ndr" - layout: "plot-throughput" +# Packet Latency - DPDK -- type: "plot" - title: "DPDK Throughput: testpmd-2n-skx-xxv710-64b-4t2c-base-ndr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-xxv710-64b-4t2c-base-ndr" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-2n-skx-xxv710-64b-4t2c-base-ndr" - layout: "plot-throughput" +# 2n-skx-xxv710 - type: "plot" - title: "DPDK Throughput: testpmd-2n-skx-xxv710-64b-2t1c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-xxv710-64b-2t1c-base-pdr" + title: "Latency: 2n-skx-xxv710-64b-2t1c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-xxv710-64b-2t1c-base-ndr-lat" data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 + include: + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-2n-skx-xxv710-64b-2t1c-base-pdr" - layout: "plot-throughput" + title: "2n-skx-xxv710-64b-2t1c-base-ndr" + layout: "plot-latency" - type: "plot" - title: "DPDK Throughput: testpmd-2n-skx-xxv710-64b-4t2c-base-pdr" - algorithm: "plot_performance_box" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-xxv710-64b-4t2c-base-pdr" + title: "Latency: 2n-skx-xxv710-64b-4t2c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-xxv710-64b-4t2c-base-ndr-lat" data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "throughput" - - "parent" - - "tags" - traces: - hoverinfo: "x+y" - boxpoints: "outliers" - whiskerwidth: 0 - layout: - title: "testpmd-2n-skx-xxv710-64b-4t2c-base-pdr" - layout: "plot-throughput" - -################################################################################ - -# Packet Latency - DPDK l3fwd 3n-hsw-x520 -- type: "plot" - title: "DPDK Latency: l3fwd-3n-hsw-x520-64b-1t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x520-64b-1t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-x520-64b-1t1c-base-ndr" + title: "2n-skx-xxv710-64b-4t2c-base-ndr" layout: "plot-latency" - type: "plot" - title: "DPDK Latency: l3fwd-3n-hsw-x520-64b-2t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x520-64b-2t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + title: "Latency: 2n-skx-xxv710-64b-8t4c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-xxv710-64b-8t4c-base-ndr-lat" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-8t4c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-8t4c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-x520-64b-2t2c-base-ndr" + title: "2n-skx-xxv710-64b-8t4c-base-ndr" layout: "plot-latency" -# Packet Latency - DPDK l3fwd 3n-hsw-x710 -- type: "plot" - title: "DPDK Latency: l3fwd-3n-hsw-x710-64b-1t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x710-64b-1t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "l3fwd-3n-hsw-x710-64b-1t1c-base-ndr" - layout: "plot-latency" +# 2n-skx-x710 - type: "plot" - title: "DPDK Latency: l3fwd-3n-hsw-x710-64b-2t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-x710-64b-2t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + title: "Latency: 2n-skx-x710-64b-2t1c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-x710-64b-2t1c-base-ndr-lat" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-x710-64b-2t2c-base-ndr" + title: "2n-skx-x710-64b-2t1c-base-ndr" layout: "plot-latency" -# Packet Latency - DPDK l3fwd 3n-hsw-xl710 - type: "plot" - title: "DPDK Latency: l3fwd-3n-hsw-xl710-64b-1t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-xl710-64b-1t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + title: "Latency: 2n-skx-x710-64b-4t2c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-x710-64b-4t2c-base-ndr-lat" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-xl710-64b-1t1c-base-ndr" + title: "2n-skx-x710-64b-4t2c-base-ndr" layout: "plot-latency" - type: "plot" - title: "DPDK Latency: l3fwd-3n-hsw-xl710-64b-2t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-hsw-xl710-64b-2t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + title: "Latency: 2n-skx-x710-64b-8t4c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/2n-skx-x710-64b-8t4c-base-ndr-lat" + data: "plot-dpdk-throughput-latency-2n-skx" + include: + - "tests.dpdk.perf.2n1l-10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-8t4c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.2n1l-10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-8t4c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-hsw-xl710-64b-2t2c-base-ndr" + title: "2n-skx-x710-64b-8t4c-base-ndr" layout: "plot-latency" -# Packet Latency - DPDK l3fwd 3n-skx-x710 -- type: "plot" - title: "DPDK Latency: l3fwd-3n-skx-x710-64b-2t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-x710-64b-2t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "l3fwd-3n-skx-x710-64b-2t1c-base-ndr" - layout: "plot-latency" +# 3n-skx-xxv710 - type: "plot" - title: "DPDK Latency: l3fwd-3n-skx-x710-64b-4t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-x710-64b-4t2c-base-ndr-lat" + title: "Latency: 3n-skx-xxv710-64b-2t1c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-xxv710-64b-2t1c-base-ndr-lat" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-x710-64b-4t2c-base-ndr" + title: "3n-skx-xxv710-64b-2t1c-base-ndr" layout: "plot-latency" -# Packet Latency - DPDK l3fwd 3n-skx-xxv710 - type: "plot" - title: "DPDK Latency: l3fwd-3n-skx-xxv710-64b-2t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-xxv710-64b-2t1c-base-ndr-lat" + title: "Latency: 3n-skx-xxv710-64b-4t2c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-xxv710-64b-4t2c-base-ndr-lat" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-xxv710-64b-2t1c-base-ndr" + title: "3n-skx-xxv710-64b-4t2c-base-ndr" layout: "plot-latency" - type: "plot" - title: "DPDK Latency: l3fwd-3n-skx-xxv710-64b-4t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-3n-skx-xxv710-64b-4t2c-base-ndr-lat" + title: "Latency: 3n-skx-xxv710-64b-8t4c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-xxv710-64b-8t4c-base-ndr-lat" data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.25ge2p1xxv710-eth-l2xcbase-testpmd-ndrpdr.64b-8t4c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.25ge2p1xxv710-ethip4-ip4base-l3fwd-ndrpdr.64b-8t4c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-3n-skx-xxv710-64b-4t2c-base-ndr" + title: "3n-skx-xxv710-64b-8t4c-base-ndr" layout: "plot-latency" -# Packet Latency - DPDK l3fwd 2n-skx-x710 -- type: "plot" - title: "DPDK Latency: l3fwd-2n-skx-x710-64b-2t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-x710-64b-2t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "l3fwd-2n-skx-x710-64b-2t1c-base-ndr" - layout: "plot-latency" +# 3n-skx-x710 - type: "plot" - title: "DPDK Latency: l3fwd-2n-skx-x710-64b-4t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-x710-64b-4t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + title: "Latency: 3n-skx-x710-64b-2t1c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-x710-64b-2t1c-base-ndr-lat" + data: "plot-dpdk-throughput-latency-3n-skx" + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-2n-skx-x710-64b-4t2c-base-ndr" + title: "3n-skx-x710-64b-2t1c-base-ndr" layout: "plot-latency" -# Packet Latency - DPDK l3fwd 2n-skx-xxv710 - type: "plot" - title: "DPDK Latency: l3fwd-2n-skx-xxv710-64b-2t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-xxv710-64b-2t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + title: "Latency: 3n-skx-x710-64b-4t2c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-x710-64b-4t2c-base-ndr-lat" + data: "plot-dpdk-throughput-latency-3n-skx" + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-4t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-2n-skx-xxv710-64b-2t1c-base-ndr" + title: "3n-skx-x710-64b-4t2c-base-ndr" layout: "plot-latency" - type: "plot" - title: "DPDK Latency: l3fwd-2n-skx-xxv710-64b-4t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/l3fwd-2n-skx-xxv710-64b-4t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - 'IP4FWD'" - parameters: - - "latency" - - "parent" - - "tags" + title: "Latency: 3n-skx-x710-64b-8t4c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-skx-x710-64b-8t4c-base-ndr-lat" + data: "plot-dpdk-throughput-latency-3n-skx" + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-8t4c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-8t4c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "l3fwd-2n-skx-xxv710-64b-4t2c-base-ndr" + title: "3n-skx-x710-64b-8t4c-base-ndr" layout: "plot-latency" -################################################################################ +# 3n-hsw-xl710 -# Packet Latency - DPDK testpmd 3n-hsw-x520 - type: "plot" - title: "DPDK Latency: testpmd-3n-hsw-x520-64b-1t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x520-64b-1t1c-base-ndr-lat" + title: "Latency: 3n-hsw-xl710-64b-1t1c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-xl710-64b-1t1c-base-ndr-lat" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.40ge2p1xl710-eth-l2xcbase-testpmd-ndrpdr.64b-1t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.40ge2p1xl710-ethip4-ip4base-l3fwd-ndrpdr.64b-1t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-x520-64b-1t1c-base-ndr" + title: "3n-hsw-xl710-64b-1t1c-base-ndr" layout: "plot-latency" - type: "plot" - title: "DPDK Latency: testpmd-3n-hsw-x520-64b-2t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x520-64b-2t2c-base-ndr-lat" + title: "Latency: 3n-hsw-xl710-64b-2t2c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-xl710-64b-2t2c-base-ndr-lat" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X520-DA2' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.40ge2p1xl710-eth-l2xcbase-testpmd-ndrpdr.64b-2t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.40ge2p1xl710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-x520-64b-2t2c-base-ndr" + title: "3n-hsw-xl710-64b-2t2c-base-ndr" layout: "plot-latency" -# Packet Latency - DPDK testpmd 3n-hsw-x710 - type: "plot" - title: "DPDK Latency: testpmd-3n-hsw-x710-64b-1t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x710-64b-1t1c-base-ndr-lat" + title: "Latency: 3n-hsw-xl710-64b-4t4c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-xl710-64b-4t4c-base-ndr-lat" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.40ge2p1xl710-eth-l2xcbase-testpmd-ndrpdr.64b-4t4c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.40ge2p1xl710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t4c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-x710-64b-1t1c-base-ndr" + title: "3n-hsw-xl710-64b-4t4c-base-ndr" layout: "plot-latency" -- type: "plot" - title: "DPDK Latency: testpmd-3n-hsw-x710-64b-2t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-x710-64b-2t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "testpmd-3n-hsw-x710-64b-2t2c-base-ndr" - layout: "plot-latency" +# 3n-hsw-x710 -# Packet Latency - DPDK testpmd 3n-hsw-xl710 - type: "plot" - title: "DPDK Latency: testpmd-3n-hsw-xl710-64b-1t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-xl710-64b-1t1c-base-ndr-lat" + title: "Latency: 3n-hsw-x710-64b-1t1c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-x710-64b-1t1c-base-ndr-lat" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '1T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-1t1c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-1t1c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-3n-hsw-xl710-64b-1t1c-base-ndr" + title: "3n-hsw-x710-64b-1t1c-base-ndr" layout: "plot-latency" - type: "plot" - title: "DPDK Latency: testpmd-3n-hsw-xl710-64b-2t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-hsw-xl710-64b-2t2c-base-ndr-lat" + title: "Latency: 3n-hsw-x710-64b-2t2c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-x710-64b-2t2c-base-ndr-lat" data: "plot-dpdk-throughput-latency-3n-hsw" - filter: "'NIC_Intel-XL710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "testpmd-3n-hsw-xl710-64b-2t2c-base-ndr" - layout: "plot-latency" - -# Packet Latency - DPDK testpmd 3n-skx-x710 -- type: "plot" - title: "DPDK Latency: testpmd-3n-skx-x710-64b-2t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-x710-64b-2t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "testpmd-3n-skx-x710-64b-2t1c-base-ndr" - layout: "plot-latency" - -- type: "plot" - title: "DPDK Latency: testpmd-3n-skx-x710-64b-4t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-x710-64b-4t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "testpmd-3n-skx-x710-64b-4t2c-base-ndr" - layout: "plot-latency" - -# Packet Latency - DPDK testpmd 3n-skx-xxv710 -- type: "plot" - title: "DPDK Latency: testpmd-3n-skx-xxv710-64b-2t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-xxv710-64b-2t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "testpmd-3n-skx-xxv710-64b-2t1c-base-ndr" - layout: "plot-latency" - -- type: "plot" - title: "DPDK Latency: testpmd-3n-skx-xxv710-64b-4t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-3n-skx-xxv710-64b-4t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-3n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "testpmd-3n-skx-xxv710-64b-4t2c-base-ndr" - layout: "plot-latency" - -# Packet Latency - DPDK testpmd 2n-skx-x710 -- type: "plot" - title: "DPDK Latency: testpmd-2n-skx-x710-64b-2t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-x710-64b-2t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "testpmd-2n-skx-x710-64b-2t1c-base-ndr" - layout: "plot-latency" - -- type: "plot" - title: "DPDK Latency: testpmd-2n-skx-x710-64b-4t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-x710-64b-4t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-X710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" - layout: - title: "testpmd-2n-skx-x710-64b-4t2c-base-ndr" - layout: "plot-latency" - -# Packet Latency - DPDK testpmd 2n-skx-xxv710 -- type: "plot" - title: "DPDK Latency: testpmd-2n-skx-xxv710-64b-2t1c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-xxv710-64b-2t1c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '2T1C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-2t2c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-2t2c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-2n-skx-xxv710-64b-2t1c-base-ndr" + title: "3n-hsw-x710-64b-2t2c-base-ndr" layout: "plot-latency" - type: "plot" - title: "DPDK Latency: testpmd-2n-skx-xxv710-64b-4t2c-base-ndr-lat" - algorithm: "plot_latency_error_bars" - output-file-type: ".html" - output-file: "{DIR[STATIC,DPDK]}/testpmd-2n-skx-xxv710-64b-4t2c-base-ndr-lat" - data: "plot-dpdk-throughput-latency-2n-skx" - filter: "'NIC_Intel-XXV710' and - '64B' and - 'BASE' and - 'NDRPDR' and - '4T2C' and - ('L2BDMACSTAT' or 'L2BDMACLRN' or 'L2XCFWD' or 'L2PATCH')" - parameters: - - "latency" - - "parent" - - "tags" + title: "Latency: 3n-hsw-x710-64b-4t4c-base-ndr-lat" + algorithm: "plot_lat_err_bars_name" + output-file: "{DIR[STATIC,DPDK]}/3n-hsw-x710-64b-4t4c-base-ndr-lat" + data: "plot-dpdk-throughput-latency-3n-hsw" + include: + - "tests.dpdk.perf.10ge2p1x710-eth-l2xcbase-testpmd-ndrpdr.64b-4t4c-eth-l2xcbase-testpmd-ndrpdr" + - "tests.dpdk.perf.10ge2p1x710-ethip4-ip4base-l3fwd-ndrpdr.64b-4t4c-ethip4-ip4base-l3fwd-ndrpdr" layout: - title: "testpmd-2n-skx-xxv710-64b-4t2c-base-ndr" + title: "3n-hsw-x710-64b-4t4c-base-ndr" layout: "plot-latency" diff --git a/resources/tools/presentation/specification_CPTA.yaml b/resources/tools/presentation/specification_CPTA.yaml index cd3b62a1e3..9363428773 100644 --- a/resources/tools/presentation/specification_CPTA.yaml +++ b/resources/tools/presentation/specification_CPTA.yaml @@ -191,7 +191,7 @@ # 3n-hsw plot-performance-trending-all-3n-hsw: csit-vpp-perf-mrr-daily-master: - start: 836 + start: 851 end: "lastCompletedBuild" skip: - 852 @@ -218,17 +218,16 @@ - 940 - 941 - 953 + - 987 csit-dpdk-perf-mrr-weekly-master: - start: 83 + start: 91 end: "lastCompletedBuild" skip: - - 87 - - 89 - - 90 + - 101 plot-performance-trending-vpp-3n-hsw: csit-vpp-perf-mrr-daily-master: - start: 836 + start: 851 end: "lastCompletedBuild" skip: - 852 @@ -255,23 +254,21 @@ - 940 - 941 - 953 + - 987 plot-performance-trending-dpdk-3n-hsw: csit-dpdk-perf-mrr-weekly-master: - start: 83 + start: 91 end: "lastCompletedBuild" skip: - - 87 - - 89 - - 90 + - 101 # 3n-skx plot-performance-trending-all-3n-skx: csit-vpp-perf-mrr-daily-master-3n-skx: - start: 587 + start: 598 end: "lastCompletedBuild" skip: - - 588 - 600 - 601 - 603 @@ -294,21 +291,20 @@ - 687 - 694 - 695 + - 696 + - 697 + - 705 csit-dpdk-perf-mrr-weekly-master-3n-skx: - start: 48 + start: 57 end: "lastCompletedBuild" skip: - - 52 - - 54 - - 55 - - 56 + - 66 plot-performance-trending-vpp-3n-skx: csit-vpp-perf-mrr-daily-master-3n-skx: - start: 587 + start: 598 end: "lastCompletedBuild" skip: - - 588 - 600 - 601 - 603 @@ -331,24 +327,23 @@ - 687 - 694 - 695 + - 696 + - 697 + - 705 plot-performance-trending-dpdk-3n-skx: csit-dpdk-perf-mrr-weekly-master-3n-skx: - start: 48 + start: 57 end: "lastCompletedBuild" skip: - - 52 - - 54 - - 55 - - 56 + - 66 # 2n-skx plot-performance-trending-all-2n-skx: csit-vpp-perf-mrr-daily-master-2n-skx: - start: 589 + start: 606 end: "lastCompletedBuild" skip: - - 605 - 608 - 609 - 612 @@ -375,20 +370,17 @@ - 701 - 702 csit-dpdk-perf-mrr-weekly-master-2n-skx: - start: 48 + start: 57 end: "lastCompletedBuild" skip: - - 52 - - 54 - - 55 - - 56 + - 64 + - 66 plot-performance-trending-vpp-2n-skx: csit-vpp-perf-mrr-daily-master-2n-skx: - start: 589 + start: 606 end: "lastCompletedBuild" skip: - - 605 - 608 - 609 - 612 @@ -417,56 +409,40 @@ plot-performance-trending-dpdk-2n-skx: csit-dpdk-perf-mrr-weekly-master-2n-skx: - start: 48 + start: 57 end: "lastCompletedBuild" skip: - - 52 - - 54 - - 55 - - 56 + - 64 + - 66 plot-performance-trending-vpp-nfv-2n-skx: csit-vpp-perf-mrr-weekly-master-2n-skx: - start: 26 + start: 30 end: "lastCompletedBuild" skip: - - 29 - 34 - 37 # 3n-tsh plot-performance-trending-all-3n-tsh: csit-vpp-perf-mrr-daily-master-3n-tsh: - start: 75 + start: 95 end: "lastCompletedBuild" skip: - - 80 - - 84 - - 89 - - 92 - - 93 - - 94 - 95 plot-performance-trending-vpp-3n-tsh: csit-vpp-perf-mrr-daily-master-3n-tsh: - start: 75 + start: 95 end: "lastCompletedBuild" skip: - - 80 - - 84 - - 89 - - 92 - - 93 - - 94 - 95 plot-performance-trending-vpp-3n-dnv: csit-vpp-perf-mrr-daily-master-3n-dnv: - start: 69 + start: 79 end: "lastCompletedBuild" skip: - - 78 - 80 - 83 - 84 @@ -511,10 +487,12 @@ - 151 - 175 - 176 + - 227 + - 228 plot-performance-trending-vpp-2n-dnv: csit-vpp-perf-mrr-daily-master-2n-dnv: - start: 67 + start: 78 end: "lastCompletedBuild" skip: - 79 @@ -560,6 +538,7 @@ - 145 - 149 - 150 + - 228 plot-layouts: @@ -658,7 +637,7 @@ # 3n-hsw csit-vpp-perf-mrr-daily-master: - start: 836 + start: 851 end: "lastCompletedBuild" skip: - 852 @@ -685,20 +664,18 @@ - 940 - 941 - 953 + - 987 csit-dpdk-perf-mrr-weekly-master: - start: 83 + start: 91 end: "lastCompletedBuild" skip: - - 87 - - 89 - - 90 + - 101 # 3n-skx csit-vpp-perf-mrr-daily-master-3n-skx: - start: 587 + start: 598 end: "lastCompletedBuild" skip: - - 588 - 600 - 601 - 603 @@ -721,21 +698,20 @@ - 687 - 694 - 695 + - 696 + - 697 + - 705 csit-dpdk-perf-mrr-weekly-master-3n-skx: - start: 48 + start: 57 end: "lastCompletedBuild" skip: - - 52 - - 54 - - 55 - - 56 + - 66 # 2n-skx csit-vpp-perf-mrr-daily-master-2n-skx: - start: 589 + start: 606 end: "lastCompletedBuild" skip: - - 605 - 608 - 609 - 612 @@ -762,40 +738,30 @@ - 701 - 702 csit-vpp-perf-mrr-weekly-master-2n-skx: - start: 26 + start: 30 end: "lastCompletedBuild" skip: - - 29 - 34 - 37 csit-dpdk-perf-mrr-weekly-master-2n-skx: - start: 48 + start: 57 end: "lastCompletedBuild" skip: - - 52 - - 54 - - 55 - - 56 + - 64 + - 66 # 3n-tsh csit-vpp-perf-mrr-daily-master-3n-tsh: - start: 75 + start: 95 end: "lastCompletedBuild" skip: - - 80 - - 84 - - 89 - - 92 - - 93 - - 94 - 95 # 3n-dnv csit-vpp-perf-mrr-daily-master-3n-dnv: - start: 69 + start: 79 end: "lastCompletedBuild" skip: - - 78 - 80 - 83 - 84 @@ -840,10 +806,12 @@ - 151 - 175 - 176 + - 227 + - 228 # 2n-dnv csit-vpp-perf-mrr-daily-master-2n-dnv: - start: 67 + start: 78 end: "lastCompletedBuild" skip: - 79 @@ -889,6 +857,7 @@ - 145 - 149 - 150 + - 228 - type: "output" @@ -987,7 +956,7 @@ - type: "table" title: "Performance trending dashboard 1t1c-3n-hsw" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-hsw" data: "plot-performance-trending-all-3n-hsw" @@ -1006,7 +975,7 @@ - type: "table" title: "Performance trending dashboard 2t2c-3n-hsw" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-hsw" data: "plot-performance-trending-all-3n-hsw" @@ -1025,7 +994,7 @@ - type: "table" title: "Performance trending dashboard 4t4c-3n-hsw" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-hsw" data: "plot-performance-trending-all-3n-hsw" @@ -1044,7 +1013,7 @@ - type: "table" title: "HTML performance trending dashboard 1t1c-3n-hsw" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-hsw.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-hsw.rst" testbed: "3n-hsw" @@ -1052,7 +1021,7 @@ - type: "table" title: "HTML performance trending dashboard 2t2c-3n-hsw" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-hsw.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-hsw.rst" testbed: "3n-hsw" @@ -1060,7 +1029,7 @@ - type: "table" title: "HTML performance trending dashboard 4t4c-3n-hsw" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-hsw.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-hsw.rst" testbed: "3n-hsw" @@ -1096,7 +1065,7 @@ - type: "table" title: "Performance trending dashboard 2t1c-3n-skx" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t1c-3n-skx" data: "plot-performance-trending-all-3n-skx" @@ -1115,7 +1084,7 @@ - type: "table" title: "Performance trending dashboard 4t2c-3n-skx" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t2c-3n-skx" data: "plot-performance-trending-all-3n-skx" @@ -1134,7 +1103,7 @@ - type: "table" title: "Performance trending dashboard 8t4c-3n-skx" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-8t4c-3n-skx" data: "plot-performance-trending-all-3n-skx" @@ -1153,7 +1122,7 @@ - type: "table" title: "HTML performance trending dashboard 2t1c-3n-skx" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t1c-3n-skx.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t1c-3n-skx.rst" testbed: "3n-skx" @@ -1161,7 +1130,7 @@ - type: "table" title: "HTML performance trending dashboard 4t2c-3n-skx" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t2c-3n-skx.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t2c-3n-skx.rst" testbed: "3n-skx" @@ -1169,7 +1138,7 @@ - type: "table" title: "HTML performance trending dashboard 8t4c-3n-skx" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-8t4c-3n-skx.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-8t4c-3n-skx.rst" testbed: "3n-skx" @@ -1205,7 +1174,7 @@ - type: "table" title: "Performance trending dashboard 2t1c-2n-skx" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t1c-2n-skx" data: "plot-performance-trending-all-2n-skx" @@ -1224,7 +1193,7 @@ - type: "table" title: "Performance trending dashboard 4t2c-2n-skx" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t2c-2n-skx" data: "plot-performance-trending-all-2n-skx" @@ -1243,7 +1212,7 @@ - type: "table" title: "Performance trending dashboard 8t4c-2n-skx" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-8t4c-2n-skx" data: "plot-performance-trending-all-2n-skx" @@ -1262,7 +1231,7 @@ - type: "table" title: "HTML performance trending dashboard 2t1c-2n-skx" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t1c-2n-skx.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t1c-2n-skx.rst" testbed: "2n-skx" @@ -1270,7 +1239,7 @@ - type: "table" title: "HTML performance trending dashboard 4t2c-2n-skx" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t2c-2n-skx.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t2c-2n-skx.rst" testbed: "2n-skx" @@ -1278,7 +1247,7 @@ - type: "table" title: "HTML performance trending dashboard 8t4c-2n-skx" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-8t4c-2n-skx.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-8t4c-2n-skx.rst" testbed: "2n-skx" @@ -1314,7 +1283,7 @@ - type: "table" title: "Performance trending dashboard 1t1c-3n-tsh" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-tsh" data: "plot-performance-trending-all-3n-tsh" @@ -1329,7 +1298,7 @@ - type: "table" title: "Performance trending dashboard 2t2c-3n-tsh" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-tsh" data: "plot-performance-trending-all-3n-tsh" @@ -1344,7 +1313,7 @@ - type: "table" title: "Performance trending dashboard 4t4c-3n-tsh" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-tsh" data: "plot-performance-trending-all-3n-tsh" @@ -1359,7 +1328,7 @@ - type: "table" title: "HTML performance trending dashboard 1t1c-3n-tsh" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-tsh.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-tsh.rst" testbed: "3n-tsh" @@ -1367,7 +1336,7 @@ - type: "table" title: "HTML performance trending dashboard 2t2c-3n-tsh" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-tsh.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-tsh.rst" testbed: "3n-tsh" @@ -1375,7 +1344,7 @@ - type: "table" title: "HTML performance trending dashboard 4t4c-3n-tsh" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-tsh.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-tsh.rst" testbed: "3n-tsh" @@ -1407,7 +1376,7 @@ - type: "table" title: "Performance trending dashboard 1t1c-3n-dnv" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-dnv" data: "plot-performance-trending-vpp-3n-dnv" @@ -1422,7 +1391,7 @@ - type: "table" title: "Performance trending dashboard 2t2c-3n-dnv" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-dnv" data: "plot-performance-trending-vpp-3n-dnv" @@ -1437,7 +1406,7 @@ - type: "table" title: "Performance trending dashboard 4t4c-3n-dnv" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-dnv" data: "plot-performance-trending-vpp-3n-dnv" @@ -1452,7 +1421,7 @@ - type: "table" title: "HTML performance trending dashboard 1t1c-3n-dnv" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-dnv.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-3n-dnv.rst" testbed: "3n-dnv" @@ -1460,7 +1429,7 @@ - type: "table" title: "HTML performance trending dashboard 2t2c-3n-dnv" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-dnv.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-3n-dnv.rst" testbed: "3n-dnv" @@ -1468,7 +1437,7 @@ - type: "table" title: "HTML performance trending dashboard 4t4c-3n-dnv" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-dnv.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-3n-dnv.rst" testbed: "3n-dnv" @@ -1500,7 +1469,7 @@ - type: "table" title: "Performance trending dashboard 1t1c-2n-dnv" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-2n-dnv" data: "plot-performance-trending-vpp-2n-dnv" @@ -1515,7 +1484,7 @@ - type: "table" title: "Performance trending dashboard 2t2c-2n-dnv" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-2n-dnv" data: "plot-performance-trending-vpp-2n-dnv" @@ -1530,7 +1499,7 @@ - type: "table" title: "Performance trending dashboard 4t4c-2n-dnv" - algorithm: "table_performance_trending_dashboard" + algorithm: "table_perf_trending_dash" output-file-ext: ".csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-2n-dnv" data: "plot-performance-trending-vpp-2n-dnv" @@ -1545,7 +1514,7 @@ - type: "table" title: "HTML performance trending dashboard 1t1c-2n-dnv" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-2n-dnv.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-1t1c-2n-dnv.rst" testbed: "2n-dnv" @@ -1553,7 +1522,7 @@ - type: "table" title: "HTML performance trending dashboard 2t2c-2n-dnv" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-2n-dnv.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-2t2c-2n-dnv.rst" testbed: "2n-dnv" @@ -1561,7 +1530,7 @@ - type: "table" title: "HTML performance trending dashboard 4t4c-2n-dnv" - algorithm: "table_performance_trending_dashboard_html" + algorithm: "table_perf_trending_dash_html" input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-2n-dnv.csv" output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c-2n-dnv.rst" testbed: "2n-dnv" diff --git a/resources/tools/presentation/specification_parser.py b/resources/tools/presentation/specification_parser.py index f99c7515fb..ef2fafccae 100644 --- a/resources/tools/presentation/specification_parser.py +++ b/resources/tools/presentation/specification_parser.py @@ -18,12 +18,14 @@ Parsing of the specification YAML file. import logging -from yaml import load, YAMLError from pprint import pformat -from errors import PresentationError -from utils import ( - get_last_successful_build_number, get_last_completed_build_number) +from yaml import load, FullLoader, YAMLError + +from pal_errors import PresentationError +from pal_utils import ( + get_last_successful_build_nr, get_last_completed_build_number +) class Specification: @@ -35,8 +37,8 @@ class Specification: # Tags are used in specification YAML file and replaced while the file is # parsed. - TAG_OPENER = "{" - TAG_CLOSER = "}" + TAG_OPENER = u"{" + TAG_CLOSER = u"}" def __init__(self, cfg_file): """Initialization. @@ -47,15 +49,17 @@ class Specification: self._cfg_file = cfg_file self._cfg_yaml = None - self._specification = {"environment": dict(), - "configuration": dict(), - "static": dict(), - "input": dict(), - "output": dict(), - "tables": list(), - "plots": list(), - "files": list(), - "cpta": dict()} + self._specification = { + u"environment": dict(), + u"configuration": dict(), + u"static": dict(), + u"input": dict(), + u"output": dict(), + u"tables": list(), + u"plots": list(), + u"files": list(), + u"cpta": dict() + } @property def specification(self): @@ -73,7 +77,7 @@ class Specification: :returns: Environment specification. :rtype: dict """ - return self._specification["environment"] + return self._specification[u"environment"] @property def configuration(self): @@ -82,7 +86,7 @@ class Specification: :returns: Configuration of PAL. :rtype: dict """ - return self._specification["configuration"] + return self._specification[u"configuration"] @property def static(self): @@ -91,7 +95,7 @@ class Specification: :returns: Static content specification. :rtype: dict """ - return self._specification["static"] + return self._specification[u"static"] @property def mapping(self): @@ -101,7 +105,7 @@ class Specification: one. :rtype: dict """ - return self._specification["configuration"]["mapping"] + return self._specification[u"configuration"][u"mapping"] @property def ignore(self): @@ -110,7 +114,7 @@ class Specification: :returns: List of ignored test cases. :rtype: list """ - return self._specification["configuration"]["ignore"] + return self._specification[u"configuration"][u"ignore"] @property def alerting(self): @@ -119,7 +123,7 @@ class Specification: :returns: Specification of alerts. :rtype: dict """ - return self._specification["configuration"]["alerting"] + return self._specification[u"configuration"][u"alerting"] @property def input(self): @@ -129,7 +133,7 @@ class Specification: :returns: Inputs. :rtype: dict """ - return self._specification["input"] + return self._specification[u"input"] @property def builds(self): @@ -138,7 +142,7 @@ class Specification: :returns: Builds defined in the specification. :rtype: dict """ - return self.input["builds"] + return self.input[u"builds"] @property def output(self): @@ -149,7 +153,7 @@ class Specification: :returns: Outputs to be generated. :rtype: dict """ - return self._specification["output"] + return self._specification[u"output"] @property def tables(self): @@ -158,7 +162,7 @@ class Specification: :returns: List of specifications of tables to be generated. :rtype: list """ - return self._specification["tables"] + return self._specification[u"tables"] @property def plots(self): @@ -167,7 +171,7 @@ class Specification: :returns: List of specifications of plots to be generated. :rtype: list """ - return self._specification["plots"] + return self._specification[u"plots"] @property def files(self): @@ -176,7 +180,7 @@ class Specification: :returns: List of specifications of files to be generated. :rtype: list """ - return self._specification["files"] + return self._specification[u"files"] @property def cpta(self): @@ -187,7 +191,7 @@ class Specification: Analysis to be generated. :rtype: list """ - return self._specification["cpta"] + return self._specification[u"cpta"] def set_input_state(self, job, build_nr, state): """Set the state of input @@ -199,17 +203,20 @@ class Specification: """ try: - for build in self._specification["input"]["builds"][job]: - if build["build"] == build_nr: - build["status"] = state + for build in self._specification[u"input"][u"builds"][job]: + if build[u"build"] == build_nr: + build[u"status"] = state break else: - raise PresentationError("Build '{}' is not defined for job '{}'" - " in specification file.". - format(build_nr, job)) + raise PresentationError( + f"Build {build_nr} is not defined for job {job} in " + f"specification file." + ) except KeyError: - raise PresentationError("Job '{}' and build '{}' is not defined in " - "specification file.".format(job, build_nr)) + raise PresentationError( + f"Job {job} and build {build_nr} is not defined in " + f"specification file." + ) def set_input_file_name(self, job, build_nr, file_name): """Set the state of input @@ -221,17 +228,20 @@ class Specification: """ try: - for build in self._specification["input"]["builds"][job]: - if build["build"] == build_nr: - build["file-name"] = file_name + for build in self._specification[u"input"][u"builds"][job]: + if build[u"build"] == build_nr: + build[u"file-name"] = file_name break else: - raise PresentationError("Build '{}' is not defined for job '{}'" - " in specification file.". - format(build_nr, job)) + raise PresentationError( + f"Build {build_nr} is not defined for job {job} in " + f"specification file." + ) except KeyError: - raise PresentationError("Job '{}' and build '{}' is not defined in " - "specification file.".format(job, build_nr)) + raise PresentationError( + f"Job {job} and build {build_nr} is not defined in " + f"specification file." + ) def _get_build_number(self, job, build_type): """Get the number of the job defined by its name: @@ -250,26 +260,27 @@ class Specification: """ # defined as a range <start, end> - if build_type == "lastSuccessfulBuild": + if build_type == u"lastSuccessfulBuild": # defined as a range <start, lastSuccessfulBuild> - ret_code, build_nr, _ = get_last_successful_build_number( - self.environment["urls"]["URL[JENKINS,CSIT]"], job) - elif build_type == "lastCompletedBuild": + ret_code, build_nr, _ = get_last_successful_build_nr( + self.environment[u"urls"][u"URL[JENKINS,CSIT]"], job) + elif build_type == u"lastCompletedBuild": # defined as a range <start, lastCompletedBuild> ret_code, build_nr, _ = get_last_completed_build_number( - self.environment["urls"]["URL[JENKINS,CSIT]"], job) + self.environment[u"urls"][u"URL[JENKINS,CSIT]"], job) else: - raise PresentationError("Not supported build type: '{0}'". - format(build_type)) + raise PresentationError(f"Not supported build type: {build_type}") if ret_code != 0: - raise PresentationError("Not possible to get the number of the " - "build number.") + raise PresentationError(u"Not possible to get the number of the " + u"build number.") try: build_nr = int(build_nr) return build_nr except ValueError as err: - raise PresentationError("Not possible to get the number of the " - "build number.\nReason: {0}".format(err)) + raise PresentationError( + f"Not possible to get the number of the build number. Reason:\n" + f"{repr(err)}" + ) def _get_type_index(self, item_type): """Get index of item type (environment, input, output, ...) in @@ -284,7 +295,7 @@ class Specification: index = 0 for item in self._cfg_yaml: - if item["type"] == item_type: + if item[u"type"] == item_type: return index index += 1 return None @@ -337,12 +348,13 @@ class Specification: data[key] = value.replace(tag, src_data[tag[1:-1]]) counter += 1 except KeyError: - raise PresentationError("Not possible to replace the " - "tag '{}'".format(tag)) + raise PresentationError( + f"Not possible to replace the tag {tag}" + ) if counter: self._replace_tags(data, src_data) else: - raise PresentationError("Replace tags: Not supported data type.") + raise PresentationError(u"Replace tags: Not supported data type.") return data @@ -350,158 +362,170 @@ class Specification: """Parse environment specification in the specification YAML file. """ - logging.info("Parsing specification file: environment ...") + logging.info(u"Parsing specification file: environment ...") - idx = self._get_type_index("environment") + idx = self._get_type_index(u"environment") if idx is None: - return None + return try: - self._specification["environment"]["configuration"] = \ - self._cfg_yaml[idx]["configuration"] + self._specification[u"environment"][u"configuration"] = \ + self._cfg_yaml[idx][u"configuration"] except KeyError: - self._specification["environment"]["configuration"] = None + self._specification[u"environment"][u"configuration"] = None try: - self._specification["environment"]["paths"] = \ - self._replace_tags(self._cfg_yaml[idx]["paths"]) + self._specification[u"environment"][u"paths"] = \ + self._replace_tags(self._cfg_yaml[idx][u"paths"]) except KeyError: - self._specification["environment"]["paths"] = None + self._specification[u"environment"][u"paths"] = None try: - self._specification["environment"]["urls"] = \ - self._cfg_yaml[idx]["urls"] + self._specification[u"environment"][u"urls"] = \ + self._cfg_yaml[idx][u"urls"] except KeyError: - self._specification["environment"]["urls"] = None + self._specification[u"environment"][u"urls"] = None try: - self._specification["environment"]["make-dirs"] = \ - self._cfg_yaml[idx]["make-dirs"] + self._specification[u"environment"][u"make-dirs"] = \ + self._cfg_yaml[idx][u"make-dirs"] except KeyError: - self._specification["environment"]["make-dirs"] = None + self._specification[u"environment"][u"make-dirs"] = None try: - self._specification["environment"]["remove-dirs"] = \ - self._cfg_yaml[idx]["remove-dirs"] + self._specification[u"environment"][u"remove-dirs"] = \ + self._cfg_yaml[idx][u"remove-dirs"] except KeyError: - self._specification["environment"]["remove-dirs"] = None + self._specification[u"environment"][u"remove-dirs"] = None try: - self._specification["environment"]["build-dirs"] = \ - self._cfg_yaml[idx]["build-dirs"] + self._specification[u"environment"][u"build-dirs"] = \ + self._cfg_yaml[idx][u"build-dirs"] except KeyError: - self._specification["environment"]["build-dirs"] = None + self._specification[u"environment"][u"build-dirs"] = None try: - self._specification["environment"]["testbeds"] = \ - self._cfg_yaml[idx]["testbeds"] + self._specification[u"environment"][u"testbeds"] = \ + self._cfg_yaml[idx][u"testbeds"] except KeyError: - self._specification["environment"]["testbeds"] = None + self._specification[u"environment"][u"testbeds"] = None + + logging.info(u"Done.") + + def _load_mapping_table(self): + """Load a mapping table if it is specified. If not, use empty list. + """ + + mapping_file_name = self._specification[u"configuration"].\ + get(u"mapping-file", None) + if mapping_file_name: + try: + with open(mapping_file_name, u'r') as mfile: + mapping = load(mfile, Loader=FullLoader) + # Make sure everything is lowercase + self._specification[u"configuration"][u"mapping"] = \ + {key.lower(): val.lower() for key, val in + mapping.items()} + logging.debug(f"Loaded mapping table:\n{mapping}") + except (YAMLError, IOError) as err: + raise PresentationError( + msg=f"An error occurred while parsing the mapping file " + f"{mapping_file_name}", + details=repr(err) + ) + else: + self._specification[u"configuration"][u"mapping"] = dict() - logging.info("Done.") + def _load_ignore_list(self): + """Load an ignore list if it is specified. If not, use empty list. + """ + + ignore_list_name = self._specification[u"configuration"].\ + get(u"ignore-list", None) + if ignore_list_name: + try: + with open(ignore_list_name, u'r') as ifile: + ignore = load(ifile, Loader=FullLoader) + # Make sure everything is lowercase + self._specification[u"configuration"][u"ignore"] = \ + [item.lower() for item in ignore] + logging.debug(f"Loaded ignore list:\n{ignore}") + except (YAMLError, IOError) as err: + raise PresentationError( + msg=f"An error occurred while parsing the ignore list file " + f"{ignore_list_name}.", + details=repr(err) + ) + else: + self._specification[u"configuration"][u"ignore"] = list() def _parse_configuration(self): """Parse configuration of PAL in the specification YAML file. """ - logging.info("Parsing specification file: configuration ...") + logging.info(u"Parsing specification file: configuration ...") idx = self._get_type_index("configuration") if idx is None: - logging.warning("No configuration information in the specification " - "file.") - return None + logging.warning( + u"No configuration information in the specification file." + ) + return try: - self._specification["configuration"] = self._cfg_yaml[idx] - + self._specification[u"configuration"] = self._cfg_yaml[idx] except KeyError: - raise PresentationError("No configuration defined.") + raise PresentationError(u"No configuration defined.") # Data sets: Replace ranges by lists - for set_name, data_set in self.configuration["data-sets"].items(): + for set_name, data_set in self.configuration[u"data-sets"].items(): if not isinstance(data_set, dict): continue for job, builds in data_set.items(): - if builds: - if isinstance(builds, dict): - build_end = builds.get("end", None) + if not builds: + continue + if isinstance(builds, dict): + build_end = builds.get(u"end", None) + try: + build_end = int(build_end) + except ValueError: + # defined as a range <start, build_type> + build_end = self._get_build_number(job, build_end) + builds = [x for x in range(builds[u"start"], + build_end + 1) + if x not in builds.get(u"skip", list())] + self.configuration[u"data-sets"][set_name][job] = builds + elif isinstance(builds, list): + for idx, item in enumerate(builds): try: - build_end = int(build_end) + builds[idx] = int(item) except ValueError: - # defined as a range <start, build_type> - build_end = self._get_build_number(job, build_end) - builds = [x for x in range(builds["start"], build_end+1) - if x not in builds.get("skip", list())] - self.configuration["data-sets"][set_name][job] = builds - elif isinstance(builds, list): - for idx, item in enumerate(builds): - try: - builds[idx] = int(item) - except ValueError: - # defined as a range <build_type> - builds[idx] = self._get_build_number(job, item) + # defined as a range <build_type> + builds[idx] = self._get_build_number(job, item) # Data sets: add sub-sets to sets (only one level): - for set_name, data_set in self.configuration["data-sets"].items(): + for set_name, data_set in self.configuration[u"data-sets"].items(): if isinstance(data_set, list): new_set = dict() for item in data_set: try: - for key, val in self.configuration["data-sets"][item].\ + for key, val in self.configuration[u"data-sets"][item].\ items(): new_set[key] = val except KeyError: raise PresentationError( - "Data set {0} is not defined in " - "the configuration section.".format(item)) - self.configuration["data-sets"][set_name] = new_set + f"Data set {item} is not defined in " + f"the configuration section." + ) + self.configuration[u"data-sets"][set_name] = new_set # Mapping table: - mapping = None - mapping_file_name = self._specification["configuration"].\ - get("mapping-file", None) - if mapping_file_name: - logging.debug("Mapping file: '{0}'".format(mapping_file_name)) - try: - with open(mapping_file_name, 'r') as mfile: - mapping = load(mfile) - logging.debug("Loaded mapping table:\n{0}".format(mapping)) - except (YAMLError, IOError) as err: - raise PresentationError( - msg="An error occurred while parsing the mapping file " - "'{0}'.".format(mapping_file_name), - details=repr(err)) - # Make sure everything is lowercase - if mapping: - self._specification["configuration"]["mapping"] = \ - {key.lower(): val.lower() for key, val in mapping.iteritems()} - else: - self._specification["configuration"]["mapping"] = dict() + self._load_mapping_table() # Ignore list: - ignore = None - ignore_list_name = self._specification["configuration"].\ - get("ignore-list", None) - if ignore_list_name: - logging.debug("Ignore list file: '{0}'".format(ignore_list_name)) - try: - with open(ignore_list_name, 'r') as ifile: - ignore = load(ifile) - logging.debug("Loaded ignore list:\n{0}".format(ignore)) - except (YAMLError, IOError) as err: - raise PresentationError( - msg="An error occurred while parsing the ignore list file " - "'{0}'.".format(ignore_list_name), - details=repr(err)) - # Make sure everything is lowercase - if ignore: - self._specification["configuration"]["ignore"] = \ - [item.lower() for item in ignore] - else: - self._specification["configuration"]["ignore"] = list() + self._load_ignore_list() - logging.info("Done.") + logging.info(u"Done.") def _parse_input(self): """Parse input specification in the specification YAML file. @@ -509,41 +533,43 @@ class Specification: :raises: PresentationError if there are no data to process. """ - logging.info("Parsing specification file: input ...") + logging.info(u"Parsing specification file: input ...") - idx = self._get_type_index("input") + idx = self._get_type_index(u"input") if idx is None: - raise PresentationError("No data to process.") + raise PresentationError(u"No data to process.") try: - for key, value in self._cfg_yaml[idx]["general"].items(): - self._specification["input"][key] = value - self._specification["input"]["builds"] = dict() + for key, value in self._cfg_yaml[idx][u"general"].items(): + self._specification[u"input"][key] = value + self._specification[u"input"][u"builds"] = dict() - for job, builds in self._cfg_yaml[idx]["builds"].items(): + for job, builds in self._cfg_yaml[idx][u"builds"].items(): if builds: if isinstance(builds, dict): - build_end = builds.get("end", None) + build_end = builds.get(u"end", None) try: build_end = int(build_end) except ValueError: # defined as a range <start, build_type> build_end = self._get_build_number(job, build_end) - builds = [x for x in range(builds["start"], build_end+1) - if x not in builds.get("skip", list())] - self._specification["input"]["builds"][job] = list() + builds = [x for x in range(builds[u"start"], + build_end + 1) + if x not in builds.get(u"skip", list())] + self._specification[u"input"][u"builds"][job] = list() for build in builds: - self._specification["input"]["builds"][job]. \ - append({"build": build, "status": None}) + self._specification[u"input"][u"builds"][job]. \ + append({u"build": build, u"status": None}) else: - logging.warning("No build is defined for the job '{}'. " - "Trying to continue without it.". - format(job)) + logging.warning( + f"No build is defined for the job {job}. Trying to " + f"continue without it." + ) except KeyError: - raise PresentationError("No data to process.") + raise PresentationError(u"No data to process.") - logging.info("Done.") + logging.info(u"Done.") def _parse_output(self): """Parse output specification in the specification YAML file. @@ -551,176 +577,226 @@ class Specification: :raises: PresentationError if there is no output defined. """ - logging.info("Parsing specification file: output ...") + logging.info(u"Parsing specification file: output ...") - idx = self._get_type_index("output") + idx = self._get_type_index(u"output") if idx is None: - raise PresentationError("No output defined.") + raise PresentationError(u"No output defined.") try: - self._specification["output"] = self._cfg_yaml[idx] + self._specification[u"output"] = self._cfg_yaml[idx] except (KeyError, IndexError): - raise PresentationError("No output defined.") + raise PresentationError(u"No output defined.") - logging.info("Done.") + logging.info(u"Done.") def _parse_static(self): """Parse specification of the static content in the specification YAML file. """ - logging.info("Parsing specification file: static content ...") + logging.info(u"Parsing specification file: static content ...") - idx = self._get_type_index("static") + idx = self._get_type_index(u"static") if idx is None: - logging.warning("No static content specified.") + logging.warning(u"No static content specified.") for key, value in self._cfg_yaml[idx].items(): if isinstance(value, str): try: self._cfg_yaml[idx][key] = self._replace_tags( - value, self._specification["environment"]["paths"]) + value, self._specification[u"environment"][u"paths"]) except KeyError: pass - self._specification["static"] = self._cfg_yaml[idx] + self._specification[u"static"] = self._cfg_yaml[idx] + + logging.info(u"Done.") + + def _parse_elements_tables(self, table): + """Parse tables from the specification YAML file. + + :param table: Table to be parsed from the specification file. + :type table: dict + :raises PresentationError: If wrong data set is used. + """ + + try: + table[u"template"] = self._replace_tags( + table[u"template"], + self._specification[u"environment"][u"paths"]) + except KeyError: + pass + + # Add data sets + try: + for item in (u"reference", u"compare"): + if table.get(item, None): + data_set = table[item].get(u"data", None) + if isinstance(data_set, str): + table[item][u"data"] = \ + self.configuration[u"data-sets"][data_set] + data_set = table[item].get(u"data-replacement", None) + if isinstance(data_set, str): + table[item][u"data-replacement"] = \ + self.configuration[u"data-sets"][data_set] + + if table.get(u"history", None): + for i in range(len(table[u"history"])): + data_set = table[u"history"][i].get(u"data", None) + if isinstance(data_set, str): + table[u"history"][i][u"data"] = \ + self.configuration[u"data-sets"][data_set] + data_set = table[u"history"][i].get( + u"data-replacement", None) + if isinstance(data_set, str): + table[u"history"][i][u"data-replacement"] = \ + self.configuration[u"data-sets"][data_set] + except KeyError: + raise PresentationError( + f"Wrong data set used in {table.get(u'title', u'')}." + ) + + self._specification[u"tables"].append(table) + + def _parse_elements_plots(self, plot): + """Parse plots from the specification YAML file. + + :param plot: Plot to be parsed from the specification file. + :type plot: dict + :raises PresentationError: If plot layout is not defined. + """ + + # Add layout to the plots: + layout = plot[u"layout"].get(u"layout", None) + if layout is not None: + plot[u"layout"].pop(u"layout") + try: + for key, val in (self.configuration[u"plot-layouts"] + [layout].items()): + plot[u"layout"][key] = val + except KeyError: + raise PresentationError( + f"Layout {layout} is not defined in the " + f"configuration section." + ) + self._specification[u"plots"].append(plot) + + def _parse_elements_files(self, file): + """Parse files from the specification YAML file. + + :param file: File to be parsed from the specification file. + :type file: dict + """ - logging.info("Done.") + try: + file[u"dir-tables"] = self._replace_tags( + file[u"dir-tables"], + self._specification[u"environment"][u"paths"]) + except KeyError: + pass + self._specification[u"files"].append(file) + + def _parse_elements_cpta(self, cpta): + """Parse cpta from the specification YAML file. + + :param cpta: cpta to be parsed from the specification file. + :type cpta: dict + :raises PresentationError: If wrong data set is used or if plot layout + is not defined. + """ + + for plot in cpta[u"plots"]: + # Add layout to the plots: + layout = plot.get(u"layout", None) + if layout is not None: + try: + plot[u"layout"] = \ + self.configuration[u"plot-layouts"][layout] + except KeyError: + raise PresentationError( + f"Layout {layout} is not defined in the " + f"configuration section." + ) + # Add data sets: + if isinstance(plot.get(u"data", None), str): + data_set = plot[u"data"] + try: + plot[u"data"] = \ + self.configuration[u"data-sets"][data_set] + except KeyError: + raise PresentationError( + f"Data set {data_set} is not defined in " + f"the configuration section." + ) + self._specification[u"cpta"] = cpta def _parse_elements(self): - """Parse elements (tables, plots) specification in the specification + """Parse elements (tables, plots, ..) specification in the specification YAML file. """ - logging.info("Parsing specification file: elements ...") + logging.info(u"Parsing specification file: elements ...") count = 1 for element in self._cfg_yaml: + + # Replace tags: try: - element["output-file"] = self._replace_tags( - element["output-file"], - self._specification["environment"]["paths"]) + element[u"output-file"] = self._replace_tags( + element[u"output-file"], + self._specification[u"environment"][u"paths"]) except KeyError: pass try: - element["input-file"] = self._replace_tags( - element["input-file"], - self._specification["environment"]["paths"]) + element[u"input-file"] = self._replace_tags( + element[u"input-file"], + self._specification[u"environment"][u"paths"]) except KeyError: pass - # add data sets to the elements: - if isinstance(element.get("data", None), str): - data_set = element["data"] + # Add data sets to the elements: + if isinstance(element.get(u"data", None), str): + data_set = element[u"data"] try: - element["data"] = self.configuration["data-sets"][data_set] + element[u"data"] = \ + self.configuration[u"data-sets"][data_set] except KeyError: - raise PresentationError("Data set {0} is not defined in " - "the configuration section.". - format(data_set)) + raise PresentationError( + f"Data set {data_set} is not defined in the " + f"configuration section." + ) - if element["type"] == "table": - logging.info(" {:3d} Processing a table ...".format(count)) - try: - element["template"] = self._replace_tags( - element["template"], - self._specification["environment"]["paths"]) - except KeyError: - pass - - # add data sets - try: - for item in ("reference", "compare"): - if element.get(item, None): - data_set = element[item].get("data", None) - if isinstance(data_set, str): - element[item]["data"] = \ - self.configuration["data-sets"][data_set] - data_set = element[item].get("data-replacement", - None) - if isinstance(data_set, str): - element[item]["data-replacement"] = \ - self.configuration["data-sets"][data_set] - - if element.get("history", None): - for i in range(len(element["history"])): - data_set = element["history"][i].get("data", None) - if isinstance(data_set, str): - element["history"][i]["data"] = \ - self.configuration["data-sets"][data_set] - data_set = element["history"][i].get( - "data-replacement", None) - if isinstance(data_set, str): - element["history"][i]["data-replacement"] = \ - self.configuration["data-sets"][data_set] + # Parse elements: + if element[u"type"] == u"table": - except KeyError: - raise PresentationError("Wrong data set used in {0}.". - format(element.get("title", ""))) - - self._specification["tables"].append(element) + logging.info(f" {count:3d} Processing a table ...") + self._parse_elements_tables(element) count += 1 - elif element["type"] == "plot": - logging.info(" {:3d} Processing a plot ...".format(count)) + elif element[u"type"] == u"plot": - # Add layout to the plots: - layout = element["layout"].get("layout", None) - if layout is not None: - element["layout"].pop("layout") - try: - for key, val in (self.configuration["plot-layouts"] - [layout].items()): - element["layout"][key] = val - except KeyError: - raise PresentationError("Layout {0} is not defined in " - "the configuration section.". - format(layout)) - self._specification["plots"].append(element) + logging.info(f" {count:3d} Processing a plot ...") + self._parse_elements_plots(element) count += 1 - elif element["type"] == "file": - logging.info(" {:3d} Processing a file ...".format(count)) - try: - element["dir-tables"] = self._replace_tags( - element["dir-tables"], - self._specification["environment"]["paths"]) - except KeyError: - pass - self._specification["files"].append(element) + elif element[u"type"] == u"file": + + logging.info(f" {count:3d} Processing a file ...") + self._parse_elements_files(element) count += 1 - elif element["type"] == "cpta": - logging.info(" {:3d} Processing Continuous Performance " - "Trending and Analysis ...".format(count)) + elif element[u"type"] == u"cpta": - for plot in element["plots"]: - # Add layout to the plots: - layout = plot.get("layout", None) - if layout is not None: - try: - plot["layout"] = \ - self.configuration["plot-layouts"][layout] - except KeyError: - raise PresentationError( - "Layout {0} is not defined in the " - "configuration section.".format(layout)) - # Add data sets: - if isinstance(plot.get("data", None), str): - data_set = plot["data"] - try: - plot["data"] = \ - self.configuration["data-sets"][data_set] - except KeyError: - raise PresentationError( - "Data set {0} is not defined in " - "the configuration section.". - format(data_set)) - self._specification["cpta"] = element + logging.info( + f" {count:3d} Processing Continuous Performance Trending " + f"and Analysis ..." + ) + self._parse_elements_cpta(element) count += 1 - logging.info("Done.") + logging.info(u"Done.") def read_specification(self): """Parse specification in the specification YAML file. @@ -729,10 +805,10 @@ class Specification: specification file. """ try: - self._cfg_yaml = load(self._cfg_file) + self._cfg_yaml = load(self._cfg_file, Loader=FullLoader) except YAMLError as err: - raise PresentationError(msg="An error occurred while parsing the " - "specification file.", + raise PresentationError(msg=u"An error occurred while parsing the " + u"specification file.", details=str(err)) self._parse_env() @@ -742,5 +818,4 @@ class Specification: self._parse_static() self._parse_elements() - logging.debug("Specification: \n{}". - format(pformat(self._specification))) + logging.debug(f"Specification: \n{pformat(self._specification)}") diff --git a/resources/tools/presentation/static_content.py b/resources/tools/presentation/static_content.py index a02330c15f..0fb81d1cbe 100644 --- a/resources/tools/presentation/static_content.py +++ b/resources/tools/presentation/static_content.py @@ -22,7 +22,7 @@ from os import makedirs from os.path import isdir from shutil import rmtree, copytree, Error -from errors import PresentationError +from pal_errors import PresentationError def prepare_static_content(spec): @@ -31,19 +31,19 @@ def prepare_static_content(spec): :param spec: Specification read from the specification file. :type spec: Specification :raises PresentationError: If it is not possible to process the static - content. + content. """ - src = spec.static.get("src-path", None) - dst = spec.static.get("dst-path", None) + src = spec.static.get(u"src-path", None) + dst = spec.static.get(u"dst-path", None) if src is None or dst is None: - logging.warning("No static content specified, skipping") + logging.warning(u"No static content specified, skipping") return # Copy all the static content to the build directory: - logging.info("Copying the static content ...") - logging.info(" Source: {0}".format(src)) - logging.info(" Destination: {0}".format(dst)) + logging.info(u"Copying the static content ...") + logging.info(f" Source: {src}") + logging.info(f" Destination: {dst}") try: if isdir(dst): @@ -51,10 +51,12 @@ def prepare_static_content(spec): copytree(src, dst) - makedirs(spec.environment["paths"]["DIR[WORKING,SRC,STATIC]"]) + makedirs(spec.environment[u"paths"][u"DIR[WORKING,SRC,STATIC]"]) except (Error, OSError) as err: - raise PresentationError("Not possible to process the static content.", - str(err)) + raise PresentationError( + u"Not possible to process the static content.", + repr(err) + ) - logging.info("Done.") + logging.info(u"Done.") |