aboutsummaryrefslogtreecommitdiffstats
path: root/resources/tools/presentation
diff options
context:
space:
mode:
authorTibor Frank <tifrank@cisco.com>2019-10-02 13:07:19 +0200
committerTibor Frank <tifrank@cisco.com>2019-10-28 09:16:30 +0100
commit465b9ba07c47e1957ac97ec44984bf682422448c (patch)
treef1d072ec6eef99eda2b1495e8adf0a6be684f32f /resources/tools/presentation
parenteaf3cde2f590ba2da28b415c5f7463af8827b1d4 (diff)
PAL: Add sortable html table for comparisons
Change-Id: I2d690685c5728e5f36b09d542e0f6d1a08cd151a Signed-off-by: Tibor Frank <tifrank@cisco.com>
Diffstat (limited to 'resources/tools/presentation')
-rw-r--r--resources/tools/presentation/generator_tables.py151
1 files changed, 134 insertions, 17 deletions
diff --git a/resources/tools/presentation/generator_tables.py b/resources/tools/presentation/generator_tables.py
index 2e6cd593e6..1a47e81361 100644
--- a/resources/tools/presentation/generator_tables.py
+++ b/resources/tools/presentation/generator_tables.py
@@ -19,6 +19,10 @@ import logging
import csv
import re
+import plotly.graph_objects as go
+import plotly.offline as ploff
+import pandas as pd
+
from string import replace
from collections import OrderedDict
from numpy import nan, isnan
@@ -250,6 +254,99 @@ def _tpc_sort_table(table):
return table
+def _tpc_generate_html_table(header, data, output_file_name):
+ """Generate html table from input data with simple sorting possibility.
+
+ :param header: Table header.
+ :param data: Input data to be included in the table. It is a list of lists.
+ Inner lists are rows in the table. All inner lists must be of the same
+ length. The length of these lists must be the same as the length of the
+ header.
+ :param output_file_name: The name (relative or full path) where the
+ generated html table is written.
+ :type header: list
+ :type data: list of lists
+ :type output_file_name: str
+ """
+
+ df = pd.DataFrame(data, columns=header)
+
+ df_sorted = [df.sort_values(
+ by=[key, header[0]], ascending=[True, True]
+ if key != header[0] else [False, True]) for key in header]
+ df_sorted_rev = [df.sort_values(
+ by=[key, header[0]], ascending=[False, True]
+ if key != header[0] else [True, True]) for key in header]
+ df_sorted.extend(df_sorted_rev)
+
+ fill_color = [["#d4e4f7" if idx % 2 else "#e9f1fb"
+ for idx in range(len(df))]]
+ table_header = dict(
+ values=["<b>{item}</b>".format(item=item) for item in header],
+ fill_color="#7eade7",
+ align=["left", "center"]
+ )
+
+ fig = go.Figure()
+
+ for table in df_sorted:
+ columns = [table.get(col) for col in header]
+ fig.add_trace(
+ go.Table(
+ columnwidth=[30, 10],
+ header=table_header,
+ cells=dict(
+ values=columns,
+ fill_color=fill_color,
+ align=["left", "right"]
+ )
+ )
+ )
+
+ buttons = list()
+ menu_items = ["<b>{0}</b> (ascending)".format(itm) for itm in header]
+ menu_items_rev = ["<b>{0}</b> (descending)".format(itm) for itm in header]
+ menu_items.extend(menu_items_rev)
+ for idx, hdr in enumerate(menu_items):
+ visible = [False, ] * len(menu_items)
+ visible[idx] = True
+ buttons.append(
+ dict(
+ label=hdr.replace(" [Mpps]", ""),
+ method="update",
+ args=[{"visible": visible}],
+ )
+ )
+
+ fig.update_layout(
+ updatemenus=[
+ go.layout.Updatemenu(
+ type="dropdown",
+ direction="down",
+ x=0.03,
+ xanchor="left",
+ y=1.045,
+ yanchor="top",
+ active=len(menu_items) - 1,
+ buttons=list(buttons)
+ )
+ ],
+ annotations=[
+ go.layout.Annotation(
+ text="<b>Sort by:</b>",
+ x=0,
+ xref="paper",
+ y=1.035,
+ yref="paper",
+ align="left",
+ showarrow=False
+ )
+ ]
+ )
+
+ ploff.plot(fig, show_link=False, auto_open=False, filename=output_file_name)
+
+
def table_performance_comparison(table, input_data):
"""Generate the table(s) with algorithm: table_performance_comparison
specified in the specification file.
@@ -297,6 +394,7 @@ def table_performance_comparison(table, input_data):
# Prepare data to the table:
tbl_dict = dict()
+ topo = ""
for job, builds in table["reference"]["data"].items():
topo = "2n-skx" if "2n-skx" in job else ""
for build in builds:
@@ -380,16 +478,16 @@ def table_performance_comparison(table, input_data):
continue
if tbl_dict[tst_name_mod].get("history", None) is None:
tbl_dict[tst_name_mod]["history"] = OrderedDict()
- if tbl_dict[tst_name_mod]["history"].get(item["title"],
- None) is None:
+ if tbl_dict[tst_name_mod]["history"].\
+ get(item["title"], None) is None:
tbl_dict[tst_name_mod]["history"][item["title"]] = \
list()
try:
# TODO: Re-work when NDRPDRDISC tests are not used
if table["include-tests"] == "MRR":
- tbl_dict[tst_name_mod]["history"][item["title"
- ]].append(tst_data["result"]["receive-rate"].
- avg)
+ tbl_dict[tst_name_mod]["history"][item[
+ "title"]].append(tst_data["result"][
+ "receive-rate"].avg)
elif table["include-tests"] == "PDR":
if tst_data["type"] == "PDR":
tbl_dict[tst_name_mod]["history"][
@@ -398,7 +496,7 @@ def table_performance_comparison(table, input_data):
elif tst_data["type"] == "NDRPDR":
tbl_dict[tst_name_mod]["history"][item[
"title"]].append(tst_data["throughput"][
- "PDR"]["LOWER"])
+ "PDR"]["LOWER"])
elif table["include-tests"] == "NDR":
if tst_data["type"] == "NDR":
tbl_dict[tst_name_mod]["history"][
@@ -407,7 +505,7 @@ def table_performance_comparison(table, input_data):
elif tst_data["type"] == "NDRPDR":
tbl_dict[tst_name_mod]["history"][item[
"title"]].append(tst_data["throughput"][
- "NDR"]["LOWER"])
+ "NDR"]["LOWER"])
else:
continue
except (TypeError, KeyError):
@@ -477,6 +575,10 @@ def table_performance_comparison(table, input_data):
"tests. See release notes."
])
+ # Generate html table:
+ _tpc_generate_html_table(header, tbl_lst,
+ "{0}.html".format(table["output-file"]))
+
def table_performance_comparison_nic(table, input_data):
"""Generate the table(s) with algorithm: table_performance_comparison
@@ -525,6 +627,7 @@ def table_performance_comparison_nic(table, input_data):
# Prepare data to the table:
tbl_dict = dict()
+ topo = ""
for job, builds in table["reference"]["data"].items():
topo = "2n-skx" if "2n-skx" in job else ""
for build in builds:
@@ -612,16 +715,16 @@ def table_performance_comparison_nic(table, input_data):
continue
if tbl_dict[tst_name_mod].get("history", None) is None:
tbl_dict[tst_name_mod]["history"] = OrderedDict()
- if tbl_dict[tst_name_mod]["history"].get(item["title"],
- None) is None:
+ if tbl_dict[tst_name_mod]["history"].\
+ get(item["title"], None) is None:
tbl_dict[tst_name_mod]["history"][item["title"]] = \
list()
try:
# TODO: Re-work when NDRPDRDISC tests are not used
if table["include-tests"] == "MRR":
- tbl_dict[tst_name_mod]["history"][item["title"
- ]].append(tst_data["result"]["receive-rate"].
- avg)
+ tbl_dict[tst_name_mod]["history"][item[
+ "title"]].append(tst_data["result"][
+ "receive-rate"].avg)
elif table["include-tests"] == "PDR":
if tst_data["type"] == "PDR":
tbl_dict[tst_name_mod]["history"][
@@ -630,7 +733,7 @@ def table_performance_comparison_nic(table, input_data):
elif tst_data["type"] == "NDRPDR":
tbl_dict[tst_name_mod]["history"][item[
"title"]].append(tst_data["throughput"][
- "PDR"]["LOWER"])
+ "PDR"]["LOWER"])
elif table["include-tests"] == "NDR":
if tst_data["type"] == "NDR":
tbl_dict[tst_name_mod]["history"][
@@ -639,7 +742,7 @@ def table_performance_comparison_nic(table, input_data):
elif tst_data["type"] == "NDRPDR":
tbl_dict[tst_name_mod]["history"][item[
"title"]].append(tst_data["throughput"][
- "NDR"]["LOWER"])
+ "NDR"]["LOWER"])
else:
continue
except (TypeError, KeyError):
@@ -709,6 +812,10 @@ def table_performance_comparison_nic(table, input_data):
"tests. See release notes."
])
+ # Generate html table:
+ _tpc_generate_html_table(header, tbl_lst,
+ "{0}.html".format(table["output-file"]))
+
def table_nics_comparison(table, input_data):
"""Generate the table(s) with algorithm: table_nics_comparison
@@ -819,6 +926,10 @@ def table_nics_comparison(table, input_data):
convert_csv_to_pretty_txt(csv_file, "{0}.txt".format(table["output-file"]))
+ # Generate html table:
+ _tpc_generate_html_table(header, tbl_lst,
+ "{0}.html".format(table["output-file"]))
+
def table_soak_vs_ndr(table, input_data):
"""Generate the table(s) with algorithm: table_soak_vs_ndr
@@ -942,6 +1053,10 @@ def table_soak_vs_ndr(table, input_data):
convert_csv_to_pretty_txt(csv_file, "{0}.txt".format(table["output-file"]))
+ # Generate html table:
+ _tpc_generate_html_table(header, tbl_lst,
+ "{0}.html".format(table["output-file"]))
+
def table_performance_trending_dashboard(table, input_data):
"""Generate the table(s) with algorithm:
@@ -1028,9 +1143,8 @@ def table_performance_trending_dashboard(table, input_data):
if classification_lst:
if isnan(rel_change_last) and isnan(rel_change_long):
continue
- if (isnan(last_avg) or
- isnan(rel_change_last) or
- isnan(rel_change_long)):
+ if isnan(last_avg) or isnan(rel_change_last) or \
+ isnan(rel_change_long):
continue
tbl_lst.append(
[tbl_dict[tst_name]["name"],
@@ -1417,6 +1531,9 @@ def table_failed_tests(table, input_data):
tbl_lst = list()
for tst_data in tbl_dict.values():
fails_nr = 0
+ fails_last_date = ""
+ fails_last_vpp = ""
+ fails_last_csit = ""
for val in tst_data["data"].values():
if val[0] == "FAIL":
fails_nr += 1