diff options
author | Tibor Frank <tifrank@cisco.com> | 2023-12-19 12:22:01 +0000 |
---|---|---|
committer | Tibor Frank <tifrank@cisco.com> | 2023-12-19 12:22:01 +0000 |
commit | d68297c9bbb2986e4cdaeaa9317e69d4371b90db (patch) | |
tree | 46be91cd500f78c0cc2a3b64cfd173704348dffb /csit.infra.dash | |
parent | 1bba1bd1351376ef79cd32d8424fb8d9c1aa6640 (diff) |
C-Dash: Fix normalized values in hover in trending
Change-Id: Ib5f7c7a8dc138fec7cf7faaba24fb430666ee3c2
Signed-off-by: Tibor Frank <tifrank@cisco.com>
Diffstat (limited to 'csit.infra.dash')
-rw-r--r-- | csit.infra.dash/app/cdash/trending/graphs.py | 71 |
1 files changed, 38 insertions, 33 deletions
diff --git a/csit.infra.dash/app/cdash/trending/graphs.py b/csit.infra.dash/app/cdash/trending/graphs.py index fa60ffdd77..57fc165cb3 100644 --- a/csit.infra.dash/app/cdash/trending/graphs.py +++ b/csit.infra.dash/app/cdash/trending/graphs.py @@ -98,7 +98,7 @@ def graph_trending( name: str, df: pd.DataFrame, color: str, - norm_factor: float + nf: float ) -> list: """Generate the trending traces for the trending graph. @@ -106,13 +106,13 @@ def graph_trending( :param name: The test name to be displayed as the graph title. :param df: Data frame with test data. :param color: The color of the trace (samples and trend line). - :param norm_factor: The factor used for normalization of the results to + :param nf: The factor used for normalization of the results to CPU frequency set to Constants.NORM_FREQUENCY. :type ttype: str :type name: str :type df: pandas.DataFrame :type color: str - :type norm_factor: float + :type nf: float :returns: Traces (samples, trending line, anomalies) :rtype: list """ @@ -121,88 +121,78 @@ def graph_trending( if df.empty: return list(), list() - x_axis = df["start_time"].tolist() - if ttype == "latency": - y_data = [(v / norm_factor) for v in df[C.VALUE[ttype]].tolist()] - else: - y_data = [(v * norm_factor) for v in df[C.VALUE[ttype]].tolist()] - units = df[C.UNIT[ttype]].unique().tolist() - - try: - anomalies, trend_avg, trend_stdev = classify_anomalies( - {k: v for k, v in zip(x_axis, y_data)} - ) - except ValueError as err: - logging.error(err) - return list(), list() - hover = list() customdata = list() customdata_samples = list() name_lst = name.split("-") - for idx, (_, row) in enumerate(df.iterrows()): + for _, row in df.iterrows(): h_tput, h_band, h_lat = str(), str(), str() if ttype in ("mrr", "mrr-bandwidth"): h_tput = ( f"tput avg [{row['result_receive_rate_rate_unit']}]: " - f"{row['result_receive_rate_rate_avg']:,.0f}<br>" + f"{row['result_receive_rate_rate_avg'] * nf:,.0f}<br>" f"tput stdev [{row['result_receive_rate_rate_unit']}]: " - f"{row['result_receive_rate_rate_stdev']:,.0f}<br>" + f"{row['result_receive_rate_rate_stdev'] * nf:,.0f}<br>" ) if pd.notna(row["result_receive_rate_bandwidth_avg"]): h_band = ( f"bandwidth avg " f"[{row['result_receive_rate_bandwidth_unit']}]: " - f"{row['result_receive_rate_bandwidth_avg']:,.0f}<br>" + f"{row['result_receive_rate_bandwidth_avg'] * nf:,.0f}" + "<br>" f"bandwidth stdev " f"[{row['result_receive_rate_bandwidth_unit']}]: " - f"{row['result_receive_rate_bandwidth_stdev']:,.0f}<br>" + f"{row['result_receive_rate_bandwidth_stdev']* nf:,.0f}" + "<br>" ) elif ttype in ("ndr", "ndr-bandwidth"): h_tput = ( f"tput [{row['result_ndr_lower_rate_unit']}]: " - f"{row['result_ndr_lower_rate_value']:,.0f}<br>" + f"{row['result_ndr_lower_rate_value'] * nf:,.0f}<br>" ) if pd.notna(row["result_ndr_lower_bandwidth_value"]): h_band = ( f"bandwidth [{row['result_ndr_lower_bandwidth_unit']}]:" - f" {row['result_ndr_lower_bandwidth_value']:,.0f}<br>" + f" {row['result_ndr_lower_bandwidth_value'] * nf:,.0f}" + "<br>" ) elif ttype in ("pdr", "pdr-bandwidth", "latency"): h_tput = ( f"tput [{row['result_pdr_lower_rate_unit']}]: " - f"{row['result_pdr_lower_rate_value']:,.0f}<br>" + f"{row['result_pdr_lower_rate_value'] * nf:,.0f}<br>" ) if pd.notna(row["result_pdr_lower_bandwidth_value"]): h_band = ( f"bandwidth [{row['result_pdr_lower_bandwidth_unit']}]:" - f" {row['result_pdr_lower_bandwidth_value']:,.0f}<br>" + f" {row['result_pdr_lower_bandwidth_value'] * nf:,.0f}" + "<br>" ) if pd.notna(row["result_latency_forward_pdr_50_avg"]): h_lat = ( f"latency " f"[{row['result_latency_forward_pdr_50_unit']}]: " - f"{row['result_latency_forward_pdr_50_avg']:,.0f}<br>" + f"{row['result_latency_forward_pdr_50_avg'] / nf:,.0f}" + "<br>" ) elif ttype in ("hoststack-cps", "hoststack-rps", "hoststack-cps-bandwidth", "hoststack-rps-bandwidth", "hoststack-latency"): h_tput = ( f"tput [{row['result_rate_unit']}]: " - f"{row['result_rate_value']:,.0f}<br>" + f"{row['result_rate_value'] * nf:,.0f}<br>" ) h_band = ( f"bandwidth [{row['result_bandwidth_unit']}]: " - f"{row['result_bandwidth_value']:,.0f}<br>" + f"{row['result_bandwidth_value'] * nf:,.0f}<br>" ) h_lat = ( f"latency [{row['result_latency_unit']}]: " - f"{row['result_latency_value']:,.0f}<br>" + f"{row['result_latency_value'] / nf:,.0f}<br>" ) elif ttype in ("hoststack-bps", ): h_band = ( f"bandwidth [{row['result_bandwidth_unit']}]: " - f"{row['result_bandwidth_value']:,.0f}<br>" + f"{row['result_bandwidth_value'] * nf:,.0f}<br>" ) hover_itm = ( f"dut: {name_lst[0]}<br>" @@ -224,6 +214,21 @@ def graph_trending( ) customdata.append({"name": name}) + x_axis = df["start_time"].tolist() + if "latency" in ttype: + y_data = [(v / nf) for v in df[C.VALUE[ttype]].tolist()] + else: + y_data = [(v * nf) for v in df[C.VALUE[ttype]].tolist()] + units = df[C.UNIT[ttype]].unique().tolist() + + try: + anomalies, trend_avg, trend_stdev = classify_anomalies( + {k: v for k, v in zip(x_axis, y_data)} + ) + except ValueError as err: + logging.error(err) + return list(), list() + hover_trend = list() for avg, stdev, (_, row) in zip(trend_avg, trend_stdev, df.iterrows()): hover_itm = ( @@ -352,7 +357,7 @@ def graph_trending( if normalize: phy = itm["phy"].split("-") topo_arch = f"{phy[0]}-{phy[1]}" if len(phy) == 4 else str() - norm_factor = (C.NORM_FREQUENCY / C.FREQUENCY[topo_arch]) \ + norm_factor = (C.NORM_FREQUENCY / C.FREQUENCY.get(topo_arch, 1.0)) \ if topo_arch else 1.0 else: norm_factor = 1.0 |