diff options
Diffstat (limited to 'resources/tools')
-rwxr-xr-x | resources/tools/trex/trex_stateless_profile.py | 210 |
1 files changed, 205 insertions, 5 deletions
diff --git a/resources/tools/trex/trex_stateless_profile.py b/resources/tools/trex/trex_stateless_profile.py index e1e56d9ffe..de29ff505a 100755 --- a/resources/tools/trex/trex_stateless_profile.py +++ b/resources/tools/trex/trex_stateless_profile.py @@ -58,8 +58,8 @@ def fmt_latency(lat_min, lat_avg, lat_max): def simple_burst(profile_file, duration, framesize, rate, warmup_time, port_0, - port_1, latency, async_start=False): - """Send the traffic and measure packet loss and latency. + port_1, latency, async_start=False, unidirection=False): + """Send traffic and measure packet loss and latency. Procedure: - reads the given traffic profile with streams, @@ -85,6 +85,7 @@ def simple_burst(profile_file, duration, framesize, rate, warmup_time, port_0, :param port_1: Port 1 on the traffic generator. :param latency: With latency stats. :param async_start: Start the traffic and exit. + :param unidirection: Traffic is unidirectional. :type profile_file: str :type framesize: int or str :type duration: float @@ -92,7 +93,45 @@ def simple_burst(profile_file, duration, framesize, rate, warmup_time, port_0, :type warmup_time: float :type port_0: int :type port_1: int - :type latency: boo; + :type latency: bool + :type async_start: bool + :type unidirection: bool + """ + + #unidirection traffic + if unidirection: + send_traffic_unidirection(profile_file, duration, framesize, rate, + warmup_time, port_0, port_1, latency, + async_start) + #bidirection traffic + else: + send_traffic_bidirection(profile_file, duration, framesize, rate, + warmup_time, port_0, port_1, latency, + async_start) + + +def send_traffic_bidirection(profile_file, duration, framesize, rate, + warmup_time, port_0, port_1, latency, + async_start=False): + """Send traffic bidirection and measure packet loss and latency. + + :param profile_file: A python module with T-rex traffic profile. + :param framesize: Frame size. + :param duration: Duration of traffic run in seconds (-1=infinite). + :param rate: Traffic rate [percentage, pps, bps]. + :param warmup_time: Traffic warm-up time in seconds, 0 = disable. + :param port_0: Port 0 on the traffic generator. + :param port_1: Port 1 on the traffic generator. + :param latency: With latency stats. + :param async_start: Start the traffic and exit. + :type profile_file: str + :type framesize: int or str + :type duration: float + :type rate: str + :type warmup_time: float + :type port_0: int + :type port_1: int + :type latency: bool :type async_start: bool """ @@ -233,6 +272,161 @@ def simple_burst(profile_file, duration, framesize, rate, warmup_time, port_0, lat_a, lat_b)) +def send_traffic_unidirection(profile_file, duration, framesize, rate, + warmup_time, port_0, port_1, latency, + async_start=False): + """Send traffic unidirection and measure packet loss and latency. + + :param profile_file: A python module with T-rex traffic profile. + :param framesize: Frame size. + :param duration: Duration of traffic run in seconds (-1=infinite). + :param rate: Traffic rate [percentage, pps, bps]. + :param warmup_time: Traffic warm-up time in seconds, 0 = disable. + :param port_0: Port 0 on the traffic generator. + :param port_1: Port 1 on the traffic generator. + :param latency: With latency stats. + :param async_start: Start the traffic and exit. + :type profile_file: str + :type framesize: int or str + :type duration: float + :type rate: str + :type warmup_time: float + :type port_0: int + :type port_1: int + :type latency: bool + :type async_start: bool + """ + + client = None + total_rcvd = 0 + total_sent = 0 + lost_a = 0 + lat_a = "-1/-1/-1" + + # Read the profile: + try: + print("### Profile file:\n{}".format(profile_file)) + profile = STLProfile.load(profile_file, direction=0, port_id=0, + framesize=framesize) + streams = profile.get_streams() + except STLError as err: + print("Error while loading profile '{0}' {1}".format(profile_file, err)) + sys.exit(1) + + try: + # Create the client: + client = STLClient(verbose_level=LoggerApi.VERBOSE_QUIET) + # Connect to server: + client.connect() + # Prepare our ports: + if port_0 == port_1: + client.reset(ports=[port_0]) + client.remove_all_streams(ports=[port_0]) + + if "macsrc" in profile_file: + client.set_port_attr(ports=[port_0], promiscuous=True, + resolve=False) + else: + client.reset(ports=[port_0, port_1]) + client.remove_all_streams(ports=[port_0, port_1]) + + if "macsrc" in profile_file: + client.set_port_attr(ports=[port_0, port_1], promiscuous=True, + resolve=False) + + if isinstance(framesize, int): + client.add_streams(streams[0], ports=[port_0]) + elif isinstance(framesize, str): + client.add_streams(streams[0:3], ports=[port_0]) + if latency: + try: + if isinstance(framesize, int): + client.add_streams(streams[2], ports=[port_0]) + elif isinstance(framesize, str): + latency = False + except STLError: + # Disable latency if NIC does not support requested stream type + print("##### FAILED to add latency streams #####") + latency = False + + # Warm-up phase: + if warmup_time > 0: + # Clear the stats before injecting: + client.clear_stats() + + # Choose rate and start traffic: + client.start(ports=[port_0], mult=rate, + duration=warmup_time) + + # Block until done: + client.wait_on_traffic(ports=[port_0], + timeout=warmup_time+30) + + if client.get_warnings(): + for warning in client.get_warnings(): + print(warning) + + # Read the stats after the test: + stats = client.get_stats() + + print("##### Warmup statistics #####") + print(json.dumps(stats, indent=4, separators=(',', ': '), + sort_keys=True)) + + lost_a = stats[port_0]["opackets"] - stats[port_1]["ipackets"] + print("\npackets lost : {0} pkts".format(lost_a)) + + # Clear the stats before injecting: + client.clear_stats() + lost_a = 0 + + # Choose rate and start traffic: + client.start(ports=[port_0], mult=rate, duration=duration) + + if not async_start: + # Block until done: + client.wait_on_traffic(ports=[port_0], timeout=duration+30) + + if client.get_warnings(): + for warning in client.get_warnings(): + print(warning) + + # Read the stats after the test + stats = client.get_stats() + + print("##### Statistics #####") + print(json.dumps(stats, indent=4, separators=(',', ': '), + sort_keys=True)) + + lost_a = stats[port_0]["opackets"] - stats[port_1]["ipackets"] + + if latency: + lat_a = fmt_latency( + str(stats["latency"][0]["latency"]["total_min"]), + str(stats["latency"][0]["latency"]["average"]), + str(stats["latency"][0]["latency"]["total_max"])) + + total_sent = stats[port_0]["opackets"] + total_rcvd = stats[port_1]["ipackets"] + + print("\npackets lost : {0} pkts".format(lost_a)) + + except STLError as err: + sys.stderr.write("{0}\n".format(err)) + sys.exit(1) + + finally: + if async_start: + if client: + client.disconnect(stop_traffic=False, release_ports=True) + else: + if client: + client.disconnect() + print("rate={0}, totalReceived={1}, totalSent={2}, " + "frameLoss={3}, latencyStream0(usec)={4}". + format(rate, total_rcvd, total_sent, lost_a, lat_a)) + + def main(): """Main function for the traffic generator using T-rex. @@ -274,7 +468,12 @@ def main(): parser.add_argument("--latency", action="store_true", default=False, - help="Add latency stream") + help="Add latency stream.") + parser.add_argument("--unidirection", + action="store_true", + default=False, + help="Send unidirection traffic.") + args = parser.parse_args() try: @@ -290,7 +489,8 @@ def main(): port_0=args.port_0, port_1=args.port_1, latency=args.latency, - async_start=args.async) + async_start=args.async, + unidirection=args.unidirection) if __name__ == '__main__': |