aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--resources/libraries/python/TrafficGenerator.py103
-rw-r--r--resources/libraries/robot/performance/performance_utils.robot31
-rwxr-xr-xresources/tools/trex/trex_stateless_profile.py210
3 files changed, 326 insertions, 18 deletions
diff --git a/resources/libraries/python/TrafficGenerator.py b/resources/libraries/python/TrafficGenerator.py
index 36a83851cb..b5558e663d 100644
--- a/resources/libraries/python/TrafficGenerator.py
+++ b/resources/libraries/python/TrafficGenerator.py
@@ -457,6 +457,85 @@ class TrafficGenerator(AbstractMeasurer):
self._latency.append(self._result.split(', ')[4].split('=')[1])
self._latency.append(self._result.split(', ')[5].split('=')[1])
+ def trex_stl_start_unidirection(self, duration, rate, framesize,
+ traffic_type, tx_port=0, rx_port=1,
+ async_call=False, latency=False,
+ warmup_time=5.0):
+ """Execute script on remote node over ssh to start unidirection traffic.
+ The purpose of this function is to support performance test that need to
+ measure unidirectional traffic, e.g. Load balancer maglev mode and l3dsr
+ mode test.
+
+ :param duration: Time expresed in seconds for how long to send traffic.
+ :param rate: Traffic rate expressed with units (pps, %)
+ :param framesize: L2 frame size to send (without padding and IPG).
+ :param traffic_type: Module name as a traffic type identifier.
+ See resources/traffic_profiles/trex for implemented modules.
+ :param tx_port: Traffic generator transmit port.
+ :param rx_port: Traffic generator receive port.
+ :param latency: With latency measurement.
+ :param async_call: If enabled then don't wait for all incomming trafic.
+ :param warmup_time: Warmup time period.
+ :type duration: float
+ :type rate: str
+ :type framesize: str
+ :type traffic_type: str
+ :type tx_port: integer
+ :type rx_port: integer
+ :type latency: bool
+ :type async_call: bool
+ :type warmup_time: float
+ :returns: Nothing
+ :raises RuntimeError: In case of TG driver issue.
+ """
+ ssh = SSH()
+ ssh.connect(self._node)
+
+ _latency = "--latency" if latency else ""
+ _async = "--async" if async_call else ""
+
+ profile_path = ("{0}/resources/traffic_profiles/trex/"
+ "{1}.py".format(Constants.REMOTE_FW_DIR,
+ traffic_type))
+ (ret, stdout, _) = ssh.exec_command(
+ "sh -c "
+ "'{0}/resources/tools/trex/trex_stateless_profile.py "
+ "--profile {1} "
+ "--duration {2} "
+ "--frame_size {3} "
+ "--rate {4} "
+ "--warmup_time {5} "
+ "--port_0 {6} "
+ "--port_1 {7} "
+ "{8} " # --async
+ "{9} " # --latency
+ "{10}'". # --unidirection
+ format(Constants.REMOTE_FW_DIR, profile_path, duration, framesize,
+ rate, warmup_time, tx_port, rx_port, _async, _latency,
+ "--unidirection"),
+ timeout=float(duration) + 60)
+
+ if int(ret) != 0:
+ raise RuntimeError('TRex unidirection runtime error')
+ elif async_call:
+ #no result
+ self._received = None
+ self._sent = None
+ self._loss = None
+ self._latency = None
+ else:
+ # last line from console output
+ line = stdout.splitlines()[-1]
+
+ self._result = line
+ logger.info('TrafficGen result: {0}'.format(self._result))
+
+ self._received = self._result.split(', ')[1].split('=')[1]
+ self._sent = self._result.split(', ')[2].split('=')[1]
+ self._loss = self._result.split(', ')[3].split('=')[1]
+ self._latency = []
+ self._latency.append(self._result.split(', ')[4].split('=')[1])
+
def stop_traffic_on_tg(self):
"""Stop all traffic on TG.
@@ -468,9 +547,9 @@ class TrafficGenerator(AbstractMeasurer):
if self._node['subtype'] == NodeSubTypeTG.TREX:
self.trex_stl_stop_remote_exec(self._node)
- def send_traffic_on_tg(self, duration, rate, framesize,
- traffic_type, warmup_time=5, async_call=False,
- latency=True):
+ def send_traffic_on_tg(self, duration, rate, framesize, traffic_type,
+ unidirection=False, tx_port=0, rx_port=1,
+ warmup_time=5, async_call=False, latency=True):
"""Send traffic from all configured interfaces on TG.
:param duration: Duration of test traffic generation in seconds.
@@ -478,6 +557,9 @@ class TrafficGenerator(AbstractMeasurer):
:param framesize: Frame size (L2) in Bytes.
:param traffic_type: Module name as a traffic type identifier.
See resources/traffic_profiles/trex for implemented modules.
+ :param unidirection: Traffic is unidirectional.
+ :param tx_port: Traffic generator transmit port.
+ :param rx_port: Traffic generator receive port.
:param warmup_time: Warmup phase in seconds.
:param async_call: Async mode.
:param latency: With latency measurement.
@@ -485,6 +567,9 @@ class TrafficGenerator(AbstractMeasurer):
:type rate: str
:type framesize: str
:type traffic_type: str
+ :type unidirection: bool
+ :type tx_port: integer
+ :type rx_port: integer
:type warmup_time: float
:type async_call: bool
:type latency: bool
@@ -505,9 +590,15 @@ class TrafficGenerator(AbstractMeasurer):
if node['subtype'] is None:
raise RuntimeError('TG subtype not defined')
elif node['subtype'] == NodeSubTypeTG.TREX:
- self.trex_stl_start_remote_exec(duration, rate, framesize,
- traffic_type, async_call, latency,
- warmup_time=warmup_time)
+ if unidirection:
+ self.trex_stl_start_unidirection(duration, rate, framesize,
+ traffic_type, tx_port,
+ rx_port, async_call, latency,
+ warmup_time)
+ else:
+ self.trex_stl_start_remote_exec(duration, rate, framesize,
+ traffic_type, async_call,
+ latency, warmup_time)
else:
raise NotImplementedError("TG subtype not supported")
diff --git a/resources/libraries/robot/performance/performance_utils.robot b/resources/libraries/robot/performance/performance_utils.robot
index 4033442833..66fd307451 100644
--- a/resources/libraries/robot/performance/performance_utils.robot
+++ b/resources/libraries/robot/performance/performance_utils.robot
@@ -391,6 +391,9 @@
| | ... | - rate - Rate for sending packets. Type: string
| | ... | - framesize - L2 Frame Size [B] or IMIX_v4_1. Type: integer/string
| | ... | - topology_type - Topology type. Type: string
+| | ... | - unidirection - False if traffic is bidirectional. Type: boolean
+| | ... | - tx_port - TX port of TG, default 0. Type: integer
+| | ... | - rx_port - RX port of TG, default 1. Type: integer
| | ... | - subsamples - How many trials in this measurement. Type:int
| | ... | - trial_duration - Duration of single trial [s]. Type: float
| | ... | - fail_no_traffic - Whether to fail on zero receive count. Type: boolean
@@ -398,14 +401,17 @@
| | ... | *Example:*
| | ...
| | ... | \| Traffic should pass with maximum rate \| 4.0mpps \| 64 \
-| | ... | \| 3-node-IPv4 \| ${1} \| ${10.0} | ${False} \|
+| | ... | \| 3-node-IPv4 \| ${False} \| ${0} | ${1} \|
+| | ... | \| ${1} \| ${10.0} \| ${False} \|
| | ...
| | [Arguments] | ${rate} | ${framesize} | ${topology_type}
+| | ... | ${unidirection}=${False} | ${tx_port}=${0} | ${rx_port}=${1}
| | ... | ${trial_duration}=${perf_trial_duration} | ${fail_no_traffic}=${True}
| | ... | ${subsamples}=${perf_trial_multiplicity}
| | ...
| | ${results} = | Send traffic at specified rate | ${trial_duration} | ${rate}
-| | ... | ${framesize} | ${topology_type} | ${subsamples}
+| | ... | ${framesize} | ${topology_type} | ${unidirection}
+| | ... | ${tx_port} | ${rx_port} | ${subsamples}
| | Set Test Message | ${\n}Maximum Receive Rate trial results
| | Set Test Message | in packets per second: ${results}
| | ... | append=yes
@@ -423,18 +429,23 @@
| | ... | - rate - Rate for sending packets. Type: string
| | ... | - framesize - L2 Frame Size [B]. Type: integer/string
| | ... | - topology_type - Topology type. Type: string
+| | ... | - unidirection - False if traffic is bidirectional. Type: boolean
+| | ... | - tx_port - TX port of TG, default 0. Type: integer
+| | ... | - rx_port - RX port of TG, default 1. Type: integer
| | ... | - subsamples - How many trials in this measurement. Type: int
| | ...
| | ... | *Example:*
| | ...
| | ... | \| Send traffic at specified rate \| ${1.0} \| 4.0mpps \| 64 \
-| | ... | \| 3-node-IPv4 \| ${10} \|
+| | ... | \| 3-node-IPv4 \| ${False} \| ${0} | ${1} \| ${10} \|
| | ...
| | [Arguments] | ${trial_duration} | ${rate} | ${framesize}
-| | ... | ${topology_type} | ${subsamples}=${1}
+| | ... | ${topology_type} | ${unidirection}=${False} | ${tx_port}=${0}
+| | ... | ${rx_port}=${1} | ${subsamples}=${1}
| | ...
| | Clear and show runtime counters with running traffic | ${trial_duration}
| | ... | ${rate} | ${framesize} | ${topology_type}
+| | ... | ${unidirection} | ${tx_port} | ${rx_port}
| | Run Keyword If | ${dut_stats}==${True} | Clear all counters on all DUTs
| | Run Keyword If | ${dut_stats}==${True} and ${pkt_trace}==${True}
| | ... | VPP Enable Traces On All DUTs | ${nodes}
@@ -443,7 +454,8 @@
| | ${results} = | Create List
| | :FOR | ${i} | IN RANGE | ${subsamples}
| | | Send traffic on tg | ${trial_duration} | ${rate} | ${framesize}
-| | | ... | ${topology_type} | warmup_time=0
+| | | ... | ${topology_type} | ${unidirection} | ${tx_port}
+| | | ... | ${rx_port} | warmup_time=0
| | | ${rx} = | Get Received
| | | ${rr} = | Evaluate | ${rx} / ${trial_duration}
| | | Append To List | ${results} | ${rr}
@@ -466,15 +478,20 @@
| | ... | - rate - Rate for sending packets. Type: string
| | ... | - framesize - L2 Frame Size [B] or IMIX_v4_1. Type: integer/string
| | ... | - topology_type - Topology type. Type: string
+| | ... | - unidirection - False if traffic is bidirectional. Type: boolean
+| | ... | - tx_port - TX port of TG, default 0. Type: integer
+| | ... | - rx_port - RX port of TG, default 1. Type: integer
| | ...
| | ... | *Example:*
| | ...
-| | ... | \| Traffic should pass with partial loss \| 10 \| 4.0mpps \| 64 \
-| | ... | \| 3-node-IPv4 \| 0.5 \| percentage \|
+| | ... | \| Clear and show runtime counters with running traffic \| 10 \
+| | ... | \| 4.0mpps \| 64 \| 3-node-IPv4 \| ${False} \| ${0} | ${1} \|
| | ...
| | [Arguments] | ${duration} | ${rate} | ${framesize} | ${topology_type}
+| | ... | ${unidirection}=${False} | ${tx_port}=${0} | ${rx_port}=${1}
| | ...
| | Send traffic on tg | -1 | ${rate} | ${framesize} | ${topology_type}
+| | ... | ${unidirection} | ${tx_port} | ${rx_port}
| | ... | warmup_time=0 | async_call=${True} | latency=${False}
| | Run Keyword If | ${dut_stats}==${True}
| | ... | Clear runtime counters on all DUTs | ${nodes}
diff --git a/resources/tools/trex/trex_stateless_profile.py b/resources/tools/trex/trex_stateless_profile.py
index e1e56d9ffe..de29ff505a 100755
--- a/resources/tools/trex/trex_stateless_profile.py
+++ b/resources/tools/trex/trex_stateless_profile.py
@@ -58,8 +58,8 @@ def fmt_latency(lat_min, lat_avg, lat_max):
def simple_burst(profile_file, duration, framesize, rate, warmup_time, port_0,
- port_1, latency, async_start=False):
- """Send the traffic and measure packet loss and latency.
+ port_1, latency, async_start=False, unidirection=False):
+ """Send traffic and measure packet loss and latency.
Procedure:
- reads the given traffic profile with streams,
@@ -85,6 +85,7 @@ def simple_burst(profile_file, duration, framesize, rate, warmup_time, port_0,
:param port_1: Port 1 on the traffic generator.
:param latency: With latency stats.
:param async_start: Start the traffic and exit.
+ :param unidirection: Traffic is unidirectional.
:type profile_file: str
:type framesize: int or str
:type duration: float
@@ -92,7 +93,45 @@ def simple_burst(profile_file, duration, framesize, rate, warmup_time, port_0,
:type warmup_time: float
:type port_0: int
:type port_1: int
- :type latency: boo;
+ :type latency: bool
+ :type async_start: bool
+ :type unidirection: bool
+ """
+
+ #unidirection traffic
+ if unidirection:
+ send_traffic_unidirection(profile_file, duration, framesize, rate,
+ warmup_time, port_0, port_1, latency,
+ async_start)
+ #bidirection traffic
+ else:
+ send_traffic_bidirection(profile_file, duration, framesize, rate,
+ warmup_time, port_0, port_1, latency,
+ async_start)
+
+
+def send_traffic_bidirection(profile_file, duration, framesize, rate,
+ warmup_time, port_0, port_1, latency,
+ async_start=False):
+ """Send traffic bidirection and measure packet loss and latency.
+
+ :param profile_file: A python module with T-rex traffic profile.
+ :param framesize: Frame size.
+ :param duration: Duration of traffic run in seconds (-1=infinite).
+ :param rate: Traffic rate [percentage, pps, bps].
+ :param warmup_time: Traffic warm-up time in seconds, 0 = disable.
+ :param port_0: Port 0 on the traffic generator.
+ :param port_1: Port 1 on the traffic generator.
+ :param latency: With latency stats.
+ :param async_start: Start the traffic and exit.
+ :type profile_file: str
+ :type framesize: int or str
+ :type duration: float
+ :type rate: str
+ :type warmup_time: float
+ :type port_0: int
+ :type port_1: int
+ :type latency: bool
:type async_start: bool
"""
@@ -233,6 +272,161 @@ def simple_burst(profile_file, duration, framesize, rate, warmup_time, port_0,
lat_a, lat_b))
+def send_traffic_unidirection(profile_file, duration, framesize, rate,
+ warmup_time, port_0, port_1, latency,
+ async_start=False):
+ """Send traffic unidirection and measure packet loss and latency.
+
+ :param profile_file: A python module with T-rex traffic profile.
+ :param framesize: Frame size.
+ :param duration: Duration of traffic run in seconds (-1=infinite).
+ :param rate: Traffic rate [percentage, pps, bps].
+ :param warmup_time: Traffic warm-up time in seconds, 0 = disable.
+ :param port_0: Port 0 on the traffic generator.
+ :param port_1: Port 1 on the traffic generator.
+ :param latency: With latency stats.
+ :param async_start: Start the traffic and exit.
+ :type profile_file: str
+ :type framesize: int or str
+ :type duration: float
+ :type rate: str
+ :type warmup_time: float
+ :type port_0: int
+ :type port_1: int
+ :type latency: bool
+ :type async_start: bool
+ """
+
+ client = None
+ total_rcvd = 0
+ total_sent = 0
+ lost_a = 0
+ lat_a = "-1/-1/-1"
+
+ # Read the profile:
+ try:
+ print("### Profile file:\n{}".format(profile_file))
+ profile = STLProfile.load(profile_file, direction=0, port_id=0,
+ framesize=framesize)
+ streams = profile.get_streams()
+ except STLError as err:
+ print("Error while loading profile '{0}' {1}".format(profile_file, err))
+ sys.exit(1)
+
+ try:
+ # Create the client:
+ client = STLClient(verbose_level=LoggerApi.VERBOSE_QUIET)
+ # Connect to server:
+ client.connect()
+ # Prepare our ports:
+ if port_0 == port_1:
+ client.reset(ports=[port_0])
+ client.remove_all_streams(ports=[port_0])
+
+ if "macsrc" in profile_file:
+ client.set_port_attr(ports=[port_0], promiscuous=True,
+ resolve=False)
+ else:
+ client.reset(ports=[port_0, port_1])
+ client.remove_all_streams(ports=[port_0, port_1])
+
+ if "macsrc" in profile_file:
+ client.set_port_attr(ports=[port_0, port_1], promiscuous=True,
+ resolve=False)
+
+ if isinstance(framesize, int):
+ client.add_streams(streams[0], ports=[port_0])
+ elif isinstance(framesize, str):
+ client.add_streams(streams[0:3], ports=[port_0])
+ if latency:
+ try:
+ if isinstance(framesize, int):
+ client.add_streams(streams[2], ports=[port_0])
+ elif isinstance(framesize, str):
+ latency = False
+ except STLError:
+ # Disable latency if NIC does not support requested stream type
+ print("##### FAILED to add latency streams #####")
+ latency = False
+
+ # Warm-up phase:
+ if warmup_time > 0:
+ # Clear the stats before injecting:
+ client.clear_stats()
+
+ # Choose rate and start traffic:
+ client.start(ports=[port_0], mult=rate,
+ duration=warmup_time)
+
+ # Block until done:
+ client.wait_on_traffic(ports=[port_0],
+ timeout=warmup_time+30)
+
+ if client.get_warnings():
+ for warning in client.get_warnings():
+ print(warning)
+
+ # Read the stats after the test:
+ stats = client.get_stats()
+
+ print("##### Warmup statistics #####")
+ print(json.dumps(stats, indent=4, separators=(',', ': '),
+ sort_keys=True))
+
+ lost_a = stats[port_0]["opackets"] - stats[port_1]["ipackets"]
+ print("\npackets lost : {0} pkts".format(lost_a))
+
+ # Clear the stats before injecting:
+ client.clear_stats()
+ lost_a = 0
+
+ # Choose rate and start traffic:
+ client.start(ports=[port_0], mult=rate, duration=duration)
+
+ if not async_start:
+ # Block until done:
+ client.wait_on_traffic(ports=[port_0], timeout=duration+30)
+
+ if client.get_warnings():
+ for warning in client.get_warnings():
+ print(warning)
+
+ # Read the stats after the test
+ stats = client.get_stats()
+
+ print("##### Statistics #####")
+ print(json.dumps(stats, indent=4, separators=(',', ': '),
+ sort_keys=True))
+
+ lost_a = stats[port_0]["opackets"] - stats[port_1]["ipackets"]
+
+ if latency:
+ lat_a = fmt_latency(
+ str(stats["latency"][0]["latency"]["total_min"]),
+ str(stats["latency"][0]["latency"]["average"]),
+ str(stats["latency"][0]["latency"]["total_max"]))
+
+ total_sent = stats[port_0]["opackets"]
+ total_rcvd = stats[port_1]["ipackets"]
+
+ print("\npackets lost : {0} pkts".format(lost_a))
+
+ except STLError as err:
+ sys.stderr.write("{0}\n".format(err))
+ sys.exit(1)
+
+ finally:
+ if async_start:
+ if client:
+ client.disconnect(stop_traffic=False, release_ports=True)
+ else:
+ if client:
+ client.disconnect()
+ print("rate={0}, totalReceived={1}, totalSent={2}, "
+ "frameLoss={3}, latencyStream0(usec)={4}".
+ format(rate, total_rcvd, total_sent, lost_a, lat_a))
+
+
def main():
"""Main function for the traffic generator using T-rex.
@@ -274,7 +468,12 @@ def main():
parser.add_argument("--latency",
action="store_true",
default=False,
- help="Add latency stream")
+ help="Add latency stream.")
+ parser.add_argument("--unidirection",
+ action="store_true",
+ default=False,
+ help="Send unidirection traffic.")
+
args = parser.parse_args()
try:
@@ -290,7 +489,8 @@ def main():
port_0=args.port_0,
port_1=args.port_1,
latency=args.latency,
- async_start=args.async)
+ async_start=args.async,
+ unidirection=args.unidirection)
if __name__ == '__main__':