summaryrefslogtreecommitdiffstats
path: root/src
diff options
context:
space:
mode:
authorIdo Barnea <ibarnea@cisco.com>2017-03-06 12:37:10 +0200
committerIdo Barnea <ibarnea@cisco.com>2017-03-08 14:21:33 +0200
commit8808908007c3d9058aced34919935a9bafc0b072 (patch)
treeea2aa93c3ebe9fea158e293882ec94ad8477c213 /src
parentf2818b3466937a91e5f5ff696a047dab59d85baa (diff)
Got rid of get_vm_one_queue(). Separated its uses to many small attributes. - second try
Signed-off-by: Ido Barnea <ibarnea@cisco.com>
Diffstat (limited to 'src')
-rwxr-xr-xsrc/bp_sim.cpp3
-rwxr-xr-xsrc/bp_sim.h34
-rw-r--r--src/debug.cpp55
-rw-r--r--src/debug.h3
-rw-r--r--src/main_dpdk.cpp541
-rw-r--r--src/main_dpdk.h15
-rw-r--r--src/pkt_gen.h3
-rw-r--r--src/pre_test.cpp6
-rw-r--r--src/pre_test.h4
9 files changed, 335 insertions, 329 deletions
diff --git a/src/bp_sim.cpp b/src/bp_sim.cpp
index 6cd0aa19..43cfed93 100755
--- a/src/bp_sim.cpp
+++ b/src/bp_sim.cpp
@@ -54,7 +54,7 @@ uint32_t CGlobalInfo::m_nodes_pool_size = 10*1024;
CParserOption CGlobalInfo::m_options;
CGlobalMemory CGlobalInfo::m_memory_cfg;
CPlatformSocketInfo CGlobalInfo::m_socket;
-
+CGlobalInfo::queues_mode CGlobalInfo::m_q_mode = CGlobalInfo::Q_MODE_NORMAL;
@@ -761,7 +761,6 @@ void CPreviewMode::Dump(FILE *fd){
fprintf(fd," vlan mode : %d\n", get_vlan_mode());
fprintf(fd," client_cfg : %d\n", (int)get_is_client_cfg_enable() );
fprintf(fd," mbuf_cache_disable : %d\n", (int)isMbufCacheDisabled() );
- fprintf(fd," vm mode : %d\n", (int)get_vm_one_queue_enable()?1:0 );
}
void CPreviewMode::set_vlan_mode_verify(uint8_t mode) {
diff --git a/src/bp_sim.h b/src/bp_sim.h
index 8c882c51..435c054b 100755
--- a/src/bp_sim.h
+++ b/src/bp_sim.h
@@ -580,12 +580,7 @@ public:
}
}
-
-
- bool get_vm_one_queue_enable(){
- return (btGetMaskBit32(m_flags1,2,2) ? true:false);
- }
-
+ // m_flags1 - bit 2 is free
void set_no_keyboard(bool enable){
btSetMaskBit32(m_flags1,5,5,enable?1:0);
}
@@ -594,10 +589,6 @@ public:
return (btGetMaskBit32(m_flags1,5,5) ? true:false);
}
- void set_vm_one_queue_enable(bool enable){
- btSetMaskBit32(m_flags1,2,2,enable?1:0);
- }
-
/* -e */
void setClientServerFlowFlipAddr(bool enable){
btSetMaskBit32(m_flags1,3,3,enable?1:0);
@@ -1250,6 +1241,14 @@ public:
class CGlobalInfo {
public:
+ typedef enum {
+ Q_MODE_NORMAL,
+ Q_MODE_ONE_QUEUE, // One RX queue and one TX queue
+ Q_MODE_RSS,
+ Q_MODE_MANY_DROP_Q // For Mellanox
+ } queues_mode;
+
+
static void init_pools(uint32_t rx_buffers);
/* for simulation */
static void free_pools();
@@ -1336,15 +1335,20 @@ public:
static void dump_pool_as_json(Json::Value &json);
static std::string dump_pool_as_json_str(void);
-
-
+ static inline int get_queues_mode() {
+ return m_q_mode;
+ }
+ static inline void set_queues_mode(queues_mode mode) {
+ m_q_mode = mode;
+ }
+
public:
static CRteMemPool m_mem_pool[MAX_SOCKETS_SUPPORTED];
-
static uint32_t m_nodes_pool_size;
static CParserOption m_options;
static CGlobalMemory m_memory_cfg;
static CPlatformSocketInfo m_socket;
+ static queues_mode m_q_mode;
};
static inline int get_is_stateless(){
@@ -1358,7 +1362,9 @@ static inline int get_is_rx_check_mode(){
static inline bool get_is_rx_filter_enable(){
uint32_t latency_rate=CGlobalInfo::m_options.m_latency_rate;
return ( ( get_is_rx_check_mode() || CGlobalInfo::is_learn_mode() || latency_rate != 0
- || get_is_stateless()) ?true:false );
+ || get_is_stateless()) && ((CGlobalInfo::get_queues_mode() != CGlobalInfo::Q_MODE_RSS)
+ || (CGlobalInfo::get_queues_mode() != CGlobalInfo::Q_MODE_ONE_QUEUE))
+ ?true:false );
}
static inline uint16_t get_rx_check_hops() {
return (CGlobalInfo::m_options.m_rx_check_hops);
diff --git a/src/debug.cpp b/src/debug.cpp
index 4abd05a0..ccef8499 100644
--- a/src/debug.cpp
+++ b/src/debug.cpp
@@ -1,5 +1,5 @@
/*
- Copyright (c) 2016-2016 Cisco Systems, Inc.
+ Copyright (c) 2016-2017 Cisco Systems, Inc.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
@@ -63,10 +63,11 @@ const uint8_t udp_pkt[] = {
0xe7
};
-CTrexDebug::CTrexDebug(CPhyEthIF m_ports_arg[12], int max_ports) {
+CTrexDebug::CTrexDebug(CPhyEthIF m_ports_arg[12], int max_ports, uint32_t rx_q_num) {
m_test = NULL;
m_ports = m_ports_arg;
m_max_ports = max_ports;
+ m_rx_q_num = rx_q_num;
}
int CTrexDebug::rcv_send(int port, int queue_id) {
@@ -340,7 +341,7 @@ struct pkt_params test_pkts[] = {
// unit test for verifying hw queues rule configuration. Can be run by:
// for stateful: --send-debug-pkt 100 -f cap2/dns.yaml -l 1
-// for stateless: --setnd-debug-pkt 100 -i
+// for stateless: --send-debug-pkt 100 -i
int CTrexDebug::verify_hw_rules(bool recv_all) {
rte_mbuf_t *m = NULL;
CPhyEthIF * lp;
@@ -361,7 +362,7 @@ int CTrexDebug::verify_hw_rules(bool recv_all) {
} else {
switch (expected_q) {
case ZERO:
- exp_q = MAIN_DPDK_DATA_Q;
+ exp_q = MAIN_DPDK_DROP_Q;
break;
case ONE:
exp_q = MAIN_DPDK_RX_Q;
@@ -371,19 +372,19 @@ int CTrexDebug::verify_hw_rules(bool recv_all) {
exp_q = MAIN_DPDK_RX_Q;
pkt_flags |= DPF_TOS_1;
} else {
- exp_q = MAIN_DPDK_DATA_Q;
+ exp_q = MAIN_DPDK_DROP_Q;
}
break;
case STF:
if ( CGlobalInfo::m_options.is_stateless() ) {
- exp_q = MAIN_DPDK_DATA_Q;
+ exp_q = MAIN_DPDK_DROP_Q;
} else {
exp_q = MAIN_DPDK_RX_Q;
pkt_flags |= DPF_TOS_1;
}
break;
default:
- exp_q = MAIN_DPDK_DATA_Q;
+ exp_q = MAIN_DPDK_DROP_Q;
break;
}
}
@@ -394,17 +395,11 @@ int CTrexDebug::verify_hw_rules(bool recv_all) {
delay(100);
- int pkt_per_q[2];
+ int pkt_per_q[16];
memset(pkt_per_q, 0, sizeof(pkt_per_q));
// We don't know which interfaces connected where, so sum all queue 1 and all queue 0
- for (int port = 0; port < m_max_ports; port++) {
- int max_q;
- if (CGlobalInfo::m_options.preview.get_vm_one_queue_enable()) {
- max_q = 0;
- } else {
- max_q = 1;
- }
- for(int queue_id = 0; queue_id <= max_q; queue_id++) {
+ for (int port = 0; port < m_max_ports; port++) {
+ for(int queue_id = 0; queue_id <= m_rx_q_num; queue_id++) {
lp = &m_ports[port];
uint16_t cnt = lp->rx_burst(queue_id, rx_pkts, 32);
pkt_per_q[queue_id] += cnt;
@@ -440,7 +435,7 @@ int CTrexDebug::test_send(uint pkt_type) {
return verify_hw_rules(true);
}
- if (! (pkt_type >= 1 && pkt_type <= 4) && !(pkt_type >= 61 && pkt_type <= 63)) {
+ if (! (pkt_type >= 1 && pkt_type <= 5) && !(pkt_type >= 61 && pkt_type <= 63)) {
printf("Unsupported packet type %d\n", pkt_type);
printf("Supported packet types are: %d(ICMP), %d(UDP), %d(TCP) %d(9k UDP)\n", 1, 2, 3, 4);
printf(" IPv6: %d(ICMP), %d(UDP), %d(TCP)\n", 61, 62, 63);
@@ -461,7 +456,7 @@ int CTrexDebug::test_send(uint pkt_type) {
} else {
ip_ver = 4;
}
- if (pkt_type > D_PKT_TYPE_ARP) {
+ if (pkt_type > D_PKT_TYPE_RSS_TEST) {
printf("Packet type not supported\n");
exit(1);
}
@@ -472,6 +467,7 @@ int CTrexDebug::test_send(uint pkt_type) {
l4_proto = IPPROTO_ICMP;
break;
case D_PKT_TYPE_UDP:
+ case D_PKT_TYPE_RSS_TEST:
l4_proto = IPPROTO_UDP;
break;
case D_PKT_TYPE_TCP:
@@ -497,13 +493,30 @@ int CTrexDebug::test_send(uint pkt_type) {
printf("Sending packet:\n");
utl_DumpBuffer(stdout, rte_pktmbuf_mtod(d, char *), 64, 0);
- test_send_pkts(d, 0, 2, 0);
- test_send_pkts(d, 0, 1, 1);
+ if (pkt_type == D_PKT_TYPE_RSS_TEST) {
+ rte_mbuf_t *mb[50];
+ char *p_d = rte_pktmbuf_mtod(d, char *);
+ IPHeader *ip;
+ for (uint32_t src = 0; src < 50; src++) {
+ mb[src] = CGlobalInfo::pktmbuf_alloc_by_port(0, d->pkt_len);
+ char *p = rte_pktmbuf_append(mb[src], d->pkt_len);
+ assert(p);
+ memcpy(p, p_d, d->pkt_len);
+ ip = (IPHeader *) (p + 14);
+ ip->updateIpDst(src + 1);
+ ip->updateIpSrc(src + 5);
+ test_send_pkts(mb[src], 0, 1, 0);
+ }
+ } else {
+ test_send_pkts(d, 0, 2, 0);
+ test_send_pkts(d, 0, 1, 1);
+ }
+ rte_pktmbuf_free(d);
delay(1000);
int j=0;
- for (j = 0; j < 2; j++) {
+ for (j = 0; j < m_rx_q_num; j++) {
printf(" =========\n");
printf(" rx queue %d \n", j);
printf(" =========\n");
diff --git a/src/debug.h b/src/debug.h
index d178e3af..5af4eba1 100644
--- a/src/debug.h
+++ b/src/debug.h
@@ -23,6 +23,7 @@ class CTrexDebug {
uint64_t m_test_drop;
CPhyEthIF *m_ports;
uint32_t m_max_ports;
+ uint32_t m_rx_q_num;
int rcv_send(int port,int queue_id);
int rcv_send_all(int queue_id);
@@ -36,7 +37,7 @@ class CTrexDebug {
int verify_hw_rules(bool recv_all);
public:
- CTrexDebug(CPhyEthIF *m_ports_arg, int max_ports);
+ CTrexDebug(CPhyEthIF *m_ports_arg, int max_ports, uint32_t rx_q_num);
int test_send(uint pkt_type);
};
diff --git a/src/main_dpdk.cpp b/src/main_dpdk.cpp
index 1a14a07d..12317d3f 100644
--- a/src/main_dpdk.cpp
+++ b/src/main_dpdk.cpp
@@ -4,7 +4,7 @@
*/
/*
- Copyright (c) 2015-2016 Cisco Systems, Inc.
+ Copyright (c) 2015-2017 Cisco Systems, Inc.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
@@ -91,12 +91,11 @@ extern "C" {
#define BP_MAX_TX_QUEUE 16
#define BP_MASTER_AND_LATENCY 2
-#define RTE_TEST_RX_DESC_DEFAULT 64
-#define RTE_TEST_RX_LATENCY_DESC_DEFAULT (1*1024)
-#define RTE_TEST_RX_DESC_DEFAULT_MLX 8
-
-#define RTE_TEST_RX_DESC_VM_DEFAULT 512
-#define RTE_TEST_TX_DESC_VM_DEFAULT 512
+#define RX_DESC_NUM_DROP_Q 64
+#define RX_DESC_NUM_DATA_Q 1024
+#define RX_DESC_NUM_DROP_Q_MLX 8
+#define RX_DESC_NUM_DATA_Q_VM 512
+#define TX_DESC_NUM 512
typedef struct rte_mbuf * (*rte_mbuf_convert_to_one_seg_t)(struct rte_mbuf *m);
struct rte_mbuf * rte_mbuf_convert_to_one_seg(struct rte_mbuf *m);
@@ -104,16 +103,9 @@ extern "C" int rte_eth_dev_get_port_by_addr(const struct rte_pci_addr *addr, uin
void set_driver();
void reorder_dpdk_ports();
-#define RTE_TEST_TX_DESC_DEFAULT 512
-#define RTE_TEST_RX_DESC_DROP 0
-
static int max_stat_hw_id_seen = 0;
static int max_stat_hw_id_seen_payload = 0;
-static inline int get_vm_one_queue_enable(){
- return (CGlobalInfo::m_options.preview.get_vm_one_queue_enable() ?1:0);
-}
-
static inline int get_is_rx_thread_enabled() {
return ((CGlobalInfo::m_options.is_rx_enabled() || CGlobalInfo::m_options.is_stateless()) ?1:0);
}
@@ -132,9 +124,6 @@ static char global_mlx5_so_id_str[50];
static char global_image_postfix[10];
#define TREX_NAME "_t-rex-64"
-
-
-
class CTRexExtendedDriverBase {
protected:
enum {
@@ -149,6 +138,8 @@ protected:
TREX_DRV_CAP_MAC_ADDR_CHG = 0x4,
/* Mellanox driver does not work well with the DPDK port reorder we do */
TREX_DRV_CAP_NO_PORT_REORDER_POSSIBLE = 0x8,
+ // Does driver support flow control change
+ TREX_DRV_FLOW_CTRL_CHG = 0x10,
} trex_drv_cap;
public:
@@ -167,11 +158,14 @@ public:
bool drop_packets_incase_of_linkdown() {
return ((m_cap & TREX_DRV_CAP_DROP_PKTS_IF_LNK_DOWN) != 0);
}
- virtual bool supports_port_reorder() {
+ bool supports_port_reorder() {
// Since only Mellanox does not support, logic here is reveresed compared to other flags.
// Put this only if not supported.
return ((m_cap & TREX_DRV_CAP_NO_PORT_REORDER_POSSIBLE) == 0);
}
+ bool flow_ctrl_chg_supp() {
+ return ((m_cap & TREX_DRV_FLOW_CTRL_CHG) != 0);
+ }
virtual int stop_queue(CPhyEthIF * _if, uint16_t q_num);
void get_extended_stats_fixed(CPhyEthIF * _if, CPhyEthIFStats *stats, int fix_i, int fix_o);
virtual void get_extended_stats(CPhyEthIF * _if,CPhyEthIFStats *stats)=0;
@@ -190,12 +184,12 @@ public:
virtual int set_rcv_all(CPhyEthIF * _if, bool set_on)=0;
virtual TRexPortAttr * create_port_attr(uint8_t port_id) = 0;
- /* Mellanox ConnectX-4 can drop only 35MPPS per Rx queue. to workaround this issue we will create multi rx queue and enable RSS. for Queue1 we will disable RSS
- return zero for disable patch and rx queues number for enable
- */
-
- virtual uint16_t enable_rss_drop_workaround(void) {
- return 0;
+ virtual void get_dpdk_drv_params(CTrexDpdkParams &p) {
+ p.rx_data_q_num = 1;
+ p.rx_drop_q_num = 1;
+ p.rx_desc_num_data_q = RX_DESC_NUM_DATA_Q;
+ p.rx_desc_num_drop_q = RX_DESC_NUM_DROP_Q;
+ p.tx_desc_num = TX_DESC_NUM;
}
protected:
@@ -208,7 +202,7 @@ class CTRexExtendedDriverBase1G : public CTRexExtendedDriverBase {
public:
CTRexExtendedDriverBase1G(){
- m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG;
+ m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG | TREX_DRV_FLOW_CTRL_CHG;
}
TRexPortAttr * create_port_attr(uint8_t port_id) {
@@ -254,6 +248,14 @@ public:
virtual int get_min_sample_rate(void){
return ( RX_CHECK_MIX_SAMPLE_RATE_1G);
}
+ virtual void get_dpdk_drv_params(CTrexDpdkParams &p) {
+ p.rx_data_q_num = 1;
+ p.rx_drop_q_num = 0;
+ p.rx_desc_num_data_q = RX_DESC_NUM_DATA_Q_VM;
+ p.rx_desc_num_drop_q = RX_DESC_NUM_DROP_Q;
+ p.tx_desc_num = TX_DESC_NUM;
+ }
+
virtual void update_configuration(port_cfg_t * cfg);
virtual int configure_rx_filter_rules(CPhyEthIF * _if);
virtual int stop_queue(CPhyEthIF * _if, uint16_t q_num);
@@ -271,8 +273,7 @@ public:
class CTRexExtendedDriverVirtio : public CTRexExtendedDriverVirtBase {
public:
CTRexExtendedDriverVirtio() {
- /* we are working in mode that we have 1 queue for rx and one queue for tx*/
- CGlobalInfo::m_options.preview.set_vm_one_queue_enable(true);
+ CGlobalInfo::set_queues_mode(CGlobalInfo::Q_MODE_ONE_QUEUE);
m_cap = /*TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG */ 0;
}
static CTRexExtendedDriverBase * create(){
@@ -284,8 +285,7 @@ public:
class CTRexExtendedDriverVmxnet3 : public CTRexExtendedDriverVirtBase {
public:
CTRexExtendedDriverVmxnet3(){
- /* we are working in mode in which we have 1 queue for rx and one queue for tx*/
- CGlobalInfo::m_options.preview.set_vm_one_queue_enable(true);
+ CGlobalInfo::set_queues_mode(CGlobalInfo::Q_MODE_ONE_QUEUE);
m_cap = /*TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG*/0;
}
@@ -299,9 +299,8 @@ public:
class CTRexExtendedDriverI40evf : public CTRexExtendedDriverVirtBase {
public:
CTRexExtendedDriverI40evf(){
- /* we are working in mode in which we have 1 queue for rx and one queue for tx*/
- CGlobalInfo::m_options.preview.set_vm_one_queue_enable(true);
- m_cap = /*TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG */0;
+ CGlobalInfo::set_queues_mode(CGlobalInfo::Q_MODE_ONE_QUEUE);
+ m_cap = /*TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG */ TREX_DRV_FLOW_CTRL_CHG;
}
virtual void get_extended_stats(CPhyEthIF * _if, CPhyEthIFStats *stats) {
get_extended_stats_fixed(_if, stats, 4, 4);
@@ -316,9 +315,8 @@ class CTRexExtendedDriverIxgbevf : public CTRexExtendedDriverI40evf {
public:
CTRexExtendedDriverIxgbevf(){
- /* we are working in mode in which we have 1 queue for rx and one queue for tx*/
- CGlobalInfo::m_options.preview.set_vm_one_queue_enable(true);
- m_cap = /*TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG */0;
+ CGlobalInfo::set_queues_mode(CGlobalInfo::Q_MODE_ONE_QUEUE);
+ m_cap = /*TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG */ TREX_DRV_FLOW_CTRL_CHG;
}
virtual void get_extended_stats(CPhyEthIF * _if, CPhyEthIFStats *stats) {
get_extended_stats_fixed(_if, stats, 4, 4);
@@ -332,7 +330,7 @@ public:
class CTRexExtendedDriverBaseE1000 : public CTRexExtendedDriverVirtBase {
CTRexExtendedDriverBaseE1000() {
// E1000 driver is only relevant in VM in our case
- CGlobalInfo::m_options.preview.set_vm_one_queue_enable(true);
+ CGlobalInfo::set_queues_mode(CGlobalInfo::Q_MODE_ONE_QUEUE);
m_cap = /*TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG */0;
}
public:
@@ -348,7 +346,7 @@ public:
class CTRexExtendedDriverBase10G : public CTRexExtendedDriverBase {
public:
CTRexExtendedDriverBase10G(){
- m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG;
+ m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG | TREX_DRV_FLOW_CTRL_CHG;
}
TRexPortAttr * create_port_attr(uint8_t port_id) {
@@ -388,7 +386,8 @@ public:
// If we want to support more counters in case of card having less interfaces, we
// Will have to identify the number of interfaces dynamically.
m_if_per_card = 4;
- m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG | TREX_DRV_CAP_DROP_PKTS_IF_LNK_DOWN;
+ m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG | TREX_DRV_CAP_DROP_PKTS_IF_LNK_DOWN
+ | TREX_DRV_FLOW_CTRL_CHG;
}
TRexPortAttr * create_port_attr(uint8_t port_id) {
@@ -437,7 +436,7 @@ private:
class CTRexExtendedDriverBaseVIC : public CTRexExtendedDriverBase {
public:
CTRexExtendedDriverBaseVIC(){
- m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG;
+ m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG | TREX_DRV_FLOW_CTRL_CHG;
}
TRexPortAttr * create_port_attr(uint8_t port_id) {
@@ -485,7 +484,8 @@ class CTRexExtendedDriverBaseMlnx5G : public CTRexExtendedDriverBase {
public:
CTRexExtendedDriverBaseMlnx5G(){
m_cap = TREX_DRV_CAP_DROP_Q | TREX_DRV_CAP_MAC_ADDR_CHG
- | TREX_DRV_CAP_NO_PORT_REORDER_POSSIBLE;
+ | TREX_DRV_CAP_NO_PORT_REORDER_POSSIBLE | TREX_DRV_FLOW_CTRL_CHG;
+ CGlobalInfo::set_queues_mode(CGlobalInfo::Q_MODE_MANY_DROP_Q);
}
TRexPortAttr * create_port_attr(uint8_t port_id) {
@@ -503,8 +503,18 @@ public:
virtual int get_min_sample_rate(void){
return (RX_CHECK_MIX_SAMPLE_RATE);
}
+ virtual void get_dpdk_drv_params(CTrexDpdkParams &p) {
+ p.rx_data_q_num = 1;
+ /* Mellanox ConnectX-4 can drop only 35MPPS per Rx queue.
+ * to workaround this issue we will create multi rx queue and enable RSS. for Queue1 we will disable RSS
+ * return zero for disable patch and rx queues number for enable.
+ */
+ p.rx_drop_q_num = 4;
+ p.rx_desc_num_data_q = RX_DESC_NUM_DATA_Q;
+ p.rx_desc_num_drop_q = RX_DESC_NUM_DROP_Q_MLX;
+ p.tx_desc_num = TX_DESC_NUM;
+ }
virtual void update_configuration(port_cfg_t * cfg);
-
virtual int configure_rx_filter_rules(CPhyEthIF * _if);
virtual void get_extended_stats(CPhyEthIF * _if,CPhyEthIFStats *stats);
virtual void clear_extended_stats(CPhyEthIF * _if);
@@ -521,10 +531,6 @@ public:
virtual CFlowStatParser *get_flow_stat_parser();
virtual int set_rcv_all(CPhyEthIF * _if, bool set_on);
- virtual uint16_t enable_rss_drop_workaround(void) {
- return (5);
- }
-
private:
virtual void add_del_rules(enum rte_filter_op op, uint8_t port_id, uint16_t type, uint16_t ip_id, uint8_t l4_proto
, int queue);
@@ -706,7 +712,7 @@ enum { OPT_HELP,
OPT_NO_SCAPY_SERVER,
OPT_ACTIVE_FLOW,
OPT_RT,
- OPT_MLX5_SO
+ OPT_MLX5_SO
};
/* these are the argument types:
@@ -1089,7 +1095,7 @@ static int parse_options(int argc, char *argv[], CParserOption* po, bool first_t
break;
case OPT_VIRT_ONE_TX_RX_QUEUE:
- po->preview.set_vm_one_queue_enable(true);
+ CGlobalInfo::set_queues_mode(CGlobalInfo::Q_MODE_ONE_QUEUE);
break;
case OPT_PREFIX:
@@ -1145,7 +1151,9 @@ static int parse_options(int argc, char *argv[], CParserOption* po, bool first_t
}
if (po->preview.get_is_rx_check_enable() || po->is_latency_enabled() || CGlobalInfo::is_learn_mode()
- || (CGlobalInfo::m_options.m_arp_ref_per != 0) || get_vm_one_queue_enable()) {
+ || (CGlobalInfo::m_options.m_arp_ref_per != 0)
+ || CGlobalInfo::get_queues_mode() == CGlobalInfo::Q_MODE_ONE_QUEUE
+ || CGlobalInfo::get_queues_mode() == CGlobalInfo::Q_MODE_RSS) {
po->set_rx_enabled();
}
@@ -1295,8 +1303,6 @@ public:
m_port_conf.rxmode.hw_strip_crc=1;
}
-
-
inline void update_var(void){
get_ex_drv()->update_configuration(this);
}
@@ -1305,40 +1311,6 @@ public:
get_ex_drv()->update_global_config_fdir(this);
}
- /* enable FDIR */
- inline void update_global_config_fdir_10g(void){
- m_port_conf.fdir_conf.mode=RTE_FDIR_MODE_PERFECT_MAC_VLAN;
- m_port_conf.fdir_conf.pballoc=RTE_FDIR_PBALLOC_64K;
- m_port_conf.fdir_conf.status=RTE_FDIR_NO_REPORT_STATUS;
- /* Offset of flexbytes field in RX packets (in 16-bit word units). */
- /* Note: divide by 2 to convert byte offset to word offset */
- if (get_is_stateless()) {
- m_port_conf.fdir_conf.flexbytes_offset = (14+4)/2;
- /* Increment offset 4 bytes for the case where we add VLAN */
- if ( CGlobalInfo::m_options.preview.get_vlan_mode() != CPreviewMode::VLAN_MODE_NONE) {
- m_port_conf.fdir_conf.flexbytes_offset += (4/2);
- }
- } else {
- if ( CGlobalInfo::m_options.preview.get_ipv6_mode_enable() ) {
- m_port_conf.fdir_conf.flexbytes_offset = (14+6)/2;
- } else {
- m_port_conf.fdir_conf.flexbytes_offset = (14+8)/2;
- }
-
- /* Increment offset 4 bytes for the case where we add VLAN */
- if ( CGlobalInfo::m_options.preview.get_vlan_mode() != CPreviewMode::VLAN_MODE_NONE ) {
- m_port_conf.fdir_conf.flexbytes_offset += (4/2);
- }
- }
- m_port_conf.fdir_conf.drop_queue=1;
- }
-
- inline void update_global_config_fdir_40g(void){
- m_port_conf.fdir_conf.mode=RTE_FDIR_MODE_PERFECT;
- m_port_conf.fdir_conf.pballoc=RTE_FDIR_PBALLOC_64K;
- m_port_conf.fdir_conf.status=RTE_FDIR_NO_REPORT_STATUS;
- }
-
struct rte_eth_conf m_port_conf;
struct rte_eth_rxconf m_rx_conf;
struct rte_eth_rxconf m_rx_drop_conf;
@@ -1531,17 +1503,10 @@ void CPhyEthIF::configure(uint16_t nb_rx_queue,
}
}
-
/*
-
- rx-queue 0 - default- all traffic not goint to queue 1
- will be drop as queue is disable
-
-
- rx-queue 1 - Latency measurement packets will go here
-
- pci_reg_write(IXGBE_L34T_IMIR(0),(1<<21));
-
+ rx-queue 0 is the default queue. All traffic not going to queue 1
+ will be dropped as queue 0 is disabled
+ rx-queue 1 - Latency measurement packets and other features that need software processing will go here.
*/
void CPhyEthIF::configure_rx_duplicate_rules(){
if ( get_is_rx_filter_enable() ){
@@ -1551,7 +1516,8 @@ void CPhyEthIF::configure_rx_duplicate_rules(){
void CPhyEthIF::stop_rx_drop_queue() {
// In debug mode, we want to see all packets. Don't want to disable any queue.
- if ( get_vm_one_queue_enable() || (CGlobalInfo::m_options.m_debug_pkt_proto != 0)) {
+ if ( (CGlobalInfo::get_queues_mode() != CGlobalInfo::Q_MODE_NORMAL)
+ || (CGlobalInfo::m_options.m_debug_pkt_proto != 0)) {
return;
}
if ( CGlobalInfo::m_options.is_rx_enabled() ) {
@@ -1560,7 +1526,9 @@ void CPhyEthIF::stop_rx_drop_queue() {
exit(1);
}
}
- get_ex_drv()->stop_queue(this, MAIN_DPDK_DATA_Q);
+ // OK to only stop MAIN_DPDK_DROP_Q here. The only driver in which there are
+ // more than 1 drop q is Mellanox. stop_queue does not work in this case anyway.
+ get_ex_drv()->stop_queue(this, MAIN_DPDK_DROP_Q);
}
@@ -2116,26 +2084,26 @@ class CCoreEthIFStateless : public CCoreEthIF {
public:
virtual int send_node_flow_stat(rte_mbuf *m, CGenNodeStateless * node_sl, CCorePerPort * lp_port
, CVirtualIFPerSideStats * lp_stats, bool is_const);
-
+
/**
* fast path version
*/
virtual int send_node(CGenNode *node);
-
+
/**
* slow path version
*/
virtual int send_node_service_mode(CGenNode *node);
-
+
protected:
template <bool SERVICE_MODE> inline int send_node_common(CGenNode *no);
-
+
inline rte_mbuf_t * generate_node_pkt(CGenNodeStateless *node_sl) __attribute__ ((always_inline));
inline int send_node_packet(CGenNodeStateless *node_sl,
rte_mbuf_t *m,
CCorePerPort *lp_port,
CVirtualIFPerSideStats *lp_stats) __attribute__ ((always_inline));
-
+
rte_mbuf_t * generate_slow_path_node_pkt(CGenNodeStateless *node_sl);
};
@@ -2277,7 +2245,7 @@ int CCoreEthIF::send_pkt(CCorePerPort * lp_port,
uint16_t len = lp_port->m_len;
lp_port->m_table[len]=m;
len++;
-
+
/* enough pkts to be sent */
if (unlikely(len == MAX_PKT_BURST)) {
send_burst(lp_port, MAX_PKT_BURST,lp_stats);
@@ -2382,10 +2350,10 @@ CCoreEthIFStateless::generate_node_pkt(CGenNodeStateless *node_sl) {
if (unlikely(node_sl->get_is_slow_path())) {
return generate_slow_path_node_pkt(node_sl);
}
-
+
/* check that we have mbuf */
rte_mbuf_t *m;
-
+
if ( likely(node_sl->is_cache_mbuf_array()) ) {
m = node_sl->cache_mbuf_array_get_cur();
rte_pktmbuf_refcnt_update(m,1);
@@ -2409,7 +2377,7 @@ CCoreEthIFStateless::send_node_packet(CGenNodeStateless *node_sl,
rte_mbuf_t *m,
CCorePerPort *lp_port,
CVirtualIFPerSideStats *lp_stats) {
-
+
if (unlikely(node_sl->is_stat_needed())) {
if ( unlikely(node_sl->is_cache_mbuf_array()) ) {
// No support for latency + cache. If user asks for cache on latency stream, we change cache to 0.
@@ -2431,26 +2399,26 @@ int CCoreEthIFStateless::send_node_service_mode(CGenNode *node) {
}
/**
- * this is the common function and it is templated
- * for two compiler evaluation for performance
- *
+ * this is the common function and it is templated
+ * for two compiler evaluation for performance
+ *
*/
template <bool SERVICE_MODE>
int CCoreEthIFStateless::send_node_common(CGenNode *node) {
CGenNodeStateless * node_sl = (CGenNodeStateless *) node;
-
+
pkt_dir_t dir = (pkt_dir_t)node_sl->get_mbuf_cache_dir();
CCorePerPort *lp_port = &m_ports[dir];
CVirtualIFPerSideStats *lp_stats = &m_stats[dir];
-
+
/* generate packet (can never fail) */
rte_mbuf_t *m = generate_node_pkt(node_sl);
-
+
/* template boolean - this will be removed at compile time */
if (SERVICE_MODE) {
TrexStatelessCaptureMngr::getInstance().handle_pkt_tx(m, lp_port->m_port->get_port_id());
}
-
+
/* send */
return send_node_packet(node_sl, m, lp_port, lp_stats);
}
@@ -2651,10 +2619,10 @@ public:
m_tx_queue_id=tx_queue;
m_rx_queue_id=rx_queue;
}
-
+
virtual int tx(rte_mbuf_t *m) {
rte_mbuf_t *tx_pkts[2];
-
+
tx_pkts[0] = m;
uint8_t vlan_mode = CGlobalInfo::m_options.preview.get_vlan_mode();
if ( likely( vlan_mode != CPreviewMode::VLAN_MODE_NONE) ) {
@@ -2681,13 +2649,13 @@ public:
return (0);
}
-
-
+
+
/* nothing special with HW implementation */
virtual int tx_latency(rte_mbuf_t *m) {
return tx(m);
}
-
+
virtual rte_mbuf_t * rx(){
rte_mbuf_t * rx_pkts[1];
uint16_t cnt=m_port->rx_burst(m_rx_queue_id,rx_pkts,1);
@@ -2698,7 +2666,7 @@ public:
}
}
-
+
virtual uint16_t rx_burst(struct rte_mbuf **rx_pkts,
uint16_t nb_pkts){
uint16_t cnt=m_port->rx_burst(m_rx_queue_id,rx_pkts,nb_pkts);
@@ -2719,18 +2687,18 @@ public:
CNodeRing *ring,
CLatencyManager *mgr,
CPhyEthIF *p) {
-
+
m_dir = (port_index % 2);
m_ring_to_dp = ring;
m_mgr = mgr;
m_port = p;
}
-
+
virtual int tx(rte_mbuf_t *m) {
return tx_common(m, false);
}
-
+
virtual int tx_latency(rte_mbuf_t *m) {
return tx_common(m, true);
}
@@ -2752,7 +2720,7 @@ public:
private:
virtual int tx_common(rte_mbuf_t *m, bool fix_timestamp) {
-
+
uint8_t vlan_mode = CGlobalInfo::m_options.preview.get_vlan_mode();
if ( likely( vlan_mode != CPreviewMode::VLAN_MODE_NONE) ) {
@@ -2769,7 +2737,7 @@ private:
if (!node) {
return (-1);
}
-
+
node->m_msg_type = CGenNodeMsgBase::LATENCY_PKT;
node->m_dir = m_dir;
node->m_pkt = m;
@@ -2780,14 +2748,14 @@ private:
} else {
node->m_update_ts = 0;
}
-
+
if ( m_ring_to_dp->Enqueue((CGenNode*)node) != 0 ){
return (-1);
}
-
+
return (0);
}
-
+
CPhyEthIF * m_port;
uint8_t m_dir;
CNodeRing * m_ring_to_dp; /* ring dp -> latency thread */
@@ -3183,7 +3151,6 @@ public:
SHUTDOWN_RPC_REQ
} shutdown_rc_e;
-
CGlobalTRex (){
m_max_ports=4;
m_max_cores=1;
@@ -3204,7 +3171,7 @@ public:
int queues_prob_init();
int ixgbe_start();
int ixgbe_rx_queue_flush();
- void ixgbe_configure_mg();
+ void rx_stf_conf();
void rx_sl_configure();
bool is_all_links_are_up(bool dump=false);
void pre_test();
@@ -3303,7 +3270,7 @@ public:
void dump_links_status(FILE *fd);
bool lookup_port_by_mac(const uint8_t *mac, uint8_t &port_id);
-
+
public:
port_cfg_t m_port_cfg;
uint32_t m_max_ports; /* active number of ports supported options are 2,4,8,10,12 */
@@ -3343,7 +3310,6 @@ private:
std::mutex m_cp_lock;
TrexMonitor m_monitor;
-
shutdown_rc_e m_mark_for_shutdown;
public:
@@ -3353,7 +3319,9 @@ public:
// Before starting, send gratuitous ARP on our addresses, and try to resolve dst MAC addresses.
void CGlobalTRex::pre_test() {
- CPretest pretest(m_max_ports);
+ CTrexDpdkParams dpdk_p;
+ get_ex_drv()->get_dpdk_drv_params(dpdk_p);
+ CPretest pretest(m_max_ports, dpdk_p.rx_data_q_num + dpdk_p.rx_drop_q_num);
bool resolve_needed = false;
uint8_t empty_mac[ETHER_ADDR_LEN] = {0,0,0,0,0,0};
bool need_grat_arp[TREX_MAX_PORTS];
@@ -3407,9 +3375,9 @@ void CGlobalTRex::pre_test() {
} else {
resolve_needed = false;
}
-
+
need_grat_arp[port_id] = CGlobalInfo::m_options.m_ip_cfg[port_id].get_ip() != 0;
-
+
pretest.add_ip(port_id, CGlobalInfo::m_options.m_ip_cfg[port_id].get_ip()
, CGlobalInfo::m_options.m_ip_cfg[port_id].get_vlan()
, CGlobalInfo::m_options.m_mac_addr[port_id].u.m_mac.src);
@@ -3521,13 +3489,13 @@ void CGlobalTRex::pre_test() {
}
}
-
+
/* for stateless only - set port mode */
if (get_is_stateless()) {
for (int port_id = 0; port_id < m_max_ports; port_id++) {
uint32_t src_ipv4 = CGlobalInfo::m_options.m_ip_cfg[port_id].get_ip();
uint32_t dg = CGlobalInfo::m_options.m_ip_cfg[port_id].get_def_gw();
- const uint8_t *dst_mac = CGlobalInfo::m_options.m_mac_addr[port_id].u.m_mac.dest;
+ const uint8_t *dst_mac = CGlobalInfo::m_options.m_mac_addr[port_id].u.m_mac.dest;
/* L3 mode */
if (src_ipv4 && dg) {
@@ -3544,7 +3512,7 @@ void CGlobalTRex::pre_test() {
}
}
-
+
}
/**
@@ -3669,7 +3637,7 @@ int CGlobalTRex::ixgbe_rx_queue_flush(){
// init stateful rx core
-void CGlobalTRex::ixgbe_configure_mg(void) {
+void CGlobalTRex::rx_stf_conf(void) {
int i;
CLatencyManagerCfg mg_cfg;
mg_cfg.m_max_ports = m_max_ports;
@@ -3688,7 +3656,7 @@ void CGlobalTRex::ixgbe_configure_mg(void) {
}
}
- if ( get_vm_one_queue_enable() ) {
+ if (CGlobalInfo::get_queues_mode() == CGlobalInfo::Q_MODE_ONE_QUEUE) {
/* vm mode, indirect queues */
for (i=0; i<m_max_ports; i++) {
CPhyEthIF * _if = &m_ports[i];
@@ -3724,8 +3692,8 @@ void CGlobalTRex::rx_sl_configure(void) {
rx_sl_cfg.m_max_ports = m_max_ports;
rx_sl_cfg.m_tx_cores = get_cores_tx();
-
- if ( get_vm_one_queue_enable() ) {
+
+ if (CGlobalInfo::get_queues_mode() == CGlobalInfo::Q_MODE_ONE_QUEUE) {
/* vm mode, indirect queues */
for (i=0; i < m_max_ports; i++) {
CPhyEthIF * _if = &m_ports[i];
@@ -3748,92 +3716,18 @@ void CGlobalTRex::rx_sl_configure(void) {
int CGlobalTRex::ixgbe_start(void){
int i;
-
for (i=0; i<m_max_ports; i++) {
socket_id_t socket_id = CGlobalInfo::m_socket.port_to_socket((port_id_t)i);
assert(CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_2048);
CPhyEthIF * _if=&m_ports[i];
_if->Create((uint8_t)i);
- uint16_t rx_rss = get_ex_drv()->enable_rss_drop_workaround();
-
- if ( get_vm_one_queue_enable() ) {
- m_port_cfg.m_port_conf.rxmode.max_rx_pkt_len = 2000;
- /* In VM case, there is one tx q and one rx q */
- _if->configure(1, 1, &m_port_cfg.m_port_conf);
- // Only 1 rx queue, so use it for everything
- m_rx_core_tx_q_id = 0;
- _if->set_rx_queue(0);
- // We usually have less memory in VM, so don't use the 9k pool. This means that large packets will
- // be received in chain of few mbufs
- _if->rx_queue_setup(0, RTE_TEST_RX_DESC_VM_DEFAULT, socket_id, &m_port_cfg.m_rx_conf,
- CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_2048);
- // 1 TX queue in VM case
- _if->tx_queue_setup(0, RTE_TEST_TX_DESC_VM_DEFAULT, socket_id, &m_port_cfg.m_tx_conf);
- } else {
- // 2 rx queues in normal case. More if rss enabled.
- // TX queues: 1 for each core handling the port pair + 1 for latency pkts + 1 for use by RX core
-
- uint16_t rx_queues;
-
- if (rx_rss==0) {
- rx_queues=2;
- }else{
- rx_queues=rx_rss;
- }
-
- _if->configure(rx_queues, m_cores_to_dual_ports + 2, &m_port_cfg.m_port_conf);
- m_rx_core_tx_q_id = m_cores_to_dual_ports;
-
- if ( rx_rss ) {
- int j=0;
- for (j=0;j<rx_rss; j++) {
- if (j==MAIN_DPDK_RX_Q){
- continue;
- }
- /* drop queue */
- _if->rx_queue_setup(j,
- // RTE_TEST_RX_DESC_DEFAULT_MLX,
- RTE_TEST_RX_DESC_DEFAULT,
- socket_id,
- &m_port_cfg.m_rx_conf,
- CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_2048);
-
-
- }
- }else{
- // setup RX drop queue
- _if->rx_queue_setup(MAIN_DPDK_DATA_Q,
- RTE_TEST_RX_DESC_DEFAULT,
- socket_id,
- &m_port_cfg.m_rx_conf,
- CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_2048);
- // setup RX filter queue
- _if->set_rx_queue(MAIN_DPDK_RX_Q);
- }
-
- _if->rx_queue_setup(MAIN_DPDK_RX_Q,
- RTE_TEST_RX_LATENCY_DESC_DEFAULT,
- socket_id,
- &m_port_cfg.m_rx_conf,
- CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_9k);
-
- for (int qid = 0; qid < m_max_queues_per_port; qid++) {
- _if->tx_queue_setup((uint16_t)qid,
- RTE_TEST_TX_DESC_DEFAULT ,
- socket_id,
- &m_port_cfg.m_tx_conf);
- }
- }
-
- if ( rx_rss ){
- _if->configure_rss_redirect_table(rx_rss,MAIN_DPDK_RX_Q);
- }
-
+ _if->conf_queues();
_if->stats_clear();
_if->start();
_if->configure_rx_duplicate_rules();
- if ( ! get_vm_one_queue_enable() && ! CGlobalInfo::m_options.preview.get_is_disable_flow_control_setting()
+ if ( get_ex_drv()->flow_ctrl_chg_supp()
+ && ! CGlobalInfo::m_options.preview.get_is_disable_flow_control_setting()
&& _if->get_port_attr()->is_fc_change_supported()) {
_if->disable_flow_control();
}
@@ -3866,8 +3760,8 @@ int CGlobalTRex::ixgbe_start(void){
ixgbe_rx_queue_flush();
if (! get_is_stateless()) {
- ixgbe_configure_mg();
- }
+ rx_stf_conf();
+ }
/* core 0 - control
@@ -3880,7 +3774,7 @@ int CGlobalTRex::ixgbe_start(void){
int port_offset=0;
uint8_t lat_q_id;
- if ( get_vm_one_queue_enable() ) {
+ if (CGlobalInfo::get_queues_mode() == CGlobalInfo::Q_MODE_ONE_QUEUE) {
lat_q_id = 0;
} else {
lat_q_id = get_cores_tx() / get_base_num_cores() + 1;
@@ -3981,16 +3875,12 @@ bool CGlobalTRex::Create(){
}
/* allocate the memory */
+ CTrexDpdkParams dpdk_p;
+ get_ex_drv()->get_dpdk_drv_params(dpdk_p);
- uint32_t rx_mbuf = 0 ;
-
- if ( get_vm_one_queue_enable() ) {
- rx_mbuf = (m_max_ports * RTE_TEST_RX_DESC_VM_DEFAULT);
- }else{
- rx_mbuf = (m_max_ports * (RTE_TEST_RX_LATENCY_DESC_DEFAULT+RTE_TEST_RX_DESC_DEFAULT));
- }
-
- CGlobalInfo::init_pools(rx_mbuf);
+ CGlobalInfo::init_pools(m_max_ports *
+ (dpdk_p.rx_data_q_num * dpdk_p.rx_desc_num_data_q +
+ dpdk_p.rx_drop_q_num * dpdk_p.rx_desc_num_drop_q));
ixgbe_start();
dump_config(stdout);
@@ -4010,7 +3900,7 @@ bool CGlobalTRex::Create(){
cfg.m_publisher = &m_zmq_publisher;
m_trex_stateless = new TrexStateless(cfg);
-
+
rx_sl_configure();
}
@@ -4107,10 +3997,10 @@ int CGlobalTRex::ixgbe_prob_init(void){
m_port_cfg.update_global_config_fdir();
}
- if ( get_vm_one_queue_enable() ) {
+ if (CGlobalInfo::get_queues_mode() == CGlobalInfo::Q_MODE_ONE_QUEUE) {
/* verify that we have only one thread/core per dual- interface */
if ( CGlobalInfo::m_options.preview.getCores()>1 ) {
- printf(" ERROR the number of cores should be 1 when the driver support only one tx queue and one rx queue \n");
+ printf("Error: the number of cores should be 1 when the driver support only one tx queue and one rx queue. Please use -c 1 \n");
exit(1);
}
}
@@ -4178,7 +4068,7 @@ bool CGlobalTRex::lookup_port_by_mac(const uint8_t *mac, uint8_t &port_id) {
return true;
}
}
-
+
return false;
}
@@ -4907,9 +4797,9 @@ int CGlobalTRex::run_in_master() {
int CGlobalTRex::run_in_rx_core(void){
CPreviewMode *lp = &CGlobalInfo::m_options.preview;
-
+
rte_thread_setname(pthread_self(), "TRex RX");
-
+
/* set RT mode if set */
if (lp->get_rt_prio_mode()) {
struct sched_param param;
@@ -4919,7 +4809,7 @@ int CGlobalTRex::run_in_rx_core(void){
exit(EXIT_FAILURE);
}
}
-
+
if (get_is_stateless()) {
m_sl_rx_running = true;
m_rx_sl.start();
@@ -4937,10 +4827,10 @@ int CGlobalTRex::run_in_rx_core(void){
int CGlobalTRex::run_in_core(virtual_thread_id_t virt_core_id){
std::stringstream ss;
CPreviewMode *lp = &CGlobalInfo::m_options.preview;
-
+
ss << "Trex DP core " << int(virt_core_id);
rte_thread_setname(pthread_self(), ss.str().c_str());
-
+
/* set RT mode if set */
if (lp->get_rt_prio_mode()) {
struct sched_param param;
@@ -4951,7 +4841,7 @@ int CGlobalTRex::run_in_core(virtual_thread_id_t virt_core_id){
}
}
-
+
if ( lp->getSingleCore() &&
(virt_core_id==2 ) &&
(lp-> getCores() ==1) ){
@@ -5154,64 +5044,117 @@ int CGlobalTRex::start_master_statefull() {
////////////////////////////////////////////
static CGlobalTRex g_trex;
+void CPhyEthIF::conf_queues() {
+ CTrexDpdkParams dpdk_p;
+ get_ex_drv()->get_dpdk_drv_params(dpdk_p);
+ uint16_t num_tx_q = (CGlobalInfo::get_queues_mode() == CGlobalInfo::Q_MODE_ONE_QUEUE) ?
+ 1 : g_trex.m_max_queues_per_port;
+ socket_id_t socket_id = CGlobalInfo::m_socket.port_to_socket((port_id_t)m_port_id);
+ assert(CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_2048);
-void CPhyEthIF::configure_rss_redirect_table(uint16_t numer_of_queues,
- uint16_t skip_queue){
+ configure(dpdk_p.rx_drop_q_num + dpdk_p.rx_data_q_num, num_tx_q, &g_trex.m_port_cfg.m_port_conf);
+ for (uint16_t qid = 0; qid < num_tx_q; qid++) {
+ tx_queue_setup(qid, dpdk_p.tx_desc_num , socket_id, &g_trex.m_port_cfg.m_tx_conf);
+ }
- struct rte_eth_dev_info dev_info;
-
- rte_eth_dev_info_get(m_port_id,&dev_info);
- assert(dev_info.reta_size>0);
+ switch (dpdk_p.rx_drop_q_num) {
+ case 0:
+ if (dpdk_p.rx_data_q_num == 1) {
+ // 1 rx rcv q. no drop q. VM mode.
+ // Only 1 rx queue, so use it for everything
+ g_trex.m_rx_core_tx_q_id = 0;
+ rx_queue_setup(0, dpdk_p.rx_desc_num_data_q, socket_id, &g_trex.m_port_cfg.m_rx_conf,
+ CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_2048);
+ set_rx_queue(0);
+ } else {
+ // no drop q. Many rcv queues. RSS mode.
+ // rss on all rcv queues. Do not skip any q.
+ configure_rss_redirect_table(dpdk_p.rx_data_q_num, 0xff);
+ g_trex.m_rx_core_tx_q_id = g_trex.m_cores_to_dual_ports;
+ for (int queue = 0; queue < dpdk_p.rx_data_q_num; queue++) {
+ rx_queue_setup(queue, dpdk_p.rx_desc_num_data_q, socket_id,
+ &g_trex.m_port_cfg.m_rx_conf, CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_9k);
+ }
+ }
+ break;
+ case 1:
+ // 1 drop q. 1 or more rx queues. Normal mode.
+ // rx core will use largest tx q
+ g_trex.m_rx_core_tx_q_id = g_trex.m_cores_to_dual_ports;
+ // configure drop q
+ rx_queue_setup(MAIN_DPDK_DROP_Q, dpdk_p.rx_desc_num_drop_q, socket_id, &g_trex.m_port_cfg.m_rx_conf,
+ CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_2048);
+ set_rx_queue(MAIN_DPDK_RX_Q);
+ rx_queue_setup(MAIN_DPDK_RX_Q, dpdk_p.rx_desc_num_data_q, socket_id,
+ &g_trex.m_port_cfg.m_rx_conf, CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_9k);
+ break;
+ default:
+ // Many drop queues. Mellanox mode.
+ g_trex.m_rx_core_tx_q_id = g_trex.m_cores_to_dual_ports;
+ // configure drop queues (all queues but MAIN_DPDK_RX_Q)
+ for (int j = 0; j < dpdk_p.rx_drop_q_num + 1; j++) {
+ if (j == MAIN_DPDK_RX_Q) {
+ continue;
+ }
+ rx_queue_setup(j, dpdk_p.rx_desc_num_drop_q, socket_id, &g_trex.m_port_cfg.m_rx_conf,
+ CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_2048);
+ }
+ rx_queue_setup(MAIN_DPDK_RX_Q, dpdk_p.rx_desc_num_data_q, socket_id,
+ &g_trex.m_port_cfg.m_rx_conf, CGlobalInfo::m_mem_pool[socket_id].m_mbuf_pool_9k);
+ // rss on all drop queues. Skip MAIN_DPDK_RX_Q
+ configure_rss_redirect_table(dpdk_p.rx_drop_q_num + 1, MAIN_DPDK_RX_Q);
+ break;
+ }
+}
- int reta_conf_size =
- std::max(1, dev_info.reta_size / RTE_RETA_GROUP_SIZE);
+void CPhyEthIF::configure_rss_redirect_table(uint16_t numer_of_queues, uint16_t skip_queue) {
+ struct rte_eth_dev_info dev_info;
- struct rte_eth_rss_reta_entry64 reta_conf[reta_conf_size];
+ rte_eth_dev_info_get(m_port_id,&dev_info);
+ assert(dev_info.reta_size > 0);
+ int reta_conf_size = std::max(1, dev_info.reta_size / RTE_RETA_GROUP_SIZE);
+ struct rte_eth_rss_reta_entry64 reta_conf[reta_conf_size];
rte_eth_dev_rss_reta_query(m_port_id,&reta_conf[0],dev_info.reta_size);
- int i,j;
-
- for (j=0; j<reta_conf_size; j++) {
- uint16_t skip=0;
- reta_conf[j].mask = ~0ULL;
- for (i=0; i<RTE_RETA_GROUP_SIZE; i++) {
+ for (int j = 0; j < reta_conf_size; j++) {
+ uint16_t skip = 0;
+ reta_conf[j].mask = ~0ULL;
+ for (int i = 0; i < RTE_RETA_GROUP_SIZE; i++) {
uint16_t q;
while (true) {
- q=(i+skip)%numer_of_queues;
- if (q!=skip_queue) {
+ q=(i + skip) % numer_of_queues;
+ if (q != skip_queue) {
break;
}
- skip+=1;
+ skip += 1;
}
- reta_conf[j].reta[i]=q;
+ reta_conf[j].reta[i] = q;
}
}
- assert (rte_eth_dev_rss_reta_update(m_port_id,&reta_conf[0],dev_info.reta_size) == 0);
-
- assert (rte_eth_dev_rss_reta_query(m_port_id,&reta_conf[0],dev_info.reta_size) == 0);
+ rte_eth_dev_rss_reta_update(m_port_id, &reta_conf[0], dev_info.reta_size);
+ rte_eth_dev_rss_reta_query(m_port_id, &reta_conf[0], dev_info.reta_size);
#if 0
/* verification */
- for (j=0; j<reta_conf_size; j++) {
- for (i=0; i<RTE_RETA_GROUP_SIZE; i++) {
+ for (j = 0; j < reta_conf_size; j++) {
+ for (i = 0; i<RTE_RETA_GROUP_SIZE; i++) {
printf(" R %d %d %d \n",j,i,reta_conf[j].reta[i]);
}
}
#endif
}
-
void CPhyEthIF::update_counters() {
get_ex_drv()->get_extended_stats(this, &m_stats);
CRXCoreIgnoreStat ign_stats;
-
+
if (get_is_stateless()) {
g_trex.m_rx_sl.get_ignore_stats(m_port_id, ign_stats, true);
} else {
g_trex.m_mg.get_ignore_stats(m_port_id, ign_stats, true);
}
-
+
m_stats.obytes -= ign_stats.get_tx_bytes();
m_stats.opackets -= ign_stats.get_tx_pkts();
m_ignore_stats.opackets += ign_stats.get_tx_pkts();
@@ -5459,7 +5402,7 @@ int update_global_info_from_platform_file(){
cg->m_mac_info[i].copy_src(( char *)CGlobalInfo::m_options.m_mac_addr[i].u.m_mac.src) ;
cg->m_mac_info[i].copy_dest(( char *)CGlobalInfo::m_options.m_mac_addr[i].u.m_mac.dest) ;
CGlobalInfo::m_options.m_mac_addr[i].u.m_mac.is_set = 1;
-
+
CGlobalInfo::m_options.m_ip_cfg[i].set_def_gw(cg->m_mac_info[i].get_def_gw());
CGlobalInfo::m_options.m_ip_cfg[i].set_ip(cg->m_mac_info[i].get_ip());
CGlobalInfo::m_options.m_ip_cfg[i].set_mask(cg->m_mac_info[i].get_mask());
@@ -5807,7 +5750,10 @@ int main_test(int argc , char * argv[]){
// For unit testing of HW rules and queues configuration. Just send some packets and exit.
if (CGlobalInfo::m_options.m_debug_pkt_proto != 0) {
- CTrexDebug debug = CTrexDebug(g_trex.m_ports, g_trex.m_max_ports);
+ CTrexDpdkParams dpdk_p;
+ get_ex_drv()->get_dpdk_drv_params(dpdk_p);
+ CTrexDebug debug = CTrexDebug(g_trex.m_ports, g_trex.m_max_ports
+ , dpdk_p.rx_data_q_num + dpdk_p.rx_drop_q_num);
int ret;
if (CGlobalInfo::m_options.m_debug_pkt_proto == D_PKT_TYPE_HW_TOGGLE_TEST) {
@@ -6039,7 +5985,6 @@ int CTRexExtendedDriverBase1G::wait_for_stable_link(){
}
void CTRexExtendedDriverBase1G::update_configuration(port_cfg_t * cfg){
-
cfg->m_tx_conf.tx_thresh.pthresh = TX_PTHRESH_1G;
cfg->m_tx_conf.tx_thresh.hthresh = TX_HTHRESH;
cfg->m_tx_conf.tx_thresh.wthresh = 0;
@@ -6323,8 +6268,31 @@ void CTRexExtendedDriverBase10G::clear_extended_stats(CPhyEthIF * _if){
_if->pci_reg_read(IXGBE_RXNFGPC);
}
-void CTRexExtendedDriverBase10G::update_global_config_fdir(port_cfg_t * cfg){
- cfg->update_global_config_fdir_10g();
+void CTRexExtendedDriverBase10G::update_global_config_fdir(port_cfg_t * cfg) {
+ cfg->m_port_conf.fdir_conf.mode = RTE_FDIR_MODE_PERFECT_MAC_VLAN;
+ cfg->m_port_conf.fdir_conf.pballoc = RTE_FDIR_PBALLOC_64K;
+ cfg->m_port_conf.fdir_conf.status = RTE_FDIR_NO_REPORT_STATUS;
+ /* Offset of flexbytes field in RX packets (in 16-bit word units). */
+ /* Note: divide by 2 to convert byte offset to word offset */
+ if (get_is_stateless()) {
+ cfg->m_port_conf.fdir_conf.flexbytes_offset = (14+4)/2;
+ /* Increment offset 4 bytes for the case where we add VLAN */
+ if ( CGlobalInfo::m_options.preview.get_vlan_mode() != CPreviewMode::VLAN_MODE_NONE) {
+ cfg->m_port_conf.fdir_conf.flexbytes_offset += (4/2);
+ }
+ } else {
+ if ( CGlobalInfo::m_options.preview.get_ipv6_mode_enable() ) {
+ cfg->m_port_conf.fdir_conf.flexbytes_offset = (14+6)/2;
+ } else {
+ cfg->m_port_conf.fdir_conf.flexbytes_offset = (14+8)/2;
+ }
+
+ /* Increment offset 4 bytes for the case where we add VLAN */
+ if ( CGlobalInfo::m_options.preview.get_vlan_mode() != CPreviewMode::VLAN_MODE_NONE ) {
+ cfg->m_port_conf.fdir_conf.flexbytes_offset += (4/2);
+ }
+ }
+ cfg->m_port_conf.fdir_conf.drop_queue = 1;
}
void CTRexExtendedDriverBase10G::update_configuration(port_cfg_t * cfg){
@@ -6541,7 +6509,9 @@ void CTRexExtendedDriverBase40G::update_configuration(port_cfg_t * cfg){
cfg->m_tx_conf.tx_thresh.pthresh = TX_PTHRESH;
cfg->m_tx_conf.tx_thresh.hthresh = TX_HTHRESH;
cfg->m_tx_conf.tx_thresh.wthresh = TX_WTHRESH;
- cfg->update_global_config_fdir_40g();
+ cfg->m_port_conf.fdir_conf.mode = RTE_FDIR_MODE_PERFECT;
+ cfg->m_port_conf.fdir_conf.pballoc = RTE_FDIR_PBALLOC_64K;
+ cfg->m_port_conf.fdir_conf.status = RTE_FDIR_NO_REPORT_STATUS;
}
// What is the type of the rule the respective hw_id counter counts.
@@ -6559,7 +6529,7 @@ void CTRexExtendedDriverBase40G::add_del_rules(enum rte_filter_op op, uint8_t po
, uint16_t ip_id, uint8_t l4_proto, int queue, uint16_t stat_idx) {
// We want to allow the use of X710 in "VM mode", for performance testing.
// In this mode, we don't want any hardware rules. Everything done by software.
- if ( get_vm_one_queue_enable())
+ if (CGlobalInfo::get_queues_mode() == CGlobalInfo::Q_MODE_ONE_QUEUE)
return;
int ret=rte_eth_dev_filter_supported(port_id, RTE_ETH_FILTER_FDIR);
@@ -6614,7 +6584,7 @@ void CTRexExtendedDriverBase40G::add_del_rules(enum rte_filter_op op, uint8_t po
ret = rte_eth_dev_filter_ctrl(port_id, RTE_ETH_FILTER_FDIR, op, (void*)&filter);
#if 0
- //????????? fix
+ //todo: fix
if ( ret != 0 ) {
rte_exit(EXIT_FAILURE, "rte_eth_dev_filter_ctrl: err=%d, port=%u\n",
ret, port_id);
@@ -6675,7 +6645,7 @@ int CTRexExtendedDriverBase40G::add_del_rx_flow_stat_rule(uint8_t port_id, enum
}
}
- add_del_rules(op, port_id, rte_type, 0, IP_ID_RESERVE_BASE + id, next_proto, MAIN_DPDK_DATA_Q, rule_id);
+ add_del_rules(op, port_id, rte_type, 0, IP_ID_RESERVE_BASE + id, next_proto, MAIN_DPDK_DROP_Q, rule_id);
return 0;
}
@@ -6771,12 +6741,12 @@ int CTRexExtendedDriverBase40G::get_rx_stats(CPhyEthIF * _if, uint32_t *pkts, ui
uint32_t hw_id = start - min + i;
add_del_rules( RTE_ETH_FILTER_ADD, port_id, fdir_hw_id_rule_params[hw_id].rule_type, 0
- , IP_ID_RESERVE_BASE + i, fdir_hw_id_rule_params[hw_id].l4_proto, MAIN_DPDK_DATA_Q
+ , IP_ID_RESERVE_BASE + i, fdir_hw_id_rule_params[hw_id].l4_proto, MAIN_DPDK_DROP_Q
, FDIR_TEMP_HW_ID);
delay(100);
rte_eth_fdir_stats_reset(port_id, &counter, hw_id, 1);
add_del_rules( RTE_ETH_FILTER_ADD, port_id, fdir_hw_id_rule_params[hw_id].rule_type, 0
- , IP_ID_RESERVE_BASE + i, fdir_hw_id_rule_params[hw_id].l4_proto, MAIN_DPDK_DATA_Q, hw_id);
+ , IP_ID_RESERVE_BASE + i, fdir_hw_id_rule_params[hw_id].l4_proto, MAIN_DPDK_DROP_Q, hw_id);
delay(100);
rte_eth_fdir_stats_reset(port_id, &temp_count, FDIR_TEMP_HW_ID, 1);
pkts[i] = counter + temp_count - prev_pkts[i];
@@ -6889,7 +6859,9 @@ void CTRexExtendedDriverBaseMlnx5G::update_configuration(port_cfg_t * cfg){
cfg->m_tx_conf.tx_thresh.pthresh = TX_PTHRESH;
cfg->m_tx_conf.tx_thresh.hthresh = TX_HTHRESH;
cfg->m_tx_conf.tx_thresh.wthresh = TX_WTHRESH;
- cfg->update_global_config_fdir_40g();
+ cfg->m_port_conf.fdir_conf.mode = RTE_FDIR_MODE_PERFECT;
+ cfg->m_port_conf.fdir_conf.pballoc = RTE_FDIR_PBALLOC_64K;
+ cfg->m_port_conf.fdir_conf.status = RTE_FDIR_NO_REPORT_STATUS;
/* update mask */
cfg->m_port_conf.fdir_conf.mask.ipv4_mask.proto=0xff;
cfg->m_port_conf.fdir_conf.mask.ipv4_mask.tos=0x01;
@@ -6898,8 +6870,8 @@ void CTRexExtendedDriverBaseMlnx5G::update_configuration(port_cfg_t * cfg){
/* enable RSS */
cfg->m_port_conf.rxmode.mq_mode =ETH_MQ_RX_RSS;
+ // This field does not do anything in case of mlx driver. Put it anyway in case it will be supported sometime.
cfg->m_port_conf.rx_adv_conf.rss_conf.rss_hf = ETH_RSS_IP;
-
}
/*
@@ -7142,8 +7114,8 @@ int CTRexExtendedDriverBaseVIC::configure_rx_filter_rules_statefull(CPhyEthIF *
// Because of some issue with VIC firmware, IPv6 UDP and ICMP go by default to q 1, so we
// need these rules to make them go to q 0.
// rule appply to all packets with 0 on tos lsb.
- add_del_rules(RTE_ETH_FILTER_ADD, port_id, RTE_ETH_FLOW_NONFRAG_IPV6_OTHER, 1, 6, 0, MAIN_DPDK_DATA_Q);
- add_del_rules(RTE_ETH_FILTER_ADD, port_id, RTE_ETH_FLOW_NONFRAG_IPV6_UDP, 1, 17, 0, MAIN_DPDK_DATA_Q);
+ add_del_rules(RTE_ETH_FILTER_ADD, port_id, RTE_ETH_FLOW_NONFRAG_IPV6_OTHER, 1, 6, 0, MAIN_DPDK_DROP_Q);
+ add_del_rules(RTE_ETH_FILTER_ADD, port_id, RTE_ETH_FLOW_NONFRAG_IPV6_UDP, 1, 17, 0, MAIN_DPDK_DROP_Q);
return 0;
}
@@ -7228,12 +7200,13 @@ CFlowStatParser *CTRexExtendedDriverBaseVIC::get_flow_stat_parser() {
/////////////////////////////////////////////////////////////////////////////////////
-void CTRexExtendedDriverVirtBase::update_configuration(port_cfg_t * cfg){
+void CTRexExtendedDriverVirtBase::update_configuration(port_cfg_t * cfg) {
cfg->m_tx_conf.tx_thresh.pthresh = TX_PTHRESH_1G;
cfg->m_tx_conf.tx_thresh.hthresh = TX_HTHRESH;
cfg->m_tx_conf.tx_thresh.wthresh = 0;
// must have this, otherwise the driver fail at init
cfg->m_tx_conf.txq_flags |= ETH_TXQ_FLAGS_NOXSUMS;
+ cfg->m_port_conf.rxmode.max_rx_pkt_len = 2000;
}
int CTRexExtendedDriverVirtBase::configure_rx_filter_rules(CPhyEthIF * _if){
@@ -7266,6 +7239,7 @@ void CTRexExtendedDriverBaseE1000::get_extended_stats(CPhyEthIF * _if,CPhyEthIFS
}
void CTRexExtendedDriverBaseE1000::update_configuration(port_cfg_t * cfg) {
+ CTRexExtendedDriverVirtBase::update_configuration(cfg);
// We configure hardware not to strip CRC. Then DPDK driver removes the CRC.
// If configuring "hardware" to remove CRC, due to bug in ESXI e1000 emulation, we got packets with CRC.
cfg->m_port_conf.rxmode.hw_strip_crc = 0;
@@ -7274,6 +7248,7 @@ void CTRexExtendedDriverBaseE1000::update_configuration(port_cfg_t * cfg) {
/////////////////////////////////////////////////////////// VMxnet3
void CTRexExtendedDriverVmxnet3::update_configuration(port_cfg_t * cfg){
+ CTRexExtendedDriverVirtBase::update_configuration(cfg);
cfg->m_tx_conf.tx_thresh.pthresh = TX_PTHRESH_1G;
cfg->m_tx_conf.tx_thresh.hthresh = TX_HTHRESH;
cfg->m_tx_conf.tx_thresh.wthresh = 0;
@@ -7283,6 +7258,7 @@ void CTRexExtendedDriverVmxnet3::update_configuration(port_cfg_t * cfg){
///////////////////////////////////////////////////////// VF
void CTRexExtendedDriverI40evf::update_configuration(port_cfg_t * cfg) {
+ CTRexExtendedDriverVirtBase::update_configuration(cfg);
cfg->m_tx_conf.tx_thresh.pthresh = TX_PTHRESH;
cfg->m_tx_conf.tx_thresh.hthresh = TX_HTHRESH;
cfg->m_tx_conf.tx_thresh.wthresh = TX_WTHRESH;
@@ -7576,4 +7552,3 @@ DpdkTRexPortAttrMlnx5G::set_link_up(bool up) {
void TrexDpdkPlatformApi::mark_for_shutdown() const {
g_trex.mark_for_shutdown(CGlobalTRex::SHUTDOWN_RPC_REQ);
}
-
diff --git a/src/main_dpdk.h b/src/main_dpdk.h
index 25b19471..e444ad2b 100644
--- a/src/main_dpdk.h
+++ b/src/main_dpdk.h
@@ -1,5 +1,5 @@
/*
- Copyright (c) 2015-2016 Cisco Systems, Inc.
+ Copyright (c) 2015-2017 Cisco Systems, Inc.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
@@ -22,10 +22,19 @@
#include "bp_sim.h"
enum {
- MAIN_DPDK_DATA_Q = 0,
+ MAIN_DPDK_DROP_Q = 0,
MAIN_DPDK_RX_Q = 1,
};
+class CTrexDpdkParams {
+ public:
+ uint16_t rx_data_q_num;
+ uint16_t rx_drop_q_num;
+ uint16_t rx_desc_num_data_q;
+ uint16_t rx_desc_num_drop_q;
+ uint16_t tx_desc_num;
+};
+
// These are statistics for packets we send, and do not expect to get back (Like ARP)
// We reduce them from general statistics we report (and report them separately, so we can keep the assumption
// that tx_pkts == rx_pkts and tx_bytes==rx_bytes
@@ -83,7 +92,7 @@ class CPhyEthIF {
void set_rx_queue(uint8_t rx_queue){
m_rx_queue=rx_queue;
}
-
+ void conf_queues();
void configure(uint16_t nb_rx_queue,
uint16_t nb_tx_queue,
const struct rte_eth_conf *eth_conf);
diff --git a/src/pkt_gen.h b/src/pkt_gen.h
index 8dcba624..6a6416d1 100644
--- a/src/pkt_gen.h
+++ b/src/pkt_gen.h
@@ -27,7 +27,8 @@ enum {
D_PKT_TYPE_UDP = 2,
D_PKT_TYPE_TCP = 3,
D_PKT_TYPE_ARP = 4,
- D_PKT_TYPE_9k_UDP = 5,
+ D_PKT_TYPE_RSS_TEST = 5,
+ D_PKT_TYPE_9k_UDP = 6,
D_PKT_TYPE_IPV6 = 60,
D_PKT_TYPE_HW_VERIFY = 100,
D_PKT_TYPE_HW_VERIFY_RCV_ALL = 101,
diff --git a/src/pre_test.cpp b/src/pre_test.cpp
index 7127645d..a72b7e8e 100644
--- a/src/pre_test.cpp
+++ b/src/pre_test.cpp
@@ -340,9 +340,9 @@ bool CPretest::resolve_all() {
// If we are on loopback, We might get requests on port even after it is in RESOLVE_DONE state
all_resolved = false;
}
- handle_rx(port, MAIN_DPDK_DATA_Q);
- if (! CGlobalInfo::m_options.preview.get_vm_one_queue_enable())
- handle_rx(port, MAIN_DPDK_RX_Q);
+ for (uint16_t queue = 0; queue < m_num_q; queue++) {
+ handle_rx(port, queue);
+ }
}
if (all_resolved) {
break;
diff --git a/src/pre_test.h b/src/pre_test.h
index 14b444cf..88616440 100644
--- a/src/pre_test.h
+++ b/src/pre_test.h
@@ -83,11 +83,12 @@ class CPretestOnePortInfo {
class CPretest {
public:
- CPretest(uint16_t max_ports) {
+ CPretest(uint16_t max_ports, uint16_t queues_per_port) {
m_max_ports = max_ports;
for (int i =0; i < max_ports; i++) {
m_port_info[i].set_port_id(i);
}
+ m_num_q = queues_per_port;
}
void add_ip(uint16_t port, uint32_t ip, uint16_t vlan, MacAddress src_mac);
void add_ip(uint16_t port, uint32_t ip, MacAddress src_mac);
@@ -115,6 +116,7 @@ class CPretest {
private:
CPretestOnePortInfo m_port_info[TREX_MAX_PORTS];
uint16_t m_max_ports;
+ uint16_t m_num_q;
};
#endif