aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorNeale Ranns <nranns@cisco.com>2019-03-12 02:34:07 -0700
committerNeale Ranns <nranns@cisco.com>2019-03-12 16:50:23 +0000
commit36abbf10045ba81733410f50a3f5a9463f7137d1 (patch)
tree3adcabaa69bcbe4cbe801c45866a1f4c46c12443
parent7bd343509fa30f9713b6efa9a4acb83e0ab86cfb (diff)
GBP: L3 out fixes
Change-Id: I0562d597fd45c7ddcb6db42cf17d3ffb569eb140 Signed-off-by: Neale Ranns <nranns@cisco.com>
-rw-r--r--src/plugins/gbp/gbp_classify_node.c89
-rw-r--r--src/plugins/gbp/gbp_endpoint.c13
-rw-r--r--src/plugins/gbp/gbp_endpoint.h5
-rw-r--r--src/plugins/gbp/gbp_ext_itf.c12
-rw-r--r--src/plugins/gbp/gbp_itf.c35
-rw-r--r--test/test_gbp.py136
-rw-r--r--test/vpp_l2.py29
7 files changed, 260 insertions, 59 deletions
diff --git a/src/plugins/gbp/gbp_classify_node.c b/src/plugins/gbp/gbp_classify_node.c
index 1b2cb0a2bc7..58b1f7ac545 100644
--- a/src/plugins/gbp/gbp_classify_node.c
+++ b/src/plugins/gbp/gbp_classify_node.c
@@ -308,6 +308,28 @@ ethertype_to_dpo_proto (const ethernet_header_t * eh0)
return (DPO_PROTO_NONE);
}
+/**
+ * per-packet trace data
+ */
+typedef struct gbp_lpm_classify_trace_t_
+{
+ sclass_t sclass;
+ index_t lbi;
+} gbp_lpm_classify_trace_t;
+
+/* packet trace format function */
+static u8 *
+format_gbp_lpm_classify_trace (u8 * s, va_list * args)
+{
+ CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
+ CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
+ gbp_lpm_classify_trace_t *t = va_arg (*args, gbp_lpm_classify_trace_t *);
+
+ s = format (s, "sclass:%d lb:%d", t->sclass, t->lbi);
+
+ return s;
+}
+
/*
* Determine the SRC EPG from a LPM
*/
@@ -335,8 +357,9 @@ gbp_lpm_classify_inline (vlib_main_t * vm,
{
u32 bi0, sw_if_index0, fib_index0, lbi0;
gbp_lpm_classify_next_t next0;
+ const ethernet_header_t *eh0;
const gbp_policy_dpo_t *gpd0;
- const gbp_ext_itf_t *gx0;
+ const gbp_endpoint_t *ge0;
const gbp_recirc_t *gr0;
const dpo_id_t *dpo0;
load_balance_t *lb0;
@@ -355,6 +378,8 @@ gbp_lpm_classify_inline (vlib_main_t * vm,
ip6_0 = NULL;
next0 = GPB_LPM_CLASSIFY_DROP;
+ lbi0 = ~0;
+ eh0 = NULL;
b0 = vlib_get_buffer (vm, bi0);
sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
@@ -366,8 +391,6 @@ gbp_lpm_classify_inline (vlib_main_t * vm,
ip6_0 = vlib_buffer_get_current (b0);
else if (DPO_PROTO_ETHERNET == dproto)
{
- const ethernet_header_t *eh0;
-
eh0 = vlib_buffer_get_current (b0);
dproto = ethertype_to_dpo_proto (eh0);
@@ -393,17 +416,61 @@ gbp_lpm_classify_inline (vlib_main_t * vm,
{
gr0 = gbp_recirc_get (sw_if_index0);
fib_index0 = gr0->gr_fib_index[dproto];
+ ge0 = NULL;
vnet_feature_next (&next0, b0);
}
else
{
- gx0 = gbp_ext_itf_get (sw_if_index0);
- fib_index0 = gx0->gx_fib_index[dproto];
+ if (NULL == eh0)
+ {
+ /* packet should be l2 */
+ sclass0 = SCLASS_INVALID;
+ goto trace;
+ }
+
+ ge0 = gbp_endpoint_find_mac (eh0->src_address,
+ vnet_buffer (b0)->l2.bd_index);
+
+ if (NULL == ge0)
+ {
+ /* packet must have come from an EP's mac */
+ sclass0 = SCLASS_INVALID;
+ goto trace;
+ }
+
+ fib_index0 = ge0->ge_fwd.gef_fib_index;
+
+ if (~0 == fib_index0)
+ {
+ sclass0 = SCLASS_INVALID;
+ goto trace;
+ }
+
+ if (DPO_PROTO_IP4 == dproto)
+ {
+ ge0 =
+ gbp_endpoint_find_ip4 (&ip4_0->src_address, fib_index0);
+ }
+ else if (DPO_PROTO_IP6 == dproto)
+ {
+ ge0 =
+ gbp_endpoint_find_ip6 (&ip6_0->src_address, fib_index0);
+ }
next0 = vnet_l2_feature_next
(b0, gscm->l2_input_feat_next[GBP_SRC_CLASSIFY_LPM],
L2INPUT_FEAT_GBP_LPM_CLASSIFY);
+
+ /*
+ * if we found the EP by IP lookup, it must be from the EP
+ * not a network behind it
+ */
+ if (NULL != ge0)
+ {
+ sclass0 = ge0->ge_fwd.gef_sclass;
+ goto trace;
+ }
}
if (DPO_PROTO_IP4 == dproto)
@@ -420,6 +487,7 @@ gbp_lpm_classify_inline (vlib_main_t * vm,
{
/* not IP so no LPM classify possible */
sclass0 = SCLASS_INVALID;
+ next0 = GPB_LPM_CLASSIFY_DROP;
goto trace;
}
lb0 = load_balance_get (lbi0);
@@ -434,7 +502,7 @@ gbp_lpm_classify_inline (vlib_main_t * vm,
{
/* could not classify => drop */
sclass0 = SCLASS_INVALID;
- next0 = GPB_LPM_CLASSIFY_DROP;
+ goto trace;
}
trace:
@@ -442,9 +510,10 @@ gbp_lpm_classify_inline (vlib_main_t * vm,
if (PREDICT_FALSE ((b0->flags & VLIB_BUFFER_IS_TRACED)))
{
- gbp_classify_trace_t *t =
+ gbp_lpm_classify_trace_t *t =
vlib_add_trace (vm, node, b0, sizeof (*t));
t->sclass = sclass0;
+ t->lbi = lbi0;
}
vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
@@ -483,7 +552,7 @@ VLIB_NODE_FN (gbp_l2_lpm_classify_node) (vlib_main_t * vm,
VLIB_REGISTER_NODE (gbp_ip4_lpm_classify_node) = {
.name = "ip4-gbp-lpm-classify",
.vector_size = sizeof (u32),
- .format_trace = format_gbp_classify_trace,
+ .format_trace = format_gbp_lpm_classify_trace,
.type = VLIB_NODE_TYPE_INTERNAL,
.n_errors = 0,
@@ -496,7 +565,7 @@ VLIB_REGISTER_NODE (gbp_ip4_lpm_classify_node) = {
VLIB_REGISTER_NODE (gbp_ip6_lpm_classify_node) = {
.name = "ip6-gbp-lpm-classify",
.vector_size = sizeof (u32),
- .format_trace = format_gbp_classify_trace,
+ .format_trace = format_gbp_lpm_classify_trace,
.type = VLIB_NODE_TYPE_INTERNAL,
.n_errors = 0,
@@ -509,7 +578,7 @@ VLIB_REGISTER_NODE (gbp_ip6_lpm_classify_node) = {
VLIB_REGISTER_NODE (gbp_l2_lpm_classify_node) = {
.name = "l2-gbp-lpm-classify",
.vector_size = sizeof (u32),
- .format_trace = format_gbp_classify_trace,
+ .format_trace = format_gbp_lpm_classify_trace,
.type = VLIB_NODE_TYPE_INTERNAL,
.n_errors = 0,
diff --git a/src/plugins/gbp/gbp_endpoint.c b/src/plugins/gbp/gbp_endpoint.c
index 36e0050c39d..2472199e320 100644
--- a/src/plugins/gbp/gbp_endpoint.c
+++ b/src/plugins/gbp/gbp_endpoint.c
@@ -254,7 +254,8 @@ gbp_endpoint_alloc (const ip46_address_t * ips,
fib_node_init (&ge->ge_node, gbp_endpoint_fib_type);
gei = gbp_endpoint_index (ge);
ge->ge_key.gek_gbd =
- ge->ge_key.gek_grd = ge->ge_fwd.gef_itf = INDEX_INVALID;
+ ge->ge_key.gek_grd =
+ ge->ge_fwd.gef_itf = ge->ge_fwd.gef_fib_index = INDEX_INVALID;
ge->ge_last_time = vlib_time_now (vlib_get_main ());
ge->ge_key.gek_gbd = gbp_bridge_domain_index (gbd);
@@ -602,7 +603,7 @@ gbb_endpoint_fwd_reset (gbp_endpoint_t * ge)
{
l2fib_del_entry (ge->ge_key.gek_mac.bytes,
gbd->gb_bd_index, gef->gef_itf);
- gbp_itf_set_l2_input_feature (gef->gef_itf, gei, (L2INPUT_FEAT_NONE));
+ gbp_itf_set_l2_input_feature (gef->gef_itf, gei, L2INPUT_FEAT_NONE);
gbp_itf_set_l2_output_feature (gef->gef_itf, gei, L2OUTPUT_FEAT_NONE);
gbp_itf_unlock (gef->gef_itf);
@@ -681,6 +682,7 @@ gbb_endpoint_fwd_recalc (gbp_endpoint_t * ge)
rewrite = NULL;
grd = gbp_route_domain_get (ge->ge_key.gek_grd);
fib_index = grd->grd_fib_index[pfx->fp_proto];
+ gef->gef_fib_index = fib_index;
bd_add_del_ip_mac (gbd->gb_bd_index, fib_proto_to_ip46 (pfx->fp_proto),
&pfx->fp_addr, &ge->ge_key.gek_mac, 1);
@@ -786,7 +788,12 @@ gbb_endpoint_fwd_recalc (gbp_endpoint_t * ge)
}
}
- if (gbp_endpoint_is_local (ge) && !gbp_endpoint_is_external (ge))
+ if (gbp_endpoint_is_external (ge))
+ {
+ gbp_itf_set_l2_input_feature (gef->gef_itf, gei,
+ L2INPUT_FEAT_GBP_LPM_CLASSIFY);
+ }
+ else if (gbp_endpoint_is_local (ge))
{
/*
* non-remote endpoints (i.e. those not arriving on iVXLAN
diff --git a/src/plugins/gbp/gbp_endpoint.h b/src/plugins/gbp/gbp_endpoint.h
index aba1f9d498c..a0d354ab8ab 100644
--- a/src/plugins/gbp/gbp_endpoint.h
+++ b/src/plugins/gbp/gbp_endpoint.h
@@ -168,6 +168,11 @@ typedef struct gbp_endpoint_fwd_t_
*/
sclass_t gef_sclass;
+ /**
+ * FIB index the EP is in
+ */
+ u32 gef_fib_index;
+
gbp_endpoint_flags_t gef_flags;
} gbp_endpoint_fwd_t;
diff --git a/src/plugins/gbp/gbp_ext_itf.c b/src/plugins/gbp/gbp_ext_itf.c
index 16cdaa87b77..be2d614e1fa 100644
--- a/src/plugins/gbp/gbp_ext_itf.c
+++ b/src/plugins/gbp/gbp_ext_itf.c
@@ -60,7 +60,6 @@ gbp_ext_itf_add (u32 sw_if_index, u32 bd_id, u32 rd_id)
if (INDEX_INVALID == gxi)
{
- gbp_bridge_domain_t *gb;
gbp_route_domain_t *gr;
fib_protocol_t fproto;
index_t gbi, gri;
@@ -81,11 +80,11 @@ gbp_ext_itf_add (u32 sw_if_index, u32 bd_id, u32 rd_id)
pool_get_zero (gbp_ext_itf_pool, gx);
gxi = gx - gbp_ext_itf_pool;
- gb = gbp_bridge_domain_get (gbi);
gr = gbp_route_domain_get (gri);
gx->gx_bd = gbi;
gx->gx_rd = gri;
+ gx->gx_itf = sw_if_index;
FOR_EACH_FIB_IP_PROTOCOL (fproto)
{
@@ -93,10 +92,6 @@ gbp_ext_itf_add (u32 sw_if_index, u32 bd_id, u32 rd_id)
gr->grd_fib_index[fib_proto_to_dpo (fproto)];
}
- gx->gx_itf = gbp_itf_add_and_lock (sw_if_index, gb->gb_bd_index);
- gbp_itf_set_l2_input_feature (gx->gx_itf, (gxi | GBP_EXT_ITF_ID),
- L2INPUT_FEAT_GBP_LPM_CLASSIFY);
-
gbp_ext_itf_db[sw_if_index] = gxi;
GBP_EXT_ITF_DBG ("add: %U", format_gbp_ext_itf, gx);
@@ -124,11 +119,6 @@ gbp_ext_itf_delete (u32 sw_if_index)
GBP_EXT_ITF_DBG ("del: %U", format_gbp_ext_itf, gx);
- gbp_itf_set_l2_input_feature (gx->gx_itf,
- (gxi | GBP_EXT_ITF_ID),
- L2INPUT_FEAT_NONE);
- gbp_itf_unlock (gx->gx_itf);
-
gbp_route_domain_unlock (gx->gx_rd);
gbp_bridge_domain_unlock (gx->gx_bd);
diff --git a/src/plugins/gbp/gbp_itf.c b/src/plugins/gbp/gbp_itf.c
index 7cfe9da0ab0..4944f278144 100644
--- a/src/plugins/gbp/gbp_itf.c
+++ b/src/plugins/gbp/gbp_itf.c
@@ -194,15 +194,48 @@ format_gbp_itf (u8 * s, va_list * args)
gi = gbp_itf_get (gii);
- s = format (s, "%U locks:%d input-feats:%U output-feats:%U",
+ s = format (s, "%U locks:%d bd-index:%d input-feats:%U output-feats:%U",
format_vnet_sw_if_index_name, vnet_get_main (),
gi->gi_sw_if_index, gi->gi_locks,
+ gi->gi_bd_index,
format_l2_input_features, gi->gi_l2_input_fb, 0,
format_l2_output_features, gi->gi_l2_output_fb, 0);
return (s);
}
+static clib_error_t *
+gbp_itf_show (vlib_main_t * vm,
+ unformat_input_t * input, vlib_cli_command_t * cmd)
+{
+ u32 gii;
+
+ vlib_cli_output (vm, "Interfaces:");
+
+ vec_foreach_index (gii, gbp_itfs)
+ {
+ vlib_cli_output (vm, " [%d] %U", gii, format_gbp_itf, gii);
+ }
+
+ return (NULL);
+}
+
+/*?
+ * Show Group Based Interfaces
+ *
+ * @cliexpar
+ * @cliexstart{show gbp contract}
+ * @cliexend
+ ?*/
+/* *INDENT-OFF* */
+VLIB_CLI_COMMAND (gbp_contract_show_node, static) = {
+ .path = "show gbp interface",
+ .short_help = "show gbp interface\n",
+ .function = gbp_itf_show,
+};
+/* *INDENT-ON* */
+
+
/*
* fd.io coding-style-patch-verification: ON
*
diff --git a/test/test_gbp.py b/test/test_gbp.py
index cf9ad251d4b..4d8328474b9 100644
--- a/test/test_gbp.py
+++ b/test/test_gbp.py
@@ -5,7 +5,7 @@ import unittest
from scapy.packet import Raw
from scapy.layers.l2 import Ether, ARP, Dot1Q
-from scapy.layers.inet import IP, UDP
+from scapy.layers.inet import IP, UDP, ICMP
from scapy.layers.inet6 import IPv6, ICMPv6ND_NS, ICMPv6NDOptSrcLLAddr, \
ICMPv6ND_NA
from scapy.utils6 import in6_getnsma, in6_getnsmac
@@ -19,7 +19,7 @@ from vpp_interface import VppInterface
from vpp_ip_route import VppIpRoute, VppRoutePath, VppIpTable, \
VppIpInterfaceAddress, VppIpInterfaceBind, find_route
from vpp_l2 import VppBridgeDomain, VppBridgeDomainPort, \
- VppBridgeDomainArpEntry, VppL2FibEntry, find_bridge_domain_port
+ VppBridgeDomainArpEntry, VppL2FibEntry, find_bridge_domain_port, VppL2Vtr
from vpp_sub_interface import VppDot1QSubint
from vpp_ip import VppIpAddress, VppIpPrefix
from vpp_papi import VppEnum, MACAddress
@@ -68,10 +68,6 @@ class VppGbpEndpoint(VppObject):
return str(self.vmac)
@property
- def mac(self):
- return self.itf.remote_mac
-
- @property
def ip4(self):
return self._ip4
@@ -3197,7 +3193,12 @@ class TestGBP(VppTestCase):
#
vlan_100 = VppDot1QSubint(self, self.pg0, 100)
vlan_100.admin_up()
- ext_itf = VppGbpExtItf(self, vlan_100, bd1, rd1)
+ VppL2Vtr(self, vlan_100, L2_VTR_OP.L2_POP_1).add_vpp_config()
+ vlan_101 = VppDot1QSubint(self, self.pg0, 101)
+ vlan_101.admin_up()
+ VppL2Vtr(self, vlan_101, L2_VTR_OP.L2_POP_1).add_vpp_config()
+
+ ext_itf = VppGbpExtItf(self, self.loop0, bd1, rd1)
ext_itf.add_vpp_config()
#
@@ -3210,15 +3211,44 @@ class TestGBP(VppTestCase):
vx_tun_l3.add_vpp_config()
#
+ # External Endpoints
+ #
+ eep1 = VppGbpEndpoint(self, vlan_100,
+ epg_220, None,
+ "10.0.0.1", "11.0.0.1",
+ "2001:10::1", "3001::1",
+ ep_flags.GBP_API_ENDPOINT_FLAG_EXTERNAL)
+ eep1.add_vpp_config()
+ eep2 = VppGbpEndpoint(self, vlan_101,
+ epg_220, None,
+ "10.0.0.2", "11.0.0.2",
+ "2001:10::2", "3001::2",
+ ep_flags.GBP_API_ENDPOINT_FLAG_EXTERNAL)
+ eep2.add_vpp_config()
+
+ #
+ # A remote endpoint
+ #
+ rep = VppGbpEndpoint(self, vx_tun_l3,
+ epg_220, None,
+ "10.0.0.101", "11.0.0.101",
+ "2001:10::101", "3001::101",
+ ep_flags.GBP_API_ENDPOINT_FLAG_REMOTE,
+ self.pg7.local_ip4,
+ self.pg7.remote_ip4,
+ mac=None)
+ rep.add_vpp_config()
+
+ #
# packets destined to unknown addresses in the BVI's subnet
# are ARP'd for
#
- p4 = (Ether(src=self.pg0.remote_mac, dst=str(self.router_mac)) /
+ p4 = (Ether(src=eep1.mac, dst=str(self.router_mac)) /
Dot1Q(vlan=100) /
IP(src="10.0.0.1", dst="10.0.0.88") /
UDP(sport=1234, dport=1234) /
Raw('\xa5' * 100))
- p6 = (Ether(src=self.pg0.remote_mac, dst=str(self.router_mac)) /
+ p6 = (Ether(src=eep1.mac, dst=str(self.router_mac)) /
Dot1Q(vlan=100) /
IPv6(src="2001:10::1", dst="2001:10::88") /
UDP(sport=1234, dport=1234) /
@@ -3244,29 +3274,6 @@ class TestGBP(VppTestCase):
self.assertTrue(inner.haslayer(ARP))
#
- # An external Endpoint
- #
- eep = VppGbpEndpoint(self, vlan_100,
- epg_220, None,
- "10.0.0.1", "11.0.0.1",
- "2001:10::1", "3001::1",
- ep_flags.GBP_API_ENDPOINT_FLAG_EXTERNAL)
- eep.add_vpp_config()
-
- #
- # A remote endpoint
- #
- rep = VppGbpEndpoint(self, vx_tun_l3,
- epg_220, None,
- "10.0.0.101", "11.0.0.101",
- "2001:10::101", "3001::101",
- ep_flags.GBP_API_ENDPOINT_FLAG_REMOTE,
- self.pg7.local_ip4,
- self.pg7.remote_ip4,
- mac=None)
- rep.add_vpp_config()
-
- #
# remote to external
#
p = (Ether(src=self.pg7.remote_mac,
@@ -3283,11 +3290,41 @@ class TestGBP(VppTestCase):
rxs = self.send_and_expect(self.pg7, p * 1, self.pg0)
#
+ # local EP pings router
+ #
+ p = (Ether(src=eep1.mac, dst=str(self.router_mac)) /
+ Dot1Q(vlan=100) /
+ IP(src=eep1.ip4.address, dst="10.0.0.128") /
+ ICMP(type='echo-request'))
+
+ rxs = self.send_and_expect(self.pg0, p * 1, self.pg0)
+
+ for rx in rxs:
+ self.assertEqual(rx[Ether].src, str(self.router_mac))
+ self.assertEqual(rx[Ether].dst, eep1.mac)
+ self.assertEqual(rx[Dot1Q].vlan, 100)
+
+ #
+ # local EP pings other local EP
+ #
+ p = (Ether(src=eep1.mac, dst=eep2.mac) /
+ Dot1Q(vlan=100) /
+ IP(src=eep1.ip4.address, dst=eep2.ip4.address) /
+ ICMP(type='echo-request'))
+
+ rxs = self.send_and_expect(self.pg0, p * 1, self.pg0)
+
+ for rx in rxs:
+ self.assertEqual(rx[Ether].src, eep1.mac)
+ self.assertEqual(rx[Ether].dst, eep2.mac)
+ self.assertEqual(rx[Dot1Q].vlan, 101)
+
+ #
# A subnet reachable through the external EP
#
ip_220 = VppIpRoute(self, "10.220.0.0", 24,
- [VppRoutePath(eep.ip4.address,
- eep.epg.bvi.sw_if_index)],
+ [VppRoutePath(eep1.ip4.address,
+ eep1.epg.bvi.sw_if_index)],
table_id=t4.table_id)
ip_220.add_vpp_config()
@@ -3311,9 +3348,40 @@ class TestGBP(VppTestCase):
rxs = self.send_and_expect(self.pg7, p * 1, self.pg0)
#
+ # A subnet reachable through the external EP
+ #
+ ip_221 = VppIpRoute(self, "10.221.0.0", 24,
+ [VppRoutePath(eep2.ip4.address,
+ eep2.epg.bvi.sw_if_index)],
+ table_id=t4.table_id)
+ ip_221.add_vpp_config()
+
+ l3o_221 = VppGbpSubnet(
+ self, rd1, "10.221.0.0", 24,
+ VppEnum.vl_api_gbp_subnet_type_t.GBP_API_SUBNET_L3_OUT,
+ epg=epg_220)
+ l3o_221.add_vpp_config()
+
+ #
+ # ping between hosts in remote subnets
+ #
+ p = (Ether(src=eep1.mac, dst=str(self.router_mac)) /
+ Dot1Q(vlan=100) /
+ IP(src="10.220.0.1", dst="10.221.0.1") /
+ ICMP(type='echo-request'))
+
+ rxs = self.send_and_expect(self.pg0, p * 1, self.pg0)
+
+ for rx in rxs:
+ self.assertEqual(rx[Ether].src, str(self.router_mac))
+ self.assertEqual(rx[Ether].dst, eep2.mac)
+ self.assertEqual(rx[Dot1Q].vlan, 101)
+
+ #
# cleanup
#
self.pg7.unconfig_ip4()
+ vlan_100.set_vtr(L2_VTR_OP.L2_DISABLED)
if __name__ == '__main__':
diff --git a/test/vpp_l2.py b/test/vpp_l2.py
index 45b7d69dd01..dc73e4238f3 100644
--- a/test/vpp_l2.py
+++ b/test/vpp_l2.py
@@ -208,3 +208,32 @@ class VppL2FibEntry(VppObject):
def object_id(self):
return "L2-Fib-Entry-%s-%s-%s" % (self.bd, self.mac, self.itf)
+
+
+class VppL2Vtr(VppObject):
+
+ def __init__(self, test, itf, op):
+ self._test = test
+ self.itf = itf
+ self.op = op
+
+ def add_vpp_config(self):
+ self.itf.set_vtr(self.op)
+ self._test.registry.register(self, self._test.logger)
+
+ def remove_vpp_config(self):
+ self.itf.set_vtr(L2_VTR_OP.L2_DISABLED)
+
+ def query_vpp_config(self):
+ ds = self._test.vapi.sw_interface_dump()
+ d = self.itf.get_interface_config_from_dump(ds)
+
+ if d is not None:
+ return (d.vtr_op == self.op)
+ return False
+
+ def __str__(self):
+ return self.object_id()
+
+ def object_id(self):
+ return "L2-vtr-%s-%d" % (str(self.itf), self.op)
/span> b) { return (u16x4) _m_punpcklwd ((__m64) a, (__m64) b); } always_inline u32x2 u32x2_interleave_hi (u32x2 a, u32x2 b) { return (u32x2) _m_punpckhdq ((__m64) a, (__m64) b); } always_inline u32x2 u32x2_interleave_lo (u32x2 a, u32x2 b) { return (u32x2) _m_punpckldq ((__m64) a, (__m64) b); } /* 128 bit packs. */ always_inline u8x16 u16x8_pack (u16x8 lo, u16x8 hi) { return (u8x16) _mm_packus_epi16 ((__m128i) lo, (__m128i) hi); } always_inline i8x16 i16x8_pack (i16x8 lo, i16x8 hi) { return (i8x16) _mm_packs_epi16 ((__m128i) lo, (__m128i) hi); } always_inline u16x8 u32x4_pack (u32x4 lo, u32x4 hi) { return (u16x8) _mm_packs_epi32 ((__m128i) lo, (__m128i) hi); } /* 64 bit packs. */ always_inline u8x8 u16x4_pack (u16x4 lo, u16x4 hi) { return (u8x8) _m_packuswb ((__m64) lo, (__m64) hi); } always_inline i8x8 i16x4_pack (i16x4 lo, i16x4 hi) { return (i8x8) _m_packsswb ((__m64) lo, (__m64) hi); } always_inline u16x4 u32x2_pack (u32x2 lo, u32x2 hi) { return (u16x4) _m_packssdw ((__m64) lo, (__m64) hi); } always_inline i16x4 i32x2_pack (i32x2 lo, i32x2 hi) { return (i16x4) _m_packssdw ((__m64) lo, (__m64) hi); } #ifndef __ICC always_inline u64x2 u64x2_read_lo (u64x2 x, u64 * a) { return (u64x2) _mm_loadl_pi ((__m128) x, (__m64 *) a); } always_inline u64x2 u64x2_read_hi (u64x2 x, u64 * a) { return (u64x2) _mm_loadh_pi ((__m128) x, (__m64 *) a); } always_inline void u64x2_write_lo (u64x2 x, u64 * a) { _mm_storel_pi ((__m64 *) a, (__m128) x); } always_inline void u64x2_write_hi (u64x2 x, u64 * a) { _mm_storeh_pi ((__m64 *) a, (__m128) x); } #endif /* Unaligned loads/stores. */ #define _(t) \ always_inline void t##_store_unaligned (t x, void * a) \ { _mm_storeu_si128 ((__m128i *) a, (__m128i) x); } \ always_inline t t##_load_unaligned (void * a) \ { return (t) _mm_loadu_si128 ((__m128i *) a); } _(u8x16) _(u16x8) _(u32x4) _(u64x2) _(i8x16) _(i16x8) _(i32x4) _(i64x2) #undef _ #define _signed_binop(n,m,f,g) \ /* Unsigned */ \ always_inline u##n##x##m \ u##n##x##m##_##f (u##n##x##m x, u##n##x##m y) \ { return (u##n##x##m) _mm_##g##n ((__m128i) x, (__m128i) y); } \ \ /* Signed */ \ always_inline i##n##x##m \ i##n##x##m##_##f (i##n##x##m x, i##n##x##m y) \ { return (i##n##x##m) _mm_##g##n ((__m128i) x, (__m128i) y); } /* Addition/subtraction with saturation. */ _signed_binop (8, 16, add_saturate, adds_epu) _signed_binop (16, 8, add_saturate, adds_epu) _signed_binop (8, 16, sub_saturate, subs_epu) _signed_binop (16, 8, sub_saturate, subs_epu) /* Multiplication. */ always_inline i16x8 i16x8_mul_lo (i16x8 x, i16x8 y) { return (i16x8) _mm_mullo_epi16 ((__m128i) x, (__m128i) y); } always_inline u16x8 u16x8_mul_lo (u16x8 x, u16x8 y) { return (u16x8) _mm_mullo_epi16 ((__m128i) x, (__m128i) y); } always_inline i16x8 i16x8_mul_hi (i16x8 x, i16x8 y) { return (i16x8) _mm_mulhi_epu16 ((__m128i) x, (__m128i) y); } always_inline u16x8 u16x8_mul_hi (u16x8 x, u16x8 y) { return (u16x8) _mm_mulhi_epu16 ((__m128i) x, (__m128i) y); } /* 128 bit shifts. */ #define _(p,a,b,c,f) \ always_inline p##a##x##b p##a##x##b##_ishift_##c (p##a##x##b x, int i) \ { return (p##a##x##b) _mm_##f##i_epi##a ((__m128i) x, i); } \ \ always_inline p##a##x##b p##a##x##b##_shift_##c (p##a##x##b x, p##a##x##b y) \ { return (p##a##x##b) _mm_##f##_epi##a ((__m128i) x, (__m128i) y); } _(u, 16, 8, left, sll) _(u, 32, 4, left, sll) _(u, 64, 2, left, sll) _(u, 16, 8, right, srl) _(u, 32, 4, right, srl) _(u, 64, 2, right, srl) _(i, 16, 8, left, sll) _(i, 32, 4, left, sll) _(i, 64, 2, left, sll) _(i, 16, 8, right, sra) _(i, 32, 4, right, sra) #undef _ /* 64 bit shifts. */ always_inline u16x4 u16x4_shift_left (u16x4 x, u16x4 i) { return (u16x4) _m_psllw ((__m64) x, (__m64) i); }; always_inline u32x2 u32x2_shift_left (u32x2 x, u32x2 i) { return (u32x2) _m_pslld ((__m64) x, (__m64) i); }; always_inline u16x4 u16x4_shift_right (u16x4 x, u16x4 i) { return (u16x4) _m_psrlw ((__m64) x, (__m64) i); }; always_inline u32x2 u32x2_shift_right (u32x2 x, u32x2 i) { return (u32x2) _m_psrld ((__m64) x, (__m64) i); }; always_inline i16x4 i16x4_shift_left (i16x4 x, i16x4 i) { return (i16x4) _m_psllw ((__m64) x, (__m64) i); }; always_inline i32x2 i32x2_shift_left (i32x2 x, i32x2 i) { return (i32x2) _m_pslld ((__m64) x, (__m64) i); }; always_inline i16x4 i16x4_shift_right (i16x4 x, i16x4 i) { return (i16x4) _m_psraw ((__m64) x, (__m64) i); }; always_inline i32x2 i32x2_shift_right (i32x2 x, i32x2 i) { return (i32x2) _m_psrad ((__m64) x, (__m64) i); }; #define u8x16_word_shift_left(a,n) (u8x16) _mm_slli_si128((__m128i) a, n) #define u8x16_word_shift_right(a,n) (u8x16) _mm_srli_si128((__m128i) a, n) #define i8x16_word_shift_left(a,n) \ ((i8x16) u8x16_word_shift_left((u8x16) (a), (n))) #define i8x16_word_shift_right(a,n) \ ((i8x16) u8x16_word_shift_right((u8x16) (a), (n))) #define u16x8_word_shift_left(a,n) \ ((u16x8) u8x16_word_shift_left((u8x16) (a), (n) * sizeof (u16))) #define i16x8_word_shift_left(a,n) \ ((u16x8) u8x16_word_shift_left((u8x16) (a), (n) * sizeof (u16))) #define u16x8_word_shift_right(a,n) \ ((u16x8) u8x16_word_shift_right((u8x16) (a), (n) * sizeof (u16))) #define i16x8_word_shift_right(a,n) \ ((i16x8) u8x16_word_shift_right((u8x16) (a), (n) * sizeof (u16))) #define u32x4_word_shift_left(a,n) \ ((u32x4) u8x16_word_shift_left((u8x16) (a), (n) * sizeof (u32))) #define i32x4_word_shift_left(a,n) \ ((u32x4) u8x16_word_shift_left((u8x16) (a), (n) * sizeof (u32))) #define u32x4_word_shift_right(a,n) \ ((u32x4) u8x16_word_shift_right((u8x16) (a), (n) * sizeof (u32))) #define i32x4_word_shift_right(a,n) \ ((i32x4) u8x16_word_shift_right((u8x16) (a), (n) * sizeof (u32))) #define u64x2_word_shift_left(a,n) \ ((u64x2) u8x16_word_shift_left((u8x16) (a), (n) * sizeof (u64))) #define i64x2_word_shift_left(a,n) \ ((u64x2) u8x16_word_shift_left((u8x16) (a), (n) * sizeof (u64))) #define u64x2_word_shift_right(a,n) \ ((u64x2) u8x16_word_shift_right((u8x16) (a), (n) * sizeof (u64))) #define i64x2_word_shift_right(a,n) \ ((i64x2) u8x16_word_shift_right((u8x16) (a), (n) * sizeof (u64))) /* SSE2 has no rotate instructions: use shifts to simulate them. */ #define _(t,n,lr1,lr2) \ always_inline t##x##n \ t##x##n##_irotate_##lr1 (t##x##n w, int i) \ { \ ASSERT (i >= 0 && i <= BITS (t)); \ return (t##x##n##_ishift_##lr1 (w, i) \ | t##x##n##_ishift_##lr2 (w, BITS (t) - i)); \ } \ \ always_inline t##x##n \ t##x##n##_rotate_##lr1 (t##x##n w, t##x##n i) \ { \ t##x##n j = t##x##n##_splat (BITS (t)); \ return (t##x##n##_shift_##lr1 (w, i) \ | t##x##n##_shift_##lr2 (w, j - i)); \ } _(u16, 8, left, right); _(u16, 8, right, left); _(u32, 4, left, right); _(u32, 4, right, left); _(u64, 2, left, right); _(u64, 2, right, left); #undef _ #ifndef __clang__ #define _(t,n,lr1,lr2) \ always_inline t##x##n \ t##x##n##_word_rotate2_##lr1 (t##x##n w0, t##x##n w1, int i) \ { \ int m = sizeof (t##x##n) / sizeof (t); \ ASSERT (i >= 0 && i < m); \ return (t##x##n##_word_shift_##lr1 (w0, i) \ | t##x##n##_word_shift_##lr2 (w1, m - i)); \ } \ \ always_inline t##x##n \ t##x##n##_word_rotate_##lr1 (t##x##n w0, int i) \ { return t##x##n##_word_rotate2_##lr1 (w0, w0, i); } _(u8, 16, left, right); _(u8, 16, right, left); _(u16, 8, left, right); _(u16, 8, right, left); _(u32, 4, left, right); _(u32, 4, right, left); _(u64, 2, left, right); _(u64, 2, right, left); #undef _ #endif always_inline int u8x16_is_all_zero (u8x16 x) { return _mm_testz_si128 ((__m128i) x, (__m128i) x); } always_inline int u16x8_is_all_zero (u16x8 x) { return _mm_testz_si128 ((__m128i) x, (__m128i) x); } always_inline int u32x4_is_all_zero (u32x4 x) { return _mm_testz_si128 ((__m128i) x, (__m128i) x); } always_inline int u64x2_is_all_zero (u64x2 x) { return _mm_testz_si128 ((__m128i) x, (__m128i) x); } #define u32x4_select(A,MASK) \ ({ \ u32x4 _x, _y; \ _x = (A); \ asm volatile ("pshufd %[mask], %[x], %[y]" \ : /* outputs */ [y] "=x" (_y) \ : /* inputs */ [x] "x" (_x), [mask] "i" (MASK)); \ _y; \ }) #define u32x4_splat_word(x,i) \ u32x4_select ((x), (((i) << (2*0)) \ | ((i) << (2*1)) \ | ((i) << (2*2)) \ | ((i) << (2*3)))) /* Extract low order 32 bit word. */ always_inline u32 u32x4_get0 (u32x4 x) { u32 result; asm volatile ("movd %[x], %[result]": /* outputs */ [result] "=r" (result) : /* inputs */ [x] "x" (x)); return result; } always_inline u32x4 u32x4_set0 (u32 x) { u32x4 result; asm volatile ("movd %[x], %[result]": /* outputs */ [result] "=x" (result) : /* inputs */ [x] "r" (x)); return result; } always_inline i32x4 i32x4_set0 (i32 x) { return (i32x4) u32x4_set0 ((u32) x); } always_inline i32 i32x4_get0 (i32x4 x) { return (i32) u32x4_get0 ((u32x4) x); } /* Converts all ones/zeros compare mask to bitmap. */ always_inline u32 u8x16_compare_byte_mask (u8x16 x) { return _mm_movemask_epi8 ((__m128i) x); } extern u8 u32x4_compare_word_mask_table[256]; always_inline u32 u32x4_compare_word_mask (u32x4 x) { u32 m = u8x16_compare_byte_mask ((u8x16) x); return (u32x4_compare_word_mask_table[(m >> 0) & 0xff] | (u32x4_compare_word_mask_table[(m >> 8) & 0xff] << 2)); } always_inline u32 u8x16_zero_byte_mask (u8x16 x) { u8x16 zero = { 0 }; return u8x16_compare_byte_mask (u8x16_is_equal (x, zero)); } always_inline u32 u16x8_zero_byte_mask (u16x8 x) { u16x8 zero = { 0 }; return u8x16_compare_byte_mask ((u8x16) u16x8_is_equal (x, zero)); } always_inline u32 u32x4_zero_byte_mask (u32x4 x) { u32x4 zero = { 0 }; return u8x16_compare_byte_mask ((u8x16) u32x4_is_equal (x, zero)); } always_inline u8x16 u8x16_max (u8x16 x, u8x16 y) { return (u8x16) _mm_max_epu8 ((__m128i) x, (__m128i) y); } always_inline u32 u8x16_max_scalar (u8x16 x) { x = u8x16_max (x, u8x16_word_shift_right (x, 8)); x = u8x16_max (x, u8x16_word_shift_right (x, 4)); x = u8x16_max (x, u8x16_word_shift_right (x, 2)); x = u8x16_max (x, u8x16_word_shift_right (x, 1)); return _mm_extract_epi16 ((__m128i) x, 0) & 0xff; } always_inline u8x16 u8x16_min (u8x16 x, u8x16 y) { return (u8x16) _mm_min_epu8 ((__m128i) x, (__m128i) y); } always_inline u8 u8x16_min_scalar (u8x16 x) { x = u8x16_min (x, u8x16_word_shift_right (x, 8)); x = u8x16_min (x, u8x16_word_shift_right (x, 4)); x = u8x16_min (x, u8x16_word_shift_right (x, 2)); x = u8x16_min (x, u8x16_word_shift_right (x, 1)); return _mm_extract_epi16 ((__m128i) x, 0) & 0xff; } always_inline i16x8 i16x8_max (i16x8 x, i16x8 y) { return (i16x8) _mm_max_epi16 ((__m128i) x, (__m128i) y); } always_inline i16 i16x8_max_scalar (i16x8 x) { x = i16x8_max (x, i16x8_word_shift_right (x, 4)); x = i16x8_max (x, i16x8_word_shift_right (x, 2)); x = i16x8_max (x, i16x8_word_shift_right (x, 1)); return _mm_extract_epi16 ((__m128i) x, 0); } always_inline i16x8 i16x8_min (i16x8 x, i16x8 y) { return (i16x8) _mm_min_epi16 ((__m128i) x, (__m128i) y); } always_inline i16 i16x8_min_scalar (i16x8 x) { x = i16x8_min (x, i16x8_word_shift_right (x, 4)); x = i16x8_min (x, i16x8_word_shift_right (x, 2)); x = i16x8_min (x, i16x8_word_shift_right (x, 1)); return _mm_extract_epi16 ((__m128i) x, 0); } #undef _signed_binop #endif /* included_vector_sse2_h */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */