#!/usr/bin/env python3 import unittest import binascii from socket import AF_INET6 from framework import VppTestCase, VppTestRunner from vpp_ip_route import VppIpRoute, VppRoutePath, FibPathProto, VppIpTable from vpp_srv6 import ( SRv6LocalSIDBehaviors, VppSRv6LocalSID, VppSRv6Policy, SRv6PolicyType, VppSRv6Steering, SRv6PolicySteeringTypes, ) import scapy.compat from scapy.packet import Raw from scapy.layers.l2 import Ether, Dot1Q from scapy.layers.inet6 import IPv6, UDP, IPv6ExtHdrSegmentRouting from scapy.layers.inet import IP, UDP from util import ppp class TestSRv6(VppTestCase): """SRv6 Test Case""" @classmethod def setUpClass(cls): super(TestSRv6, cls).setUpClass() @classmethod def tearDownClass(cls): super(TestSRv6, cls).tearDownClass() def setUp(self): """Perform test setup before each test case.""" super(TestSRv6, self).setUp() # packet sizes, inclusive L2 overhead self.pg_packet_sizes = [64, 512, 1518, 9018] # reset packet_infos self.reset_packet_infos() def tearDown(self): """Clean up test setup after each test case.""" self.teardown_interfaces() super(TestSRv6, self).tearDown() def configure_interface( self, interface, ipv6=False, ipv4=False, ipv6_table_id=0, ipv4_table_id=0 ): """Configure interface. :param ipv6: configure IPv6 on interface :param ipv4: configure IPv4 on interface :param ipv6_table_id: FIB table_id for IPv6 :param ipv4_table_id: FIB table_id for IPv4 """ self.logger.debug("Configuring interface %s" % (interface.name)) if ipv6: self.logger.debug("Configuring IPv6") interface.set_table_ip6(ipv6_table_id) interface.config_ip6() interface.resolve_ndp(timeout=5) if ipv4: self.logger.debug("Configuring IPv4") interface.set_table_ip4(ipv4_table_id) interface.config_ip4() interface.resolve_arp() interface.admin_up() def setup_interfaces(self, ipv6=[], ipv4=[], ipv6_table_id=[], ipv4_table_id=[]): """Create and configure interfaces. :param ipv6: list of interface IPv6 capabilities :param ipv4: list of interface IPv4 capabilities :param ipv6_table_id: list of intf IPv6 FIB table_ids :param ipv4_table_id: list of intf IPv4 FIB table_ids :returns: List of created interfaces. """ # how many interfaces? if len(ipv6): count = len(ipv6) else: count = len(ipv4) self.logger.debug("Creating and configuring %d interfaces" % (count)) # fill up ipv6 and ipv4 lists if needed # not enabled (False) is the default if len(ipv6) < count: ipv6 += (count - len(ipv6)) * [False] if len(ipv4) < count: ipv4 += (count - len(ipv4)) * [False] # fill up table_id lists if needed # table_id 0 (global) is the default if len(ipv6_table_id) < count: ipv6_table_id += (count - len(ipv6_table_id)) * [0] if len(ipv4_table_id) < count: ipv4_table_id += (count - len(ipv4_table_id)) * [0] # create 'count' pg interfaces self.create_pg_interfaces(range(count)) # setup all interfaces for i in range(count): intf = self.pg_interfaces[i] self.configure_interface( intf, ipv6[i], ipv4[i], ipv6_table_id[i], ipv4_table_id[i] ) if any(ipv6): self.logger.debug(self.vapi.cli("show ip6 neighbors")) if any(ipv4): self.logger.debug(self.vapi.cli("show ip4 neighbors")) self.logger.debug(self.vapi.cli("show interface")) self.logger.debug(self.vapi.cli("show hardware")) return self.pg_interfaces def teardown_interfaces(self): """Unconfigure and bring down interface.""" self.logger.debug("Tearing down interfaces") # tear down all interfaces # AFAIK they cannot be deleted for i in self.pg_interfaces: self.logger.debug("Tear down interface %s" % (i.name)) i.admin_down() i.unconfig() i.set_table_ip4(0) i.set_table_ip6(0) @unittest.skipUnless(0, "PC to fix") def test_SRv6_T_Encaps(self): """Test SRv6 Transit.Encaps behavior for IPv6.""" # send traffic to one destination interface # source and destination are IPv6 only self.setup_interfaces(ipv6=[True, True]) # configure FIB entries route = VppIpRoute( self, "a4::", 64, [VppRoutePath(self.pg1.remote_ip6, self.pg1.sw_if_index)] ) route.add_vpp_config() # configure encaps IPv6 source address # needs to be done before SR Policy config # TODO: API? self.vapi.cli("set sr encaps source addr a3::") bsid = "a3::9999:1" # configure SRv6 Policy # Note: segment list order: first -> last sr_policy = VppSRv6Policy( self, bsid=bsid, is_encap=1, sr_type=SRv6PolicyType.SR_POLICY_TYPE_DEFAULT, weight=1, fib_table=0, segments=["a4::", "a5::", "a6::c7"], source="a3::", ) sr_policy.add_vpp_config() self.sr_policy = sr_policy # log the sr policies self.logger.info(self.vapi.cli("show sr policies")) # steer IPv6 traffic to a7::/64 into SRv6 Policy # use the bsid of the above self.sr_policy pol_steering = VppSRv6Steering( self, bsid=self.sr_policy.bsid, prefix="a7::", mask_width=64, traffic_type=SRv6PolicySteeringTypes.SR_STEER_IPV6, sr_policy_index=0, table_id=0, sw_if_index=0, ) pol_steering.add_vpp_config() # log the sr steering policies self.logger.info(self.vapi.cli("show sr steering policies")) # create packets count = len(self.pg_packet_sizes) dst_inner = "a7::1234" pkts = [] # create IPv6 packets without SRH packet_header = self.create_packet_header_IPv6(dst_inner) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # create IPv6 packets with SRH # packets with segments-left 1, active segment a7:: packet_header = self.create_packet_header_IPv6_SRH( sidlist=["a8::", "a7::", "a6::"], segleft=1 ) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # create IPv6 packets with SRH and IPv6 # packets with segments-left 1, active segment a7:: packet_header = self.create_packet_header_IPv6_SRH_IPv6( dst_inner, sidlist=["a8::", "a7::", "a6::"], segleft=1 ) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # send packets and verify received packets self.send_and_verify_pkts( self.pg0, pkts, self.pg1, self.compare_rx_tx_packet_T_Encaps ) # log the localsid counters self.logger.info(self.vapi.cli("show sr localsid")) # remove SR steering pol_steering.remove_vpp_config() self.logger.info(self.vapi.cli("show sr steering policies")) # remove SR Policies self.sr_policy.remove_vpp_config() self.logger.info(self.vapi.cli("show sr policies")) # remove FIB entries # done by tearDown # cleanup interfaces self.teardown_interfaces() @unittest.skipUnless(0, "PC to fix") def test_SRv6_T_Insert(self): """Test SRv6 Transit.Insert behavior (IPv6 only).""" # send traffic to one destination interface # source and destination are IPv6 only self.setup_interfaces(ipv6=[True, True]) # configure FIB entries route = VppIpRoute( self, "a4::", 64, [VppRoutePath(self.pg1.remote_ip6, self.pg1.sw_if_index)] ) route.add_vpp_config() # configure encaps IPv6 source address # needs to be done before SR Policy config # TODO: API? self.vapi.cli("set sr encaps source addr a3::") bsid = "a3::9999:1" # configure SRv6 Policy # Note: segment list order: first -> last sr_policy = VppSRv6Policy( self, bsid=bsid, is_encap=0, sr_type=SRv6PolicyType.SR_POLICY_TYPE_DEFAULT, weight=1, fib_table=0, segments=["a4::", "a5::", "a6::c7"], source="a3::", ) sr_policy.add_vpp_config() self.sr_policy = sr_policy # log the sr policies self.logger.info(self.vapi.cli("show sr policies")) # steer IPv6 traffic to a7::/64 into SRv6 Policy # use the bsid of the above self.sr_policy pol_steering = VppSRv6Steering( self, bsid=self.sr_policy.bsid, prefix="a7::", mask_width=64, traffic_type=SRv6PolicySteeringTypes.SR_STEER_IPV6, sr_policy_index=0, table_id=0, sw_if_index=0, ) pol_steering.add_vpp_config() # log the sr steering policies self.logger.info(self.vapi.cli("show sr steering policies")) # create packets count = len(self.pg_packet_sizes) dst_inner = "a7::1234" pkts = [] # create IPv6 packets without SRH packet_header = self.create_packet_header_IPv6(dst_inner) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # create IPv6 packets with SRH # packets with segments-left 1, active segment a7:: packet_header = self.create_packet_header_IPv6_SRH( sidlist=["a8::", "a7::", "a6::"], segleft=1 ) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # send packets and verify received packets self.send_and_verify_pkts( self.pg0, pkts, self.pg1, self.compare_rx_tx_packet_T_Insert ) # log the localsid counters self.logger.info(self.vapi.cli("show sr localsid")) # remove SR steering pol_steering.remove_vpp_config() self.logger.info(self.vapi.cli("show sr steering policies")) # remove SR Policies self.sr_policy.remove_vpp_config() self.logger.info(self.vapi.cli("show sr policies")) # remove FIB entries # done by tearDown # cleanup interfaces self.teardown_interfaces() @unittest.skipUnless(0, "PC to fix") def test_SRv6_T_Encaps_IPv4(self): """Test SRv6 Transit.Encaps behavior for IPv4.""" # send traffic to one destination interface # source interface is IPv4 only # destination interface is IPv6 only self.setup_interfaces(ipv6=[False, True], ipv4=[True, False]) # configure FIB entries route = VppIpRoute( self, "a4::", 64, [VppRoutePath(self.pg1.remote_ip6, self.pg1.sw_if_index)] ) route.add_vpp_config() # configure encaps IPv6 source address # needs to be done before SR Policy config # TODO: API? self.vapi.cli("set sr encaps source addr a3::") bsid = "a3::9999:1" # configure SRv6 Policy # Note: segment list order: first -> last sr_policy = VppSRv6Policy( self, bsid=bsid, is_encap=1, sr_type=SRv6PolicyType.SR_POLICY_TYPE_DEFAULT, weight=1, fib_table=0, segments=["a4::", "a5::", "a6::c7"], source="a3::", ) sr_policy.add_vpp_config() self.sr_policy = sr_policy # log the sr policies self.logger.info(self.vapi.cli("show sr policies")) # steer IPv4 traffic to 7.1.1.0/24 into SRv6 Policy # use the bsid of the above self.sr_policy pol_steering = VppSRv6Steering( self, bsid=self.sr_policy.bsid, prefix="7.1.1.0", mask_width=24, traffic_type=SRv6PolicySteeringTypes.SR_STEER_IPV4, sr_policy_index=0, table_id=0, sw_if_index=0, ) pol_steering.add_vpp_config() # log the sr steering policies self.logger.info(self.vapi.cli("show sr steering policies")) # create packets count = len(self.pg_packet_sizes) dst_inner = "7.1.1.123" pkts = [] # create IPv4 packets packet_header = self.create_packet_header_IPv4(dst_inner) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # send packets and verify received packets self.send_and_verify_pkts( self.pg0, pkts, self.pg1, self.compare_rx_tx_packet_T_Encaps_IPv4 ) # log the localsid counters self.logger.info(self.vapi.cli("show sr localsid")) # remove SR steering pol_steering.remove_vpp_config() self.logger.info(self.vapi.cli("show sr steering policies")) # remove SR Policies self.sr_policy.remove_vpp_config() self.logger.info(self.vapi.cli("show sr policies")) # remove FIB entries # done by tearDown # cleanup interfaces self.teardown_interfaces() @unittest.skip("VPP crashes after running this test") def test_SRv6_T_Encaps_L2(self): """Test SRv6 Transit.Encaps behavior for L2.""" # send traffic to one destination interface # source interface is IPv4 only TODO? # destination interface is IPv6 only self.setup_interfaces(ipv6=[False, True], ipv4=[False, False]) # configure FIB entries route = VppIpRoute( self, "a4::", 64, [VppRoutePath(self.pg1.remote_ip6, self.pg1.sw_if_index)] ) route.add_vpp_config() # configure encaps IPv6 source address # needs to be done before SR Policy config # TODO: API? self.vapi.cli("set sr encaps source addr a3::") bsid = "a3::9999:1" # configure SRv6 Policy # Note: segment list order: first -> last sr_policy = VppSRv6Policy( self, bsid=bsid, is_encap=1, sr_type=SRv6PolicyType.SR_POLICY_TYPE_DEFAULT, weight=1, fib_table=0, segments=["a4::", "a5::", "a6::c7"], source="a3::", ) sr_policy.add_vpp_config() self.sr_policy = sr_policy # log the sr policies self.logger.info(self.vapi.cli("show sr policies")) # steer L2 traffic into SRv6 Policy # use the bsid of the above self.sr_policy pol_steering = VppSRv6Steering( self, bsid=self.sr_policy.bsid, prefix="::", mask_width=0, traffic_type=SRv6PolicySteeringTypes.SR_STEER_L2, sr_policy_index=0, table_id=0, sw_if_index=self.pg0.sw_if_index, ) pol_steering.add_vpp_config() # log the sr steering policies self.logger.info(self.vapi.cli("show sr steering policies")) # create packets count = len(self.pg_packet_sizes) pkts = [] # create L2 packets without dot1q header packet_header = self.create_packet_header_L2() # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # create L2 packets with dot1q header packet_header = self.create_packet_header_L2(vlan=123) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # send packets and verify received packets self.send_and_verify_pkts( self.pg0, pkts, self.pg1, self.compare_rx_tx_packet_T_Encaps_L2 ) # log the localsid counters self.logger.info(self.vapi.cli("show sr localsid")) # remove SR steering pol_steering.remove_vpp_config() self.logger.info(self.vapi.cli("show sr steering policies")) # remove SR Policies self.sr_policy.remove_vpp_config() self.logger.info(self.vapi.cli("show sr policies")) # remove FIB entries # done by tearDown # cleanup interfaces self.teardown_interfaces() def test_SRv6_End(self): """Test SRv6 End (without PSP) behavior.""" # send traffic to one destination interface # source and destination interfaces are IPv6 only self.setup_interfaces(ipv6=[True, True]) # configure FIB entries route = VppIpRoute( self, "a4::", 64, [VppRoutePath(self.pg1.remote_ip6, self.pg1.sw_if_index)] ) route.add_vpp_config() # configure SRv6 localSID End without PSP behavior localsid = VppSRv6LocalSID( self, localsid="A3::0", behavior=SRv6LocalSIDBehaviors.SR_BEHAVIOR_END, nh_addr=0, end_psp=0, sw_if_index=0, vlan_index=0, fib_table=0, ) localsid.add_vpp_config() # log the localsids self.logger.debug(self.vapi.cli("show sr localsid")) # create IPv6 packets with SRH (SL=2, SL=1, SL=0) # send one packet per SL value per packet size # SL=0 packet with localSID End with USP needs 2nd SRH count = len(self.pg_packet_sizes) dst_inner = "a4::1234" pkts = [] # packets with segments-left 2, active segment a3:: packet_header = self.create_packet_header_IPv6_SRH_IPv6( dst_inner, sidlist=["a5::", "a4::", "a3::"], segleft=2 ) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # packets with segments-left 1, active segment a3:: packet_header = self.create_packet_header_IPv6_SRH_IPv6( dst_inner, sidlist=["a4::", "a3::", "a2::"], segleft=1 ) # add to traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # TODO: test behavior with SL=0 packet (needs 2*SRH?) expected_count = len(pkts) # packets without SRH (should not crash) packet_header = self.create_packet_header_IPv6("a3::") # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # send packets and verify received packets self.send_and_verify_pkts( self.pg0, pkts, self.pg1, self.compare_rx_tx_packet_End, expected_count=expected_count, ) # log the localsid counters self.logger.info(self.vapi.cli("show sr localsid")) # remove SRv6 localSIDs localsid.remove_vpp_config() # remove FIB entries # done by tearDown # cleanup interfaces self.teardown_interfaces() def test_SRv6_End_with_PSP(self): """Test SRv6 End with PSP behavior.""" # send traffic to one destination interface # source and destination interfaces are IPv6 only self.setup_interfaces(ipv6=[True, True]) # configure FIB entries route = VppIpRoute( self, "a4::", 64, [VppRoutePath(self.pg1.remote_ip6, self.pg1.sw_if_index)] ) route.add_vpp_config() # configure SRv6 localSID End with PSP behavior localsid = VppSRv6LocalSID( self, localsid="A3::0", behavior=SRv6LocalSIDBehaviors.SR_BEHAVIOR_END, nh_addr=0, end_psp=1, sw_if_index=0, vlan_index=0, fib_table=0, ) localsid.add_vpp_config() # log the localsids self.logger.debug(self.vapi.cli("show sr localsid")) # create IPv6 packets with SRH (SL=2, SL=1) # send one packet per SL value per packet size # SL=0 packet with localSID End with PSP is dropped count = len(self.pg_packet_sizes) dst_inner = "a4::1234" pkts = [] # packets with segments-left 2, active segment a3:: packet_header = self.create_packet_header_IPv6_SRH_IPv6( dst_inner, sidlist=["a5::", "a4::", "a3::"], segleft=2 ) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # packets with segments-left 1, active segment a3:: packet_header = self.create_packet_header_IPv6_SRH_IPv6( dst_inner, sidlist=["a4::", "a3::", "a2::"], segleft=1 ) # add to traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # send packets and verify received packets self.send_and_verify_pkts( self.pg0, pkts, self.pg1, self.compare_rx_tx_packet_End_PSP ) # log the localsid counters self.logger.info(self.vapi.cli("show sr localsid")) # remove SRv6 localSIDs localsid.remove_vpp_config() # remove FIB entries # done by tearDown # cleanup interfaces self.teardown_interfaces() def test_SRv6_End_X(self): """Test SRv6 End.X (without PSP) behavior.""" # create three interfaces (1 source, 2 destinations) # source and destination interfaces are IPv6 only self.setup_interfaces(ipv6=[True, True, True]) # configure FIB entries # a4::/64 via pg1 and pg2 route = VppIpRoute( self, "a4::", 64, [ VppRoutePath(self.pg1.remote_ip6, self.pg1.sw_if_index), VppRoutePath(self.pg2.remote_ip6, self.pg2.sw_if_index), ], ) route.add_vpp_config() self.logger.debug(self.vapi.cli("show ip6 fib")) # configure SRv6 localSID End.X without PSP behavior # End.X points to interface pg1 localsid = VppSRv6LocalSID( self, localsid="A3::C4", behavior=SRv6LocalSIDBehaviors.SR_BEHAVIOR_X, nh_addr=self.pg1.remote_ip6, end_psp=0, sw_if_index=self.pg1.sw_if_index, vlan_index=0, fib_table=0, ) localsid.add_vpp_config() # log the localsids self.logger.debug(self.vapi.cli("show sr localsid")) # create IPv6 packets with SRH (SL=2, SL=1) # send one packet per SL value per packet size # SL=0 packet with localSID End with PSP is dropped count = len(self.pg_packet_sizes) dst_inner = "a4::1234" pkts = [] # packets with segments-left 2, active segment a3::c4 packet_header = self.create_packet_header_IPv6_SRH_IPv6( dst_inner, sidlist=["a5::", "a4::", "a3::c4"], segleft=2 ) # create traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # packets with segments-left 1, active segment a3::c4 packet_header = self.create_packet_header_IPv6_SRH_IPv6( dst_inner, sidlist=["a4::", "a3::c4", "a2::"], segleft=1 ) # add to traffic stream pg0->pg1 pkts.extend( self.create_stream( self.pg0, self.pg1, packet_header, self.pg_packet_sizes, count ) ) # send packets and verify received packets # using same comparison function as End (no PSP) self.send_and_verify_pkts( self.pg0, pkts, self.pg1, self.compare_rx_tx_packet_End ) # assert nothing was received on the other interface (pg2) self.pg2.assert_nothing_captured(remark="mis-directed packet(s)") # log the localsid counters self.logger.info(self.vapi.cli("show sr localsid")) # remove SRv6 localSIDs localsid.remove_vpp_config() # remove FIB entries # done by tearDown # cleanup interfaces self.teardown_interfaces() def test_SRv6_End_X_with_PSP(self): """Test SRv6 End.X with PSP behavior.""" # create three interfaces (1 source, 2 destinations) # source and destination interfaces are IPv6 only self.setup_interfaces(ipv6=[True, True, True]) # configure FIB entries # a4::/64 via pg1 and pg2 route = VppIpRoute( self, "a4::", 64, [ VppRoutePath(self.pg1.remote_ip6, self.pg1.sw_if_index), VppRoutePath(self.pg2.remote_ip6, self.pg2.sw_if_index), ], ) route.add_vpp_config() # configure SRv6 localSID End with PSP behavior localsid = VppSRv6LocalSID( self, localsid="A3::C4", behavior=SRv6LocalSIDBehaviors.SR_BEHAVIOR_X, nh_addr=self.pg1.remote_ip6, end_psp=1, sw_if_index=self.pg1.sw
/*
* Copyright (c) 2018 Cisco and/or its affiliates.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <plugins/gbp/gbp_vxlan.h>
#include <plugins/gbp/gbp_learn.h>
#include <plugins/gbp/gbp_bridge_domain.h>
#include <plugins/gbp/gbp_route_domain.h>
#include <vnet/vxlan-gbp/vxlan_gbp.h>
#include <vlibmemory/api.h>
#include <vnet/fib/fib_table.h>
#include <vlib/punt.h>
/**
* A reference to a VXLAN-GBP tunnel created as a child/dependent tunnel
* of the tempplate GBP-VXLAN tunnel
*/
typedef struct vxlan_tunnel_ref_t_
{
gbp_itf_hdl_t vxr_itf;
u32 vxr_sw_if_index;
index_t vxr_parent;
gbp_vxlan_tunnel_layer_t vxr_layer;
} vxlan_tunnel_ref_t;
/**
* DB of added tunnels
*/
uword *gv_db;
/**
* Logger
*/
static vlib_log_class_t gt_logger;
/**
* Pool of template tunnels
*/
static gbp_vxlan_tunnel_t *gbp_vxlan_tunnel_pool;
/**
* Pool of child tunnels
*/
static vxlan_tunnel_ref_t *vxlan_tunnel_ref_pool;
/**
* DB of template interfaces by SW interface index
*/
static index_t *gbp_vxlan_tunnel_db;
/**
* DB of child interfaces by SW interface index
*/
static index_t *vxlan_tunnel_ref_db;
/**
* handle registered with the ;unt infra
*/
static vlib_punt_hdl_t punt_hdl;
static char *gbp_vxlan_tunnel_layer_strings[] = {
#define _(n,s) [GBP_VXLAN_TUN_##n] = s,
forecah_gbp_vxlan_tunnel_layer
#undef _
};
#define GBP_VXLAN_TUN_DBG(...) \
vlib_log_debug (gt_logger, __VA_ARGS__);
gbp_vxlan_tunnel_t *
gbp_vxlan_tunnel_get (index_t gti)
{
return (pool_elt_at_index (gbp_vxlan_tunnel_pool, gti));
}
static vxlan_tunnel_ref_t *
vxlan_tunnel_ref_get (index_t vxri)
{
return (pool_elt_at_index (vxlan_tunnel_ref_pool, vxri));
}
static u8 *
format_vxlan_tunnel_ref (u8 * s, va_list * args)
{
index_t vxri = va_arg (*args, u32);
vxlan_tunnel_ref_t *vxr;
vxr = vxlan_tunnel_ref_get (vxri);
s = format (s, "[%U]", format_gbp_itf_hdl, vxr->vxr_itf);
return (s);
}
static void
gdb_vxlan_dep_del (u32 sw_if_index)
{
vxlan_tunnel_ref_t *vxr;
gbp_vxlan_tunnel_t *gt;
index_t vxri;
u32 pos;
vxr = vxlan_tunnel_ref_get (vxlan_tunnel_ref_db[sw_if_index]);
vxri = vxr - vxlan_tunnel_ref_pool;
gt = gbp_vxlan_tunnel_get (vxr->vxr_parent);
GBP_VXLAN_TUN_DBG ("del-dep:%U", format_vxlan_tunnel_ref, vxri);
vxlan_tunnel_ref_db[vxr->vxr_sw_if_index] = INDEX_INVALID;
pos = vec_search (gt->gt_tuns, vxri);
ASSERT (~0 != pos);
vec_del1 (gt->gt_tuns, pos);
vnet_vxlan_gbp_tunnel_del (vxr->vxr_sw_if_index);
pool_put (vxlan_tunnel_ref_pool, vxr);
}
static gbp_itf_hdl_t
gdb_vxlan_dep_add (gbp_vxlan_tunnel_t * gt,
const ip46_address_t * src, const ip46_address_t * dst)
{
vnet_vxlan_gbp_tunnel_add_del_args_t args = {
.is_add = 1,
.is_ip6 = !ip46_address_is_ip4 (src),
.vni = gt->gt_vni,
.src = *src,
.dst = *dst,
.instance = ~0,
.mode = (GBP_VXLAN_TUN_L2 == gt->gt_layer ?
VXLAN_GBP_TUNNEL_MODE_L2 : VXLAN_GBP_TUNNEL_MODE_L3),
};
vxlan_tunnel_ref_t *vxr;
u32 sw_if_index;
index_t vxri;
int rv;
sw_if_index = ~0;
rv = vnet_vxlan_gbp_tunnel_add_del (&args, &sw_if_index);
if (VNET_API_ERROR_TUNNEL_EXIST == rv)
{
vxri = vxlan_tunnel_ref_db[sw_if_index];
vxr = vxlan_tunnel_ref_get (vxri);
gbp_itf_lock (vxr->vxr_itf);
}
else if (0 == rv)
{
ASSERT (~0 != sw_if_index);
GBP_VXLAN_TUN_DBG ("add-dep:%U %U %U %d", format_vnet_sw_if_index_name,
vnet_get_main (), sw_if_index,
format_ip46_address, src, IP46_TYPE_ANY,
format_ip46_address, dst, IP46_TYPE_ANY, gt->gt_vni);
pool_get_zero (vxlan_tunnel_ref_pool, vxr);
vxri = (vxr - vxlan_tunnel_ref_pool);
vxr->vxr_parent = gt - gbp_vxlan_tunnel_pool;
vxr->vxr_sw_if_index = sw_if_index;
vxr->vxr_layer = gt->gt_layer;
/*
* store the child both on the parent's list and the global DB
*/
vec_add1 (gt->gt_tuns, vxri);
vec_validate_init_empty (vxlan_tunnel_ref_db,
vxr->vxr_sw_if_index, INDEX_INVALID);
vxlan_tunnel_ref_db[vxr->vxr_sw_if_index] = vxri;
if (GBP_VXLAN_TUN_L2 == vxr->vxr_layer)
{
l2output_feat_masks_t ofeat;
l2input_feat_masks_t ifeat;
gbp_bridge_domain_t *gbd;
gbd = gbp_bridge_domain_get (gt->gt_gbd);
vxr->vxr_itf = gbp_itf_l2_add_and_lock_w_free
(vxr->vxr_sw_if_index, gt->gt_gbd, gdb_vxlan_dep_del);
ofeat = L2OUTPUT_FEAT_GBP_POLICY_MAC;
ifeat = L2INPUT_FEAT_NONE;
if (!(gbd->gb_flags & GBP_BD_FLAG_DO_NOT_LEARN))
ifeat |= L2INPUT_FEAT_GBP_LEARN;
gbp_itf_l2_set_output_feature (vxr->vxr_itf, ofeat);
gbp_itf_l2_set_input_feature (vxr->vxr_itf, ifeat);
}
else
{
vxr->vxr_itf = gbp_itf_l3_add_and_lock_w_free
(vxr->vxr_sw_if_index, gt->gt_grd, gdb_vxlan_dep_del);
gbp_itf_l3_set_input_feature (vxr->vxr_itf, GBP_ITF_L3_FEAT_LEARN);
}
}
else
{
return (GBP_ITF_HDL_INVALID);
}
return (vxr->vxr_itf);
}
u32
vxlan_gbp_tunnel_get_parent (u32 sw_if_index)
{
ASSERT ((sw_if_index < vec_len (vxlan_tunnel_ref_db)) &&
(INDEX_INVALID != vxlan_tunnel_ref_db[sw_if_index]));
gbp_vxlan_tunnel_t *gt;
vxlan_tunnel_ref_t *vxr;
vxr = vxlan_tunnel_ref_get (vxlan_tunnel_ref_db[sw_if_index]);
gt = gbp_vxlan_tunnel_get (vxr->vxr_parent);
return (gt->gt_sw_if_index);
}
gbp_itf_hdl_t
vxlan_gbp_tunnel_lock_itf (u32 sw_if_index)
{
ASSERT ((sw_if_index < vec_len (vxlan_tunnel_ref_db)) &&
(INDEX_INVALID != vxlan_tunnel_ref_db[sw_if_index]));
vxlan_tunnel_ref_t *vxr;
vxr = vxlan_tunnel_ref_get (vxlan_tunnel_ref_db[sw_if_index]);
gbp_itf_lock (vxr->vxr_itf);
return (vxr->vxr_itf);
}
gbp_vxlan_tunnel_type_t
gbp_vxlan_tunnel_get_type (u32 sw_if_index)
{
if (sw_if_index < vec_len (vxlan_tunnel_ref_db) &&
INDEX_INVALID != vxlan_tunnel_ref_db[sw_if_index])
{
return (VXLAN_GBP_TUNNEL);
}
else if (sw_if_index < vec_len (gbp_vxlan_tunnel_db) &&
INDEX_INVALID != gbp_vxlan_tunnel_db[sw_if_index])
{
return (GBP_VXLAN_TEMPLATE_TUNNEL);
}
ASSERT (0);
return (GBP_VXLAN_TEMPLATE_TUNNEL);
}
gbp_itf_hdl_t
gbp_vxlan_tunnel_clone_and_lock (u32 sw_if_index,
const ip46_address_t * src,
const ip46_address_t * dst)
{
gbp_vxlan_tunnel_t *gt;
index_t gti;
gti = gbp_vxlan_tunnel_db[sw_if_index];
if (INDEX_INVALID == gti)
return (GBP_ITF_HDL_INVALID);
gt = pool_elt_at_index (gbp_vxlan_tunnel_pool, gti);
return (gdb_vxlan_dep_add (gt, src, dst));
}
void
vxlan_gbp_tunnel_unlock (u32 sw_if_index)
{
/* vxlan_tunnel_ref_t *vxr; */
/* index_t vxri; */
/* vxri = vxlan_tunnel_ref_db[sw_if_index]; */
/* ASSERT (vxri != INDEX_INVALID); */
/* vxr = vxlan_tunnel_ref_get (vxri); */
/* gdb_vxlan_dep_del (vxri); */
}
void
gbp_vxlan_walk (gbp_vxlan_cb_t cb, void *ctx)
{
gbp_vxlan_tunnel_t *gt;
/* *INDENT-OFF* */
pool_foreach (gt, gbp_vxlan_tunnel_pool,
({
if (WALK_CONTINUE != cb(gt, ctx))
break;
}));
/* *INDENT-ON* */
}
static walk_rc_t
gbp_vxlan_tunnel_show_one (gbp_vxlan_tunnel_t * gt, void *ctx)
{
vlib_cli_output (ctx, "%U", format_gbp_vxlan_tunnel,
gt - gbp_vxlan_tunnel_pool);
return (WALK_CONTINUE);
}
static u8 *
format_gbp_vxlan_tunnel_name (u8 * s, va_list * args)
{
u32 dev_instance = va_arg (*args, u32);
return format (s, "gbp-vxlan-%d", dev_instance);
}
u8 *
format_gbp_vxlan_tunnel_layer (u8 * s, va_list * args)
{
gbp_vxlan_tunnel_layer_t gl = va_arg (*args, gbp_vxlan_tunnel_layer_t);
s = format (s, "%s", gbp_vxlan_tunnel_layer_strings[gl]);
return (s);
}
u8 *
format_gbp_vxlan_tunnel (u8 * s, va_list * args)
{
u32 dev_instance = va_arg (*args, u32);
CLIB_UNUSED (int verbose) = va_arg (*args, int);
gbp_vxlan_tunnel_t *gt = gbp_vxlan_tunnel_get (dev_instance);
index_t *vxri;
s = format (s, " [%d] gbp-vxlan-tunnel: hw:%d sw:%d vni:%d %U",
dev_instance, gt->gt_hw_if_index,
gt->gt_sw_if_index, gt->gt_vni,
format_gbp_vxlan_tunnel_layer, gt->gt_layer);
if (GBP_VXLAN_TUN_L2 == gt->gt_layer)
s = format (s, " BD:%d gbd-index:%d", gt->gt_bd_rd_id, gt->gt_gbd);
else
s = format (s, " RD:%d grd-index:%d", gt->gt_bd_rd_id, gt->gt_grd);
s = format (s, " dependents:");
vec_foreach (vxri, gt->gt_tuns)
{
s = format (s, "\n %U, ", format_vxlan_tunnel_ref, *vxri);
}
return s;
}
typedef struct gbp_vxlan_tx_trace_t_
{
u32 vni;
} gbp_vxlan_tx_trace_t;
u8 *
format_gbp_vxlan_tx_trace (u8 * s, va_list * args)
{
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
gbp_vxlan_tx_trace_t *t = va_arg (*args, gbp_vxlan_tx_trace_t *);
s = format (s, "GBP-VXLAN: vni:%d", t->vni);
return (s);
}
clib_error_t *
gbp_vxlan_interface_admin_up_down (vnet_main_t * vnm,
u32 hw_if_index, u32 flags)
{
vnet_hw_interface_t *hi;
u32 ti;
hi = vnet_get_hw_interface (vnm, hw_if_index);
if (NULL == gbp_vxlan_tunnel_db ||
hi->sw_if_index >= vec_len (gbp_vxlan_tunnel_db))
return (NULL);
ti = gbp_vxlan_tunnel_db[hi->sw_if_index];
if (~0 == ti)
/* not one of ours */
return (NULL);
if (flags & VNET_SW_INTERFACE_FLAG_ADMIN_UP)
vnet_hw_interface_set_flags (vnm, hw_if_index,
VNET_HW_INTERFACE_FLAG_LINK_UP);
else
vnet_hw_interface_set_flags (vnm, hw_if_index, 0);
return (NULL);
}
static uword
gbp_vxlan_interface_tx (vlib_main_t * vm,
vlib_node_runtime_t * node, vlib_frame_t * frame)
{
clib_warning ("you shouldn't be here, leaking buffers...");
return frame->n_vectors;
}
/* *INDENT-OFF* */
VNET_DEVICE_CLASS (gbp_vxlan_device_class) = {
.name = "GBP VXLAN tunnel-template",
.format_device_name = format_gbp_vxlan_tunnel_name,
.format_device = format_gbp_vxlan_tunnel,
.format_tx_trace = format_gbp_vxlan_tx_trace,
.admin_up_down_function = gbp_vxlan_interface_admin_up_down,
.tx_function = gbp_vxlan_interface_tx,
};
VNET_HW_INTERFACE_CLASS (gbp_vxlan_hw_interface_class) = {
.name = "GBP-VXLAN",
.flags = VNET_HW_INTERFACE_CLASS_FLAG_P2P,
};
/* *INDENT-ON* */
int
gbp_vxlan_tunnel_add (u32 vni, gbp_vxlan_tunnel_layer_t layer,
u32 bd_rd_id,
<