#!/usr/bin/env python
import socket
from util import ip4n_range, ip4_range, reassemble4
import unittest
from framework import VppTestCase, VppTestRunner
from template_bd import BridgeDomain
from scapy.layers.l2 import Ether, Raw
from scapy.layers.inet import IP, UDP
from scapy.layers.vxlan import VXLAN
from scapy.utils import atol
from vpp_ip_route import VppIpRoute, VppRoutePath
from vpp_ip import INVALID_INDEX
class TestVxlan(BridgeDomain, VppTestCase):
""" VXLAN Test Case """
def __init__(self, *args):
BridgeDomain.__init__(self)
VppTestCase.__init__(self, *args)
def encapsulate(self, pkt, vni):
"""
Encapsulate the original payload frame by adding VXLAN header with its
UDP, IP and Ethernet fields
"""
return (Ether(src=self.pg0.remote_mac, dst=self.pg0.local_mac) /
IP(src=self.pg0.remote_ip4, dst=self.pg0.local_ip4) /
UDP(sport=self.dport, dport=self.dport, chksum=0) /
VXLAN(vni=vni, flags=self.flags) /
pkt)
def ip_range(self, start, end):
""" range of remote ip's """
return ip4_range(self.pg0.remote_ip4, start, end)
def encap_mcast(self, pkt, src_ip, src_mac, vni):
"""
Encapsulate the original payload frame by adding VXLAN header with its
UDP, IP and Ethernet fields
"""
return (Ether(src=src_mac, dst=self.mcast_mac) /
IP(src=src_ip, dst=self.mcast_ip4) /
UDP(sport=self.dport, dport=self.dport, chksum=0) /
VXLAN(vni=vni, flags=self.flags) /
pkt)
def decapsulate(self, pkt):
"""
Decapsulate the original payload frame by removing VXLAN header
"""
# check if is set I flag
self.assertEqual(pkt[VXLAN].flags, int('0x8', 16))
return pkt[VXLAN].payload
# Method for checking VXLAN encapsulation.
#
def check_encapsulation(self, pkt, vni, local_only=False, mcast_pkt=False):
# TODO: add error messages
# Verify source MAC is VPP_MAC and destination MAC is MY_MAC resolved
# by VPP using ARP.
self.assertEqual(pkt[Ether].src, self.pg0.local_mac)
if not local_only:
if not mcast_pkt:
self.assertEqual(pkt[Ether].dst, self.pg0.remote_mac)
else:
self.assertEqual(pkt[Ether].dst, type(self).mcast_mac)
# Verify VXLAN tunnel source IP is VPP_IP and destination IP is MY_IP.
self.assertEqual(pkt[IP].src, self.pg0.local_ip4)
if not local_only:
if not mcast_pkt:
self.assertEqual(pkt[IP].dst, self.pg0.remote_ip4)
else:
self.assertEqual(pkt[IP].dst, type(self).mcast_ip4)
# Verify UDP destination port is VXLAN 4789, source UDP port could be
# arbitrary.
self.assertEqual(pkt[UDP].dport, type(self).dport)
# TODO: checksum check
# Verify VNI
self.assertEqual(pkt[VXLAN].vni, vni)
@classmethod
def create_vxlan_flood_test_bd(cls, vni, n_ucast_tunnels):
# Create 10 ucast vxlan tunnels under bd
ip_range_start = 10
ip_range_end = ip_range_start + n_ucast_tunnels
next_hop_address = cls.pg0.remote_ip4
for dest_ip4 in ip4_range(next_hop_address, ip_range_start,
ip_range_end):
# add host route so dest_ip4n will not be resolved
rip = VppIpRoute(cls, dest_ip4, 32,
[VppRoutePath(next_hop_address,
INVALID_INDEX)],
register=False)
rip.add_vpp_config()
dest_ip4n = socket.inet_pton(socket.AF_INET, dest_ip4)
r = cls.vapi.vxlan_add_del_tunnel(src_address=cls.pg0.local_ip4n,
dst_address=dest_ip4n, vni=vni)
cls.vapi.sw_interface_set_l2_bridge(r.sw_if_index, bd_id=vni)
@classmethod
def add_del_shared_mcast_dst_load(cls, is_add):
"""
add or del tunnels sharing the same mcast dst
to test vxlan ref_count mechanism
"""
n_shared_dst_tunnels = 20
vni_start = 10000
vni_end = vni_start + n_shared_dst_tunnels
for vni in range(vni_start, vni_end):
r = cls.vapi.vxlan_add_del_tunnel(src_address=cls.pg0.local_ip4n,
dst_address=cls.mcast_ip4n,
mcast_sw_if_index=1,
is_add=is_add, vni=vni)
if r.sw_if_index == 0xffffffff:
raise ValueError("bad sw_if_index: ~0")
@classmethod
def add_shared_mcast_dst_load(cls):
cls.add_del_shared_mcast_dst_load(is_add=1)
@classmethod
def del_shared_mcast_dst_load(cls):
cls.add_del_shared_mcast_dst_load(is_add=0)
@classmethod
def add_del_mcast_tunnels_load(cls, is_add):
"""
add or del tunnels to test vxlan stability
"""
n_distinct_dst_tunnels = 200
ip_range_start = 10
ip_range_end = ip_range_start + n_distinct_dst_tunnels
for dest_ip4n in ip4n_range(cls.mcast_ip4n, ip_range_start,
ip_range_end):
vni = bytearray(dest_ip4n)[3]
cls.vapi.vxlan_add_del_tunnel(src_address=cls.pg0.local_ip4n,
dst_address=dest_ip4n,
mcast_sw_if_index=1, is_add=is_add,
vni=vni)
@classmethod
def add_mcast_tunnels_load(cls):
cls.add_del_mcast_tunnels_load(is_add=1)
@classmethod
def del_mcast_tunnels_load(cls):
cls.add_del_mcast_tunnels_load(is_add=0)
# Class method to start the VXLAN test case.
# Overrides setUpClass method in VppTestCase class.
# Python try..except statement is used to ensure that the tear down of
# the class will be executed even if exception is raised.
# @param cls The class pointer.
@classmethod
def setUpClass(cls):
super(TestVxlan, /*
* Copyright (c) 2015 Cisco and/or its affiliates.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/*
* node_funcs.h: processing nodes global functions/inlines
*
* Copyright (c) 2008 Eliot Dresselhaus
*
* Permission is hereby granted, free of charge, to any person obtaining
* a copy of this software and associated documentation files (the
* "Software"), to deal in the Software without restriction, including
* without limitation the rights to use, copy, modify, merge, publish,
* distribute, sublicense, and/or sell copies of the Software, and to
* permit persons to whom the Software is furnished to do so, subject to
* the following conditions:
*
* The above copyright notice and this permission notice shall be
* included in all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
/** \file
vlib node functions
*/
#ifndef included_vlib_node_funcs_h
#define included_vlib_node_funcs_h
#include <vppinfra/fifo.h>
#include <vppinfra/tw_timer_1t_3w_1024sl_ov.h>
#ifdef CLIB_SANITIZE_ADDR
#include <sanitizer/asan_interface.h>
#endif
static_always_inline void
vlib_process_start_switch_stack (vlib_main_t * vm, vlib_process_t * p)
{
#ifdef CLIB_SANITIZE_ADDR
void *stack = p ? (void *) p->stack : vlib_thread_stacks[vm->thread_index];
u32 stack_bytes = p ? p->log2_n_stack_bytes : VLIB_THREAD_STACK_SIZE;
__sanitizer_start_switch_fiber (&vm->asan_stack_save, stack, stack_bytes);
#endif
}
static_always_inline void
vlib_process_finish_switch_stack (vlib_main_t * vm)
{
#ifdef CLIB_SANITIZE_ADDR
const void *bottom_old;
size_t size_old;
__sanitizer_finish_switch_fiber (&vm->asan_stack_save, &bottom_old,
&size_old);
#endif
}
/** \brief Get vlib node by index.
@warning This function will ASSERT if @c i is out of range.
@param vm vlib_main_t pointer, varies by thread
@param i node index.
@return pointer to the requested vlib_node_t.
*/
always_inline vlib_node_t *
vlib_get_node (vlib_main_t * vm, u32 i)
{
return vec_elt (vm->node_main.nodes, i);
}
/** \brief Get vlib node by graph arc (next) index.
@param vm vlib_main_t pointer, varies by thread
@param node_index index of original node
@param next_index graph arc index
@return pointer to the vlib_node_t at the end of the indicated arc
*/
always_inline vlib_node_t *
vlib_get_next_node (vlib_main_t * vm, u32 node_index, u32 next_index)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n;
n = vec_elt (nm->nodes, node_index);
ASSERT (next_index < vec_len (n->next_nodes));
return vlib_get_node (vm, n->next_nodes[next_index]);
}
/** \brief Get node runtime by node index.
@param vm vlib_main_t pointer, varies by thread
@param node_index index of node
@return pointer to the indicated vlib_node_runtime_t
*/
always_inline vlib_node_runtime_t *
vlib_node_get_runtime (vlib_main_t * vm, u32 node_index)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n = vec_elt (nm->nodes, node_index);
vlib_process_t *p;
if (n->type != VLIB_NODE_TYPE_PROCESS)
return vec_elt_at_index (nm->nodes_by_type[n->type], n->runtime_index);
else
{
p = vec_elt (nm->processes, n->runtime_index);
return &p->node_runtime;
}
}
/** \brief Get node runtime private data by node index.
@param vm vlib_main_t pointer, varies by thread
@param node_index index of the node
@return pointer to the indicated vlib_node_runtime_t private data
*/
always_inline void *
vlib_node_get_runtime_data (vlib_main_t * vm, u32 node_index)
{
vlib_node_runtime_t *r = vlib_node_get_runtime (vm, node_index);
return r->runtime_data;
}
/** \brief Set node runtime private data.
@param vm vlib_main_t pointer, varies by thread
@param node_index index of the node
@param runtime_data arbitrary runtime private data
@param n_runtime_data_bytes size of runtime private data
*/
always_inline void
vlib_node_set_runtime_data (vlib_main_t * vm, u32 node_index,
void *runtime_data, u32 n_runtime_data_bytes)
{
vlib_node_t *n = vlib_get_node (vm, node_index);
vlib_node_runtime_t *r = vlib_node_get_runtime (vm, node_index);
n->runtime_data_bytes = n_runtime_data_bytes;
vec_free (n->runtime_data);
vec_add (n->runtime_data, runtime_data, n_runtime_data_bytes);
ASSERT (vec_len (n->runtime_data) <= sizeof (vlib_node_runtime_t) -
STRUCT_OFFSET_OF (vlib_node_runtime_t, runtime_data));
if (vec_len (n->runtime_data) > 0)
clib_memcpy_fast (r->runtime_data, n->runtime_data,
vec_len (n->runtime_data));
}
/** \brief Set node dispatch state.
@param vm vlib_main_t pointer, varies by thread
@param node_index index of the node
@param new_state new state for node, see vlib_node_state_t
*/
always_inline void
vlib_node_set_state (vlib_main_t * vm, u32 node_index,
vlib_node_state_t new_state)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n;
vlib_node_runtime_t *r;
n = vec_elt (nm->nodes, node_index);
if (n->type == VLIB_NODE_TYPE_PROCESS)
{
vlib_process_t *p = vec_elt (nm->processes, n->runtime_index);
r = &p->node_runtime;
/* When disabling make sure flags are cleared. */
p->flags &= ~(VLIB_PROCESS_RESUME_PENDING
| VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_CLOCK
| VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_EVENT);
}
else
r = vec_elt_at_index (nm->nodes_by_type[n->type], n->runtime_index);
ASSERT (new_state < VLIB_N_NODE_STATE);
if (n->type == VLIB_NODE_TYPE_INPUT)
{
ASSERT (nm->input_node_counts_by_state[n->state] > 0);
nm->input_node_counts_by_state[n->state] -= 1;
nm->input_node_counts_by_state[new_state] += 1;
}
if (PREDICT_FALSE (r->state == VLIB_NODE_STATE_DISABLED))
vlib_node_runtime_perf_counter (vm, r, 0, 0, 0,
VLIB_NODE_RUNTIME_PERF_RESET);
n->state = new_state;
r->state = new_state;
}
/** \brief Get node dispatch state.
@param vm vlib_main_t pointer, varies by thread
@param node_index index of the node
@return state for node, see vlib_node_state_t
*/
always_inline vlib_node_state_t
vlib_node_get_state (vlib_main_t * vm, u32 node_index)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n;
n = vec_elt (nm->nodes, node_index);
return n->state;
}
always_inline void
vlib_node_set_interrupt_pending_with_data (vlib_main_t * vm, u32 node_index,
u32 data)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n = vec_elt (nm->nodes, node_index);
vlib_node_interrupt_t *i;
ASSERT (n->type == VLIB_NODE_TYPE_INPUT);
if (vm == vlib_get_main ())
{
/* local thread */
vec_add2 (nm->pending_local_interrupts, i, 1);
i->node_runtime_index = n->runtime_index;
i->data = data;
}
else
{
/* remote thread */
clib_spinlock_lock (&nm->pending_interrupt_lock);
vec_add2 (nm->pending_remote_interrupts, i, 1);
i->node_runtime_index = n->runtime_index;
i->data = data;
*nm->pending_remote_interrupts_notify = 1;
clib_spinlock_unlock (&nm->pending_interrupt_lock);
}
}
always_inline void
vlib_node_set_interrupt_pending (vlib_main_t * vm, u32 node_index)
{
vlib_node_set_interrupt_pending_with_data (vm, node_index, 0);
}
always_inline vlib_process_t *
vlib_get_process_from_node (vlib_main_t * vm, vlib_node_t * node)
{
vlib_node_main_t *nm = &vm->node_main;
ASSERT (node->type == VLIB_NODE_TYPE_PROCESS);
return vec_elt (nm->processes, node->runtime_index);
}
always_inline vlib_frame_t *
vlib_get_frame (vlib_main_t * vm, vlib_frame_t * f)
{
ASSERT (f != NULL);
ASSERT (f->frame_flags & VLIB_FRAME_IS_ALLOCATED);
return f;
}
always_inline void
vlib_frame_no_append (vlib_frame_t * f)
{
f->frame_flags |= VLIB_FRAME_NO_APPEND;
}
/* Byte alignment for vector arguments. */
#define VLIB_FRAME_VECTOR_ALIGN (1 << 4)
always_inline u32
vlib_frame_vector_byte_offset (u32 scalar_size)
{
return round_pow2 (sizeof (vlib_frame_t) + scalar_size,
VLIB_FRAME_VECTOR_ALIGN);
}
/** \brief Get pointer to frame vector data.
@param f vlib_frame_t pointer
@return pointer to first vector element in frame
*/
always_inline void *
vlib_frame_vector_args (vlib_frame_t * f)
{
return (void *) f + vlib_frame_vector_byte_offset (f->scalar_size);
}
/** \brief Get pointer to frame scalar data.
@param f vlib_frame_t pointer
@return arbitrary node scalar data
@sa vlib_frame_vector_args
*/
always_inline void *
vlib_frame_scalar_args (vlib_frame_t * f)
{
return vlib_frame_vector_args (f) - f->scalar_size;
}
always_inline vlib_next_frame_t *
vlib_node_runtime_get_next_frame (vlib_main_t * vm,
vlib_node_runtime_t * n, u32 next_index)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_next_frame_t *nf;
ASSERT (next_index < n->n_next_nodes);
nf = vec_elt_at_index (nm->next_frames, n->next_frame_index + next_index);
if (CLIB_DEBUG > 0)
{
vlib_node_t *node, *next;
node = vec_elt (nm->nodes, n->node_index);
next = vec_elt (nm->nodes, node->next_nodes[next_index]);
ASSERT (nf->node_runtime_index == next->runtime_index);
}
return nf;
}
/** \brief Get pointer to frame by (@c node_index, @c next_index).
@warning This is not a function that you should call directly.
See @ref vlib_get_next_frame instead.
@param vm vlib_main_t pointer, varies by thread
@param node_index index of the node
@param next_index graph arc index
@return pointer to the requested vlib_next_frame_t
@sa vlib_get_next_frame
*/
always_inline vlib_next_frame_t *
vlib_node_get_next_frame (vlib_main_t * vm, u32 node_index, u32 next_index)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n;
vlib_node_runtime_t *r;
n = vec_elt (nm->nodes, node_index);
r = vec_elt_at_index (nm->nodes_by_type[n->type], n->runtime_index);
return vlib_node_runtime_get_next_frame (vm, r, next_index);
}
vlib_frame_t *vlib_get_next_frame_internal (vlib_main_t * vm,
vlib_node_runtime_t * node,
u32 next_index,
u32 alloc_new_frame);
#define vlib_get_next_frame_macro(vm,node,next_index,vectors,n_vectors_left,alloc_new_frame) \
do { \
vlib_frame_t * _f \
= vlib_get_next_frame_internal ((vm), (node), (next_index), \
(alloc_new_frame)); \
u32 _n = _f->n_vectors; \
(vectors) = vlib_frame_vector_args (_f) + _n * sizeof ((vectors)[0]); \
(n_vectors_left) = VLIB_FRAME_SIZE - _n; \
} while (0)
/** \brief Get pointer to next frame vector data by
(@c vlib_node_runtime_t, @c next_index).
Standard single/dual loop boilerplate element.
@attention This is a MACRO, with SIDE EFFECTS.
@param vm vlib_main_t pointer, varies by thread
@param node current node vlib_node_runtime_t pointer
@param next_index requested graph arc index
@return @c vectors -- pointer to next available vector slot
@return @c n_vectors_left -- number of vector slots available
*/
#define vlib_get_next_frame(vm,node,next_index,vectors,n_vectors_left) \
vlib_get_next_frame_macro (vm, node, next_index, \
vectors, n_vectors_left, \
/* alloc new frame */ 0)
#define vlib_get_new_next_frame(vm,node,next_index,vectors,n_vectors_left) \
vlib_get_next_frame_macro (vm, node, next_index, \
vectors, n_vectors_left, \
/* alloc new frame */ 1)
/** \brief Release pointer to next frame vector data.
Standard single/dual loop boilerplate element.
@param vm vlib_main_t pointer, varies by thread
@param r current node vlib_node_runtime_t pointer
@param next_index graph arc index
@param n_packets_left number of slots still available in vector
*/
void
vlib_put_next_frame (vlib_main_t * vm,
vlib_node_runtime_t * r,
u32 next_index, u32 n_packets_left);
/* Combination get plus put. Returns vector argument just added. */
#define vlib_set_next_frame(vm,node,next_index,v) \
({ \
uword _n_left; \
vlib_get_next_frame ((vm), (node), (next_index), (v), _n_left); \
ASSERT (_n_left > 0); \
vlib_put_next_frame ((vm), (node), (next_index), _n_left - 1); \
(v); \
})
always_inline void
vlib_set_next_frame_buffer (vlib_main_t * vm,
vlib_node_runtime_t * node,
u32 next_index, u32 buffer_index)
{
u32 *p;
p = vlib_set_next_frame (vm, node, next_index, p);
p[0] = buffer_index;
}
vlib_frame_t *vlib_get_frame_to_node (vlib_main_t * vm, u32 to_node_index);
void vlib_put_frame_to_node (vlib_main_t * vm, u32 to_node_index,
vlib_frame_t * f);
always_inline uword
vlib_in_process_context (vlib_main_t * vm)
{
return vm->node_main.current_process_index != ~0;
}
always_inline vlib_process_t *
vlib_get_current_process (vlib_main_t * vm)
{
vlib_node_main_t *nm = &vm->node_main;
if (vlib_in_process_context (vm))
return vec_elt (nm->processes, nm->current_process_index);
return 0;
}
always_inline uword
vlib_current_process (vlib_main_t * vm)
{
return vlib_get_current_process (vm)->node_runtime.node_index;
}
always_inline u32
vlib_get_current_process_node_index (vlib_main_t * vm)
{
vlib_process_t *process = vlib_get_current_process (vm);
return process->node_runtime.node_index;
}
/** Returns TRUE if a process suspend time is less than 10us
@param dt - remaining poll time in seconds
@returns 1 if dt < 10e-6, 0 otherwise
*/
always_inline uword
vlib_process_suspend_time_is_zero (f64 dt)
{
return dt < 10e-6;
}
/** Suspend a vlib cooperative multi-tasking thread for a period of time
@param vm - vlib_main_t *
@param dt - suspend interval in seconds
@returns VLIB_PROCESS_RESUME_LONGJMP_RESUME, routinely ignored
*/
always_inline uword
vlib_process_suspend (vlib_main_t * vm, f64 dt)
{
uword r;
vlib_node_main_t *nm = &vm->node_main;
vlib_process_t *p = vec_elt (nm->processes, nm->current_process_index);
if (vlib_process_suspend_time_is_zero (dt))
return VLIB_PROCESS_RESUME_LONGJMP_RESUME;
p->flags |= VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_CLOCK;
r = clib_setjmp (&p->resume_longjmp, VLIB_PROCESS_RESUME_LONGJMP_SUSPEND);
if (r == VLIB_PROCESS_RESUME_LONGJMP_SUSPEND)
{
/* expiration time in 10us ticks */
p->resume_clock_interval = dt * 1e5;
vlib_process_start_switch_stack (vm, 0);
clib_longjmp (&p->return_longjmp, VLIB_PROCESS_RETURN_LONGJMP_SUSPEND);
}
else
vlib_process_finish_switch_stack (vm);
return r;
}
always_inline void
vlib_process_free_event_type (vlib_process_t * p, uword t,
uword is_one_time_event)
{
ASSERT (!pool_is_free_index (p->event_type_pool, t));
pool_put_index (p->event_type_pool, t);
if (is_one_time_event)
p->one_time_event_type_bitmap =
clib_bitmap_andnoti (p->one_time_event_type_bitmap, t);
}
always_inline void
vlib_process_maybe_free_event_type (vlib_process_t * p, uword t)
{
ASSERT (!pool_is_free_index (p->event_type_pool, t));
if (clib_bitmap_get (p->one_time_event_type_bitmap, t))
vlib_process_free_event_type (p, t, /* is_one_time_event */ 1);
}
always_inline void *
vlib_process_get_event_data (vlib_main_t * vm,
uword * return_event_type_opaque)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_process_t *p;
vlib_process_event_type_t *et;
uword t;
void *event_data_vector;
p = vec_elt (nm->processes, nm->current_process_index);
/* Find first type with events ready.
Return invalid type when there's nothing there. */
t = clib_bitmap_first_set (p->non_empty_event_type_bitmap);
if (t == ~0)
return 0;
p->non_empty_event_type_bitmap =
clib_bitmap_andnoti (p->non_empty_event_type_bitmap, t);
ASSERT (_vec_len (p->pending_event_data_by_type_index[t]) > 0);
event_data_vector = p->pending_event_data_by_type_index[t];
p->pending_event_data_by_type_index[t] = 0;
et = pool_elt_at_index (p->event_type_pool, t);
/* Return user's opaque value and possibly index. */
*return_event_type_opaque = et->opaque;
vlib_process_maybe_free_event_type (p, t);
return event_data_vector;
}
/* Return event data vector for later reuse. We reuse event data to avoid
repeatedly allocating event vectors in cases where we care about speed. */
always_inline void
vlib_process_put_event_data (vlib_main_t * vm, void *event_data)
{
vlib_node_main_t *nm = &vm->node_main;
vec_add1 (nm->recycled_event_data_vectors, event_data);
}
/** Return the first event type which has occurred and a vector of per-event
data of that type, or a timeout indication
@param vm - vlib_main_t pointer
@param data_vector - pointer to a (uword *) vector to receive event data
@returns either an event type and a vector of per-event instance data,
or ~0 to indicate a timeout.
*/
always_inline uword
vlib_process_get_events (vlib_main_t * vm, uword ** data_vector)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_process_t *p;
vlib_process_event_type_t *et;
uword r, t, l;
p = vec_elt (nm->processes, nm->current_process_index);
/* Find first type with events ready.
Return invalid type when there's nothing there. */
t = clib_bitmap_first_set (p->non_empty_event_type_bitmap);
if (t == ~0)
return t;
p->non_empty_event_type_bitmap =
clib_bitmap_andnoti (p->non_empty_event_type_bitmap, t);
l = _vec_len (p->pending_event_data_by_type_index[t]);
if (data_vector)
vec_add (*data_vector, p->pending_event_data_by_type_index[t], l);
_vec_len (p->pending_event_data_by_type_index[t]) = 0;
et = pool_elt_at_index (p->event_type_pool, t);
/* Return user's opaque value. */
r = et->opaque;
vlib_process_maybe_free_event_type (p, t);
return r;
}
always_inline uword
vlib_process_get_events_helper (vlib_process_t * p, uword t,
uword ** data_vector)
{
uword l;
p->non_empty_event_type_bitmap =
clib_bitmap_andnoti (p->non_empty_event_type_bitmap, t);
l = _vec_len (p->pending_event_data_by_type_index[t]);
if (data_vector)
vec_add (*data_vector, p->pending_event_data_by_type_index[t], l);
_vec_len (p->pending_event_data_by_type_index[t]) = 0;
vlib_process_maybe_free_event_type (p, t);
return l;
}
/* As above but query as specified type of event. Returns number of
events found. */
always_inline uword
vlib_process_get_events_with_type (vlib_main_t * vm, uword ** data_vector,
uword with_type_opaque)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_process_t *p;
uword t, *h;
p = vec_elt (nm->processes, nm->current_process_index);
h = hash_get (p->event_type_index_by_type_opaque, with_type_opaque);
if (!h)
/* This can happen when an event has not yet been
signaled with given opaque type. */
return 0;
t = h[0];
if (!clib_bitmap_get (p->non_empty_event_type_bitmap, t))
return 0;
return vlib_process_get_events_helper (p, t, data_vector);
}
always_inline uword *
vlib_process_wait_for_event (vlib_main_t * vm)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_process_t *p;
uword r;
p = vec_elt (nm->processes, nm->current_process_index);
if (clib_bitmap_is_zero (p->non_empty_event_type_bitmap))
{
p->flags |= VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_EVENT;
r =
clib_setjmp (&p->resume_longjmp, VLIB_PROCESS_RESUME_LONGJMP_SUSPEND);
if (r == VLIB_PROCESS_RESUME_LONGJMP_SUSPEND)
{
vlib_process_start_switch_stack (vm, 0);
clib_longjmp (&p->return_longjmp,
VLIB_PROCESS_RETURN_LONGJMP_SUSPEND);
}
else
vlib_process_finish_switch_stack (vm);
}
return p->non_empty_event_type_bitmap;
}
always_inline uword
vlib_process_wait_for_one_time_event (vlib_main_t * vm,
uword ** data_vector,
uword with_type_index)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_process_t *p;
uword r;
p = vec_elt (nm->processes, nm->current_process_index);
ASSERT (!pool_is_free_index (p->event_type_pool, with_type_index));
while (!clib_bitmap_get (p->non_empty_event_type_bitmap, with_type_index))
{
p->flags |= VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_EVENT;
r =
clib_setjmp (&p->resume_longjmp, VLIB_PROCESS_RESUME_LONGJMP_SUSPEND);
if (r == VLIB_PROCESS_RESUME_LONGJMP_SUSPEND)
{
vlib_process_start_switch_stack (vm, 0);
clib_longjmp (&p->return_longjmp,
VLIB_PROCESS_RETURN_LONGJMP_SUSPEND);
}
else
vlib_process_finish_switch_stack (vm);
}
return vlib_process_get_events_helper (p, with_type_index, data_vector);
}
always_inline uword
vlib_process_wait_for_event_with_type (vlib_main_t * vm,
uword ** data_vector,
uword with_type_opaque)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_process_t *p;
uword r, *h;
p = vec_elt (nm->processes, nm->current_process_index);
h = hash_get (p->event_type_index_by_type_opaque, with_type_opaque);
while (!h || !clib_bitmap_get (p->non_empty_event_type_bitmap, h[0]))
{
p->flags |= VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_EVENT;
r =
clib_setjmp (&p->resume_longjmp, VLIB_PROCESS_RESUME_LONGJMP_SUSPEND);
if (r == VLIB_PROCESS_RESUME_LONGJMP_SUSPEND)
{
vlib_process_start_switch_stack (vm, 0);
clib_longjmp (&p->return_longjmp,
VLIB_PROCESS_RETURN_LONGJMP_SUSPEND);
}
else
vlib_process_finish_switch_stack (vm);
/* See if unknown event type has been signaled now. */
if (!h)
h = hash_get (p->event_type_index_by_type_opaque, with_type_opaque);
}
return vlib_process_get_events_helper (p, h[0], data_vector);
}
/** Suspend a cooperative multi-tasking thread
Waits for an event, or for the indicated number of seconds to elapse
@param vm - vlib_main_t pointer
@param dt - timeout, in seconds.
@returns the remaining time interval
*/
always_inline f64
vlib_process_wait_for_event_or_clock (vlib_main_t * vm, f64 dt)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_process_t *p;
f64 wakeup_time;
uword r;
p = vec_elt (nm->processes, nm->current_process_index);
if (vlib_process_suspend_time_is_zero (dt)
|| !clib_bitmap_is_zero (p->non_empty_event_type_bitmap))
return dt;
wakeup_time = vlib_time_now (vm) + dt;
/* Suspend waiting for both clock and event to occur. */
p->flags |= (VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_EVENT
| VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_CLOCK);
r = clib_setjmp (&p->resume_longjmp, VLIB_PROCESS_RESUME_LONGJMP_SUSPEND);
if (r == VLIB_PROCESS_RESUME_LONGJMP_SUSPEND)
{
p->resume_clock_interval = dt * 1e5;
vlib_process_start_switch_stack (vm, 0);
clib_longjmp (&p->return_longjmp, VLIB_PROCESS_RETURN_LONGJMP_SUSPEND);
}
else
vlib_process_finish_switch_stack (vm);
/* Return amount of time still left to sleep.
If <= 0 then we've been waken up by the clock (and not an event). */
return wakeup_time - vlib_time_now (vm);
}
always_inline vlib_process_event_type_t *
vlib_process_new_event_type (vlib_process_t * p, uword with_type_opaque)
{
vlib_process_event_type_t *et;
pool_get (p->event_type_pool, et);
et->opaque = with_type_opaque;
return et;
}
always_inline uword
vlib_process_create_one_time_event (vlib_main_t * vm, uword node_index,
uword with_type_opaque)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n = vlib_get_node (vm, node_index);
vlib_process_t *p = vec_elt (nm->processes, n->runtime_index);
vlib_process_event_type_t *et;
uword t;
et = vlib_process_new_event_type (p, with_type_opaque);
t = et - p->event_type_pool;
p->one_time_event_type_bitmap =
clib_bitmap_ori (p->one_time_event_type_bitmap, t);
return t;
}
always_inline void
vlib_process_delete_one_time_event (vlib_main_t * vm, uword node_index,
uword t)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n = vlib_get_node (vm, node_index);
vlib_process_t *p = vec_elt (nm->processes, n->runtime_index);
ASSERT (clib_bitmap_get (p->one_time_event_type_bitmap, t));
vlib_process_free_event_type (p, t, /* is_one_time_event */ 1);
}
always_inline void *
vlib_process_signal_event_helper (vlib_node_main_t * nm,
vlib_node_t * n,
vlib_process_t * p,
uword t,
uword n_data_elts, uword n_data_elt_bytes)
{
uword p_flags, add_to_pending, delete_from_wheel;
void *data_to_be_written_by_caller;
ASSERT (n->type == VLIB_NODE_TYPE_PROCESS);
ASSERT (!pool_is_free_index (p->event_type_pool, t));
vec_validate (p->pending_event_data_by_type_index, t);
/* Resize data vector and return caller's data to be written. */
{
void *data_vec = p->pending_event_data_by_type_index[t];
uword l;
if (!data_vec && vec_len (nm->recycled_event_data_vectors))
{
data_vec = vec_pop (nm->recycled_event_data_vectors);
_vec_len (data_vec) = 0;
}
l = vec_len (data_vec);
data_vec = _vec_resize (data_vec,
/* length_increment */ n_data_elts,
/* total size after increment */
(l + n_data_elts) * n_data_elt_bytes,
/* header_bytes */ 0, /* data_align */ 0);
p->pending_event_data_by_type_index[t] = data_vec;
data_to_be_written_by_caller = data_vec + l * n_data_elt_bytes;
}
p->non_empty_event_type_bitmap =
clib_bitmap_ori (p->non_empty_event_type_bitmap, t);
p_flags = p->flags;
/* Event was already signalled? */
add_to_pending = (p_flags & VLIB_PROCESS_RESUME_PENDING) == 0;
/* Process will resume when suspend time elapses? */
delete_from_wheel = 0;
if (p_flags & VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_CLOCK)
{
/* Waiting for both event and clock? */
if (p_flags & VLIB_PROCESS_IS_SUSPENDED_WAITING_FOR_EVENT)
{
if (!TW (tw_timer_handle_is_free)
((TWT (tw_timer_wheel) *) nm->timing_wheel,
p->stop_timer_handle))
delete_from_wheel = 1;
else
/* timer just popped so process should already be on the list */
add_to_pending = 0;
}
else
/* Waiting only for clock. Event will be queue and may be
handled when timer expires. */
add_to_pending = 0;
}
/* Never add current process to pending vector since current process is
already running. */
add_to_pending &= nm->current_process_index != n->runtime_index;
if (add_to_pending)
{
u32 x = vlib_timing_wheel_data_set_suspended_process (n->runtime_index);
p->flags = p_flags | VLIB_PROCESS_RESUME_PENDING;
vec_add1 (nm->data_from_advancing_timing_wheel, x);
if (delete_from_wheel)
TW (tw_timer_stop) ((TWT (tw_timer_wheel) *) nm->timing_wheel,
p->stop_timer_handle);
}
return data_to_be_written_by_caller;
}
always_inline void *
vlib_process_signal_event_data (vlib_main_t * vm,
uword node_index,
uword type_opaque,
uword n_data_elts, uword n_data_elt_bytes)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n = vlib_get_node (vm, node_index);
vlib_process_t *p = vec_elt (nm->processes, n->runtime_index);
uword *h, t;
/* Must be in main thread */
ASSERT (vlib_get_thread_index () == 0);
h = hash_get (p->event_type_index_by_type_opaque, type_opaque);
if (!h)
{
vlib_process_event_type_t *et =
vlib_process_new_event_type (p, type_opaque);
t = et - p->event_type_pool;
hash_set (p->event_type_index_by_type_opaque, type_opaque, t);
}
else
t = h[0];
return vlib_process_signal_event_helper (nm, n, p, t, n_data_elts,
n_data_elt_bytes);
}
always_inline void *
vlib_process_signal_event_at_time (vlib_main_t * vm,
f64 dt,
uword node_index,
uword type_opaque,
uword n_data_elts, uword n_data_elt_bytes)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n = vlib_get_node (vm, node_index);
vlib_process_t *p = vec_elt (nm->processes, n->runtime_index);
uword *h, t;
h = hash_get (p->event_type_index_by_type_opaque, type_opaque);
if (!h)
{
vlib_process_event_type_t *et =
vlib_process_new_event_type (p, type_opaque);
t = et - p->event_type_pool;
hash_set (p->event_type_index_by_type_opaque, type_opaque, t);
}
else
t = h[0];
if (vlib_process_suspend_time_is_zero (dt))
return vlib_process_signal_event_helper (nm, n, p, t, n_data_elts,
n_data_elt_bytes);
else
{
vlib_signal_timed_event_data_t *te;
pool_get_aligned (nm->signal_timed_event_data_pool, te, sizeof (te[0]));
te->n_data_elts = n_data_elts;
te->n_data_elt_bytes = n_data_elt_bytes;
te->n_data_bytes = n_data_elts * n_data_elt_bytes;
/* Assert that structure fields are big enough. */
ASSERT (te->n_data_elts == n_data_elts);
ASSERT (te->n_data_elt_bytes == n_data_elt_bytes);
ASSERT (te->n_data_bytes == n_data_elts * n_data_elt_bytes);
te->process_node_index = n->runtime_index;
te->event_type_index = t;
p->stop_timer_handle =
TW (tw_timer_start) ((TWT (tw_timer_wheel) *) nm->timing_wheel,
vlib_timing_wheel_data_set_timed_event
(te - nm->signal_timed_event_data_pool),
0 /* timer_id */ ,
(vlib_time_now (vm) + dt) * 1e5);
/* Inline data big enough to hold event? */
if (te->n_data_bytes < sizeof (te->inline_event_data))
return te->inline_event_data;
else
{
te->event_data_as_vector = 0;
vec_resize (te->event_data_as_vector, te->n_data_bytes);
return te->event_data_as_vector;
}
}
}
always_inline void *
vlib_process_signal_one_time_event_data (vlib_main_t * vm,
uword node_index,
uword type_index,
uword n_data_elts,
uword n_data_elt_bytes)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n = vlib_get_node (vm, node_index);
vlib_process_t *p = vec_elt (nm->processes, n->runtime_index);
return vlib_process_signal_event_helper (nm, n, p, type_index, n_data_elts,
n_data_elt_bytes);
}
always_inline void
vlib_process_signal_event (vlib_main_t * vm,
uword node_index, uword type_opaque, uword data)
{
uword *d = vlib_process_signal_event_data (vm, node_index, type_opaque,
1 /* elts */ , sizeof (uword));
d[0] = data;
}
always_inline void
vlib_process_signal_event_pointer (vlib_main_t * vm,
uword node_index,
uword type_opaque, void *data)
{
void **d = vlib_process_signal_event_data (vm, node_index, type_opaque,
1 /* elts */ , sizeof (data));
d[0] = data;
}
/**
* Signal event to process from any thread.
*
* When in doubt, use this.
*/
always_inline void
vlib_process_signal_event_mt (vlib_main_t * vm,
uword node_index, uword type_opaque, uword data)
{
if (vlib_get_thread_index () != 0)
{
vlib_process_signal_event_mt_args_t args = {
.node_index = node_index,
.type_opaque = type_opaque,
.data = data,
};
vlib_rpc_call_main_thread (vlib_process_signal_event_mt_helper,
(u8 *) & args, sizeof (args));
}
else
vlib_process_signal_event (vm, node_index, type_opaque, data);
}
always_inline void
vlib_process_signal_one_time_event (vlib_main_t * vm,
uword node_index,
uword type_index, uword data)
{
uword *d =
vlib_process_signal_one_time_event_data (vm, node_index, type_index,
1 /* elts */ , sizeof (uword));
d[0] = data;
}
always_inline void
vlib_signal_one_time_waiting_process (vlib_main_t * vm,
vlib_one_time_waiting_process_t * p)
{
vlib_process_signal_one_time_event (vm, p->node_index, p->one_time_event,
/* data */ ~0);
clib_memset (p, ~0, sizeof (p[0]));
}
always_inline void
vlib_signal_one_time_waiting_process_vector (vlib_main_t * vm,
vlib_one_time_waiting_process_t
** wps)
{
vlib_one_time_waiting_process_t *wp;
vec_foreach (wp, *wps) vlib_signal_one_time_waiting_process (vm, wp);
vec_free (*wps);
}
always_inline void
vlib_current_process_wait_for_one_time_event (vlib_main_t * vm,
vlib_one_time_waiting_process_t
* p)
{
p->node_index = vlib_current_process (vm);
p->one_time_event = vlib_process_create_one_time_event (vm, p->node_index, /* type opaque */
~0);
vlib_process_wait_for_one_time_event (vm,
/* don't care about data */ 0,
p->one_time_event);
}
always_inline void
vlib_current_process_wait_for_one_time_event_vector (vlib_main_t * vm,
vlib_one_time_waiting_process_t
** wps)
{
vlib_one_time_waiting_process_t *wp;
vec_add2 (*wps, wp, 1);
vlib_current_process_wait_for_one_time_event (vm, wp);
}
always_inline u32
vlib_node_runtime_update_main_loop_vector_stats (vlib_main_t * vm,
vlib_node_runtime_t * node,
uword n_vectors)
{
u32 i, d, vi0, vi1;
u32 i0, i1;
ASSERT (is_pow2 (ARRAY_LEN (node->main_loop_vector_stats)));
i = ((vm->main_loop_count >> VLIB_LOG2_MAIN_LOOPS_PER_STATS_UPDATE)
& (ARRAY_LEN (node->main_loop_vector_stats) - 1));
i0 = i ^ 0;
i1 = i ^ 1;
d = ((vm->main_loop_count >> VLIB_LOG2_MAIN_LOOPS_PER_STATS_UPDATE)
-
(node->main_loop_count_last_dispatch >>
VLIB_LOG2_MAIN_LOOPS_PER_STATS_UPDATE));
vi0 = node->main_loop_vector_stats[i0];
vi1 = node->main_loop_vector_stats[i1];
vi0 = d == 0 ? vi0 : 0;
vi1 = d <= 1 ? vi1 : 0;
vi0 += n_vectors;
node->main_loop_vector_stats[i0] = vi0;
node->main_loop_vector_stats[i1] = vi1;
node->main_loop_count_last_dispatch = vm->main_loop_count;
/* Return previous counter. */
return node->main_loop_vector_stats[i1];
}
always_inline f64
vlib_node_vectors_per_main_loop_as_float (vlib_main_t * vm, u32 node_index)
{
vlib_node_runtime_t *rt = vlib_node_get_runtime (vm, node_index);
u32 v;
v = vlib_node_runtime_update_main_loop_vector_stats (vm, rt, /* n_vectors */
0);
return (f64) v / (1 << VLIB_LOG2_MAIN_LOOPS_PER_STATS_UPDATE);
}
always_inline u32
vlib_node_vectors_per_main_loop_as_integer (vlib_main_t * vm, u32 node_index)
{
vlib_node_runtime_t *rt = vlib_node_get_runtime (vm, node_index);
u32 v;
v = vlib_node_runtime_update_main_loop_vector_stats (vm, rt, /* n_vectors */
0);
return v >> VLIB_LOG2_MAIN_LOOPS_PER_STATS_UPDATE;
}
void
vlib_frame_free (vlib_main_t * vm, vlib_node_runtime_t * r, vlib_frame_t * f);
/* Return the edge index if present, ~0 otherwise */
uword vlib_node_get_next (vlib_main_t * vm, uword node, uword next_node);
/* Add next node to given node in given slot. */
uword
vlib_node_add_next_with_slot (vlib_main_t * vm,
uword node, uword next_node, uword slot);
/* As above but adds to end of node's next vector. */
always_inline uword
vlib_node_add_next (vlib_main_t * vm, uword node, uword next_node)
{
return vlib_node_add_next_with_slot (vm, node, next_node, ~0);
}
/* Add next node to given node in given slot. */
uword
vlib_node_add_named_next_with_slot (vlib_main_t * vm,
uword node, char *next_name, uword slot);
/* As above but adds to end of node's next vector. */
always_inline uword
vlib_node_add_named_next (vlib_main_t * vm, uword node, char *name)
{
return vlib_node_add_named_next_with_slot (vm, node, name, ~0);
}
/**
* Get list of nodes
*/
void
vlib_node_get_nodes (vlib_main_t * vm, u32 max_threads, int include_stats,
int barrier_sync, vlib_node_t **** node_dupsp,
vlib_main_t *** stat_vmsp);
/* Query node given name. */
vlib_node_t *vlib_get_node_by_name (vlib_main_t * vm, u8 * name);
/* Rename a node. */
void vlib_node_rename (vlib_main_t * vm, u32 node_index, char *fmt, ...);
/* Register new packet processing node. Nodes can be registered
dynamically via this call or statically via the VLIB_REGISTER_NODE
macro. */
u32 vlib_register_node (vlib_main_t * vm, vlib_node_registration_t * r);
/* Register all static nodes registered via VLIB_REGISTER_NODE. */
void vlib_register_all_static_nodes (vlib_main_t * vm);
/* Start a process. */
void vlib_start_process (vlib_main_t * vm, uword process_index);
/* Sync up runtime and main node stats. */
void vlib_node_sync_stats (vlib_main_t * vm, vlib_node_t * n);
/* Node graph initialization function. */
clib_error_t *vlib_node_main_init (vlib_main_t * vm);
format_function_t format_vlib_node_graph;
format_function_t format_vlib_node_name;
format_function_t format_vlib_next_node_name;
format_function_t format_vlib_node_and_next;
format_function_t format_vlib_cpu_time;
format_function_t format_vlib_time;
/* Parse node name -> node index. */
unformat_function_t unformat_vlib_node;
always_inline void
vlib_node_increment_counter (vlib_main_t * vm, u32 node_index,
u32 counter_index, u64 increment)
{
vlib_node_t *n = vlib_get_node (vm, node_index);
vlib_error_main_t *em = &vm->error_main;
u32 node_counter_base_index = n->error_heap_index;
em->counters[node_counter_base_index + counter_index] += increment;
}
/** @brief Create a vlib process
* @param vm &vlib_global_main
* @param f the process node function
* @param log2_n_stack_bytes size of the process stack, defaults to 16K
* @return newly-create node index
* @warning call only on the main thread. Barrier sync required
*/
u32 vlib_process_create (vlib_main_t * vm, char *name,
vlib_node_function_t * f, u32 log2_n_stack_bytes);
#endif /* included_vlib_node_funcs_h */
/*
* fd.io coding-style-patch-verification: ON
*
* Local Variables:
* eval: (c-set-style "gnu")
* End:
*/