aboutsummaryrefslogtreecommitdiffstats
path: root/src
diff options
context:
space:
mode:
authorDamjan Marion <damarion@cisco.com>2017-11-02 17:07:59 +0100
committerDamjan Marion <dmarion.lists@gmail.com>2018-03-22 15:54:43 +0000
commit5c37ce3e0264c0bec75610837c5819ff4407bd5c (patch)
tree445c8d5f9fa2ee7bdb1f8fb6e1228c1bcbf633d9 /src
parent32f4e18c59f368e9c43f4483de12353280c2149b (diff)
memif: version 2
In version 1 of the protocol sender was always ring producer and receiver was consumer. In version 2 slave is always producer, and in case of master-to-slave rings, slave is responsible for populating ring with empty buffers. As this is major change, we need to bump version number. In addition, descriptor size is reduced to 16 bytes. This change allows zero-copy-slave operation (to be privided in the separate patch). Change-Id: I02115d232f455ffc05c0bd247f7d03f47252cfaf Signed-off-by: Damjan Marion <damarion@cisco.com> Signed-off-by: Jakub Grajciar <jgrajcia@cisco.com>
Diffstat (limited to 'src')
-rw-r--r--src/plugins/memif/cli.c8
-rw-r--r--src/plugins/memif/device.c265
-rw-r--r--src/plugins/memif/memif.c39
-rw-r--r--src/plugins/memif/memif.h14
-rw-r--r--src/plugins/memif/node.c685
-rw-r--r--src/plugins/memif/private.h33
6 files changed, 558 insertions, 486 deletions
diff --git a/src/plugins/memif/cli.c b/src/plugins/memif/cli.c
index 6a149f9fb94..ed2b1b7e6d0 100644
--- a/src/plugins/memif/cli.c
+++ b/src/plugins/memif/cli.c
@@ -392,17 +392,17 @@ format_memif_descriptor (u8 * s, va_list * args)
s = format (s, "%Udescriptor table:\n", format_white_space, indent);
s =
format (s,
- "%Uid flags buf len desc len address offset user address\n",
+ "%Uid flags len address offset user address\n",
format_white_space, indent);
s =
format (s,
- "%U===== ===== ======= ======== ================== ====== ==================\n",
+ "%U===== ===== ======== ================== ====== ==================\n",
format_white_space, indent);
for (slot = 0; slot < ring_size; slot++)
{
- s = format (s, "%U%-5d %-5d %-7d %-7d 0x%016lx %-6d 0x%016lx\n",
+ s = format (s, "%U%-5d %-5d %-7d 0x%016lx %-6d 0x%016lx\n",
format_white_space, indent, slot,
- ring->desc[slot].flags, ring->desc[slot].buffer_length,
+ ring->desc[slot].flags,
ring->desc[slot].length,
mif->regions[ring->desc[slot].region].shm,
ring->desc[slot].offset, memif_get_buffer (mif, ring,
diff --git a/src/plugins/memif/device.c b/src/plugins/memif/device.c
index 22f9753dc42..112db57b4b4 100644
--- a/src/plugins/memif/device.c
+++ b/src/plugins/memif/device.c
@@ -30,8 +30,7 @@
#define foreach_memif_tx_func_error \
_(NO_FREE_SLOTS, "no free tx slots") \
-_(TRUNC_PACKET, "packet > buffer size -- truncated in tx ring") \
-_(PENDING_MSGS, "pending msgs in tx ring")
+_(ROLLBACK, "no enough space in tx buffers")
typedef enum
{
@@ -86,77 +85,15 @@ format_memif_tx_trace (u8 * s, va_list * args)
}
static_always_inline void
-memif_prefetch_buffer_and_data (vlib_main_t * vm, u32 bi)
+memif_add_copy_op (memif_per_thread_data_t * ptd, void *data, u32 len,
+ u16 buffer_offset, u16 buffer_vec_index)
{
- vlib_buffer_t *b = vlib_get_buffer (vm, bi);
- vlib_prefetch_buffer_header (b, LOAD);
- CLIB_PREFETCH (b->data, CLIB_CACHE_LINE_BYTES, LOAD);
-}
-
-/**
- * @brief Copy buffer to tx ring
- *
- * @param * vm (in)
- * @param * node (in)
- * @param * mif (in) pointer to memif interface
- * @param bi (in) vlib buffer index
- * @param * ring (in) pointer to memif ring
- * @param * head (in/out) ring head
- * @param mask (in) ring size - 1
- */
-static_always_inline void
-memif_copy_buffer_to_tx_ring (vlib_main_t * vm, vlib_node_runtime_t * node,
- memif_if_t * mif, u32 bi, memif_ring_t * ring,
- u16 * head, u16 mask)
-{
- vlib_buffer_t *b0;
- void *mb0;
- u32 total = 0, len;
- u16 slot = (*head) & mask;
-
- mb0 = memif_get_buffer (mif, ring, slot);
- ring->desc[slot].flags = 0;
- do
- {
- b0 = vlib_get_buffer (vm, bi);
- len = b0->current_length;
- if (PREDICT_FALSE (ring->desc[slot].buffer_length < (total + len)))
- {
- if (PREDICT_TRUE (total))
- {
- ring->desc[slot].length = total;
- total = 0;
- ring->desc[slot].flags |= MEMIF_DESC_FLAG_NEXT;
- (*head)++;
- slot = (*head) & mask;
- mb0 = memif_get_buffer (mif, ring, slot);
- ring->desc[slot].flags = 0;
- }
- }
- if (PREDICT_TRUE (ring->desc[slot].buffer_length >= (total + len)))
- {
- clib_memcpy (mb0 + total, vlib_buffer_get_current (b0),
- CLIB_CACHE_LINE_BYTES);
- if (len > CLIB_CACHE_LINE_BYTES)
- clib_memcpy (mb0 + CLIB_CACHE_LINE_BYTES + total,
- vlib_buffer_get_current (b0) + CLIB_CACHE_LINE_BYTES,
- len - CLIB_CACHE_LINE_BYTES);
- total += len;
- }
- else
- {
- vlib_error_count (vm, node->node_index, MEMIF_TX_ERROR_TRUNC_PACKET,
- 1);
- break;
- }
- }
- while ((bi = (b0->flags & VLIB_BUFFER_NEXT_PRESENT) ? b0->next_buffer : 0));
-
- if (PREDICT_TRUE (total))
- {
- ring->desc[slot].length = total;
- (*head)++;
- }
+ memif_copy_op_t *co;
+ vec_add2_aligned (ptd->copy_ops, co, 1, CLIB_CACHE_LINE_BYTES);
+ co->data = data;
+ co->data_len = len;
+ co->buffer_offset = buffer_offset;
+ co->buffer_vec_index = buffer_vec_index;
}
static_always_inline uword
@@ -168,13 +105,18 @@ memif_interface_tx_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
memif_ring_t *ring;
u32 *buffers = vlib_frame_args (frame);
u32 n_left = frame->n_vectors;
- u16 ring_size, mask;
- u16 head, tail;
- u16 free_slots;
+ u32 n_copy_op;
+ u16 ring_size, mask, slot, free_slots;
u32 thread_index = vlib_get_thread_index ();
+ memif_per_thread_data_t *ptd = vec_elt_at_index (memif_main.per_thread_data,
+ thread_index);
u8 tx_queues = vec_len (mif->tx_queues);
memif_queue_t *mq;
int n_retries = 5;
+ vlib_buffer_t *b0, *b1, *b2, *b3;
+ memif_copy_op_t *co;
+ memif_region_index_t last_region = ~0;
+ void *last_region_shm = 0;
if (tx_queues < vec_len (vlib_mains))
{
@@ -189,49 +131,158 @@ memif_interface_tx_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
ring = mq->ring;
ring_size = 1 << mq->log2_ring_size;
mask = ring_size - 1;
-retry:
-
- /* free consumed buffers */
- head = ring->head;
- tail = ring->tail;
+retry:
- free_slots = ring_size - head + tail;
+ free_slots = ring->tail - mq->last_tail;
+ mq->last_tail += free_slots;
+ slot = (type == MEMIF_RING_S2M) ? ring->head : ring->tail;
- while (n_left > 5 && free_slots > 1)
- {
- CLIB_PREFETCH (memif_get_buffer (mif, ring, (head + 2) & mask),
- CLIB_CACHE_LINE_BYTES, STORE);
- CLIB_PREFETCH (memif_get_buffer (mif, ring, (head + 3) & mask),
- CLIB_CACHE_LINE_BYTES, STORE);
- CLIB_PREFETCH (&ring->desc[(head + 4) & mask], CLIB_CACHE_LINE_BYTES,
- STORE);
- CLIB_PREFETCH (&ring->desc[(head + 5) & mask], CLIB_CACHE_LINE_BYTES,
- STORE);
- memif_prefetch_buffer_and_data (vm, buffers[2]);
- memif_prefetch_buffer_and_data (vm, buffers[3]);
-
- memif_copy_buffer_to_tx_ring (vm, node, mif, buffers[0], ring, &head,
- mask);
- memif_copy_buffer_to_tx_ring (vm, node, mif, buffers[1], ring, &head,
- mask);
-
- buffers += 2;
- n_left -= 2;
- free_slots -= 2;
- }
+ if (type == MEMIF_RING_S2M)
+ free_slots = ring_size - ring->head + mq->last_tail;
+ else
+ free_slots = ring->head - ring->tail;
while (n_left && free_slots)
{
- memif_copy_buffer_to_tx_ring (vm, node, mif, buffers[0], ring, &head,
- mask);
+ memif_desc_t *d0;
+ void *mb0;
+ i32 src_off;
+ u32 bi0, dst_off, src_left, dst_left, bytes_to_copy;
+ u32 saved_ptd_copy_ops_len = _vec_len (ptd->copy_ops);
+ u32 saved_ptd_buffers_len = _vec_len (ptd->buffers);
+ u16 saved_slot = slot;
+
+ CLIB_PREFETCH (&ring->desc[(slot + 8) & mask], CLIB_CACHE_LINE_BYTES,
+ LOAD);
+
+ d0 = &ring->desc[slot & mask];
+ if (PREDICT_FALSE (last_region != d0->region))
+ {
+ last_region_shm = mif->regions[d0->region].shm;
+ last_region = d0->region;
+ }
+ mb0 = last_region_shm + d0->offset;
+
+ dst_off = 0;
+
+ /* slave is the producer, so it should be able to reset buffer length */
+ dst_left = (type == MEMIF_RING_S2M) ? mif->run.buffer_size : d0->length;
+
+ if (PREDICT_TRUE (n_left >= 4))
+ vlib_prefetch_buffer_header (vlib_get_buffer (vm, buffers[3]), LOAD);
+ bi0 = buffers[0];
+
+ next_in_chain:
+
+ b0 = vlib_get_buffer (vm, bi0);
+ src_off = b0->current_data;
+ src_left = b0->current_length;
+
+ while (src_left)
+ {
+ if (PREDICT_FALSE (dst_left == 0))
+ {
+ if (free_slots)
+ {
+ slot++;
+ free_slots--;
+ d0->flags = MEMIF_DESC_FLAG_NEXT;
+ d0 = &ring->desc[slot & mask];
+ dst_off = 0;
+ dst_left =
+ (type ==
+ MEMIF_RING_S2M) ? mif->run.buffer_size : d0->length;
+
+ if (PREDICT_FALSE (last_region != d0->region))
+ {
+ last_region_shm = mif->regions[d0->region].shm;
+ last_region = d0->region;
+ }
+ mb0 = last_region_shm + d0->offset;
+ }
+ else
+ {
+ /* we need to rollback vectors before bailing out */
+ _vec_len (ptd->buffers) = saved_ptd_buffers_len;
+ _vec_len (ptd->copy_ops) = saved_ptd_copy_ops_len;
+ vlib_error_count (vm, node->node_index,
+ MEMIF_TX_ERROR_ROLLBACK, 1);
+ slot = saved_slot;
+ goto no_free_slots;
+ }
+ }
+ bytes_to_copy = clib_min (src_left, dst_left);
+ memif_add_copy_op (ptd, mb0 + dst_off, bytes_to_copy, src_off,
+ vec_len (ptd->buffers));
+ vec_add1_aligned (ptd->buffers, bi0, CLIB_CACHE_LINE_BYTES);
+ src_off += bytes_to_copy;
+ dst_off += bytes_to_copy;
+ src_left -= bytes_to_copy;
+ dst_left -= bytes_to_copy;
+ }
+
+ if (PREDICT_FALSE (b0->flags & VLIB_BUFFER_NEXT_PRESENT))
+ {
+ bi0 = b0->next_buffer;
+ goto next_in_chain;
+ }
+
+ d0->length = dst_off;
+ d0->flags = 0;
+
+ free_slots -= 1;
+ slot += 1;
+
buffers++;
n_left--;
- free_slots--;
}
+no_free_slots:
+
+ /* copy data */
+ n_copy_op = vec_len (ptd->copy_ops);
+ co = ptd->copy_ops;
+ while (n_copy_op >= 8)
+ {
+ CLIB_PREFETCH (co[4].data, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (co[5].data, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (co[6].data, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (co[7].data, CLIB_CACHE_LINE_BYTES, LOAD);
+
+ b0 = vlib_get_buffer (vm, ptd->buffers[co[0].buffer_vec_index]);
+ b1 = vlib_get_buffer (vm, ptd->buffers[co[1].buffer_vec_index]);
+ b2 = vlib_get_buffer (vm, ptd->buffers[co[2].buffer_vec_index]);
+ b3 = vlib_get_buffer (vm, ptd->buffers[co[3].buffer_vec_index]);
+
+ clib_memcpy (co[0].data, b0->data + co[0].buffer_offset,
+ co[0].data_len);
+ clib_memcpy (co[1].data, b1->data + co[1].buffer_offset,
+ co[1].data_len);
+ clib_memcpy (co[2].data, b2->data + co[2].buffer_offset,
+ co[2].data_len);
+ clib_memcpy (co[3].data, b3->data + co[3].buffer_offset,
+ co[3].data_len);
+
+ co += 4;
+ n_copy_op -= 4;
+ }
+ while (n_copy_op)
+ {
+ b0 = vlib_get_buffer (vm, ptd->buffers[co[0].buffer_vec_index]);
+ clib_memcpy (co[0].data, b0->data + co[0].buffer_offset,
+ co[0].data_len);
+ co += 1;
+ n_copy_op -= 1;
+ }
+
+ vec_reset_length (ptd->copy_ops);
+ vec_reset_length (ptd->buffers);
CLIB_MEMORY_STORE_BARRIER ();
- ring->head = head;
+ if (type == MEMIF_RING_S2M)
+ ring->head = slot;
+ else
+ ring->tail = slot;
if (n_left && n_retries--)
goto retry;
diff --git a/src/plugins/memif/memif.c b/src/plugins/memif/memif.c
index d630f2a6182..c83a955090c 100644
--- a/src/plugins/memif/memif.c
+++ b/src/plugins/memif/memif.c
@@ -240,6 +240,8 @@ memif_connect (memif_if_t * mif)
if (rxmode == VNET_HW_INTERFACE_RX_MODE_POLLING)
mq->ring->flags |= MEMIF_RING_FLAG_MASK_INT;
+ else
+ vnet_device_input_set_interrupt_pending (vnm, mif->hw_if_index, i);
}
}
@@ -308,7 +310,7 @@ memif_init_regions_and_queues (memif_if_t * mif)
ring->desc[j].region = 0;
ring->desc[j].offset =
buffer_offset + (u32) (slot * mif->run.buffer_size);
- ring->desc[j].buffer_length = mif->run.buffer_size;
+ ring->desc[j].length = mif->run.buffer_size;
}
}
for (i = 0; i < mif->run.num_m2s_rings; i++)
@@ -323,7 +325,7 @@ memif_init_regions_and_queues (memif_if_t * mif)
ring->desc[j].region = 0;
ring->desc[j].offset =
buffer_offset + (u32) (slot * mif->run.buffer_size);
- ring->desc[j].buffer_length = mif->run.buffer_size;
+ ring->desc[j].length = mif->run.buffer_size;
}
}
@@ -722,6 +724,35 @@ memif_create_if (vlib_main_t * vm, memif_create_if_args_t * args)
DBG ("initializing socket file %s", msf->filename);
}
+ if (mm->per_thread_data == 0)
+ {
+ int i;
+ vlib_buffer_free_list_t *fl;
+
+ vec_validate_aligned (mm->per_thread_data, tm->n_vlib_mains - 1,
+ CLIB_CACHE_LINE_BYTES);
+
+ fl =
+ vlib_buffer_get_free_list (vm, VLIB_BUFFER_DEFAULT_FREE_LIST_INDEX);
+ for (i = 0; i < tm->n_vlib_mains; i++)
+ {
+ memif_per_thread_data_t *ptd =
+ vec_elt_at_index (mm->per_thread_data, i);
+ vlib_buffer_t *bt = &ptd->buffer_template;
+ vlib_buffer_init_for_free_list (bt, fl);
+ bt->flags = VLIB_BUFFER_TOTAL_LENGTH_VALID;
+ bt->total_length_not_including_first_buffer = 0;
+ vnet_buffer (bt)->sw_if_index[VLIB_TX] = (u32) ~ 0;
+
+ /* initially prealloc copy_ops so we can use
+ _vec_len instead of vec_elen */
+ vec_validate_aligned (ptd->copy_ops, 0, CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (ptd->copy_ops);
+ vec_validate_aligned (ptd->buffers, 0, CLIB_CACHE_LINE_BYTES);
+ vec_reset_length (ptd->buffers);
+ }
+ }
+
pool_get (mm->interfaces, mif);
memset (mif, 0, sizeof (*mif));
mif->dev_instance = mif - mm->interfaces;
@@ -860,7 +891,6 @@ static clib_error_t *
memif_init (vlib_main_t * vm)
{
memif_main_t *mm = &memif_main;
- vlib_thread_main_t *tm = vlib_get_thread_main ();
u8 *filename;
memset (mm, 0, sizeof (memif_main_t));
@@ -878,9 +908,6 @@ memif_init (vlib_main_t * vm)
MEMIF_DEFAULT_SOCKET_FILENAME, 0);
memif_add_socket_file (0, filename);
- vec_validate_aligned (mm->rx_buffers, tm->n_vlib_mains - 1,
- CLIB_CACHE_LINE_BYTES);
-
return 0;
}
diff --git a/src/plugins/memif/memif.h b/src/plugins/memif/memif.h
index 11918eabcde..38b54029e46 100644
--- a/src/plugins/memif/memif.h
+++ b/src/plugins/memif/memif.h
@@ -22,8 +22,8 @@
#define MEMIF_CACHELINE_SIZE 64
#endif
-#define MEMIF_COOKIE 0x3E31F10
-#define MEMIF_VERSION_MAJOR 1
+#define MEMIF_COOKIE 0x3E31F20
+#define MEMIF_VERSION_MAJOR 2
#define MEMIF_VERSION_MINOR 0
#define MEMIF_VERSION ((MEMIF_VERSION_MAJOR << 8) | MEMIF_VERSION_MINOR)
@@ -58,7 +58,7 @@ typedef enum
} memif_interface_mode_t;
typedef uint16_t memif_region_index_t;
-typedef uint64_t memif_region_offset_t;
+typedef uint32_t memif_region_offset_t;
typedef uint64_t memif_region_size_t;
typedef uint16_t memif_ring_index_t;
typedef uint32_t memif_interface_id_t;
@@ -148,15 +148,13 @@ typedef struct __attribute__ ((packed))
uint16_t flags;
#define MEMIF_DESC_FLAG_NEXT (1 << 0)
memif_region_index_t region;
- uint32_t buffer_length;
uint32_t length;
- uint8_t reserved[4];
memif_region_offset_t offset;
- uint64_t metadata;
+ uint32_t metadata;
} memif_desc_t;
-_Static_assert (sizeof (memif_desc_t) == 32,
- "Size of memif_dsct_t must be 32");
+_Static_assert (sizeof (memif_desc_t) == 16,
+ "Size of memif_dsct_t must be 16 bytes");
#define MEMIF_CACHELINE_ALIGN_MARK(mark) \
uint8_t mark[0] __attribute__((aligned(MEMIF_CACHELINE_SIZE)))
diff --git a/src/plugins/memif/node.c b/src/plugins/memif/node.c
index 83f1277637e..e8223d12435 100644
--- a/src/plugins/memif/node.c
+++ b/src/plugins/memif/node.c
@@ -31,6 +31,7 @@
#include <memif/private.h>
#define foreach_memif_input_error \
+ _(BUFFER_ALLOC_FAIL, "buffer allocation failed") \
_(NOT_IP, "not ip packet")
typedef enum
@@ -69,155 +70,75 @@ format_memif_input_trace (u8 * s, va_list * args)
return s;
}
-static_always_inline void
-memif_prefetch (vlib_main_t * vm, u32 bi)
-{
- vlib_buffer_t *b = vlib_get_buffer (vm, bi);
- vlib_prefetch_buffer_header (b, STORE);
- CLIB_PREFETCH (b->data, CLIB_CACHE_LINE_BYTES, STORE);
-}
-
-static_always_inline void
-memif_buffer_add_to_chain (vlib_main_t * vm, u32 bi, u32 first_bi,
- u32 prev_bi)
+static_always_inline u32
+memif_next_from_ip_hdr (vlib_node_runtime_t * node, vlib_buffer_t * b)
{
- vlib_buffer_t *b = vlib_get_buffer (vm, bi);
- vlib_buffer_t *first_b = vlib_get_buffer (vm, first_bi);
- vlib_buffer_t *prev_b = vlib_get_buffer (vm, prev_bi);
-
- /* update first buffer */
- first_b->total_length_not_including_first_buffer += b->current_length;
+ u8 *ptr = vlib_buffer_get_current (b);
+ u8 v = *ptr & 0xf0;
- /* update previous buffer */
- prev_b->next_buffer = bi;
- prev_b->flags |= VLIB_BUFFER_NEXT_PRESENT;
+ if (PREDICT_TRUE (v == 0x40))
+ return VNET_DEVICE_INPUT_NEXT_IP4_NCS_INPUT;
+ else if (PREDICT_TRUE (v == 0x60))
+ return VNET_DEVICE_INPUT_NEXT_IP6_INPUT;
- /* update current buffer */
- b->next_buffer = 0;
+ b->error = node->errors[MEMIF_INPUT_ERROR_NOT_IP];
+ return VNET_DEVICE_INPUT_NEXT_DROP;
}
-/**
- * @brief Copy buffer from rx ring
- *
- * @param * vm (in)
- * @param * mif (in) pointer to memif interface
- * @param * ring (in) pointer to memif ring
- * @param * rd (in) pointer to ring data
- * @param ring_size (in) ring size
- * @param * n_free_bufs (in/out) the number of free vlib buffers available
- * @param ** first_b (out) the first vlib buffer pointer
- * @param * first_bi (out) the first vlib buffer index
- * @param * bi (in/out) the current buffer index
- * #param * num_slots (in/out) the number of descriptors available to read
- *
- * @return total bytes read from rx ring also written to vlib buffers
- */
-static_always_inline uword
-memif_copy_buffer_from_rx_ring (vlib_main_t * vm, memif_if_t * mif,
- memif_ring_t * ring, memif_queue_t * mq,
- u16 ring_size, u32 n_buffer_bytes,
- u32 * n_free_bufs, vlib_buffer_t ** first_b,
- u32 * first_bi, u32 * bi, u16 * num_slots)
+static_always_inline void
+memif_trace_buffer (vlib_main_t * vm, vlib_node_runtime_t * node,
+ memif_if_t * mif, vlib_buffer_t * b, u32 next, u16 qid,
+ uword * n_tracep)
{
- memif_main_t *nm = &memif_main;
- u32 thread_index = vlib_get_thread_index ();
- u32 total_bytes = 0, offset = 0;
- u32 data_len;
- u32 bytes_to_copy;
- void *mb;
- vlib_buffer_t *b;
- u16 mask = ring_size - 1;
- u32 prev_bi;
- u16 last_head;
-
- while (*num_slots)
- {
- data_len = ring->desc[mq->last_head & mask].length;
- while (data_len && (*n_free_bufs))
- {
- /* get empty buffer */
- u32 last_buf = vec_len (nm->rx_buffers[thread_index]) - 1;
- prev_bi = *bi;
- *bi = nm->rx_buffers[thread_index][last_buf];
- b = vlib_get_buffer (vm, *bi);
- /* Clear the error first to ensure following node forget setting it */
- /* It will cause null-node error counter increasement instead of potential crash */
- b->error = 0x0;
- _vec_len (nm->rx_buffers[thread_index]) = last_buf;
- (*n_free_bufs)--;
- if (PREDICT_FALSE (*n_free_bufs == 0))
- {
- *n_free_bufs +=
- vlib_buffer_alloc (vm,
- &nm->rx_buffers[thread_index]
- [*n_free_bufs], ring_size);
- _vec_len (nm->rx_buffers[thread_index]) = *n_free_bufs;
- }
-
- if (last_buf > 4)
- {
- memif_prefetch (vm, nm->rx_buffers[thread_index][last_buf - 2]);
- memif_prefetch (vm, nm->rx_buffers[thread_index][last_buf - 3]);
- }
-
- /* copy buffer */
- bytes_to_copy =
- data_len > n_buffer_bytes ? n_buffer_bytes : data_len;
- b->current_data = 0;
- mb = memif_get_buffer (mif, ring, mq->last_head & mask);
- clib_memcpy (vlib_buffer_get_current (b), mb + offset,
- CLIB_CACHE_LINE_BYTES);
- if (bytes_to_copy > CLIB_CACHE_LINE_BYTES)
- clib_memcpy (vlib_buffer_get_current (b) + CLIB_CACHE_LINE_BYTES,
- mb + CLIB_CACHE_LINE_BYTES + offset,
- bytes_to_copy - CLIB_CACHE_LINE_BYTES);
-
- /* fill buffer header */
- b->current_length = bytes_to_copy;
-
- if (total_bytes == 0)
- {
- /* fill buffer metadata */
- b->total_length_not_including_first_buffer = 0;
- b->flags = VLIB_BUFFER_TOTAL_LENGTH_VALID;
- vnet_buffer (b)->sw_if_index[VLIB_RX] = mif->sw_if_index;
- vnet_buffer (b)->sw_if_index[VLIB_TX] = (u32) ~ 0;
- *first_bi = *bi;
- *first_b = vlib_get_buffer (vm, *first_bi);
- }
- else
- memif_buffer_add_to_chain (vm, *bi, *first_bi, prev_bi);
+ VLIB_BUFFER_TRACE_TRAJECTORY_INIT (b);
- offset += bytes_to_copy;
- total_bytes += bytes_to_copy;
- data_len -= bytes_to_copy;
- }
- last_head = mq->last_head;
- /* Advance to next descriptor */
- mq->last_head++;
- offset = 0;
- (*num_slots)--;
- if ((ring->desc[last_head & mask].flags & MEMIF_DESC_FLAG_NEXT) == 0)
- break;
+ if (PREDICT_TRUE (b != 0))
+ {
+ memif_input_trace_t *tr;
+ vlib_trace_buffer (vm, node, next, b, /* follow_chain */ 0);
+ vlib_set_trace_count (vm, node, --(*n_tracep));
+ tr = vlib_add_trace (vm, node, b, sizeof (*tr));
+ tr->next_index = next;
+ tr->hw_if_index = mif->hw_if_index;
+ tr->ring = qid;
}
-
- return (total_bytes);
}
+static_always_inline void
+memif_add_copy_op (memif_per_thread_data_t * ptd, void *data, u32 len,
+ u16 buffer_offset, u16 buffer_vec_index)
+{
+ memif_copy_op_t *co;
+ vec_add2_aligned (ptd->copy_ops, co, 1, CLIB_CACHE_LINE_BYTES);
+ co->data = data;
+ co->data_len = len;
+ co->buffer_offset = buffer_offset;
+ co->buffer_vec_index = buffer_vec_index;
+}
-static_always_inline u32
-memif_next_from_ip_hdr (vlib_node_runtime_t * node, vlib_buffer_t * b)
+static_always_inline void
+memif_add_to_chain (vlib_main_t * vm, vlib_buffer_t * b, u32 * buffers,
+ u32 buffer_size)
{
- u8 *ptr = vlib_buffer_get_current (b);
- u8 v = *ptr & 0xf0;
+ vlib_buffer_t *seg = b;
+ i32 bytes_left = b->current_length - buffer_size + b->current_data;
- if (PREDICT_TRUE (v == 0x40))
- return VNET_DEVICE_INPUT_NEXT_IP4_NCS_INPUT;
- else if (PREDICT_TRUE (v == 0x60))
- return VNET_DEVICE_INPUT_NEXT_IP6_INPUT;
+ if (PREDICT_TRUE (bytes_left <= 0))
+ return;
- b->error = node->errors[MEMIF_INPUT_ERROR_NOT_IP];
- return VNET_DEVICE_INPUT_NEXT_DROP;
+ b->current_length -= bytes_left;
+ b->total_length_not_including_first_buffer = bytes_left;
+
+ while (bytes_left)
+ {
+ seg->flags |= VLIB_BUFFER_NEXT_PRESENT;
+ seg->next_buffer = buffers[0];
+ seg = vlib_get_buffer (vm, buffers[0]);
+ buffers++;
+ seg->current_data = 0;
+ seg->current_length = clib_min (buffer_size, bytes_left);
+ bytes_left -= seg->current_length;
+ }
}
static_always_inline uword
@@ -227,301 +148,347 @@ memif_device_input_inline (vlib_main_t * vm, vlib_node_runtime_t * node,
memif_interface_mode_t mode)
{
vnet_main_t *vnm = vnet_get_main ();
+ memif_main_t *mm = &memif_main;
memif_ring_t *ring;
memif_queue_t *mq;
- u16 head;
+ u16 buffer_size = VLIB_BUFFER_DEFAULT_FREE_LIST_BYTES;
u32 next_index;
uword n_trace = vlib_get_trace_count (vm, node);
- memif_main_t *nm = &memif_main;
- u32 n_rx_packets = 0;
- u32 n_rx_bytes = 0;
- u32 *to_next = 0;
- u32 n_free_bufs;
- u32 b0_total, b1_total;
+ u32 n_rx_packets = 0, n_rx_bytes = 0;
+ u32 n_left, *to_next = 0;
+ u32 bi0, bi1, bi2, bi3;
+ vlib_buffer_t *b0, *b1, *b2, *b3;
u32 thread_index = vlib_get_thread_index ();
- u16 ring_size, mask, num_slots;
- u32 n_buffer_bytes = vlib_buffer_free_list_buffer_size (vm,
- VLIB_BUFFER_DEFAULT_FREE_LIST_INDEX);
+ memif_per_thread_data_t *ptd = vec_elt_at_index (mm->per_thread_data,
+ thread_index);
+ vlib_buffer_t *bt = &ptd->buffer_template;
+ u16 cur_slot, last_slot, ring_size, n_slots, mask;
+ i16 start_offset;
+ u16 n_buffers = 0, n_alloc;
+ memif_copy_op_t *co;
+ memif_packet_op_t *po;
+ memif_region_index_t last_region = ~0;
+ void *last_region_shm = 0;
mq = vec_elt_at_index (mif->rx_queues, qid);
ring = mq->ring;
ring_size = 1 << mq->log2_ring_size;
mask = ring_size - 1;
- if (mode == MEMIF_INTERFACE_MODE_IP)
+ next_index = (mode == MEMIF_INTERFACE_MODE_IP) ?
+ VNET_DEVICE_INPUT_NEXT_IP6_INPUT : VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT;
+
+ /* asume that somebody will want to add ethernet header on the packet
+ so start with IP header at offset 14 */
+ start_offset = (mode == MEMIF_INTERFACE_MODE_IP) ? 14 : 0;
+
+
+ /* for S2M rings, we are consumers of packet buffers, and for M2S rings we
+ are producers of empty buffers */
+ cur_slot = (type == MEMIF_RING_S2M) ? mq->last_head : mq->last_tail;
+ last_slot = (type == MEMIF_RING_S2M) ? ring->head : ring->tail;
+ if (cur_slot == last_slot)
+ goto refill;
+ n_slots = last_slot - cur_slot;
+
+ /* construct copy and packet vector out of ring slots */
+ while (n_slots && n_rx_packets < MEMIF_RX_VECTOR_SZ)
+ {
+ u32 dst_off, src_off, n_bytes_left;
+ u16 s0;
+ memif_desc_t *d0;
+ void *mb0;
+ po = ptd->packet_ops + n_rx_packets;
+ n_rx_packets++;
+ po->first_buffer_vec_index = n_buffers++;
+ po->packet_len = 0;
+ src_off = 0;
+ dst_off = start_offset;
+
+ next_slot:
+ CLIB_PREFETCH (&ring->desc[(cur_slot + 8) & mask],
+ CLIB_CACHE_LINE_BYTES, LOAD);
+ s0 = cur_slot & mask;
+ d0 = &ring->desc[s0];
+ n_bytes_left = d0->length;
+
+ /* slave resets buffer length,
+ * so it can produce full size buffer for master
+ */
+ if (type == MEMIF_RING_M2S)
+ d0->length = mif->run.buffer_size;
+
+ po->packet_len += n_bytes_left;
+ if (PREDICT_FALSE (last_region != d0->region))
+ {
+ last_region_shm = mif->regions[d0->region].shm;
+ last_region = d0->region;
+ }
+ mb0 = last_region_shm + d0->offset;
+
+ do
+ {
+ u32 dst_free = buffer_size - dst_off;
+ if (dst_free == 0)
+ {
+ dst_off = 0;
+ dst_free = buffer_size;
+ n_buffers++;
+ }
+ u32 bytes_to_copy = clib_min (dst_free, n_bytes_left);
+ memif_add_copy_op (ptd, mb0 + src_off, bytes_to_copy, dst_off,
+ n_buffers - 1);
+ n_bytes_left -= bytes_to_copy;
+ src_off += bytes_to_copy;
+ dst_off += bytes_to_copy;
+ }
+ while (PREDICT_FALSE (n_bytes_left));
+
+ cur_slot++;
+ n_slots--;
+ if ((d0->flags & MEMIF_DESC_FLAG_NEXT) && n_slots)
+ {
+ src_off = 0;
+ goto next_slot;
+ }
+ }
+
+ /* allocate free buffers */
+ vec_validate_aligned (ptd->buffers, n_buffers - 1, CLIB_CACHE_LINE_BYTES);
+ n_alloc = vlib_buffer_alloc (vm, ptd->buffers, n_buffers);
+ if (PREDICT_FALSE (n_alloc != n_buffers))
{
- next_index = VNET_DEVICE_INPUT_NEXT_IP6_INPUT;
+ if (n_alloc)
+ vlib_buffer_free (vm, ptd->buffers, n_alloc);
+ vlib_error_count (vm, node->node_index,
+ MEMIF_INPUT_ERROR_BUFFER_ALLOC_FAIL, 1);
+ goto refill;
}
- else
+
+ /* copy data */
+ n_left = vec_len (ptd->copy_ops);
+ co = ptd->copy_ops;
+ while (n_left >= 8)
+ {
+ CLIB_PREFETCH (co[4].data, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (co[5].data, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (co[6].data, CLIB_CACHE_LINE_BYTES, LOAD);
+ CLIB_PREFETCH (co[7].data, CLIB_CACHE_LINE_BYTES, LOAD);
+
+ b0 = vlib_get_buffer (vm, ptd->buffers[co[0].buffer_vec_index]);
+ b1 = vlib_get_buffer (vm, ptd->buffers[co[1].buffer_vec_index]);
+ b2 = vlib_get_buffer (vm, ptd->buffers[co[2].buffer_vec_index]);
+ b3 = vlib_get_buffer (vm, ptd->buffers[co[3].buffer_vec_index]);
+
+ clib_memcpy (b0->data + co[0].buffer_offset, co[0].data,
+ co[0].data_len);
+ clib_memcpy (b1->data + co[1].buffer_offset, co[1].data,
+ co[1].data_len);
+ clib_memcpy (b2->data + co[2].buffer_offset, co[2].data,
+ co[2].data_len);
+ clib_memcpy (b3->data + co[3].buffer_offset, co[3].data,
+ co[3].data_len);
+
+ co += 4;
+ n_left -= 4;
+ }
+ while (n_left)
{
- next_index = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT;
+ b0 = vlib_get_buffer (vm, ptd->buffers[co[0].buffer_vec_index]);
+ clib_memcpy (b0->data + co[0].buffer_offset, co[0].data,
+ co[0].data_len);
+ co += 1;
+ n_left -= 1;
}
- n_free_bufs = vec_len (nm->rx_buffers[thread_index]);
- if (PREDICT_FALSE (n_free_bufs < ring_size))
+ /* release slots from the ring */
+ if (type == MEMIF_RING_S2M)
{
- vec_validate (nm->rx_buffers[thread_index],
- ring_size + n_free_bufs - 1);
- n_free_bufs +=
- vlib_buffer_alloc (vm, &nm->rx_buffers[thread_index][n_free_bufs],
- ring_size);
- _vec_len (nm->rx_buffers[thread_index]) = n_free_bufs;
+ CLIB_MEMORY_STORE_BARRIER ();
+ ring->tail = mq->last_head = cur_slot;
+ }
+ else
+ {
+ mq->last_tail = cur_slot;
}
- head = ring->head;
- mq->last_head = ring->tail;
- if (head == mq->last_head)
- return 0;
+ u32 n_from = n_rx_packets;
+ po = ptd->packet_ops;
- num_slots = head - mq->last_head;
+ vnet_buffer (bt)->sw_if_index[VLIB_RX] = mif->sw_if_index;
+ bt->current_data = start_offset;
- while (num_slots)
+ while (n_from)
{
u32 n_left_to_next;
- u32 next0 = next_index;
- u32 next1 = next_index;
- vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+ u32 next0, next1, next2, next3;
- while (num_slots > 11 && n_left_to_next > 2)
+ vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
+ while (n_from >= 8 && n_left_to_next >= 4)
{
- CLIB_PREFETCH (memif_get_buffer
- (mif, ring, (mq->last_head + 2) & mask),
- CLIB_CACHE_LINE_BYTES, LOAD);
- CLIB_PREFETCH (memif_get_buffer
- (mif, ring, (mq->last_head + 3) & mask),
- CLIB_CACHE_LINE_BYTES, LOAD);
- CLIB_PREFETCH (&ring->desc[(mq->last_head + 4) & mask],
- CLIB_CACHE_LINE_BYTES, LOAD);
- CLIB_PREFETCH (&ring->desc[(mq->last_head + 5) & mask],
- CLIB_CACHE_LINE_BYTES, LOAD);
-
- vlib_buffer_t *first_b0 = 0;
- u32 bi0 = 0, first_bi0 = 0;
- b0_total = memif_copy_buffer_from_rx_ring (vm, mif, ring, mq,
- ring_size,
- n_buffer_bytes,
- &n_free_bufs, &first_b0,
- &first_bi0, &bi0,
- &num_slots);
-
- vlib_buffer_t *first_b1 = 0;
- u32 bi1 = 0, first_bi1 = 0;
- b1_total = memif_copy_buffer_from_rx_ring (vm, mif, ring, mq,
- ring_size,
- n_buffer_bytes,
- &n_free_bufs, &first_b1,
- &first_bi1, &bi1,
- &num_slots);
-
- if (PREDICT_FALSE (!first_bi0 || !first_bi1))
- {
- goto _invalid_pkt01;
- }
- /* enqueue buffer */
- to_next[0] = first_bi0;
- to_next[1] = first_bi1;
- to_next += 2;
- n_left_to_next -= 2;
+ b0 = vlib_get_buffer (vm, po[4].first_buffer_vec_index);
+ b1 = vlib_get_buffer (vm, po[5].first_buffer_vec_index);
+ b2 = vlib_get_buffer (vm, po[6].first_buffer_vec_index);
+ b3 = vlib_get_buffer (vm, po[7].first_buffer_vec_index);
+ vlib_prefetch_buffer_header (b0, STORE);
+ vlib_prefetch_buffer_header (b1, STORE);
+ vlib_prefetch_buffer_header (b2, STORE);
+ vlib_prefetch_buffer_header (b3, STORE);
+ /* enqueue buffer */
+ u32 fbvi0 = po[0].first_buffer_vec_index;
+ u32 fbvi1 = po[1].first_buffer_vec_index;
+ u32 fbvi2 = po[2].first_buffer_vec_index;
+ u32 fbvi3 = po[3].first_buffer_vec_index;
+ to_next[0] = bi0 = ptd->buffers[fbvi0];
+ to_next[1] = bi1 = ptd->buffers[fbvi1];
+ to_next[2] = bi2 = ptd->buffers[fbvi2];
+ to_next[3] = bi3 = ptd->buffers[fbvi3];
+ to_next += 4;
+ n_left_to_next -= 4;
+
+ b0 = vlib_get_buffer (vm, bi0);
+ b1 = vlib_get_buffer (vm, bi1);
+ b2 = vlib_get_buffer (vm, bi2);
+ b3 = vlib_get_buffer (vm, bi3);
+
+ clib_memcpy64_x4 (b0, b1, b2, b3, bt);
+
+ b0->current_length = po[0].packet_len;
+ b1->current_length = po[1].packet_len;
+ b2->current_length = po[2].packet_len;
+ b3->current_length = po[3].packet_len;
+
+ memif_add_to_chain (vm, b0, ptd->buffers + fbvi0 + 1, buffer_size);
+ memif_add_to_chain (vm, b1, ptd->buffers + fbvi1 + 1, buffer_size);
+ memif_add_to_chain (vm, b2, ptd->buffers + fbvi2 + 1, buffer_size);
+ memif_add_to_chain (vm, b3, ptd->buffers + fbvi3 + 1, buffer_size);
if (mode == MEMIF_INTERFACE_MODE_IP)
{
- next0 = memif_next_from_ip_hdr (node, first_b0);
- next1 = memif_next_from_ip_hdr (node, first_b1);
+ next0 = memif_next_from_ip_hdr (node, b0);
+ next1 = memif_next_from_ip_hdr (node, b1);
+ next2 = memif_next_from_ip_hdr (node, b2);
+ next3 = memif_next_from_ip_hdr (node, b3);
}
else if (mode == MEMIF_INTERFACE_MODE_ETHERNET)
{
if (PREDICT_FALSE (mif->per_interface_next_index != ~0))
- next0 = next1 = mif->per_interface_next_index;
- else
- /* redirect if feature path
- * enabled */
- vnet_feature_start_device_input_x2 (mif->sw_if_index,
- &next0, &next1,
- first_b0, first_b1);
- }
-
- /* trace */
- VLIB_BUFFER_TRACE_TRAJECTORY_INIT (first_b0);
- VLIB_BUFFER_TRACE_TRAJECTORY_INIT (first_b1);
-
- if (PREDICT_FALSE (n_trace > 0))
- {
- /* b0 */
- if (PREDICT_TRUE (first_b0 != 0))
{
- memif_input_trace_t *tr;
- vlib_trace_buffer (vm, node, next0, first_b0,
- /* follow_chain */ 0);
- vlib_set_trace_count (vm, node, --n_trace);
- tr = vlib_add_trace (vm, node, first_b0, sizeof (*tr));
- tr->next_index = next0;
- tr->hw_if_index = mif->hw_if_index;
- tr->ring = qid;
+ next0 = mif->per_interface_next_index;
+ next1 = mif->per_interface_next_index;
+ next2 = mif->per_interface_next_index;
+ next3 = mif->per_interface_next_index;
}
- if (n_trace)
+ else
{
- /* b1 */
- if (PREDICT_TRUE (first_b1 != 0))
- {
- memif_input_trace_t *tr;
- vlib_trace_buffer (vm, node, next1, first_b1,
- /* follow_chain */ 0);
- vlib_set_trace_count (vm, node, --n_trace);
- tr = vlib_add_trace (vm, node, first_b1, sizeof (*tr));
- tr->next_index = next1;
- tr->hw_if_index = mif->hw_if_index;
- tr->ring = qid;
- }
+ next0 = next1 = next2 = next3 = next_index;
+ /* redirect if feature path enabled */
+ vnet_feature_start_device_input_x1 (mif->sw_if_index,
+ &next0, b0);
+ vnet_feature_start_device_input_x1 (mif->sw_if_index,
+ &next1, b1);
+ vnet_feature_start_device_input_x1 (mif->sw_if_index,
+ &next2, b2);
+ vnet_feature_start_device_input_x1 (mif->sw_if_index,
+ &next3, b3);
}
}
- /* enqueue */
- vlib_validate_buffer_enqueue_x2 (vm, node, next_index, to_next,
- n_left_to_next, first_bi0,
- first_bi1, next0, next1);
-
- /* next packet */
- n_rx_packets += 2;
- n_rx_bytes += b0_total + b1_total;
-
- continue;
- _invalid_pkt01:
- if (!first_bi0 && !first_bi1)
- {
- continue;
- }
- if (first_bi1)
- {
- first_bi0 = first_bi1;
- first_b0 = first_b1;
- bi0 = bi1;
- b0_total = b1_total;
- }
-
- if (mode == MEMIF_INTERFACE_MODE_IP)
- {
- next0 = memif_next_from_ip_hdr (node, first_b0);
- }
- else if (mode == MEMIF_INTERFACE_MODE_ETHERNET)
- {
- if (PREDICT_FALSE (mif->per_interface_next_index != ~0))
- next0 = mif->per_interface_next_index;
- else
- /* redirect if feature path
- * enabled */
- vnet_feature_start_device_input_x1 (mif->sw_if_index, &next0,
- first_b0);
- }
-
/* trace */
- VLIB_BUFFER_TRACE_TRAJECTORY_INIT (first_b0);
-
if (PREDICT_FALSE (n_trace > 0))
{
- if (PREDICT_TRUE (first_b0 != 0))
- {
- memif_input_trace_t *tr;
- vlib_trace_buffer (vm, node, next0, first_b0,
- /* follow_chain */ 0);
- vlib_set_trace_count (vm, node, --n_trace);
- tr = vlib_add_trace (vm, node, first_b0, sizeof (*tr));
- tr->next_index = next0;
- tr->hw_if_index = mif->hw_if_index;
- tr->ring = qid;
- }
+ memif_trace_buffer (vm, node, mif, b0, next0, qid, &n_trace);
+ if (PREDICT_FALSE (n_trace > 0))
+ memif_trace_buffer (vm, node, mif, b1, next1, qid, &n_trace);
+ if (PREDICT_FALSE (n_trace > 0))
+ memif_trace_buffer (vm, node, mif, b2, next2, qid, &n_trace);
+ if (PREDICT_FALSE (n_trace > 0))
+ memif_trace_buffer (vm, node, mif, b3, next3, qid, &n_trace);
}
+ /* enqueue */
+ vlib_validate_buffer_enqueue_x4 (vm, node, next_index, to_next,
+ n_left_to_next, bi0, bi1, bi2, bi3,
+ next0, next1, next2, next3);
+
+ /* next */
+ n_from -= 4;
+ po += 4;
+ }
+ while (n_from && n_left_to_next)
+ {
/* enqueue buffer */
- to_next[0] = first_bi0;
+ u32 fbvi0 = po->first_buffer_vec_index;
+ to_next[0] = bi0 = ptd->buffers[fbvi0];
to_next += 1;
n_left_to_next--;
- /* enqueue */
- vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
- n_left_to_next, first_bi0, next0);
+ b0 = vlib_get_buffer (vm, bi0);
+ clib_memcpy (b0, bt, 64);
+ b0->current_length = po->packet_len;
- /* next packet */
- n_rx_packets++;
- n_rx_bytes += b0_total;
- }
- while (num_slots && n_left_to_next)
- {
- vlib_buffer_t *first_b0 = 0;
- u32 bi0 = 0, first_bi0 = 0;
- b0_total = memif_copy_buffer_from_rx_ring (vm, mif, ring, mq,
- ring_size,
- n_buffer_bytes,
- &n_free_bufs, &first_b0,
- &first_bi0, &bi0,
- &num_slots);
- if (PREDICT_FALSE (!first_bi0))
- {
- goto _invalid_pkt0;
- }
+ memif_add_to_chain (vm, b0, ptd->buffers + fbvi0 + 1, buffer_size);
if (mode == MEMIF_INTERFACE_MODE_IP)
{
- next0 = memif_next_from_ip_hdr (node, first_b0);
+ next0 = memif_next_from_ip_hdr (node, b0);
}
else if (mode == MEMIF_INTERFACE_MODE_ETHERNET)
{
if (PREDICT_FALSE (mif->per_interface_next_index != ~0))
next0 = mif->per_interface_next_index;
else
- /* redirect if feature path
- * enabled */
- vnet_feature_start_device_input_x1 (mif->sw_if_index,
- &next0, first_b0);
- }
-
- /* trace */
- VLIB_BUFFER_TRACE_TRAJECTORY_INIT (first_b0);
-
- if (PREDICT_FALSE (n_trace > 0))
- {
- if (PREDICT_TRUE (first_b0 != 0))
{
- memif_input_trace_t *tr;
- vlib_trace_buffer (vm, node, next0, first_b0,
- /* follow_chain */ 0);
- vlib_set_trace_count (vm, node, --n_trace);
- tr = vlib_add_trace (vm, node, first_b0, sizeof (*tr));
- tr->next_index = next0;
- tr->hw_if_index = mif->hw_if_index;
- tr->ring = qid;
+ next0 = next_index;
+ /* redirect if feature path enabled */
+ vnet_feature_start_device_input_x1 (mif->sw_if_index,
+ &next0, b0);
}
+
}
- /* enqueue buffer */
- to_next[0] = first_bi0;
- to_next += 1;
- n_left_to_next--;
+ /* trace */
+ if (PREDICT_FALSE (n_trace > 0))
+ memif_trace_buffer (vm, node, mif, b0, next0, qid, &n_trace);
/* enqueue */
vlib_validate_buffer_enqueue_x1 (vm, node, next_index, to_next,
- n_left_to_next, first_bi0, next0);
-
- /* next packet */
- n_rx_packets++;
- n_rx_bytes += b0_total;
- continue;
- _invalid_pkt0:
- ;
- }
- if (PREDICT_TRUE (n_rx_packets != 0))
- {
- vlib_put_next_frame (vm, node, next_index, n_left_to_next);
+ n_left_to_next, bi0, next0);
+
+ /* next */
+ n_from--;
+ po++;
}
+ vlib_put_next_frame (vm, node, next_index, n_left_to_next);
}
- CLIB_MEMORY_STORE_BARRIER ();
- ring->tail = mq->last_head;
vlib_increment_combined_counter (vnm->interface_main.combined_sw_if_counters
+ VNET_INTERFACE_COUNTER_RX, thread_index,
mif->hw_if_index, n_rx_packets,
n_rx_bytes);
+ /* refill ring with empty buffers */
+refill:
+ vec_reset_length (ptd->buffers);
+ vec_reset_length (ptd->copy_ops);
+
+ if (type == MEMIF_RING_M2S)
+ {
+ u16 head = ring->head;
+ n_slots = ring_size - head + mq->last_tail;
+
+ while (n_slots--)
+ {
+ u16 s = head++ & mask;
+ memif_desc_t *d = &ring->desc[s];
+ d->length = mif->run.buffer_size;
+ }
+
+ CLIB_MEMORY_STORE_BARRIER ();
+ ring->head = head;
+ }
+
return n_rx_packets;
}
@@ -531,14 +498,14 @@ CLIB_MULTIARCH_FN (memif_input_fn) (vlib_main_t * vm,
vlib_frame_t * frame)
{
u32 n_rx = 0;
- memif_main_t *nm = &memif_main;
+ memif_main_t *mm = &memif_main;
vnet_device_input_runtime_t *rt = (void *) node->runtime_data;
vnet_device_and_queue_t *dq;
foreach_device_and_queue (dq, rt->devices_and_queues)
{
memif_if_t *mif;
- mif = vec_elt_at_index (nm->interfaces, dq->dev_instance);
+ mif = vec_elt_at_index (mm->interfaces, dq->dev_instance);
if ((mif->flags & MEMIF_IF_FLAG_ADMIN_UP) &&
(mif->flags & MEMIF_IF_FLAG_CONNECTED))
{
diff --git a/src/plugins/memif/private.h b/src/plugins/memif/private.h
index b7c18c93e2a..f4ace4cee5d 100644
--- a/src/plugins/memif/private.h
+++ b/src/plugins/memif/private.h
@@ -176,6 +176,35 @@ typedef struct
typedef struct
{
+ u32 packet_len;
+ u16 first_buffer_vec_index;
+} memif_packet_op_t;
+
+typedef struct
+{
+ void *data;
+ u32 data_len;
+ i16 buffer_offset;
+ u16 buffer_vec_index;
+} memif_copy_op_t;
+
+#define MEMIF_RX_VECTOR_SZ VLIB_FRAME_SIZE
+
+typedef struct
+{
+ CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
+
+ /* copy vector */
+ memif_packet_op_t packet_ops[MEMIF_RX_VECTOR_SZ];
+ memif_copy_op_t *copy_ops;
+ u32 *buffers;
+
+ /* buffer template */
+ vlib_buffer_t buffer_template;
+} memif_per_thread_data_t;
+
+typedef struct
+{
CLIB_CACHE_LINE_ALIGN_MARK (cacheline0);
/** API message ID base */
@@ -188,8 +217,8 @@ typedef struct
memif_socket_file_t *socket_files;
uword *socket_file_index_by_sock_id; /* map user socket id to pool idx */
- /* rx buffer cache */
- u32 **rx_buffers;
+ /* per thread data */
+ memif_per_thread_data_t *per_thread_data;
} memif_main_t;