aboutsummaryrefslogtreecommitdiffstats
path: root/src/svm/svm_fifo.c
diff options
context:
space:
mode:
authorFlorin Coras <fcoras@cisco.com>2019-12-12 12:09:29 -0800
committerDave Barach <openvpp@barachs.net>2020-02-25 19:18:49 +0000
commitb020806806c0e6c54886cdb4347a5fd1f19504b0 (patch)
tree77322fd6c17967df0626e99c98076b5dac71d89d /src/svm/svm_fifo.c
parent5cd31ec9405d2bb2fbc8152a08c4cfb64f2a8e73 (diff)
svm: refactor fifo chunk tracking
Avoid tracking with rbtrees all of the chunks associated to a fifo. Instead, only track chunks when doing out-of-order operations (peek or ooo enqueue). Type: refactor Change-Id: I9f8bd266211746637d98e6a12ffc4b2d6346950a Signed-off-by: Florin Coras <fcoras@cisco.com>
Diffstat (limited to 'src/svm/svm_fifo.c')
-rw-r--r--src/svm/svm_fifo.c338
1 files changed, 281 insertions, 57 deletions
diff --git a/src/svm/svm_fifo.c b/src/svm/svm_fifo.c
index 975a82026f7..3df67135cfd 100644
--- a/src/svm/svm_fifo.c
+++ b/src/svm/svm_fifo.c
@@ -413,8 +413,8 @@ svm_fifo_init_chunks (svm_fifo_t * f)
return;
f->flags |= SVM_FIFO_F_MULTI_CHUNK;
- rb_tree_init (&f->chunk_lookup);
- rb_tree_add2 (&f->chunk_lookup, 0, pointer_to_uword (f->start_chunk));
+ rb_tree_init (&f->ooo_enq_lookup);
+ rb_tree_init (&f->ooo_deq_lookup);
f->start_chunk->start_byte = 0;
prev = f->start_chunk;
@@ -423,7 +423,6 @@ svm_fifo_init_chunks (svm_fifo_t * f)
while (c != f->start_chunk)
{
c->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, c->start_byte, pointer_to_uword (c));
prev = c;
c = c->next;
}
@@ -458,6 +457,7 @@ svm_fifo_create (u32 data_size_in_bytes)
c->next = c;
c->start_byte = 0;
c->length = data_size_in_bytes;
+ c->rb_index = RBTREE_TNIL_INDEX;
f->start_chunk = f->end_chunk = c;
svm_fifo_init (f, data_size_in_bytes);
@@ -491,6 +491,60 @@ svm_fifo_chunk_includes_pos (svm_fifo_chunk_t * c, u32 pos)
return (pos >= c->start_byte && pos < c->start_byte + c->length);
}
+static rb_node_t *
+svm_fifo_find_node_rbtree (rb_tree_t * rt, u32 pos)
+{
+ rb_node_t *cur, *prev;
+
+ cur = rb_node (rt, rt->root);
+ if (PREDICT_FALSE (rb_node_is_tnil (rt, cur)))
+ return 0;
+
+ while (pos != cur->key)
+ {
+ prev = cur;
+ if (pos < cur->key)
+ {
+ cur = rb_node_left (rt, cur);
+ if (rb_node_is_tnil (rt, cur))
+ {
+ cur = rb_tree_predecessor (rt, prev);
+ break;
+ }
+ }
+ else
+ {
+ cur = rb_node_right (rt, cur);
+ if (rb_node_is_tnil (rt, cur))
+ {
+ cur = prev;
+ break;
+ }
+ }
+ }
+
+ if (rb_node_is_tnil (rt, cur))
+ return 0;
+
+ return cur;
+}
+
+static svm_fifo_chunk_t *
+svm_fifo_find_chunk_rbtree (rb_tree_t * rt, u32 pos)
+{
+ svm_fifo_chunk_t *c;
+ rb_node_t *n;
+
+ n = svm_fifo_find_node_rbtree (rt, pos);
+ if (!n)
+ return 0;
+ c = uword_to_pointer (n->opaque, svm_fifo_chunk_t *);
+ if (svm_fifo_chunk_includes_pos (c, pos))
+ return c;
+
+ return 0;
+}
+
/**
* Find chunk for given byte position
*
@@ -502,44 +556,170 @@ svm_fifo_chunk_includes_pos (svm_fifo_chunk_t * c, u32 pos)
static svm_fifo_chunk_t *
svm_fifo_find_chunk (svm_fifo_t * f, u32 pos)
{
- rb_tree_t *rt = &f->chunk_lookup;
- rb_node_t *cur, *prev;
svm_fifo_chunk_t *c;
- cur = rb_node (rt, rt->root);
- while (pos != cur->key)
+ c = f->start_chunk;
+ do
{
- prev = cur;
- if (pos < cur->key)
- cur = rb_node_left (rt, cur);
- else
- cur = rb_node_right (rt, cur);
+ if (svm_fifo_chunk_includes_pos (c, pos))
+ return c;
+ c = c->next;
+ }
+ while (c != f->start_chunk);
+
+ return 0;
+}
+
+static void
+svm_fifo_update_ooo_enq (svm_fifo_t * f, u32 ref_pos, u32 start_pos,
+ u32 end_pos)
+{
+ rb_tree_t *rt = &f->ooo_enq_lookup;
+ svm_fifo_chunk_t *c;
+ rb_node_t *cur;
+
+ if (svm_fifo_chunk_includes_pos (f->ooo_enq, start_pos)
+ && svm_fifo_chunk_includes_pos (f->ooo_enq, end_pos)
+ && ref_pos < start_pos)
+ return;
- if (rb_node_is_tnil (rt, cur))
+ if (rt->root == RBTREE_TNIL_INDEX)
+ {
+ c = f->tail_chunk;
+ c->rb_index = rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+ }
+ else
+ {
+ cur = svm_fifo_find_node_rbtree (rt, start_pos);
+ c = uword_to_pointer (cur->opaque, svm_fifo_chunk_t *);
+ if (ref_pos > start_pos && c->start_byte > start_pos)
{
- /* Hit tnil as a left child. Find predecessor */
- if (pos < prev->key)
- {
- cur = rb_tree_predecessor (rt, prev);
- if (rb_node_is_tnil (rt, cur))
- return 0;
- c = uword_to_pointer (cur->opaque, svm_fifo_chunk_t *);
- if (svm_fifo_chunk_includes_pos (c, pos))
- return c;
- return 0;
- }
- /* Hit tnil as a right child. Check if this is the one */
- c = uword_to_pointer (prev->opaque, svm_fifo_chunk_t *);
- if (svm_fifo_chunk_includes_pos (c, pos))
- return c;
+ c = f->end_chunk;
+ ASSERT (c->rb_index != RBTREE_TNIL_INDEX);
+ }
+ }
- return 0;
+ if (svm_fifo_chunk_includes_pos (c, start_pos))
+ f->ooo_enq = c;
+
+ if (svm_fifo_chunk_includes_pos (c, end_pos) && ref_pos < end_pos)
+ return;
+
+ do
+ {
+ c = c->next;
+ if (c->rb_index != RBTREE_TNIL_INDEX)
+ break;
+
+ c->rb_index = rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+
+ if (svm_fifo_chunk_includes_pos (c, start_pos))
+ f->ooo_enq = c;
+
+ }
+ while (!svm_fifo_chunk_includes_pos (c, end_pos));
+}
+
+static void
+svm_fifo_update_ooo_deq (svm_fifo_t * f, u32 ref_pos, u32 start_pos,
+ u32 end_pos)
+{
+ rb_tree_t *rt = &f->ooo_deq_lookup;
+ rb_node_t *cur;
+ svm_fifo_chunk_t *c;
+
+ if (svm_fifo_chunk_includes_pos (f->ooo_deq, start_pos)
+ && svm_fifo_chunk_includes_pos (f->ooo_deq, end_pos)
+ && ref_pos < start_pos)
+ return;
+
+ if (rt->root == RBTREE_TNIL_INDEX)
+ {
+ c = f->head_chunk;
+ c->rb_index = rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+ }
+ else
+ {
+ cur = svm_fifo_find_node_rbtree (rt, start_pos);
+ c = uword_to_pointer (cur->opaque, svm_fifo_chunk_t *);
+ if (ref_pos > start_pos && c->start_byte > start_pos)
+ {
+ c = f->end_chunk;
+ ASSERT (c->rb_index != RBTREE_TNIL_INDEX);
}
}
- if (!rb_node_is_tnil (rt, cur))
- return uword_to_pointer (cur->opaque, svm_fifo_chunk_t *);
- return 0;
+ if (svm_fifo_chunk_includes_pos (c, start_pos))
+ f->ooo_deq = c;
+
+ if (svm_fifo_chunk_includes_pos (c, end_pos) && ref_pos < end_pos)
+ return;
+
+ do
+ {
+ c = c->next;
+ if (c->rb_index != RBTREE_TNIL_INDEX)
+ break;
+
+ c->rb_index = rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+
+ if (svm_fifo_chunk_includes_pos (c, start_pos))
+ f->ooo_deq = c;
+
+ }
+ while (!svm_fifo_chunk_includes_pos (c, end_pos));
+}
+
+void
+svm_fifo_ooo_deq_track (svm_fifo_t * f, u32 start_pos, u32 end_pos)
+{
+ rb_tree_t *rt = &f->ooo_deq_lookup;
+ svm_fifo_chunk_t *c;
+
+ if (svm_fifo_chunk_includes_pos (f->ooo_deq, end_pos)
+ && start_pos < end_pos)
+ return;
+
+ c = f->ooo_deq->next;
+ do
+ {
+ ASSERT (c->rb_index == RBTREE_TNIL_INDEX);
+ rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+
+ c = c->next;
+ }
+ while (!svm_fifo_chunk_includes_pos (c, end_pos));
+}
+
+static svm_fifo_chunk_t *
+svm_fifo_lookup_clear_chunks (svm_fifo_t * f, rb_tree_t * rt,
+ svm_fifo_chunk_t * start, u32 start_pos,
+ u32 end_pos)
+{
+ svm_fifo_chunk_t *c;
+ rb_node_t *n;
+
+ /* Nothing to do if still in the same chunk and not wrapped */
+ if (svm_fifo_chunk_includes_pos (start, end_pos) && start_pos < end_pos)
+ return start;
+
+ c = start;
+ do
+ {
+ if (c->rb_index == RBTREE_TNIL_INDEX)
+ {
+ c = c->next;
+ continue;
+ }
+
+ n = rb_node (rt, c->rb_index);
+ rb_tree_del_node (rt, n);
+ c->rb_index = RBTREE_TNIL_INDEX;
+ c = c->next;
+ }
+ while (!svm_fifo_chunk_includes_pos (c, end_pos));
+
+ return c;
}
static inline void
@@ -585,8 +765,8 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
if (!(f->flags & SVM_FIFO_F_MULTI_CHUNK))
{
ASSERT (f->start_chunk->next == f->start_chunk);
- rb_tree_init (&f->chunk_lookup);
- rb_tree_add2 (&f->chunk_lookup, 0, pointer_to_uword (f->start_chunk));
+ rb_tree_init (&f->ooo_enq_lookup);
+ rb_tree_init (&f->ooo_deq_lookup);
f->flags |= SVM_FIFO_F_MULTI_CHUNK;
}
@@ -608,8 +788,7 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
while (cur)
{
cur->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, cur->start_byte,
- pointer_to_uword (cur));
+ cur->rb_index = RBTREE_TNIL_INDEX;
prev = cur;
cur = cur->next;
}
@@ -637,7 +816,12 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
while (cur != f->start_chunk)
{
/* remove any existing rb_tree entry */
- rb_tree_del (&f->chunk_lookup, cur->start_byte);
+ if (cur->rb_index != RBTREE_TNIL_INDEX)
+ {
+ rb_tree_del (&f->ooo_enq_lookup, cur->start_byte);
+ rb_tree_del (&f->ooo_deq_lookup, cur->start_byte);
+ }
+ cur->rb_index = RBTREE_TNIL_INDEX;
cur = cur->next;
}
@@ -647,8 +831,7 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
{
add_bytes += c->length;
c->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, c->start_byte,
- pointer_to_uword (c));
+ cur->rb_index = RBTREE_TNIL_INDEX;
prev = c;
c = c->next;
@@ -660,8 +843,6 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
while (cur != f->start_chunk)
{
cur->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, cur->start_byte,
- pointer_to_uword (cur));
prev = cur;
cur = cur->next;
}
@@ -694,8 +875,7 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
while (cur)
{
cur->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, cur->start_byte,
- pointer_to_uword (cur));
+ cur->rb_index = RBTREE_TNIL_INDEX;
prev = cur;
cur = cur->next;
}
@@ -723,7 +903,11 @@ svm_fifo_collect_chunks (svm_fifo_t * f)
cur = list;
while (cur)
{
- rb_tree_del (&f->chunk_lookup, cur->start_byte);
+ if (cur->rb_index != RBTREE_TNIL_INDEX)
+ {
+ rb_tree_del (&f->ooo_enq_lookup, cur->start_byte);
+ rb_tree_del (&f->ooo_deq_lookup, cur->start_byte);
+ }
cur = cur->next;
}
@@ -846,7 +1030,8 @@ svm_fifo_reduce_size (svm_fifo_t * f, u32 len, u8 try_shrink)
void
svm_fifo_free_chunk_lookup (svm_fifo_t * f)
{
- rb_tree_free_nodes (&f->chunk_lookup);
+ rb_tree_free_nodes (&f->ooo_enq_lookup);
+ rb_tree_free_nodes (&f->ooo_deq_lookup);
}
void
@@ -910,8 +1095,10 @@ svm_fifo_enqueue (svm_fifo_t * f, u32 len, const u8 * src)
if (PREDICT_FALSE (f->ooos_list_head != OOO_SEGMENT_INVALID_INDEX))
{
len += ooo_segment_try_collect (f, len, &tail);
- if (!svm_fifo_chunk_includes_pos (f->tail_chunk, tail))
- f->tail_chunk = svm_fifo_find_chunk (f, tail);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ f->tail_chunk = svm_fifo_lookup_clear_chunks (f, &f->ooo_enq_lookup,
+ f->tail_chunk, f->tail,
+ tail);
}
/* store-rel: producer owned index (paired with load-acq in consumer) */
@@ -949,8 +1136,9 @@ svm_fifo_enqueue_with_offset (svm_fifo_t * f, u32 offset, u32 len, u8 * src)
ooo_segment_add (f, offset, head, tail, len);
tail_idx = (tail + offset) % f->size;
- if (!svm_fifo_chunk_includes_pos (f->ooo_enq, tail_idx))
- f->ooo_enq = svm_fifo_find_chunk (f, tail_idx);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ svm_fifo_update_ooo_enq (f, f->tail, tail_idx,
+ (tail_idx + len) % f->size);
svm_fifo_copy_to_chunk (f, f->ooo_enq, tail_idx, src, len, &f->ooo_enq);
@@ -970,8 +1158,10 @@ svm_fifo_enqueue_nocopy (svm_fifo_t * f, u32 len)
tail = f->tail;
tail = (tail + len) % f->size;
- if (!svm_fifo_chunk_includes_pos (f->tail_chunk, tail))
- f->tail_chunk = svm_fifo_find_chunk (f, tail);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ f->tail_chunk = svm_fifo_lookup_clear_chunks (f, &f->ooo_enq_lookup,
+ f->tail_chunk, f->tail,
+ tail);
/* store-rel: producer owned index (paired with load-acq in consumer) */
clib_atomic_store_rel_n (&f->tail, tail);
@@ -1018,8 +1208,9 @@ svm_fifo_peek (svm_fifo_t * f, u32 offset, u32 len, u8 * dst)
len = clib_min (cursize - offset, len);
head_idx = (head + offset) % f->size;
- if (!svm_fifo_chunk_includes_pos (f->ooo_deq, head_idx))
- f->ooo_deq = svm_fifo_find_chunk (f, head_idx);
+
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ svm_fifo_update_ooo_deq (f, head, head_idx, (head_idx + len) % f->size);
svm_fifo_copy_from_chunk (f, f->ooo_deq, head_idx, dst, len, &f->ooo_deq);
return len;
@@ -1045,8 +1236,10 @@ svm_fifo_dequeue_drop (svm_fifo_t * f, u32 len)
/* move head */
head = (head + total_drop_bytes) % f->size;
- if (!svm_fifo_chunk_includes_pos (f->head_chunk, head))
- f->head_chunk = svm_fifo_find_chunk (f, head);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ f->head_chunk = svm_fifo_lookup_clear_chunks (f, &f->ooo_deq_lookup,
+ f->head_chunk, f->head,
+ head);
if (PREDICT_FALSE (f->flags & SVM_FIFO_F_GROW))
svm_fifo_try_grow (f, head);
@@ -1063,8 +1256,10 @@ svm_fifo_dequeue_drop_all (svm_fifo_t * f)
/* consumer foreign index */
u32 tail = clib_atomic_load_acq_n (&f->tail);
- if (!svm_fifo_chunk_includes_pos (f->head_chunk, tail))
- f->head_chunk = svm_fifo_find_chunk (f, tail);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ f->head_chunk = svm_fifo_lookup_clear_chunks (f, &f->ooo_deq_lookup,
+ f->head_chunk, tail,
+ tail - 1);
if (PREDICT_FALSE (f->flags & SVM_FIFO_F_GROW))
svm_fifo_try_grow (f, tail);
@@ -1220,6 +1415,35 @@ svm_fifo_is_sane (svm_fifo_t * f)
return 0;
if (prev && (prev->start_byte + prev->length != c->start_byte))
return 0;
+
+ if (c->rb_index != RBTREE_TNIL_INDEX)
+ {
+ u8 found = 0;
+
+ tmp = svm_fifo_find_chunk_rbtree (&f->ooo_enq_lookup,
+ c->start_byte);
+ if (tmp)
+ {
+ found = 1;
+ if (tmp != c)
+ return 0;
+ }
+
+ tmp = svm_fifo_find_chunk_rbtree (&f->ooo_deq_lookup,
+ c->start_byte);
+ if (tmp)
+ {
+ if (found)
+ return 0;
+
+ found = 1;
+ if (tmp != c)
+ return 0;
+ }
+ if (!found)
+ return 0;
+ }
+
size += c->length;
prev = c;
c = c->next;