aboutsummaryrefslogtreecommitdiffstats
path: root/src/svm
diff options
context:
space:
mode:
authorFlorin Coras <fcoras@cisco.com>2019-12-12 12:09:29 -0800
committerDave Barach <openvpp@barachs.net>2020-02-25 19:18:49 +0000
commitb020806806c0e6c54886cdb4347a5fd1f19504b0 (patch)
tree77322fd6c17967df0626e99c98076b5dac71d89d /src/svm
parent5cd31ec9405d2bb2fbc8152a08c4cfb64f2a8e73 (diff)
svm: refactor fifo chunk tracking
Avoid tracking with rbtrees all of the chunks associated to a fifo. Instead, only track chunks when doing out-of-order operations (peek or ooo enqueue). Type: refactor Change-Id: I9f8bd266211746637d98e6a12ffc4b2d6346950a Signed-off-by: Florin Coras <fcoras@cisco.com>
Diffstat (limited to 'src/svm')
-rw-r--r--src/svm/fifo_segment.c32
-rw-r--r--src/svm/svm_fifo.c338
-rw-r--r--src/svm/svm_fifo.h6
3 files changed, 294 insertions, 82 deletions
diff --git a/src/svm/fifo_segment.c b/src/svm/fifo_segment.c
index 58c11383e43..bc46aba7bf6 100644
--- a/src/svm/fifo_segment.c
+++ b/src/svm/fifo_segment.c
@@ -412,6 +412,7 @@ fs_try_alloc_fifo_batch (fifo_segment_header_t * fsh,
c = (svm_fifo_chunk_t *) (fmem + sizeof (*f));
c->start_byte = 0;
c->length = rounded_data_size;
+ c->rb_index = RBTREE_TNIL_INDEX;
c->next = fss->free_chunks[fl_index];
fss->free_chunks[fl_index] = c;
fmem += hdrs + rounded_data_size;
@@ -506,11 +507,7 @@ fifo_segment_alloc_fifo_w_slice (fifo_segment_t * fs, u32 slice_index,
/* Initialize chunks and rbtree for multi-chunk fifos */
if (f->start_chunk->next != f->start_chunk)
- {
- void *oldheap = ssvm_push_heap (fsh->ssvm_sh);
- svm_fifo_init_chunks (f);
- ssvm_pop_heap (oldheap);
- }
+ svm_fifo_init_chunks (f);
/* If rx fifo type add to active fifos list. When cleaning up segment,
* we need a list of active sessions that should be disconnected. Since
@@ -541,7 +538,6 @@ fifo_segment_free_fifo (fifo_segment_t * fs, svm_fifo_t * f)
fifo_segment_header_t *fsh = fs->h;
svm_fifo_chunk_t *cur, *next;
fifo_segment_slice_t *fss;
- void *oldheap;
int fl_index;
ASSERT (f->refcnt > 0);
@@ -576,6 +572,7 @@ fifo_segment_free_fifo (fifo_segment_t * fs, svm_fifo_t * f)
fl_index = fs_freelist_for_size (cur->length);
ASSERT (fl_index < vec_len (fss->free_chunks));
cur->next = fss->free_chunks[fl_index];
+ cur->rb_index = RBTREE_TNIL_INDEX;
fss->free_chunks[fl_index] = cur;
fss->n_fl_chunk_bytes += fs_freelist_index_to_size (fl_index);
cur = next;
@@ -585,9 +582,7 @@ fifo_segment_free_fifo (fifo_segment_t * fs, svm_fifo_t * f)
f->start_chunk = f->end_chunk = f->new_chunks = 0;
f->head_chunk = f->tail_chunk = f->ooo_enq = f->ooo_deq = 0;
- oldheap = ssvm_push_heap (fsh->ssvm_sh);
svm_fifo_free_chunk_lookup (f);
- ssvm_pop_heap (oldheap);
/* not allocated on segment heap */
svm_fifo_free_ooo_data (f);
@@ -768,25 +763,21 @@ fifo_segment_grow_fifo (fifo_segment_t * fs, svm_fifo_t * f, u32 chunk_size)
fl_index = fs_freelist_for_size (chunk_size);
fss = fsh_slice_get (fsh, f->slice_index);
- oldheap = ssvm_push_heap (fsh->ssvm_sh);
-
c = fss->free_chunks[fl_index];
if (!c)
{
fsh_check_mem (fsh);
if (fsh_n_free_bytes (fsh) < chunk_size)
- {
- ssvm_pop_heap (oldheap);
- return -1;
- }
+ return -1;
+ oldheap = ssvm_push_heap (fsh->ssvm_sh);
c = svm_fifo_chunk_alloc (chunk_size);
+ ssvm_pop_heap (oldheap);
+
if (!c)
- {
- ssvm_pop_heap (oldheap);
- return -1;
- }
+ return -1;
+
fsh_free_bytes_sub (fsh, chunk_size + sizeof (*c));
}
else
@@ -798,7 +789,6 @@ fifo_segment_grow_fifo (fifo_segment_t * fs, svm_fifo_t * f, u32 chunk_size)
svm_fifo_add_chunk (f, c);
- ssvm_pop_heap (oldheap);
return 0;
}
@@ -808,10 +798,8 @@ fifo_segment_collect_fifo_chunks (fifo_segment_t * fs, svm_fifo_t * f)
fifo_segment_header_t *fsh = fs->h;
svm_fifo_chunk_t *cur, *next;
fifo_segment_slice_t *fss;
- void *oldheap;
int fl_index;
- oldheap = ssvm_push_heap (fsh->ssvm_sh);
cur = svm_fifo_collect_chunks (f);
fss = fsh_slice_get (fsh, f->slice_index);
@@ -825,8 +813,6 @@ fifo_segment_collect_fifo_chunks (fifo_segment_t * fs, svm_fifo_t * f)
cur = next;
}
- ssvm_pop_heap (oldheap);
-
return 0;
}
diff --git a/src/svm/svm_fifo.c b/src/svm/svm_fifo.c
index 975a82026f7..3df67135cfd 100644
--- a/src/svm/svm_fifo.c
+++ b/src/svm/svm_fifo.c
@@ -413,8 +413,8 @@ svm_fifo_init_chunks (svm_fifo_t * f)
return;
f->flags |= SVM_FIFO_F_MULTI_CHUNK;
- rb_tree_init (&f->chunk_lookup);
- rb_tree_add2 (&f->chunk_lookup, 0, pointer_to_uword (f->start_chunk));
+ rb_tree_init (&f->ooo_enq_lookup);
+ rb_tree_init (&f->ooo_deq_lookup);
f->start_chunk->start_byte = 0;
prev = f->start_chunk;
@@ -423,7 +423,6 @@ svm_fifo_init_chunks (svm_fifo_t * f)
while (c != f->start_chunk)
{
c->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, c->start_byte, pointer_to_uword (c));
prev = c;
c = c->next;
}
@@ -458,6 +457,7 @@ svm_fifo_create (u32 data_size_in_bytes)
c->next = c;
c->start_byte = 0;
c->length = data_size_in_bytes;
+ c->rb_index = RBTREE_TNIL_INDEX;
f->start_chunk = f->end_chunk = c;
svm_fifo_init (f, data_size_in_bytes);
@@ -491,6 +491,60 @@ svm_fifo_chunk_includes_pos (svm_fifo_chunk_t * c, u32 pos)
return (pos >= c->start_byte && pos < c->start_byte + c->length);
}
+static rb_node_t *
+svm_fifo_find_node_rbtree (rb_tree_t * rt, u32 pos)
+{
+ rb_node_t *cur, *prev;
+
+ cur = rb_node (rt, rt->root);
+ if (PREDICT_FALSE (rb_node_is_tnil (rt, cur)))
+ return 0;
+
+ while (pos != cur->key)
+ {
+ prev = cur;
+ if (pos < cur->key)
+ {
+ cur = rb_node_left (rt, cur);
+ if (rb_node_is_tnil (rt, cur))
+ {
+ cur = rb_tree_predecessor (rt, prev);
+ break;
+ }
+ }
+ else
+ {
+ cur = rb_node_right (rt, cur);
+ if (rb_node_is_tnil (rt, cur))
+ {
+ cur = prev;
+ break;
+ }
+ }
+ }
+
+ if (rb_node_is_tnil (rt, cur))
+ return 0;
+
+ return cur;
+}
+
+static svm_fifo_chunk_t *
+svm_fifo_find_chunk_rbtree (rb_tree_t * rt, u32 pos)
+{
+ svm_fifo_chunk_t *c;
+ rb_node_t *n;
+
+ n = svm_fifo_find_node_rbtree (rt, pos);
+ if (!n)
+ return 0;
+ c = uword_to_pointer (n->opaque, svm_fifo_chunk_t *);
+ if (svm_fifo_chunk_includes_pos (c, pos))
+ return c;
+
+ return 0;
+}
+
/**
* Find chunk for given byte position
*
@@ -502,44 +556,170 @@ svm_fifo_chunk_includes_pos (svm_fifo_chunk_t * c, u32 pos)
static svm_fifo_chunk_t *
svm_fifo_find_chunk (svm_fifo_t * f, u32 pos)
{
- rb_tree_t *rt = &f->chunk_lookup;
- rb_node_t *cur, *prev;
svm_fifo_chunk_t *c;
- cur = rb_node (rt, rt->root);
- while (pos != cur->key)
+ c = f->start_chunk;
+ do
{
- prev = cur;
- if (pos < cur->key)
- cur = rb_node_left (rt, cur);
- else
- cur = rb_node_right (rt, cur);
+ if (svm_fifo_chunk_includes_pos (c, pos))
+ return c;
+ c = c->next;
+ }
+ while (c != f->start_chunk);
+
+ return 0;
+}
+
+static void
+svm_fifo_update_ooo_enq (svm_fifo_t * f, u32 ref_pos, u32 start_pos,
+ u32 end_pos)
+{
+ rb_tree_t *rt = &f->ooo_enq_lookup;
+ svm_fifo_chunk_t *c;
+ rb_node_t *cur;
+
+ if (svm_fifo_chunk_includes_pos (f->ooo_enq, start_pos)
+ && svm_fifo_chunk_includes_pos (f->ooo_enq, end_pos)
+ && ref_pos < start_pos)
+ return;
- if (rb_node_is_tnil (rt, cur))
+ if (rt->root == RBTREE_TNIL_INDEX)
+ {
+ c = f->tail_chunk;
+ c->rb_index = rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+ }
+ else
+ {
+ cur = svm_fifo_find_node_rbtree (rt, start_pos);
+ c = uword_to_pointer (cur->opaque, svm_fifo_chunk_t *);
+ if (ref_pos > start_pos && c->start_byte > start_pos)
{
- /* Hit tnil as a left child. Find predecessor */
- if (pos < prev->key)
- {
- cur = rb_tree_predecessor (rt, prev);
- if (rb_node_is_tnil (rt, cur))
- return 0;
- c = uword_to_pointer (cur->opaque, svm_fifo_chunk_t *);
- if (svm_fifo_chunk_includes_pos (c, pos))
- return c;
- return 0;
- }
- /* Hit tnil as a right child. Check if this is the one */
- c = uword_to_pointer (prev->opaque, svm_fifo_chunk_t *);
- if (svm_fifo_chunk_includes_pos (c, pos))
- return c;
+ c = f->end_chunk;
+ ASSERT (c->rb_index != RBTREE_TNIL_INDEX);
+ }
+ }
- return 0;
+ if (svm_fifo_chunk_includes_pos (c, start_pos))
+ f->ooo_enq = c;
+
+ if (svm_fifo_chunk_includes_pos (c, end_pos) && ref_pos < end_pos)
+ return;
+
+ do
+ {
+ c = c->next;
+ if (c->rb_index != RBTREE_TNIL_INDEX)
+ break;
+
+ c->rb_index = rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+
+ if (svm_fifo_chunk_includes_pos (c, start_pos))
+ f->ooo_enq = c;
+
+ }
+ while (!svm_fifo_chunk_includes_pos (c, end_pos));
+}
+
+static void
+svm_fifo_update_ooo_deq (svm_fifo_t * f, u32 ref_pos, u32 start_pos,
+ u32 end_pos)
+{
+ rb_tree_t *rt = &f->ooo_deq_lookup;
+ rb_node_t *cur;
+ svm_fifo_chunk_t *c;
+
+ if (svm_fifo_chunk_includes_pos (f->ooo_deq, start_pos)
+ && svm_fifo_chunk_includes_pos (f->ooo_deq, end_pos)
+ && ref_pos < start_pos)
+ return;
+
+ if (rt->root == RBTREE_TNIL_INDEX)
+ {
+ c = f->head_chunk;
+ c->rb_index = rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+ }
+ else
+ {
+ cur = svm_fifo_find_node_rbtree (rt, start_pos);
+ c = uword_to_pointer (cur->opaque, svm_fifo_chunk_t *);
+ if (ref_pos > start_pos && c->start_byte > start_pos)
+ {
+ c = f->end_chunk;
+ ASSERT (c->rb_index != RBTREE_TNIL_INDEX);
}
}
- if (!rb_node_is_tnil (rt, cur))
- return uword_to_pointer (cur->opaque, svm_fifo_chunk_t *);
- return 0;
+ if (svm_fifo_chunk_includes_pos (c, start_pos))
+ f->ooo_deq = c;
+
+ if (svm_fifo_chunk_includes_pos (c, end_pos) && ref_pos < end_pos)
+ return;
+
+ do
+ {
+ c = c->next;
+ if (c->rb_index != RBTREE_TNIL_INDEX)
+ break;
+
+ c->rb_index = rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+
+ if (svm_fifo_chunk_includes_pos (c, start_pos))
+ f->ooo_deq = c;
+
+ }
+ while (!svm_fifo_chunk_includes_pos (c, end_pos));
+}
+
+void
+svm_fifo_ooo_deq_track (svm_fifo_t * f, u32 start_pos, u32 end_pos)
+{
+ rb_tree_t *rt = &f->ooo_deq_lookup;
+ svm_fifo_chunk_t *c;
+
+ if (svm_fifo_chunk_includes_pos (f->ooo_deq, end_pos)
+ && start_pos < end_pos)
+ return;
+
+ c = f->ooo_deq->next;
+ do
+ {
+ ASSERT (c->rb_index == RBTREE_TNIL_INDEX);
+ rb_tree_add2 (rt, c->start_byte, pointer_to_uword (c));
+
+ c = c->next;
+ }
+ while (!svm_fifo_chunk_includes_pos (c, end_pos));
+}
+
+static svm_fifo_chunk_t *
+svm_fifo_lookup_clear_chunks (svm_fifo_t * f, rb_tree_t * rt,
+ svm_fifo_chunk_t * start, u32 start_pos,
+ u32 end_pos)
+{
+ svm_fifo_chunk_t *c;
+ rb_node_t *n;
+
+ /* Nothing to do if still in the same chunk and not wrapped */
+ if (svm_fifo_chunk_includes_pos (start, end_pos) && start_pos < end_pos)
+ return start;
+
+ c = start;
+ do
+ {
+ if (c->rb_index == RBTREE_TNIL_INDEX)
+ {
+ c = c->next;
+ continue;
+ }
+
+ n = rb_node (rt, c->rb_index);
+ rb_tree_del_node (rt, n);
+ c->rb_index = RBTREE_TNIL_INDEX;
+ c = c->next;
+ }
+ while (!svm_fifo_chunk_includes_pos (c, end_pos));
+
+ return c;
}
static inline void
@@ -585,8 +765,8 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
if (!(f->flags & SVM_FIFO_F_MULTI_CHUNK))
{
ASSERT (f->start_chunk->next == f->start_chunk);
- rb_tree_init (&f->chunk_lookup);
- rb_tree_add2 (&f->chunk_lookup, 0, pointer_to_uword (f->start_chunk));
+ rb_tree_init (&f->ooo_enq_lookup);
+ rb_tree_init (&f->ooo_deq_lookup);
f->flags |= SVM_FIFO_F_MULTI_CHUNK;
}
@@ -608,8 +788,7 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
while (cur)
{
cur->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, cur->start_byte,
- pointer_to_uword (cur));
+ cur->rb_index = RBTREE_TNIL_INDEX;
prev = cur;
cur = cur->next;
}
@@ -637,7 +816,12 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
while (cur != f->start_chunk)
{
/* remove any existing rb_tree entry */
- rb_tree_del (&f->chunk_lookup, cur->start_byte);
+ if (cur->rb_index != RBTREE_TNIL_INDEX)
+ {
+ rb_tree_del (&f->ooo_enq_lookup, cur->start_byte);
+ rb_tree_del (&f->ooo_deq_lookup, cur->start_byte);
+ }
+ cur->rb_index = RBTREE_TNIL_INDEX;
cur = cur->next;
}
@@ -647,8 +831,7 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
{
add_bytes += c->length;
c->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, c->start_byte,
- pointer_to_uword (c));
+ cur->rb_index = RBTREE_TNIL_INDEX;
prev = c;
c = c->next;
@@ -660,8 +843,6 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
while (cur != f->start_chunk)
{
cur->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, cur->start_byte,
- pointer_to_uword (cur));
prev = cur;
cur = cur->next;
}
@@ -694,8 +875,7 @@ svm_fifo_add_chunk (svm_fifo_t * f, svm_fifo_chunk_t * c)
while (cur)
{
cur->start_byte = prev->start_byte + prev->length;
- rb_tree_add2 (&f->chunk_lookup, cur->start_byte,
- pointer_to_uword (cur));
+ cur->rb_index = RBTREE_TNIL_INDEX;
prev = cur;
cur = cur->next;
}
@@ -723,7 +903,11 @@ svm_fifo_collect_chunks (svm_fifo_t * f)
cur = list;
while (cur)
{
- rb_tree_del (&f->chunk_lookup, cur->start_byte);
+ if (cur->rb_index != RBTREE_TNIL_INDEX)
+ {
+ rb_tree_del (&f->ooo_enq_lookup, cur->start_byte);
+ rb_tree_del (&f->ooo_deq_lookup, cur->start_byte);
+ }
cur = cur->next;
}
@@ -846,7 +1030,8 @@ svm_fifo_reduce_size (svm_fifo_t * f, u32 len, u8 try_shrink)
void
svm_fifo_free_chunk_lookup (svm_fifo_t * f)
{
- rb_tree_free_nodes (&f->chunk_lookup);
+ rb_tree_free_nodes (&f->ooo_enq_lookup);
+ rb_tree_free_nodes (&f->ooo_deq_lookup);
}
void
@@ -910,8 +1095,10 @@ svm_fifo_enqueue (svm_fifo_t * f, u32 len, const u8 * src)
if (PREDICT_FALSE (f->ooos_list_head != OOO_SEGMENT_INVALID_INDEX))
{
len += ooo_segment_try_collect (f, len, &tail);
- if (!svm_fifo_chunk_includes_pos (f->tail_chunk, tail))
- f->tail_chunk = svm_fifo_find_chunk (f, tail);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ f->tail_chunk = svm_fifo_lookup_clear_chunks (f, &f->ooo_enq_lookup,
+ f->tail_chunk, f->tail,
+ tail);
}
/* store-rel: producer owned index (paired with load-acq in consumer) */
@@ -949,8 +1136,9 @@ svm_fifo_enqueue_with_offset (svm_fifo_t * f, u32 offset, u32 len, u8 * src)
ooo_segment_add (f, offset, head, tail, len);
tail_idx = (tail + offset) % f->size;
- if (!svm_fifo_chunk_includes_pos (f->ooo_enq, tail_idx))
- f->ooo_enq = svm_fifo_find_chunk (f, tail_idx);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ svm_fifo_update_ooo_enq (f, f->tail, tail_idx,
+ (tail_idx + len) % f->size);
svm_fifo_copy_to_chunk (f, f->ooo_enq, tail_idx, src, len, &f->ooo_enq);
@@ -970,8 +1158,10 @@ svm_fifo_enqueue_nocopy (svm_fifo_t * f, u32 len)
tail = f->tail;
tail = (tail + len) % f->size;
- if (!svm_fifo_chunk_includes_pos (f->tail_chunk, tail))
- f->tail_chunk = svm_fifo_find_chunk (f, tail);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ f->tail_chunk = svm_fifo_lookup_clear_chunks (f, &f->ooo_enq_lookup,
+ f->tail_chunk, f->tail,
+ tail);
/* store-rel: producer owned index (paired with load-acq in consumer) */
clib_atomic_store_rel_n (&f->tail, tail);
@@ -1018,8 +1208,9 @@ svm_fifo_peek (svm_fifo_t * f, u32 offset, u32 len, u8 * dst)
len = clib_min (cursize - offset, len);
head_idx = (head + offset) % f->size;
- if (!svm_fifo_chunk_includes_pos (f->ooo_deq, head_idx))
- f->ooo_deq = svm_fifo_find_chunk (f, head_idx);
+
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ svm_fifo_update_ooo_deq (f, head, head_idx, (head_idx + len) % f->size);
svm_fifo_copy_from_chunk (f, f->ooo_deq, head_idx, dst, len, &f->ooo_deq);
return len;
@@ -1045,8 +1236,10 @@ svm_fifo_dequeue_drop (svm_fifo_t * f, u32 len)
/* move head */
head = (head + total_drop_bytes) % f->size;
- if (!svm_fifo_chunk_includes_pos (f->head_chunk, head))
- f->head_chunk = svm_fifo_find_chunk (f, head);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ f->head_chunk = svm_fifo_lookup_clear_chunks (f, &f->ooo_deq_lookup,
+ f->head_chunk, f->head,
+ head);
if (PREDICT_FALSE (f->flags & SVM_FIFO_F_GROW))
svm_fifo_try_grow (f, head);
@@ -1063,8 +1256,10 @@ svm_fifo_dequeue_drop_all (svm_fifo_t * f)
/* consumer foreign index */
u32 tail = clib_atomic_load_acq_n (&f->tail);
- if (!svm_fifo_chunk_includes_pos (f->head_chunk, tail))
- f->head_chunk = svm_fifo_find_chunk (f, tail);
+ if (f->flags & SVM_FIFO_F_MULTI_CHUNK)
+ f->head_chunk = svm_fifo_lookup_clear_chunks (f, &f->ooo_deq_lookup,
+ f->head_chunk, tail,
+ tail - 1);
if (PREDICT_FALSE (f->flags & SVM_FIFO_F_GROW))
svm_fifo_try_grow (f, tail);
@@ -1220,6 +1415,35 @@ svm_fifo_is_sane (svm_fifo_t * f)
return 0;
if (prev && (prev->start_byte + prev->length != c->start_byte))
return 0;
+
+ if (c->rb_index != RBTREE_TNIL_INDEX)
+ {
+ u8 found = 0;
+
+ tmp = svm_fifo_find_chunk_rbtree (&f->ooo_enq_lookup,
+ c->start_byte);
+ if (tmp)
+ {
+ found = 1;
+ if (tmp != c)
+ return 0;
+ }
+
+ tmp = svm_fifo_find_chunk_rbtree (&f->ooo_deq_lookup,
+ c->start_byte);
+ if (tmp)
+ {
+ if (found)
+ return 0;
+
+ found = 1;
+ if (tmp != c)
+ return 0;
+ }
+ if (!found)
+ return 0;
+ }
+
size += c->length;
prev = c;
c = c->next;
diff --git a/src/svm/svm_fifo.h b/src/svm/svm_fifo.h
index 0b7c70790c0..2b6e8542cdf 100644
--- a/src/svm/svm_fifo.h
+++ b/src/svm/svm_fifo.h
@@ -60,6 +60,7 @@ typedef struct svm_fifo_chunk_
u32 start_byte; /**< chunk start byte */
u32 length; /**< length of chunk in bytes */
struct svm_fifo_chunk_ *next; /**< pointer to next chunk in linked-lists */
+ rb_node_index_t rb_index; /**< node index if chunk in rbtree */
u8 data[0]; /**< start of chunk data */
} svm_fifo_chunk_t;
@@ -80,8 +81,8 @@ typedef struct _svm_fifo
u32 nitems; /**< usable size (size-1) */
svm_fifo_chunk_t *start_chunk;/**< first chunk in fifo chunk list */
svm_fifo_chunk_t *end_chunk; /**< end chunk in fifo chunk list */
- svm_fifo_chunk_t *new_chunks; /**< chunks yet to be added to list */
- rb_tree_t chunk_lookup; /**< rbtree for chunk lookup */
+ rb_tree_t ooo_enq_lookup; /**< rbtree for ooo enq chunk lookup */
+ rb_tree_t ooo_deq_lookup; /**< rbtree for ooo deq chunk lookup */
u8 flags; /**< fifo flags */
u8 slice_index; /**< segment slice for fifo */
@@ -96,6 +97,7 @@ typedef struct _svm_fifo
u32 segment_index; /**< segment index in segment manager */
struct _svm_fifo *next; /**< next in freelist/active chain */
struct _svm_fifo *prev; /**< prev in active chain */
+ svm_fifo_chunk_t *new_chunks; /**< chunks yet to be added to list */
u32 size_decrement; /**< bytes to remove from fifo */
CLIB_CACHE_LINE_ALIGN_MARK (consumer);