/*
 * replication.h : packet replication
 *
 * Copyright (c) 2013 Cisco and/or its affiliates.
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at:
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

#ifndef included_replication_h
#define included_replication_h


#include <vlib/vlib.h>
#include <vnet/vnet.h>
#include <vnet/replication.h>


typedef struct
{
  /* The entire vnet buffer header restored for each replica */
  u8 vnet_buffer[40];		/* 16B aligned to allow vector unit copy */
  u8 reserved[24];		/* space for future expansion of vnet buffer header */

  /* feature state used during this replication */
  u64 feature_replicas;		/* feature's id for its set of replicas */
  u32 feature_counter;		/* feature's current index into set of replicas */
  u32 recycle_node_index;	/* feature's recycle node index */

  /*
   * data saved from the start of replication and restored
   * at the end of replication
   */
  vlib_buffer_free_list_index_t saved_free_list_index;	/* from vlib buffer */

  /* data saved from the original packet and restored for each replica */
  u64 l2_header[3];		/*  24B (must be at least 22B for l2 packets) */
  u32 flags;			/* vnet buffer flags */
  u16 ip_tos;			/* v4 and v6 */
  u16 ip4_checksum;		/* needed for v4 only */

  /* data saved from the vlib buffer header and restored for each replica */
  i16 current_data;		/* offset of first byte of packet in packet data */
  u8 pad[2];			/* to 64B */
  u8 l2_packet;			/* flag for l2 vs l3 packet data */

} replication_context_t;	/* 128B */


typedef struct
{

  u32 recycle_list_index;

  /* per-thread pools of replication contexts */
  replication_context_t **contexts;

  vlib_main_t *vlib_main;
  vnet_main_t *vnet_main;

} replication_main_t;


extern replication_main_t replication_main;


/* Return 1 if this buffer just came from the replication recycle handler. */
always_inline u32
replication_is_recycled (vlib_buffer_t * b0)
{
  return b0->flags & VLIB_BUFFER_IS_RECYCLED;
}

/*
 * Clear the recycle flag. If buffer came from the replication recycle
 * handler, this flag must be cleared before the packet is transmitted again.
 */
always_inline void
replication_clear_recycled (vlib_buffer_t * b0)
{
  b0->flags &= ~VLIB_BUFFER_IS_RECYCLED;
}

/*
 * Return the active replication context if this buffer has
 * been recycled, otherwise return 0. (Note that this essentially
 * restricts access to the replication context to the replication
 * feature's prep and recycle nodes.)
 */
always_inline replication_context_t *
replication_get_ctx (vlib_buffer_t * b0)
{
  replication_main_t *rm = &replication_main;

  return replication_is_recycled (b0) ?
    pool_elt_at_index (rm->contexts[vlib_get_thread_index ()],
		       b0->recycle_count) : 0;
}

/* Prefetch the replication context for this buffer, if it exists */
always_inline void
replication_prefetch_ctx (vlib_buffer_t * b0)
{
  replication_context_t *ctx = replication_get_ctx (b0);

  if (ctx)
    {
      CLIB_PREFETCH (ctx, (2 * CLIB_CACHE_LINE_BYTES), STORE);
    }
}

replication_context_t *replication_prep (vlib_main_t * vm,
					 vlib_buffer_t * b0,
					 u32 recycle_node_index,
					 u32 l2_packet);

replication_context_t *replication_recycle (vlib_main_t * vm,
					    vlib_buffer_t * b0, u32 is_last);


#endif

/*
 * fd.io coding-style-patch-verification: ON
 *
 * Local Variables:
 * eval: (c-set-style "gnu")
 * End:
 */