/* *------------------------------------------------------------------ * Copyright (c) 2018 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. *------------------------------------------------------------------ */ #include #include #include #include #include #include /* enumerate all vlib messages */ #define vl_typedefs /* define message structures */ #include #undef vl_typedefs /* instantiate all the print functions we know about */ #define vl_printfun #include #undef vl_printfun /* instantiate all the endian swap functions we know about */ #define vl_endianfun #include #undef vl_endianfun volatile int **vl_api_queue_cursizes; static void memclnt_queue_callback (vlib_main_t * vm) { int i; api_main_t *am = vlibapi_get_main (); int have_pending_rpcs; if (PREDICT_FALSE (vec_len (vl_api_queue_cursizes) != 1 + vec_len (am->vlib_private_rps))) { vl_shmem_hdr_t *shmem_hdr = am->shmem_hdr; svm_queue_t *q; if (shmem_hdr == 0) return; q = shmem_hdr->vl_input_queue; if (q == 0) return; vec_add1 (vl_api_queue_cursizes, &q->cursize); for (i = 0; i < vec_len (am->vlib_private_rps); i++) { svm_region_t *vlib_rp = am->vlib_private_rps[i]; shmem_hdr = (void *) vlib_rp->user_ctx; q = shmem_hdr->vl_input_queue; vec_add1 (vl_api_queue_cursizes, &q->cursize); } } for (i = 0; i < vec_len (vl_api_queue_cursizes); i++) { if (*vl_api_queue_cursizes[i]) { vm->queue_signal_pending = 1; vm->api_queue_nonempty = 1; vlib_process_signal_event (vm, vl_api_clnt_node.index, /* event_type */ QUEUE_SIGNAL_EVENT, /* event_data */ 0); break; } } clib_spinlock_lock_if_init (&vm->pending_rpc_lock); have_pending_rpcs = vec_len (vm->pending_rpc_requests) > 0; clib_spinlock_unlock_if_init (&vm->pending_rpc_lock); if (have_pending_rpcs) { vm->queue_signal_pending = 1; vm->api_queue_nonempty = 1; vlib_process_signal_event (vm, vl_api_clnt_node.index, /* event_type */ QUEUE_SIGNAL_EVENT, /* event_data */ 0); } } /* * vl_api_memclnt_create_internal */ u32 vl_api_memclnt_create_internal (char *name, svm_queue_t * q) { vl_api_registration_t **regpp; vl_api_registration_t *regp; void *oldheap; api_main_t *am = vlibapi_get_main (); ASSERT (vlib_get_thread_index () == 0); pool_get (am->vl_clients, regpp); oldheap = vl_msg_push_heap (); *regpp = clib_mem_alloc (sizeof (vl_api_registration_t)); regp = *regpp; clib_memset (regp, 0, sizeof (*regp)); regp->registration_type = REGISTRATION_TYPE_SHMEM; regp->vl_api_registration_pool_index = regpp - am->vl_clients; regp->vlib_rp = am->vlib_rp; regp->shmem_hdr = am->shmem_hdr; regp->vl_input_queue = q; regp->name = format (0, "%s%c", name, 0); vl_msg_pop_heap (oldheap); return vl_msg_api_handle_from_index_and_epoch (regp->vl_api_registration_pool_index, am->shmem_hdr->application_restarts); } /* * vl_api_memclnt_create_t_handler */ void vl_api_memclnt_create_t_handler (vl_api_memclnt_create_t * mp) { vl_api_registration_t **regpp; vl_api_registration_t *regp; vl_api_memclnt_create_reply_t *rp; svm_queue_t *q; int rv = 0; void *oldheap; api_main_t *am = vlibapi_get_main (); u8 *msg_table; /* * This is tortured. Maintain a vlib-address-space private * pool of client registrations. We use the shared-memory virtual * address of client structure as a handle, to allow direct * manipulation of context quota vbls from the client library. * * This scheme causes trouble w/ API message trace replay, since * some random VA from clib_mem_alloc() certainly won't * occur in the Linux sim. The (very) few places * that care need to use the pool index. * * Putting the registration object(s) into a pool in shared memory and * using the pool index as a handle seems like a great idea. * Unfortunately, each and every reference to that pool would need * to be protected by a mutex: * * Client VLIB * ------ ---- * convert pool index to * pointer. * * expand pool * * kaboom! */ pool_get (am->vl_clients, regpp); oldheap = vl_msg_push_heap (); *regpp = clib_mem_alloc (sizeof (vl_api_registration_t)); regp = *regpp; clib_memset (regp, 0, sizeof (*regp)); regp->registration_type = REGISTRATION_TYPE_SHMEM; regp->vl_api_registration_pool_index = regpp - am->vl_clients; regp->vlib_rp = am->vlib_rp; regp->shmem_hdr = am->shmem_hdr; regp->clib_file_index = am->shmem_hdr->clib_file_index; q = regp->vl_input_queue = (svm_queue_t *) (uword) mp->input_queue; VL_MSG_API_SVM_QUEUE_UNPOISON (q); regp->name = format (0, "%s", mp->name); vec_add1 (regp->name, 0); regp->keepalive = true; if (am->serialized_message_table_in_shmem == 0) am->serialized_message_table_in_shmem = vl_api_serialize_message_table (am, 0); if (am->vlib_rp != am->vlib_primary_rp) msg_table = vl_api_serialize_message_table (am, 0); else msg_table = am->serialized_message_table_in_shmem; vl_msg_pop_heap (oldheap); rp = vl_msg_api_alloc (sizeof (*rp)); rp->_vl_msg_id = ntohs (VL_API_MEMCLNT_CREATE_REPLY); rp->handle = (uword) regp; rp->index = vl_msg_api_handle_from_index_and_epoch (regp->vl_api_registration_pool_index, am->shmem_hdr->application_restarts); rp->context = mp->context; rp->response = ntohl (rv); rp->message_table = pointer_to_uword (msg_table); vl_msg_api_send_shmem (q, (u8 *) & rp); } void vl_api_memclnt_create_v2_t_handler (vl_api_memclnt_create_v2_t *mp) { vl_api_registration_t **regpp; vl_api_registration_t *regp; vl_api_memclnt_create_v2_reply_t *rp; svm_queue_t *q; int rv = 0; void *oldheap; api_main_t *am = vlibapi_get_main (); u8 *msg_table; /* * This is tortured. Maintain a vlib-address-space private * pool of client registrations. We use the shared-memory virtual * address of client structure as a handle, to allow direct * manipulation of context quota vbls from the client library. * * This scheme causes trouble w/ API message trace replay, since * some random VA from clib_mem_alloc() certainly won't * occur in the Linux sim. The (very) few places * that care need to use the pool index. * * Putting the registration object(s) into a pool in shared memory and * using the pool index as a handle seems like a great idea. * Unfortunately, each and every reference to that pool would need * to be protected by a mutex: * * Client VLIB * ------ ---- * convert pool index to * pointer. * * expand pool * * kaboom! */ pool_get (am->vl_clients, regpp); oldheap = vl_msg_push_heap (); *regpp = clib_mem_alloc (sizeof (vl_api_registration_t)); regp = *regpp; clib_memset (regp, 0, sizeof (*regp)); regp->registration_type = REGISTRATION_TYPE_SHMEM; regp->vl_api_registration_pool_index = regpp - am->vl_clients; regp->vlib_rp = am->vlib_rp; regp->shmem_hdr = am->shmem_hdr; regp->clib_file_index = am->shmem_hdr->clib_file_index; q = regp->vl_input_queue = (svm_queue_t *) (uword) mp->input_queue; VL_MSG_API_SVM_QUEUE_UNPOISON (q); regp->name = format (0, "%s", mp->name); vec_add1 (regp->name, 0); regp->keepalive = mp->keepalive; if (am->serialized_message_table_in_shmem == 0) am->serialized_message_table_in_shmem = vl_api_serialize_message_table (am, 0); if (am->vlib_rp != am->vlib_primary_rp) msg_table = vl_api_serialize_message_table (am, 0); else msg_table = am->serialized_message_table_in_shmem; vl_msg_pop_heap (oldheap); rp = vl_msg_api_alloc (sizeof (*rp)); rp->_vl_msg_id = ntohs (VL_API_MEMCLNT_CREATE_V2_REPLY); rp->handle = (uword) regp; rp->index = vl_msg_api_handle_from_index_and_epoch ( regp->vl_api_registration_pool_index, am->shmem_hdr->application_restarts); rp->context = mp->context; rp->response = ntohl (rv); rp->message_table = pointer_to_uword (msg_table); vl_msg_api_send_shmem (q, (u8 *) &rp); } void vl_api_call_reaper_functions (u32 client_index) { clib_error_t *error = 0; _vl_msg_api_function_list_elt_t *i; i = vlibapi_get_main ()->reaper_function_registrations; while (i) { error = i->f (client_index); if (error) clib_error_report (error); i = i->next_init_function; } } /* * vl_api_memclnt_delete_t_handler */ void vl_api_memclnt_delete_t_handler (vl_api_memclnt_delete_t * mp) { vl_api_registration_t **regpp; vl_api_registration_t *regp; vl_api_memclnt_delete_reply_t *rp; void *oldheap; api_main_t *am = vlibapi_get_main (); u32 handle, client_index, epoch; handle = mp->index; vl_api_call_reaper_functions (handle); epoch = vl_msg_api_handle_get_epoch (handle); client_index = vl_msg_api_handle_get_index (handle); if (epoch != (am->shmem_hdr->application_restarts & VL_API_EPOCH_MASK)) { clib_warning ("Stale clnt delete index %d old epoch %d cur epoch %d", client_index, epoch, (am->shmem_hdr->application_restarts & VL_API_EPOCH_MASK)); return; } regpp = pool_elt_at_index (am->vl_clients, client_index); if (!pool_is_free (am->vl_clients, regpp)) { int i; regp = *regpp; int private_registration = 0; /* Send reply unless client asked us to do the cleanup */ if (!mp->do_cleanup) { /* * Note: the API message handling path will set am->vlib_rp * as appropriate for pairwise / private memory segments */ rp = vl_msg_api_alloc (sizeof (*rp)); rp->_vl_msg_id = ntohs (VL_API_MEMCLNT_DELETE_REPLY); rp->handle = mp->handle; rp->response = 1; vl_msg_api_send_shmem (regp->vl_input_queue, (u8 *) & rp); if (client_index != regp->vl_api_registration_pool_index) { clib_warning ("mismatch client_index %d pool_index %d", client_index, regp->vl_api_registration_pool_index); vl_msg_api_free (rp); return; } } /* No dangling references, please */ *regpp = 0; /* For horizontal scaling, add a hash table... */ for (i = 0; i < vec_len (am->vlib_private_rps); i++) { /* Is this a pairwise / private API segment? */ if (am->vlib_private_rps[i] == am->vlib_rp) { /* Note: account for the memfd header page */ uword virtual_base = am->vlib_rp->virtual_base - MMAP_PAGESIZE; uword virtual_size = am->vlib_rp->virtual_size + MMAP_PAGESIZE; /* * Kill the registration pool element before we make * the index vanish forever */ pool_put_index (am->vl_clients, regp->vl_api_registration_pool_index); vec_delete (am->vlib_private_rps, 1, i); /* Kill it, accounting for the memfd header page */ if (munmap ((void *) virtual_base, virtual_size) < 0) clib_unix_warning ("munmap"); /* Reset the queue-length-address cache */ vec_reset_length (vl_api_queue_cursizes); private_registration = 1; break; } } if (private_registration == 0) { pool_put_index (am->vl_clients, regp->vl_api_registration_pool_index); oldheap = vl_msg_push_heap (); if (mp->do_cleanup) svm_queue_free (regp->vl_input_queue); vec_free (regp->name); /* Poison the old registration */ clib_memset (regp, 0xF1, sizeof (*regp)); clib_mem_free (regp); vl_msg_pop_heap (oldheap); /* * These messages must be freed manually, since they're set up * as "bounce" messages. In the private_registration == 1 case, * we kill the shared-memory segment which contains the message * with munmap. */ vl_msg_api_free (mp); } } else { clib_warning ("unknown client ID %d", mp->index); } } /** * client answered a ping, stave off the grim reaper... */ void vl_api_memclnt_keepalive_reply_t_handler (vl_api_memclnt_keepalive_reply_t * mp) { vl_api_registration_t *regp; vlib_main_t *vm = vlib_get_main (); regp = vl_api_client_index_to_registration (mp->context); if (regp) { regp->last_heard = vlib_time_now (vm); regp->unanswered_pings = 0; } else clib_warning ("BUG: anonymous memclnt_keepalive_reply"); } /** * We can send ourselves these messages if someone uses the * builtin binary api test tool... */ static void vl_api_memclnt_keepalive_t_handler (vl_api_memclnt_keepalive_t * mp) { vl_api_memclnt_keepalive_reply_t *rmp; api_main_t *am; vl_shmem_hdr_t *shmem_hdr; am = vlibapi_get_main (); shmem_hdr = am->shmem_hdr; rmp = vl_msg_api_alloc_as_if_client (sizeof (*rmp)); clib_memset (rmp, 0, sizeof (*rmp)); rmp->_vl_msg_id = ntohs (VL_API_MEMCLNT_KEEPALIVE_REPLY); rmp->context = mp->context; vl_msg_api_send_shmem (shmem_hdr->vl_input_queue, (u8 *) & rmp); } /* * To avoid filling the API trace buffer with boring messages, * don't trace memclnt_keepalive[_reply] msgs */ #define foreach_vlib_api_msg \ _ (MEMCLNT_CREATE, memclnt_create, 0) \ _ (MEMCLNT_CREATE_V2, memclnt_create_v2, 0) \ _ (MEMCLNT_DELETE, memclnt_delete, 0) \ _ (MEMCLNT_KEEPALIVE, memclnt_keepalive, 0) \ _ (MEMCLNT_KEEPALIVE_REPLY, memclnt_keepalive_reply, 0) /* * memory_api_init */ int vl_mem_api_init (const char *region_name) { int rv; api_main_t *am = vlibapi_get_main (); vl_msg_api_msg_config_t cfg; vl_msg_api_msg_config_t *c = &cfg; vl_shmem_hdr_t *shm; vlib_main_t *vm = vlib_get_main (); clib_memset (c, 0, sizeof (*c)); if ((rv = vl_map_shmem (region_name, 1 /* is_vlib */ )) < 0) return rv; #define _(N, n, t) \ do \ { \ c->id = VL_API_##N; \ c->name = #n; \ c->handler = vl_api_##n##_t_handler; \ c->endian = vl_api_##n##_t_endian; \ c->format_fn = vl_api_##n##_t_format; \ c->size = sizeof (vl_api_##n##_t); \ c->traced = t; /* trace, so these msgs print */ \ c->replay = 0; /* don't replay client create/delete msgs */ \ c->message_bounce = 0; /* don't bounce this message */ \ vl_msg_api_config (c); \ } \ while (0); foreach_vlib_api_msg; #undef _ #define vl_msg_name_crc_list #include #undef vl_msg_name_crc_list #define _(id, n, crc) vl_msg_api_add_msg_name_crc (am, #n "_" #crc, id); foreach_vl_msg_name_crc_memclnt; #undef _ /* * special-case freeing of memclnt_delete messages, so we can * simply munmap pairwise / private API segments... */ am->msg_data[VL_API_MEMCLNT_DELETE].bounce = 1; vl_api_set_msg_thread_safe (am, VL_API_MEMCLNT_KEEPALIVE_REPLY, 1); vl_api_set_msg_thread_safe (am, VL_API_MEMCLNT_KEEPALIVE, 1); vlib_set_queue_signal_callback (vm, memclnt_queue_callback); shm = am->shmem_hdr; ASSERT (shm && shm->vl_input_queue); /* Make a note so we can always find the primary region easily */ am->vlib_primary_rp = am->vlib_rp; return 0; } clib_error_t * map_api_segment_init (vlib_main_t * vm) { api_main_t *am = vlibapi_get_main (); int rv; if ((rv = vl_mem_api_init (am->region_name)) < 0) { return clib_error_return (0, "vl_mem_api_init (%s) failed", am->region_name); } return 0; } static void send_memclnt_keepalive (vl_api_registration_t * regp, f64 now) { vl_api_memclnt_keepalive_t *mp; svm_queue_t *q; api_main_t *am = vlibapi_get_main (); q = regp->vl_input_queue; /* * If the queue head is moving, assume that the client is processing * messages and skip the ping. This heuristic may fail if the queue * is in the same position as last time, net of wrapping; in which * case, the client will receive a keepalive. */ if (regp->last_queue_head != q->head) { regp->last_heard = now; regp->unanswered_pings = 0; regp->last_queue_head = q->head; return; } /* * push/pop shared memory segment, so this routine * will work with "normal" as well as "private segment" * memory clients.. */ mp = vl_mem_api_alloc_as_if_client_w_reg (regp, sizeof (*mp)); clib_memset (mp, 0, sizeof (*mp)); mp->_vl_msg_id = clib_host_to_net_u16 (VL_API_MEMCLNT_KEEPALIVE); mp->context = mp->client_index = vl_msg_api_handle_from_index_and_epoch (regp->vl_api_registration_pool_index, am->shmem_hdr->application_restarts); regp->unanswered_pings++; /* Failure-to-send due to a stuffed queue is absolutely expected */ if (svm_queue_add (q, (u8 *) & mp, 1 /* nowait */ )) vl_msg_api_free_w_region (regp->vlib_rp, mp); } static void vl_mem_send_client_keepalive_w_reg (api_main_t * am, f64 now, vl_api_registration_t ** regpp, u32 ** dead_indices, u32 ** confused_indices) { vl_api_registration_t *regp = *regpp; if (regp) { /* If we haven't heard from this client recently... */ if (regp->last_heard < (now - 10.0)) { if (regp->unanswered_pings == 2) { svm_queue_t *q; q = regp->vl_input_queue; if (kill (q->consumer_pid, 0) >= 0) { clib_warning ("REAPER: lazy binary API client '%s'", regp->name); regp->unanswered_pings = 0; regp->last_heard = now; } else { clib_warning ("REAPER: binary API client '%s' died", regp->name); vec_add1 (*dead_indices, regpp - am->vl_clients); } } else send_memclnt_keepalive (regp, now); } else regp->unanswered_pings = 0; } else { clib_warning ("NULL client registration index %d", regpp - am->vl_clients); vec_add1 (*confused_indices, regpp - am->vl_clients); } } void vl_mem_api_dead_client_scan (api_main_t * am, vl_shmem_hdr_t * shm, f64 now) { vl_api_registration_t **regpp; static u32 *dead_indices; static u32 *confused_indices; vec_reset_length (dead_indices); vec_reset_length (confused_indices); pool_foreach (regpp, am->vl_clients) { if (!(*regpp)->keepalive) continue; vl_mem_send_client_keepalive_w_reg (am, now, regpp, &dead_indices, &confused_indices); } /* This should "never happen," but if it does, fix it... */ if (PREDICT_FALSE (vec_len (confused_indices) > 0)) { int i; for (i = 0; i < vec_len (confused_indices); i++) { pool_put_index (am->vl_clients, confused_indices[i]); } } if (PREDICT_FALSE (vec_len (dead_indices) > 0)) { int i; void *oldheap; /* Allow the application to clean up its registrations */ for (i = 0; i < vec_len (dead_indices); i++) { regpp = pool_elt_at_index (am->vl_clients, dead_indices[i]); if (regpp) { u32 handle; handle = vl_msg_api_handle_from_index_and_epoch (dead_indices[i], shm->application_restarts); vl_api_call_reaper_functions (handle); } } oldheap = vl_msg_push_heap (); for (i = 0; i < vec_len (dead_indices); i++) { regpp = pool_elt_at_index (am->vl_clients, dead_indices[i]); if (regpp) { /* Is this a pairwise SVM segment? */ if ((*regpp)->vlib_rp != am->vlib_rp) { int i; svm_region_t *dead_rp = (*regpp)->vlib_rp; /* Note: account for the memfd header page */ uword virtual_base = dead_rp->virtual_base - MMAP_PAGESIZE; uword virtual_size = dead_rp->virtual_size + MMAP_PAGESIZE; /* For horizontal scaling, add a hash table... */ for (i = 0; i < vec_len (am->vlib_private_rps); i++) if (am->vlib_private_rps[i] == dead_rp) { vec_delete (am->vlib_private_rps, 1, i); goto found; } svm_pop_heap (oldheap); clib_warning ("private rp %llx AWOL", dead_rp); oldheap = svm_push_data_heap (am->vlib_rp); found: /* Kill it, accounting for the memfd header page */ svm_pop_heap (oldheap); if (munmap ((void *) virtual_base, virtual_size) < 0) clib_unix_warning ("munmap"); /* Reset the queue-length-address cache */ vec_reset_length (vl_api_queue_cursizes); oldheap = svm_push_data_heap (am->vlib_rp); } else { /* Poison the old registration */ clib_memset (*regpp, 0xF3, sizeof (**regpp)); clib_mem_free (*regpp); } /* no dangling references, please */ *regpp = 0; } else { svm_pop_heap (oldheap); clib_warning ("Duplicate free, client index %d", regpp - am->vl_clients); oldheap = svm_push_data_heap (am->vlib_rp); } } svm_client_scan_this_region_nolock (am->vlib_rp); vl_msg_pop_heap (oldheap); for (i = 0; i < vec_len (dead_indices); i++) pool_put_index (am->vl_clients, dead_indices[i]); } } void (*vl_mem_api_fuzz_hook) (u16, void *); /* This is only to be called from a vlib/vnet app */ static void vl_mem_api_handler_with_vm_node (api_main_t *am, svm_region_t *vlib_rp, void *the_msg, vlib_main_t *vm, vlib_node_runtime_t *node, u8 is_private) { u16 id = clib_net_to_host_u16 (*((u16 *) the_msg)); vl_api_msg_data_t *m = vl_api_get_msg_data (am, id); u8 *(*handler) (void *, void *, void *); svm_region_t *old_vlib_rp; void *save_shmem_hdr; int is_mp_safe = 1; if (PREDICT_FALSE (am->elog_trace_api_messages)) { ELOG_TYPE_DECLARE (e) = { .format = "api-msg: %s", .format_args = "T4", }; struct { u32 c; } * ed; ed = ELOG_DATA (am->elog_main, e); if (m && m->name) ed->c = elog_string (am->elog_main, (char *) m->name); else ed->c = elog_string (am->elog_main, "BOGUS"); } if (m && m->handler) { handler = (void *) m->handler; if (PREDICT_FALSE (am->rx_trace && am->rx_trace->enabled)) vl_msg_api_trace (am, am->rx_trace, the_msg); if (PREDICT_FALSE (am->msg_print_flag)) { fformat (stdout, "[%d]: %s\n", id, m->name); fformat (stdout, "%U", format_vl_api_msg_text, am, id, the_msg); } is_mp_safe = am->msg_data[id].is_mp_safe; if (!is_mp_safe) { vl_msg_api_barrier_trace_context (am->msg_data[id].name); vl_msg_api_barrier_sync (); } if (is_private) { old_vlib_rp = am->vlib_rp; save_shmem_hdr = am->shmem_hdr; am->vlib_rp = vlib_rp; am->shmem_hdr = (void *) vlib_rp->user_ctx; } if (PREDICT_FALSE (vl_mem_api_fuzz_hook != 0)) (*vl_mem_api_fuzz_hook) (id, the_msg); if (m->is_autoendian) { void (*endian_fp) (void *); endian_fp = am->msg_data[id].endian_handler; (*endian_fp) (the_msg); } if (PREDICT_FALSE (vec_len (am->perf_counter_cbs) != 0)) clib_call_callbacks (am->perf_counter_cbs, am, id, 0 /* before */); (*handler) (the_msg, vm, node); if (PREDICT_FALSE (vec_len (am->perf_counter_cbs) != 0)) clib_call_callbacks (am->perf_counter_cbs, am, id, 1 /* after */); if (is_private) { am->vlib_rp = old_vlib_rp; am->shmem_hdr = save_shmem_hdr; } if (!is_mp_safe) vl_msg_api_barrier_release (); } else { clib_warning ("no handler for msg id %d", id); } /* * Special-case, so we can e.g. bounce messages off the vnet * main thread without copying them... */ if (!m || !m->bounce) { if (is_private) { old_vlib_rp = am->vlib_rp; save_shmem_hdr = am->shmem_hdr; am->vlib_rp = vlib_rp; am->shmem_hdr = (void *) vlib_rp->user_ctx; } vl_msg_api_free (the_msg); if (is_private) { am->vlib_rp = old_vlib_rp; am->shmem_hdr = save_shmem_hdr; } } if (PREDICT_FALSE (am->elog_trace_api_messages)) { ELOG_TYPE_DECLARE (e) = { .format = "api-msg-done(%s): %s", .format_args = "t4T4", .n_enum_strings = 2, .enum_strings = { "barrier", "mp-safe", } }; struct { u32 barrier; u32 c; } * ed; ed = ELOG_DATA (am->elog_main, e); if (m && m->name) ed->c = elog_string (am->elog_main, (char *) m->name); else ed->c = elog_string (am->elog_main, "BOGUS"); ed->barrier = is_mp_safe; } } static inline int void_mem_api_handle_msg_i (api_main_t * am, svm_region_t * vlib_rp, vlib_main_t * vm, vlib_node_runtime_t * node, u8 is_private) { svm_queue_t *q; uword mp; q = ((vl_shmem_hdr_t *) (void *) vlib_rp->user_ctx)->vl_input_queue; if (!svm_queue_sub2 (q, (u8 *) & mp)) { VL_MSG_API_UNPOISON ((void *) mp); vl_mem_api_handler_with_vm_node (am, vlib_rp, (void *) mp, vm, node, is_private); return 0; } return -1; } int vl_mem_api_handle_msg_main (vlib_main_t * vm, vlib_node_runtime_t * node) { api_main_t *am = vlibapi_get_main (); return void_mem_api_handle_msg_i (am, am->vlib_rp, vm, node, 0 /* is_private */ ); } int vl_mem_api_handle_rpc (vlib_main_t * vm, vlib_node_runtime_t * node) { api_main_t *am = vlibapi_get_main (); int i; uword *tmp, mp; /* * Swap pending and processing vectors, then process the RPCs * Avoid deadlock conditions by construction. */ clib_spinlock_lock_if_init (&vm->pending_rpc_lock); tmp = vm->processing_rpc_requests; vec_reset_length (tmp); vm->processing_rpc_requests = vm->pending_rpc_requests; vm->pending_rpc_requests = tmp; clib_spinlock_unlock_if_init (&vm->pending_rpc_lock); /* * RPCs are used to reflect function calls to thread 0 * when the underlying code is not thread-safe. * * Grabbing the thread barrier across a set of RPCs * greatly increases efficiency, and avoids * running afoul of the barrier sync holddown timer. * The barrier sync code supports recursive locking. * * We really need to rewrite RPC-based code... */ if (PREDICT_TRUE (vec_len (vm->processing_rpc_requests))) { vl_msg_api_barrier_sync (); for (i = 0; i < vec_len (vm->processing_rpc_requests); i++) { mp = vm->processing_rpc_requests[i]; vl_mem_api_handler_with_vm_node (am, am->vlib_rp, (void *) mp, vm, node, 0 /* is_private */); } vl_msg_api_barrier_release (); } return 0; } int vl_mem_api_handle_msg_private (vlib_main_t * vm, vlib_node_runtime_t * node, u32 reg_index) { api_main_t *am = vlibapi_get_main (); return void_mem_api_handle_msg_i (am, am->vlib_private_rps[reg_index], vm, node, 1 /* is_private */ ); } vl_api_registration_t * vl_mem_api_client_index_to_registration (u32 handle) { vl_api_registration_t **regpp; vl_api_registration_t *regp; api_main_t *am = vlibapi_get_main (); vl_shmem_hdr_t *shmem_hdr; u32 index; index = vl_msg_api_handle_get_index (handle); regpp = am->vl_clients + index; if (pool_is_free (am->vl_clients, regpp)) { vl_msg_api_increment_missing_client_counter (); return 0; } regp = *regpp; shmem_hdr = (vl_shmem_hdr_t *) regp->shmem_hdr; if (!vl_msg_api_handle_is_valid (handle, shmem_hdr->application_restarts)) { vl_msg_api_increment_missing_client_counter (); return 0; } return (regp); } svm_queue_t * vl_api_client_index_to_input_queue (u32 index) { vl_api_registration_t *regp; api_main_t *am = vlibapi_get_main (); /* Special case: vlib trying to send itself a message */ if (index == (u32) ~ 0) return (am->shmem_hdr->vl_input_queue); regp = vl_mem_api_client_index_to_registration (index); if (!regp) return 0; return (regp->vl_input_queue); } static clib_error_t * setup_memclnt_exit (vlib_main_t * vm) { atexit (vl_unmap_shmem); return 0; } VLIB_INIT_FUNCTION (setup_memclnt_exit); u8 * format_api_message_rings (u8 * s, va_list * args) { api_main_t *am = va_arg (*args, api_main_t *); vl_shmem_hdr_t *shmem_hdr = va_arg (*args, vl_shmem_hdr_t *); int main_segment = va_arg (*args, int); ring_alloc_t *ap; int i; if (shmem_hdr == 0) return format (s, "%8s %8s %8s %8s %8s\n", "Owner", "Size", "Nitems", "Hits", "Misses"); ap = shmem_hdr->vl_rings; for (i = 0; i < vec_len (shmem_hdr->vl_rings); i++) { s = format (s, "%8s %8d %8d %8d %8d\n", "vlib", ap->size, ap->nitems, ap->hits, ap->misses); ap++; } ap = shmem_hdr->client_rings; for (i = 0; i < vec_len (shmem_hdr->client_rings); i++) { s = format (s, "%8s %8d %8d %8d %8d\n", "clnt", ap->size, ap->nitems, ap->hits, ap->misses); ap++; } if (main_segment) { s = format (s, "%d ring miss fallback allocations\n", am->ring_misses); s = format (s, "%d application restarts, %d reclaimed msgs, %d garbage collects\n", shmem_hdr->application_restarts, shmem_hdr->restart_reclaims, shmem_hdr->garbage_collects); } return s; } static clib_error_t * vl_api_ring_command (vlib_main_t * vm, unformat_input_t * input, vlib_cli_command_t * cli_cmd) { int i; vl_shmem_hdr_t *shmem_hdr; api_main_t *am = vlibapi_get_main (); /* First, dump the primary region rings.. */ if (am->vlib_primary_rp == 0 || am->vlib_primary_rp->user_ctx == 0) { vlib_cli_output (vm, "Shared memory segment not initialized...\n"); return 0; } shmem_hdr = (void *) am->vlib_primary_rp->user_ctx; vlib_cli_output (vm, "Main API segment rings:"); vlib_cli_output (vm, "%U", format_api_message_rings, am, 0 /* print header */ , 0 /* notused */ ); vlib_cli_output (vm, "%U", format_api_message_rings, am, shmem_hdr, 1 /* main segment */ ); for (i = 0; i < vec_len (am->vlib_private_rps); i++) { svm_region_t *vlib_rp = am->vlib_private_rps[i]; shmem_hdr = (void *) vlib_rp->user_ctx; vl_api_registration_t **regpp; vl_api_registration_t *regp = 0; /* For horizontal scaling, add a hash table... */ pool_foreach (regpp, am->vl_clients) { regp = *regpp; if (regp && regp->vlib_rp == vlib_rp) { vlib_cli_output (vm, "%s segment rings:", regp->name); goto found; } } vlib_cli_output (vm, "regp %llx not found?", regp); continue; found: vlib_cli_output (vm, "%U", format_api_message_rings, am, 0 /* print header */ , 0 /* notused */ ); vlib_cli_output (vm, "%U", format_api_message_rings, am, shmem_hdr, 0 /* main segment */ ); } return 0; } /*? * Display binary api message allocation ring statistics ?*/ VLIB_CLI_COMMAND (cli_show_api_ring_command, static) = { .path = "show api ring-stats", .short_help = "Message ring statistics", .function = vl_api_ring_command, }; clib_error_t * vlibmemory_init (vlib_main_t * vm) { api_main_t *am = vlibapi_get_main (); svm_map_region_args_t _a, *a = &_a; u8 *remove_path1, *remove_path2; void vlibsocket_reference (void); vlibsocket_reference (); /* * By popular request / to avoid support fires, remove any old api segment * files Right Here. */ if (am->root_path == 0) { remove_path1 = format (0, "/dev/shm/global_vm%c", 0); remove_path2 = format (0, "/dev/shm/vpe-api%c", 0); } else { remove_path1 = format (0, "/dev/shm/%s-global_vm%c", am->root_path, 0); remove_path2 = format (0, "/dev/shm/%s-vpe-api%c", am->root_path, 0); } (void) unlink ((char *) remove_path1); (void) unlink ((char *) remove_path2); vec_free (remove_path1); vec_free (remove_path2); clib_memset (a, 0, sizeof (*a)); a->root_path = am->root_path; a->name = SVM_GLOBAL_REGION_NAME; a->baseva = (am->global_baseva != 0) ? am->global_baseva : +svm_get_global_region_base_va (); a->size = (am->global_size != 0) ? am->global_size : SVM_GLOBAL_REGION_SIZE; a->flags = SVM_FLAGS_NODATA; a->uid = am->api_uid; a->gid = am->api_gid; a->pvt_heap_size = (am->global_pvt_heap_size != 0) ? am->global_pvt_heap_size : SVM_PVT_MHEAP_SIZE; svm_region_init_args (a); return 0; } void vl_set_memory_region_name (const char *name) { api_main_t *am = vlibapi_get_main (); am->region_name = name; } /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */