/* * Copyright (c) 2015 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /* * physmem.h: virtual <-> physical memory mapping for VLIB buffers * * Copyright (c) 2008 Eliot Dresselhaus * * Permission is hereby granted, free of charge, to any person obtaining * a copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sublicense, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice shall be * included in all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ #ifndef included_vlib_physmem_funcs_h #define included_vlib_physmem_funcs_h #include <vppinfra/clib.h> #include <vppinfra/clib_error.h> #include <vlib/physmem.h> #include <vlib/main.h> clib_error_t *vlib_physmem_init (vlib_main_t * vm); clib_error_t *vlib_physmem_shared_map_create (vlib_main_t * vm, char *name, uword size, u32 log2_page_sz, u32 numa_node, u32 * map_index); vlib_physmem_map_t *vlib_physmem_get_map (vlib_main_t * vm, u32 index); always_inline void * vlib_physmem_alloc_aligned (vlib_main_t * vm, uword n_bytes, uword alignment) { clib_pmalloc_main_t *pm = vm->physmem_main.pmalloc_main; return clib_pmalloc_alloc_aligned (pm, n_bytes, alignment); } always_inline void * vlib_physmem_alloc_aligned_on_numa (vlib_main_t * vm, uword n_bytes, uword alignment, u32 numa_node) { clib_pmalloc_main_t *pm = vm->physmem_main.pmalloc_main; return clib_pmalloc_alloc_aligned_on_numa (pm, n_bytes, alignment, numa_node); } /* By default allocate I/O memory with cache line alignment. */ always_inline void * vlib_physmem_alloc (vlib_main_t * vm, uword n_bytes) { return vlib_physmem_alloc_aligned (vm, n_bytes, CLIB_CACHE_LINE_BYTES); } always_inline void * vlib_physmem_alloc_from_map (vlib_main_t * vm, u32 physmem_map_index, uword n_bytes, uword alignment) { clib_pmalloc_main_t *pm = vm->physmem_main.pmalloc_main; vlib_physmem_map_t *map = vlib_physmem_get_map (vm, physmem_map_index); return clib_pmalloc_alloc_from_arena (pm, map->base, n_bytes, CLIB_CACHE_LINE_BYTES); } always_inline void vlib_physmem_free (vlib_main_t * vm, void *p) { if (p) clib_pmalloc_free (vm->physmem_main.pmalloc_main, p); } always_inline u64 vlib_physmem_get_page_index (vlib_main_t * vm, void *mem) { clib_pmalloc_main_t *pm = vm->physmem_main.pmalloc_main; return clib_pmalloc_get_page_index (pm, mem); } always_inline u64 vlib_physmem_get_pa (vlib_main_t * vm, void *mem) { clib_pmalloc_main_t *pm = vm->physmem_main.pmalloc_main; return clib_pmalloc_get_pa (pm, mem); } always_inline clib_error_t * vlib_physmem_last_error (struct vlib_main_t * vm) { clib_pmalloc_main_t *pm = vm->physmem_main.pmalloc_main; return pm->error; } #endif /* included_vlib_physmem_funcs_h */ /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */