/* * Copyright (c) 2017 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #define _GNU_SOURCE #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #ifndef F_LINUX_SPECIFIC_BASE #define F_LINUX_SPECIFIC_BASE 1024 #endif #ifndef F_ADD_SEALS #define F_ADD_SEALS (F_LINUX_SPECIFIC_BASE + 9) #define F_GET_SEALS (F_LINUX_SPECIFIC_BASE + 10) #define F_SEAL_SEAL 0x0001 /* prevent further seals from being set */ #define F_SEAL_SHRINK 0x0002 /* prevent file from shrinking */ #define F_SEAL_GROW 0x0004 /* prevent file from growing */ #define F_SEAL_WRITE 0x0008 /* prevent writes */ #endif #ifndef MFD_HUGETLB #define MFD_HUGETLB 0x0004U #endif #ifndef MAP_HUGE_SHIFT #define MAP_HUGE_SHIFT 26 #endif #ifndef MAP_FIXED_NOREPLACE #define MAP_FIXED_NOREPLACE 0x100000 #endif uword clib_mem_get_default_hugepage_size (void) { unformat_input_t input; static u32 size = 0; int fd; if (size) goto done; /* * If the kernel doesn't support hugepages, /proc/meminfo won't * say anything about it. Use the regular page size as a default. */ size = clib_mem_get_page_size () / 1024; if ((fd = open ("/proc/meminfo", 0)) == -1) return 0; unformat_init_clib_file (&input, fd); while (unformat_check_input (&input) != UNFORMAT_END_OF_INPUT) { if (unformat (&input, "Hugepagesize:%_%u kB", &size)) ; else unformat_skip_line (&input); } unformat_free (&input); close (fd); done: return 1024ULL * size; } static clib_mem_page_sz_t legacy_get_log2_default_hugepage_size (void) { clib_mem_page_sz_t log2_page_size = CLIB_MEM_PAGE_SZ_UNKNOWN; FILE *fp; char tmp[33] = { }; if ((fp = fopen ("/proc/meminfo", "r")) == NULL) return CLIB_MEM_PAGE_SZ_UNKNOWN; while (fscanf (fp, "%32s", tmp) > 0) if (strncmp ("Hugepagesize:", tmp, 13) == 0) { u32 size; if (fscanf (fp, "%u", &size) > 0) log2_page_size = 10 + min_log2 (size); break; } fclose (fp); return log2_page_size; } void clib_mem_main_init () { clib_mem_main_t *mm = &clib_mem_main; uword page_size; void *va; int fd; if (mm->log2_page_sz != CLIB_MEM_PAGE_SZ_UNKNOWN) return; /* system page size */ page_size = sysconf (_SC_PAGESIZE); mm->log2_page_sz = min_log2 (page_size); /* default system hugeppage size */ if ((fd = memfd_create ("test", MFD_HUGETLB)) != -1) { mm->log2_default_hugepage_sz = clib_mem_get_fd_log2_page_size (fd); close (fd); } else /* likely kernel older than 4.14 */ mm->log2_default_hugepage_sz = legacy_get_log2_default_hugepage_size (); /* numa nodes */ va = mmap (0, page_size, PROT_READ | PROT_WRITE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); if (va == MAP_FAILED) return; if (mlock (va, page_size)) goto done; for (int i = 0; i < CLIB_MAX_NUMAS; i++) { int status; if (move_pages (0, 1, &va, &i, &status, 0) == 0) mm->numa_node_bitmap |= 1ULL << i; } done: munmap (va, page_size); } u64 clib_mem_get_fd_page_size (int fd) { struct stat st = { 0 }; if (fstat (fd, &st) == -1) return 0; return st.st_blksize; } clib_mem_page_sz_t clib_mem_get_fd_log2_page_size (int fd) { uword page_size = clib_mem_get_fd_page_size (fd); return page_size ? min_log2 (page_size) : CLIB_MEM_PAGE_SZ_UNKNOWN; } void clib_mem_vm_randomize_va (uword * requested_va, clib_mem_page_sz_t log2_page_size) { u8 bit_mask = 15; if (log2_page_size <= 12) bit_mask = 15; else if (log2_page_size > 12 && log2_page_size <= 16) bit_mask = 3; else bit_mask = 0; *requested_va += (clib_cpu_time_now () & bit_mask) * (1ull << log2_page_size); } clib_error_t * clib_mem_create_fd (char *name, int *fdp) { int fd; ASSERT (name); if ((fd = memfd_create (name, MFD_ALLOW_SEALING)) == -1) return clib_error_return_unix (0, "memfd_create"); if ((fcntl (fd, F_ADD_SEALS, F_SEAL_SHRINK)) == -1) { close (fd); return clib_error_return_unix (0, "fcntl (F_ADD_SEALS)"); } *fdp = fd; return 0; } clib_error_t * clib_mem_create_hugetlb_fd (char *name, int *fdp) { clib_error_t *err = 0; int fd = -1; static int memfd_hugetlb_supported = 1; char *mount_dir; char template[] = "/tmp/hugepage_mount.XXXXXX"; u8 *filename; ASSERT (name); if (memfd_hugetlb_supported) { if ((fd = memfd_create (name, MFD_HUGETLB)) != -1) goto done; /* avoid further tries if memfd MFD_HUGETLB is not supported */ if (errno == EINVAL && strnlen (name, 256) <= 249) memfd_hugetlb_supported = 0; } mount_dir = mkdtemp (template); if (mount_dir == 0) return clib_error_return_unix (0, "mkdtemp \'%s\'", template); if (mount ("none", (char *) mount_dir, "hugetlbfs", 0, NULL)) { rmdir ((char *) mount_dir); err = clib_error_return_unix (0, "mount hugetlb directory '%s'", mount_dir); } filename = format (0, "%s/%s%c", mount_dir, name, 0); fd = open ((char *) filename, O_CREAT | O_RDWR, 0755); umount2 ((char *) mount_dir, MNT_DETACH); rmdir ((char *) mount_dir); if (fd == -1) err = clib_error_return_unix (0, "open"); done: if (fd != -1) fdp[0] = fd; return err; } clib_error_t * clib_mem_vm_ext_alloc (clib_mem_vm_alloc_t * a) { int fd = -1; clib_error_t *err = 0; void *addr = 0; u8 *filename = 0; int mmap_flags = 0; int log2_page_size; int n_pages; int old_mpol = -1; long unsigned int old_mask[16] = { 0 }; /* save old numa mem policy if needed */ if (a->flags & (CLIB_MEM_VM_F_NUMA_PREFER | CLIB_MEM_VM_F_NUMA_FORCE)) { int rv; rv = get_mempolicy (&old_mpol, old_mask, sizeof (old_mask) * 8 + 1, 0, 0); if (rv == -1) { if (a->numa_node != 0 && (a->flags & CLIB_MEM_VM_F_NUMA_FORCE) != 0) { err = clib_error_return_unix (0, "get_mempolicy"); goto error; } else old_mpol = -1; } } if (a->flags & CLIB_MEM_VM_F_LOCKED) mmap_flags |= MAP_LOCKED; /* if we are creating shared segment, we need file descriptor */ if (a->flags & CLIB_MEM_VM_F_SHARED) { mmap_flags |= MAP_SHARED; /* if hugepages are needed we need to create mount point */ if (a->flags & CLIB_MEM_VM_F_HUGETLB) { if ((err = clib_mem_create_hugetlb_fd (a->name, &fd))) goto error; mmap_flags |= MAP_LOCKED; } else { if ((err = clib_mem_create_fd (a->name, &fd))) goto error; } log2_page_size = clib_mem_get_fd_log2_page_size (fd); if (log2_page_size == 0) { err = clib_error_return_unix (0, "cannot determine page size"); goto error; } if (a->requested_va) { clib_mem_vm_randomize_va (&a->requested_va, log2_page_size); mmap_flags |= MAP_FIXED; } } else /* not CLIB_MEM_VM_F_SHARED */ { mmap_flags |= MAP_PRIVATE | MAP_ANONYMOUS; if (a->flags & CLIB_MEM_VM_F_HUGETLB) { mmap_flags |= MAP_HUGETLB; log2_page_size = 21; } else { log2_page_size = min_log2 (sysconf (_SC_PAGESIZE)); } } n_pages = ((a->size - 1) >> log2_page_size) + 1; if (a->flags & CLIB_MEM_VM_F_HUGETLB_PREALLOC) { err = clib_sysfs_prealloc_hugepages (a->numa_node, log2_page_size, n_pages); if (err) goto error; } if (fd != -1) if ((ftruncate (fd, (u64) n_pages * (1 << log2_page_size))) == -1) { err = clib_error_return_unix (0, "ftruncate"); goto error; } if (old_mpol != -1) { int rv; long unsigned int mask[16] = { 0 }; mask[0] = 1 << a->numa_node; rv = set_mempolicy (MPOL_BIND, mask, sizeof (mask) * 8 + 1); if (rv == -1 && a->numa_node != 0 && (a->flags & CLIB_MEM_VM_F_NUMA_FORCE) != 0) { err = clib_error_return_unix (0, "set_mempolicy"); goto error; } } addr = mmap (uword_to_pointer (a->requested_va, void *), a->size, (PROT_READ | PROT_WRITE), mmap_flags, fd, 0); if (addr == MAP_FAILED) { err = clib_error_return_unix (0, "mmap"); goto error; } /* re-apply old numa memory policy */ if (old_mpol != -1 && set_mempolicy (old_mpol, old_mask, sizeof (old_mask) * 8 + 1) == -1) { err = clib_error_return_unix (0, "set_mempolicy"); goto error; } a->log2_page_size = log2_page_size; a->n_pages = n_pages; a->addr = addr; a->fd = fd; CLIB_MEM_UNPOISON (addr, a->size); goto done; error: if (fd != -1) close (fd); done: vec_free (filename); return err; } void clib_mem_vm_ext_free (clib_mem_vm_alloc_t * a) { if (a != 0) { clib_mem_vm_free (a->addr, 1ull << a->log2_page_size); if (a->fd != -1) close (a->fd); } } uword clib_mem_vm_reserve (uword start, uword size, clib_mem_page_sz_t log2_page_sz) { clib_mem_main_t *mm = &clib_mem_main; uword pagesize = 1ULL << log2_page_sz; uword sys_page_sz = 1ULL << mm->log2_page_sz; uword n_bytes; void *base = 0, *p; size = round_pow2 (size, pagesize); /* in adition of requested reservation, we also rserve one system page * (typically 4K) adjacent to the start off reservation */ if (start) { /* start address is provided, so we just need to make sure we are not * replacing existing map */ if (start & pow2_mask (log2_page_sz)) return ~0; base = (void *) start - sys_page_sz; base = mmap (base, size + sys_page_sz, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS | MAP_FIXED_NOREPLACE, -1, 0); return (base == MAP_FAILED) ? ~0 : start; } /* to make sure that we get reservation aligned to page_size we need to * request one additional page as mmap will return us address which is * aligned only to system page size */ base = mmap (0, size + pagesize, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); if (base == MAP_FAILED) return ~0; /* return additional space at the end of allocation */ p = base + size + pagesize; n_bytes = (uword) p & pow2_mask (log2_page_sz); if (n_bytes) { p -= n_bytes; munmap (p, n_bytes); } /* return additional space at the start of allocation */ n_bytes = pagesize - sys_page_sz - n_bytes; if (n_bytes) { munmap (base, n_bytes); base += n_bytes; } return (uword) base + sys_page_sz; } clib_mem_vm_map_hdr_t * clib_mem_vm_get_next_map_hdr (clib_mem_vm_map_hdr_t * hdr) { clib_mem_main_t *mm = &clib_mem_main; uword sys_page_sz = 1 << mm->log2_page_sz; clib_mem_vm_map_hdr_t *next; if (hdr == 0) { hdr = mm->first_map; if (hdr) mprotect (hdr, sys_page_sz, PROT_READ); return hdr; } next = hdr->next; mprotect (hdr, sys_page_sz, PROT_NONE); if (next) mprotect (next, sys_page_sz, PROT_READ); return next; } void * clib_mem_vm_map_internal (void *base, clib_mem_page_sz_t log2_page_sz, uword size, int fd, uword offset, char *name) { clib_mem_main_t *mm = &clib_mem_main; clib_mem_vm_map_hdr_t *hdr; uword sys_page_sz = 1 << mm->log2_page_sz; int mmap_flags = MAP_FIXED, is_huge = 0; if (fd != -1) { mmap_flags |= MAP_SHARED; log2_page_sz = clib_mem_get_fd_log2_page_size (fd); if (log2_page_sz > mm->log2_page_sz) is_huge = 1; } else { mmap_flags |= MAP_PRIVATE | MAP_ANONYMOUS; if (log2_page_sz == mm->log2_page_sz) log2_page_sz = CLIB_MEM_PAGE_SZ_DEFAULT; switch (log2_page_sz) { case CLIB_MEM_PAGE_SZ_UNKNOWN: /* will fail later */ break; case CLIB_MEM_PAGE_SZ_DEFAULT: log2_page_sz = mm->log2_page_sz; break; case CLIB_MEM_PAGE_SZ_DEFAULT_HUGE: mmap_flags |= MAP_HUGETLB; log2_page_sz = mm->log2_default_hugepage_sz; is_huge = 1; break; default: mmap_flags |= MAP_HUGETLB; mmap_flags |= log2_page_sz << MAP_HUGE_SHIFT; is_huge = 1; } } if (log2_page_sz == CLIB_MEM_PAGE_SZ_UNKNOWN) return CLIB_MEM_VM_MAP_FAILED; size = round_pow2 (size, 1 << log2_page_sz); base = (void *) clib_mem_vm_reserve ((uword) base, size, log2_page_sz); if (base == (void *) ~0) return CLIB_MEM_VM_MAP_FAILED; base = mmap (base, size, PROT_READ | PROT_WRITE, mmap_flags, fd, offset); if (base == MAP_FAILED) return CLIB_MEM_VM_MAP_FAILED; if (is_huge && (mlock (base, size) != 0)) { munmap (base, size); return CLIB_MEM_VM_MAP_FAILED; } hdr = mmap (base - sys_page_sz, sys_page_sz, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_PRIVATE | MAP_FIXED, -1, 0); if (hdr != base - sys_page_sz) { munmap (base, size); return CLIB_MEM_VM_MAP_FAILED; } if (mm->last_map) { mprotect (mm->last_map, sys_page_sz, PROT_READ | PROT_WRITE); mm->last_map->next = hdr; mprotect (mm->last_map, sys_page_sz, PROT_NONE); } else mm->first_map = hdr; hdr->next = 0; hdr->prev = mm->last_map; mm->last_map = hdr; hdr->base_addr = (uword) base; hdr->log2_page_sz = log2_page_sz; hdr->num_pages = size >> log2_page_sz; hdr->fd = fd; snprintf (hdr->name, CLIB_VM_MAP_HDR_NAME_MAX_LEN - 1, "%s", (char *) name); hdr->name[CLIB_VM_MAP_HDR_NAME_MAX_LEN - 1] = 0; mprotect (hdr, sys_page_sz, PROT_NONE); CLIB_MEM_UNPOISON (base, size); return base; } int clib_mem_vm_unmap (void *base) { clib_mem_main_t *mm = &clib_mem_main; uword size, sys_page_sz = 1 << mm->log2_page_sz; clib_mem_vm_map_hdr_t *hdr = base - sys_page_sz;; if (mprotect (hdr, sys_page_sz, PROT_READ | PROT_WRITE) != 0) return -1; size = hdr->num_pages << hdr->log2_page_sz; if (munmap ((void *) hdr->base_addr, size) != 0) return -1; if (hdr->next) { mprotect (hdr->next, sys_page_sz, PROT_READ | PROT_WRITE); hdr->next->prev = hdr->prev; mprotect (hdr->next, sys_page_sz, PROT_NONE); } else mm->last_map = hdr->prev; if (hdr->prev) { mprotect (hdr->prev, sys_page_sz, PROT_READ | PROT_WRITE); hdr->prev->next = hdr->next; mprotect (hdr->prev, sys_page_sz, PROT_NONE); } else mm->first_map = hdr->next; if (munmap (hdr, sys_page_sz) != 0) return -1; return 0; } void clib_mem_get_page_stats (void *start, clib_mem_page_sz_t log2_page_size, uword n_pages, clib_mem_page_stats_t * stats) { int i, *status = 0; void **ptr = 0; log2_page_size = clib_mem_log2_page_size_validate (log2_page_size); vec_validate (status, n_pages - 1); vec_validate (ptr, n_pages - 1); for (i = 0; i < n_pages; i++) ptr[i] = start + (i << log2_page_size); clib_memset (stats, 0, sizeof (clib_mem_page_stats_t)); if (move_pages (0, n_pages, ptr, 0, status, 0) != 0) { stats->unknown = n_pages; return; } for (i = 0; i < n_pages; i++) { if (status[i] >= 0 && status[i] < CLIB_MAX_NUMAS) { stats->mapped++; stats->per_numa[status[i]]++; } else if (status[i] == -EFAULT) stats->not_mapped++; else stats->unknown++; } } u64 * clib_mem_vm_get_paddr (void *mem, clib_mem_page_sz_t log2_page_size, int n_pages) { int pagesize = sysconf (_SC_PAGESIZE); int fd; int i; u64 *r = 0; log2_page_size = clib_mem_log2_page_size_validate (log2_page_size); if ((fd = open ((char *) "/proc/self/pagemap", O_RDONLY)) == -1) return 0; for (i = 0; i < n_pages; i++) { u64 seek, pagemap = 0; uword vaddr = pointer_to_uword (mem) + (((u64) i) << log2_page_size); seek = ((u64) vaddr / pagesize) * sizeof (u64); if (lseek (fd, seek, SEEK_SET) != seek) goto done; if (read (fd, &pagemap, sizeof (pagemap)) != (sizeof (pagemap))) goto done; if ((pagemap & (1ULL << 63)) == 0) goto done; pagemap &= pow2_mask (55); vec_add1 (r, pagemap * pagesize); } done: close (fd); if (vec_len (r) != n_pages) { vec_free (r); return 0; } return r; } clib_error_t * clib_mem_vm_ext_map (clib_mem_vm_map_t * a) { long unsigned int old_mask[16] = { 0 }; int mmap_flags = MAP_SHARED; clib_error_t *err = 0; int old_mpol = -1; void *addr; int rv; if (a->numa_node) { rv = get_mempolicy (&old_mpol, old_mask, sizeof (old_mask) * 8 + 1, 0, 0); if (rv == -1) { err = clib_error_return_unix (0, "get_mempolicy"); goto done; } } if (a->requested_va) mmap_flags |= MAP_FIXED; if (old_mpol != -1) { long unsigned int mask[16] = { 0 }; mask[0] = 1 << a->numa_node; rv = set_mempolicy (MPOL_BIND, mask, sizeof (mask) * 8 + 1); if (rv == -1) { err = clib_error_return_unix (0, "set_mempolicy"); goto done; } } addr = (void *) mmap (uword_to_pointer (a->requested_va, void *), a->size, PROT_READ | PROT_WRITE, mmap_flags, a->fd, 0); if (addr == MAP_FAILED) return clib_error_return_unix (0, "mmap"); /* re-apply old numa memory policy */ if (old_mpol != -1 && set_mempolicy (old_mpol, old_mask, sizeof (old_mask) * 8 + 1) == -1) { err = clib_error_return_unix (0, "set_mempolicy"); goto done; } a->addr = addr; CLIB_MEM_UNPOISON (addr, a->size); done: return err; } /* * fd.io coding-style-patch-verification: ON * * Local Variables: * eval: (c-set-style "gnu") * End: */