/tools/testing/selftests/vm/ |
D | mremap_dontunmap.c | 22 unsigned long page_size; variable 49 void *source_mapping = mmap(NULL, num_pages * page_size, PROT_NONE, in kernel_support_for_mremap_dontunmap() 56 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in kernel_support_for_mremap_dontunmap() 61 BUG_ON(munmap(dest_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap() 65 BUG_ON(munmap(source_mapping, num_pages * page_size) == -1, in kernel_support_for_mremap_dontunmap() 74 BUG_ON(size & (page_size - 1), in check_region_contains_byte() 76 BUG_ON((unsigned long)addr & (page_size - 1), in check_region_contains_byte() 79 memset(page_buffer, byte, page_size); in check_region_contains_byte() 81 unsigned long num_pages = size / page_size; in check_region_contains_byte() 87 memcmp(addr + (i * page_size), page_buffer, page_size); in check_region_contains_byte() [all …]
|
D | mlock2-tests.c | 195 unsigned long page_size = getpagesize(); in test_mlock_lock() local 197 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_mlock_lock() 204 if (mlock2_(map, 2 * page_size, 0)) { in test_mlock_lock() 217 if (munlock(map, 2 * page_size)) { in test_mlock_lock() 225 munmap(map, 2 * page_size); in test_mlock_lock() 243 unsigned long page_size = getpagesize(); in unlock_onfault_check() local 246 is_vma_lock_on_fault((unsigned long)map + page_size)) { in unlock_onfault_check() 258 unsigned long page_size = getpagesize(); in test_mlock_onfault() local 260 map = mmap(NULL, 2 * page_size, PROT_READ | PROT_WRITE, in test_mlock_onfault() 267 if (mlock2_(map, 2 * page_size, MLOCK_ONFAULT)) { in test_mlock_onfault() [all …]
|
D | map_fixed_noreplace.c | 50 unsigned long flags, addr, size, page_size; in main() local 53 page_size = sysconf(_SC_PAGE_SIZE); in main() 56 size = 5 * page_size; in main() 68 size = 5 * page_size; in main() 80 if (munmap((void *)addr, 5 * page_size) != 0) { in main() 88 addr = base_addr + page_size; in main() 89 size = 3 * page_size; in main() 109 size = 5 * page_size; in main() 129 addr = base_addr + (2 * page_size); in main() 130 size = page_size; in main() [all …]
|
D | ksm_tests.c | 242 static int check_ksm_merge(int mapping, int prot, long page_count, int timeout, size_t page_size) in check_ksm_merge() argument 253 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); in check_ksm_merge() 257 if (ksm_merge_pages(map_ptr, page_size * page_count, start_time, timeout)) in check_ksm_merge() 263 munmap(map_ptr, page_size * page_count); in check_ksm_merge() 269 munmap(map_ptr, page_size * page_count); in check_ksm_merge() 273 static int check_ksm_unmerge(int mapping, int prot, int timeout, size_t page_size) in check_ksm_unmerge() argument 285 map_ptr = allocate_memory(NULL, prot, mapping, '*', page_size * page_count); in check_ksm_unmerge() 289 if (ksm_merge_pages(map_ptr, page_size * page_count, start_time, timeout)) in check_ksm_unmerge() 294 memset(map_ptr + page_size, '+', 1); in check_ksm_unmerge() 303 munmap(map_ptr, page_size * page_count); in check_ksm_unmerge() [all …]
|
D | userfaultfd.c | 66 static unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; variable 109 ((pthread_mutex_t *) ((___area) + (___nr)*page_size)) 116 ((___area) + (___nr)*page_size + \ 208 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) in anon_release_pages() 214 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in anon_allocate_area() 230 rel_area == huge_fd_off0 ? 0 : nr_pages * page_size, in hugetlb_release_pages() 231 nr_pages * page_size)) in hugetlb_release_pages() 240 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in hugetlb_allocate_area() 245 *alloc_area == area_src ? 0 : nr_pages * page_size); in hugetlb_allocate_area() 250 area_alias = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, in hugetlb_allocate_area() [all …]
|
D | memfd_secret.c | 37 static unsigned long page_size; variable 65 if (len % page_size != 0) in test_mlock_limit() 66 len = (len/page_size) * page_size; in test_mlock_limit() 184 mem = mmap(NULL, page_size, prot, mode, fd, 0); in test_remote_access() 190 ftruncate(fd, page_size); in test_remote_access() 191 memset(mem, PATTERN, page_size); in test_remote_access() 236 page_size = sysconf(_SC_PAGE_SIZE); in prepare() 237 if (!page_size) in prepare() 249 page_size, mlock_limit_cur, mlock_limit_max); in prepare() 251 if (page_size > mlock_limit_cur) in prepare() [all …]
|
D | khugepaged.c | 20 static unsigned long page_size; variable 455 for (i = start / page_size; i < end / page_size; i++) in fill_memory() 456 p[i * page_size / sizeof(*p)] = i + 0xdead0000; in fill_memory() 463 for (i = start / page_size; i < end / page_size; i++) { in validate_memory() 464 if (p[i * page_size / sizeof(*p)] != i + 0xdead0000) { in validate_memory() 466 i, p[i * page_size / sizeof(*p)]); in validate_memory() 522 madvise(p, page_size, MADV_DONTNEED); in alloc_at_fault() 566 fill_memory(p, 0, page_size); in collapse_single_pte_entry() 573 validate_memory(p, 0, page_size); in collapse_single_pte_entry() 588 fill_memory(p, 0, (hpage_pmd_nr - max_ptes_none - 1) * page_size); in collapse_max_ptes_none() [all …]
|
D | mremap_test.c | 53 #define PTE page_size 341 int page_size; in main() local 352 page_size = sysconf(_SC_PAGESIZE); in main() 355 test_cases[0] = MAKE_TEST(page_size, page_size, page_size, in main() 359 test_cases[1] = MAKE_TEST(page_size, page_size/4, page_size, in main() 362 test_cases[2] = MAKE_TEST(page_size/4, page_size, page_size, in main() 395 perf_test_cases[0] = MAKE_TEST(page_size, page_size, _1GB, NON_OVERLAPPING, EXPECT_SUCCESS, in main()
|
D | hmm-tests.c | 59 unsigned int page_size; in FIXTURE() local 67 unsigned int page_size; in FIXTURE() local 86 self->page_size = sysconf(_SC_PAGE_SIZE); in FIXTURE_SETUP() 87 self->page_shift = ffs(self->page_size) - 1; in FIXTURE_SETUP() 95 self->page_size = sysconf(_SC_PAGE_SIZE); in FIXTURE_SETUP() 96 self->page_shift = ffs(self->page_size) - 1; in FIXTURE_SETUP() 234 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; in TEST_F() 256 i = 2 * self->page_size / sizeof(*ptr); in TEST_F() 265 val = *(int *)(buffer->ptr + self->page_size); in TEST_F() 276 for (i = 0; i < 2 * self->page_size / sizeof(*ptr); ++i) in TEST_F() [all …]
|
/tools/testing/selftests/mincore/ |
D | mincore_selftest.c | 33 int page_size; in TEST() local 37 page_size = sysconf(_SC_PAGESIZE); in TEST() 45 retval = mincore(NULL, page_size, vec); in TEST() 50 addr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, in TEST() 58 retval = mincore(addr + 1, page_size, vec); in TEST() 70 retval = mincore(addr, page_size, NULL); in TEST() 73 munmap(addr, page_size); in TEST() 88 int page_size; in TEST() local 90 page_size = sysconf(_SC_PAGESIZE); in TEST() 94 addr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, in TEST() [all …]
|
/tools/testing/selftests/bpf/prog_tests/ |
D | ringbuf.c | 90 int page_size = getpagesize(); in test_ringbuf() local 97 skel->maps.ringbuf.max_entries = page_size; in test_ringbuf() 105 mmap_ptr = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, rb_fd, 0); in test_ringbuf() 107 tmp_ptr = mremap(mmap_ptr, page_size, 2 * page_size, MREMAP_MAYMOVE); in test_ringbuf() 110 ASSERT_ERR(mprotect(mmap_ptr, page_size, PROT_EXEC), "exec_cons_pos_protect"); in test_ringbuf() 111 ASSERT_OK(munmap(mmap_ptr, page_size), "unmap_rw"); in test_ringbuf() 114 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, page_size); in test_ringbuf() 120 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, 2 * page_size); in test_ringbuf() 124 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, 3 * page_size); in test_ringbuf() 126 mmap_ptr = mmap(NULL, 2 * page_size, PROT_WRITE, MAP_SHARED, rb_fd, 2 * page_size); in test_ringbuf() [all …]
|
D | mmap.c | 12 long page_size = sysconf(_SC_PAGE_SIZE); in roundup_page() local 13 return (sz + page_size - 1) / page_size * page_size; in roundup_page() 21 const long page_size = sysconf(_SC_PAGE_SIZE); in test_mmap() local 36 err = bpf_map__set_max_entries(skel->maps.rdonly_map, page_size); in test_mmap() 42 4 * (page_size / sizeof(u64))); in test_mmap() 55 tmp1 = mmap(NULL, page_size, PROT_READ | PROT_WRITE, MAP_SHARED, rdmap_fd, 0); in test_mmap() 57 munmap(tmp1, page_size); in test_mmap() 61 tmp1 = mmap(NULL, page_size, PROT_READ, MAP_SHARED, rdmap_fd, 0); in test_mmap() 200 tmp0 = mmap(NULL, 4 * page_size, PROT_READ, MAP_SHARED | MAP_ANONYMOUS, in test_mmap() 206 tmp1 = mmap(tmp0, 3 * page_size, PROT_READ, MAP_SHARED | MAP_FIXED, in test_mmap() [all …]
|
D | map_ptr.c | 15 int page_size = getpagesize(); in test_map_ptr() local 21 err = bpf_map__set_max_entries(skel->maps.m_ringbuf, page_size); in test_map_ptr() 29 skel->bss->page_size = page_size; in test_map_ptr()
|
D | ringbuf_multi.c | 46 int page_size = getpagesize(); in test_ringbuf_multi() local 53 err = bpf_map__set_max_entries(skel->maps.ringbuf1, page_size); in test_ringbuf_multi() 57 err = bpf_map__set_max_entries(skel->maps.ringbuf2, page_size); in test_ringbuf_multi() 61 err = bpf_map__set_max_entries(bpf_map__inner_map(skel->maps.ringbuf_arr), page_size); in test_ringbuf_multi() 65 proto_fd = bpf_create_map(BPF_MAP_TYPE_RINGBUF, 0, 0, page_size, 0); in test_ringbuf_multi()
|
/tools/testing/selftests/powerpc/primitives/ |
D | load_unaligned_zeropad.c | 38 static int page_size; variable 43 if (mprotect(mem_region + page_size, page_size, PROT_NONE)) { in protect_region() 53 if (mprotect(mem_region + page_size, page_size, PROT_READ|PROT_WRITE)) { in unprotect_region() 125 page_size = getpagesize(); in test_body() 126 mem_region = mmap(NULL, page_size * 2, PROT_READ|PROT_WRITE, in test_body() 131 for (i = 0; i < page_size; i++) in test_body() 134 memset(mem_region+page_size, 0, page_size); in test_body() 138 for (i = 0; i < page_size; i++) in test_body()
|
/tools/power/acpi/os_specific/service_layers/ |
D | osunixmap.c | 67 acpi_size page_size; in acpi_os_map_memory() local 78 page_size = acpi_os_get_page_size(); in acpi_os_map_memory() 79 offset = where % page_size; in acpi_os_map_memory() 112 acpi_size page_size; in acpi_os_unmap_memory() local 114 page_size = acpi_os_get_page_size(); in acpi_os_unmap_memory() 115 offset = ACPI_TO_INTEGER(where) % page_size; in acpi_os_unmap_memory()
|
/tools/testing/selftests/kvm/lib/s390x/ |
D | processor.c | 18 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", in virt_pgd_alloc() 19 vm->page_size); in virt_pgd_alloc() 26 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_pgd_alloc() 43 memset(addr_gpa2hva(vm, taddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_alloc_region() 55 TEST_ASSERT((gva % vm->page_size) == 0, in virt_pg_map() 58 gva, vm->page_size); in virt_pg_map() 63 TEST_ASSERT((gpa % vm->page_size) == 0, in virt_pg_map() 66 gva, vm->page_size); in virt_pg_map() 70 gva, vm->max_gfn, vm->page_size); in virt_pg_map() 94 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", in addr_gva2gpa() [all …]
|
/tools/testing/selftests/powerpc/copyloops/ |
D | exc_validate.c | 81 int page_size; in test_copy_exception() local 85 page_size = getpagesize(); in test_copy_exception() 86 p = mmap(NULL, page_size * 2, PROT_READ|PROT_WRITE, in test_copy_exception() 94 memset(p, 0, page_size); in test_copy_exception() 98 if (mprotect(p + page_size, page_size, PROT_NONE)) { in test_copy_exception() 103 q = p + page_size - MAX_LEN; in test_copy_exception()
|
/tools/testing/selftests/powerpc/mm/ |
D | stack_expansion_ldst.c | 157 static void test_one_type(enum access_type type, unsigned long page_size, unsigned long rlim_cur) in test_one_type() argument 162 for (delta = page_size; delta <= rlim_cur; delta += page_size) in test_one_type() 173 unsigned long page_size; in test() local 176 page_size = getpagesize(); in test() 181 test_one_type(LOAD, page_size, rlimit.rlim_cur); in test() 183 test_one_type(STORE, page_size, rlimit.rlim_cur); in test()
|
D | bad_accesses.c | 69 unsigned long i, j, addr, region_shift, page_shift, page_size; in test() local 82 page_size = sysconf(_SC_PAGESIZE); in test() 83 if (page_size == (64 * 1024)) in test() 88 if (page_size == (64 * 1024) || !hash_mmu) { in test() 93 } else if (page_size == (4 * 1024) && hash_mmu) { in test()
|
D | large_vm_fork_separation.c | 27 unsigned long page_size; in test() local 30 page_size = sysconf(_SC_PAGESIZE); in test() 31 SKIP_IF(page_size != 65536); in test() 34 p = mmap((void *)(512ul << 40), page_size, PROT_READ | PROT_WRITE, in test()
|
/tools/perf/arch/x86/util/ |
D | intel-bts.c | 154 opts->auxtrace_mmap_pages = MiB(4) / page_size; in intel_bts_recording_options() 156 opts->auxtrace_mmap_pages = KiB(128) / page_size; in intel_bts_recording_options() 158 opts->mmap_pages = KiB(256) / page_size; in intel_bts_recording_options() 162 opts->mmap_pages = KiB(256) / page_size; in intel_bts_recording_options() 166 opts->auxtrace_mmap_pages * (size_t)page_size; in intel_bts_recording_options() 170 sz = round_up(sz, page_size) / page_size; in intel_bts_recording_options() 174 opts->auxtrace_mmap_pages * (size_t)page_size) { in intel_bts_recording_options() 177 opts->auxtrace_mmap_pages * (size_t)page_size); in intel_bts_recording_options() 191 opts->auxtrace_mmap_pages = MiB(4) / page_size; in intel_bts_recording_options() 193 opts->auxtrace_mmap_pages = KiB(128) / page_size; in intel_bts_recording_options() [all …]
|
/tools/testing/selftests/kvm/lib/aarch64/ |
D | processor.c | 21 return (v + vm->page_size) & ~(vm->page_size - 1); in page_align() 81 page_align(vm, ptrs_per_pgd(vm) * 8) / vm->page_size, in virt_pgd_alloc() 94 TEST_ASSERT((vaddr % vm->page_size) == 0, in _virt_pg_map() 96 " vaddr: 0x%lx vm->page_size: 0x%x", vaddr, vm->page_size); in _virt_pg_map() 100 TEST_ASSERT((paddr % vm->page_size) == 0, in _virt_pg_map() 102 " paddr: 0x%lx vm->page_size: 0x%x", paddr, vm->page_size); in _virt_pg_map() 106 paddr, vm->max_gfn, vm->page_size); in _virt_pg_map() 172 return pte_addr(vm, *ptep) + (gva & (vm->page_size - 1)); in addr_gva2gpa() 296 size_t stack_size = vm->page_size == 4096 ? in aarch64_vcpu_add_default() 297 DEFAULT_STACK_PGS * vm->page_size : in aarch64_vcpu_add_default() [all …]
|
/tools/testing/selftests/arm64/mte/ |
D | check_child_memory.c | 23 static size_t page_size; variable 149 page_size = getpagesize(); in main() 150 if (!page_size) { in main() 154 sizes[item - 3] = page_size - 1; in main() 155 sizes[item - 2] = page_size; in main() 156 sizes[item - 1] = page_size + 1; in main()
|
/tools/testing/selftests/futex/functional/ |
D | futex_wait_uninitialized_heap.c | 69 long page_size; in main() local 89 page_size = sysconf(_SC_PAGESIZE); in main() 91 buf = mmap(NULL, page_size, PROT_READ|PROT_WRITE, in main()
|