| /linux/tools/testing/selftests/mm/ |
| A D | mremap_dontunmap.c | 19 unsigned long page_size; variable 53 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in kernel_support_for_mremap_dontunmap() 71 BUG_ON(size & (page_size - 1), in check_region_contains_byte() 84 memcmp(addr + (i * page_size), page_buffer, page_size); in check_region_contains_byte() 108 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in mremap_dontunmap_simple() 150 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in mremap_dontunmap_simple_shmem() 202 mremap(source_mapping, num_pages * page_size, num_pages * page_size, in mremap_dontunmap_simple_fixed() 254 mremap(source_mapping + (5 * page_size), 5 * page_size, in mremap_dontunmap_partial_mapping() 255 5 * page_size, in mremap_dontunmap_partial_mapping() 264 (source_mapping + (5 * page_size), 5 * page_size, 0) != 0, in mremap_dontunmap_partial_mapping() [all …]
|
| A D | mseal_test.c | 284 ret = sys_munmap(ptr + page_size, 2 * page_size); in test_seal_unmapped_middle() 299 ret = sys_mseal(ptr + 3 * page_size, page_size); in test_seal_unmapped_middle() 345 ret = sys_mprotect(ptr + page_size, 2 * page_size, in test_seal_multiple_vmas() 354 ret = sys_mprotect(ptr + page_size, 2 * page_size, in test_seal_multiple_vmas() 384 ret = sys_mseal(ptr + page_size, 3 * page_size); in test_seal_split_start() 405 ret = sys_mseal(ptr + 3 * page_size, page_size); in test_seal_split_end() 444 ret = sys_mseal(ptr - page_size, 5 * page_size); in test_seal_invalid_input() 786 ret = sys_mseal(ptr + page_size, page_size); in test_seal_mprotect_partial_mprotect_tail() 1034 ret = sys_mseal(ptr + page_size, page_size); in test_seal_munmap_partial_across_vmas() 1564 ret = sys_mseal(ptr + page_size, page_size); in test_seal_merge_and_split() [all …]
|
| A D | mlock2-tests.c | 187 munmap(map, 2 * page_size); in test_mlock_lock() 195 munmap(map, 2 * page_size); in test_mlock_lock() 200 munmap(map, 2 * page_size); in test_mlock_lock() 238 munmap(map, 2 * page_size); in test_mlock_onfault() 246 munmap(map, 2 * page_size); in test_mlock_onfault() 251 munmap(map, 2 * page_size); in test_mlock_onfault() 267 munmap(map, 2 * page_size); in test_lock_onfault_of_present() 274 munmap(map, 2 * page_size); in test_lock_onfault_of_present() 300 munmap(map, 2 * page_size); in test_munlockall0() 340 munmap(map, 2 * page_size); in test_munlockall1() [all …]
|
| A D | map_fixed_noreplace.c | 53 size = 5 * page_size; in main() 60 size = 5 * page_size; in main() 73 addr = base_addr + page_size; in main() 74 size = 3 * page_size; in main() 92 size = 5 * page_size; in main() 111 size = page_size; in main() 129 size = 2 * page_size; in main() 147 size = 2 * page_size; in main() 165 size = page_size; in main() 183 size = page_size; in main() [all …]
|
| A D | ksm_tests.c | 377 munmap(map_ptr, page_size * page_count); in check_ksm_merge() 402 memset(map_ptr + page_size, '+', 1); in check_ksm_unmerge() 519 memset(numa1_map_ptr, '*', page_size); in check_ksm_numa_merge() 520 memset(numa2_map_ptr, '*', page_size); in check_ksm_numa_merge() 538 numa_free(numa1_map_ptr, page_size); in check_ksm_numa_merge() 539 numa_free(numa2_map_ptr, page_size); in check_ksm_numa_merge() 544 numa_free(numa1_map_ptr, page_size); in check_ksm_numa_merge() 545 numa_free(numa2_map_ptr, page_size); in check_ksm_numa_merge() 901 ksm_scan_limit_sec, page_size); in main() 906 page_size); in main() [all …]
|
| A D | uffd-unit-tests.c | 196 page_size = psize(); in uffd_setup_environment() 587 page_size); in uffd_minor_test_common() 615 nr_pages * page_size / in uffd_minor_test_common() 720 page_size, false); in faulting_process() 743 area_dst = mremap(area_dst, nr_pages * page_size, nr_pages * page_size, in faulting_process() 920 if (res != page_size) in do_uffdio_zeropage() 1000 else if (res != page_size) in do_uffdio_poison() 1017 offset &= ~(page_size-1); in uffd_poison_handle_fault() 1020 if (offset & page_size) in uffd_poison_handle_fault() 1145 if (chunk_size > page_size) { in uffd_move_test_common() [all …]
|
| A D | memfd_secret.c | 36 static unsigned long page_size; variable 64 if (len % page_size != 0) in test_mlock_limit() 65 len = (len/page_size) * page_size; in test_mlock_limit() 107 memset(mem, PATTERN, page_size); in test_vmsplice() 110 iov.iov_len = page_size; in test_vmsplice() 118 munmap(mem, page_size); in test_vmsplice() 228 memset(mem, PATTERN, page_size); in test_remote_access() 274 if (!page_size) in prepare() 289 mlock_limit_cur = page_size; in prepare() 291 mlock_limit_max = page_size; in prepare() [all …]
|
| A D | pagemap_ioctl.c | 37 int page_size; variable 354 mem[(mem_size/page_size - 1) * page_size]++; in sanity_tests_sd() 377 ret = pagemap_ioctl(mem + page_size, 2 * page_size, vec, 1, in sanity_tests_sd() 454 mem[i * page_size]++; in sanity_tests_sd() 722 memset(mem + page_size, 0, mem_size - (2 * page_size)); in base_tests() 765 written = pagemap_ioctl(&mem[vec_size/2 * page_size], 2 * page_size, vec, 1, 0, in base_tests() 1016 map = aligned_alloc(page_size, page_size); in test_simple() 1020 wp_init(map, page_size); in test_simple() 1042 wp_free(map, page_size); in test_simple() 1257 mem2 = mmap(mem + page_size, page_size, PROT_READ|PROT_WRITE, in mprotect_tests() [all …]
|
| A D | mremap_test.c | 60 #define PTE page_size 253 munmap(start + page_size, page_size); in mremap_expand_merge() 254 remap = mremap(start, page_size, 2 * page_size, 0); in mremap_expand_merge() 257 munmap(start, page_size); in mremap_expand_merge() 258 munmap(start + 2 * page_size, page_size); in mremap_expand_merge() 294 munmap(start + 2 * page_size, page_size); in mremap_expand_merge_offset() 295 remap = mremap(start + page_size, page_size, 2 * page_size, 0); in mremap_expand_merge_offset() 726 int page_size; in main() local 766 test_cases[0] = MAKE_TEST(page_size, page_size, page_size, in main() 770 test_cases[1] = MAKE_TEST(page_size, page_size/4, page_size, in main() [all …]
|
| A D | seal_elf.c | 46 unsigned long page_size = getpagesize(); in seal_support() local 48 ptr = mmap(NULL, page_size, PROT_READ, MAP_ANONYMOUS | MAP_PRIVATE, -1, 0); in seal_support() 52 ret = sys_mseal(ptr, page_size); in seal_support() 69 unsigned long page_size = getpagesize(); in test_seal_elf() local 76 if (((unsigned long long)ptr % page_size) != 0) in test_seal_elf() 77 ptr = (unsigned long long)ptr & ~(page_size - 1); in test_seal_elf() 81 ret = sys_mprotect((void *)ptr, page_size, PROT_READ|PROT_WRITE); in test_seal_elf() 85 ret = sys_mprotect((void *)ptr, page_size, PROT_READ); in test_seal_elf() 115 ret = sys_mprotect((void *)ptr, page_size, PROT_READ | PROT_WRITE); in test_seal_elf()
|
| A D | uffd-stress.c | 158 copy_page_retry(uffd, page_nr * page_size); in background_thread() 167 nr_pages_per_cpu * page_size, true); in background_thread() 174 copy_page_retry(uffd, page_nr * page_size); in background_thread() 249 uint64_t mem_size = nr_pages * page_size; in userfaultfd_stress() 260 if (posix_memalign(&area, page_size, page_size)) in userfaultfd_stress() 263 bzero(zeropage, page_size); in userfaultfd_stress() 338 nr_pages * page_size, false); in userfaultfd_stress() 398 page_size = default_huge_page_size(); in parse_test_type_arg() 400 page_size = sysconf(_SC_PAGE_SIZE); in parse_test_type_arg() 402 if (!page_size) in parse_test_type_arg() [all …]
|
| /linux/tools/testing/selftests/kvm/ |
| A D | guest_memfd_test.c | 64 page_size - 1, page_size); in test_fallocate() 78 total_size + page_size, page_size); in test_fallocate() 82 page_size, page_size - 1); in test_fallocate() 86 page_size, page_size); in test_fallocate() 89 ret = fallocate(fd, FALLOC_FL_KEEP_SIZE, page_size, page_size); in test_fallocate() 100 {0, page_size - 1}, in test_invalid_punch_hole() 105 {1, page_size}, in test_invalid_punch_hole() 108 {page_size, 1}, in test_invalid_punch_hole() 109 {page_size, page_size - 1}, in test_invalid_punch_hole() 110 {page_size, page_size + 1}, in test_invalid_punch_hole() [all …]
|
| /linux/tools/testing/selftests/mincore/ |
| A D | mincore_selftest.c | 33 int page_size; in TEST() local 73 munmap(addr, page_size); in TEST() 88 int page_size; in TEST() local 107 mlock(addr, page_size); in TEST() 118 munlock(addr, page_size); in TEST() 125 munmap(addr, page_size); in TEST() 144 int page_size; in TEST() local 165 mlock(addr, page_size); in TEST() 173 munmap(addr, page_size); in TEST() 193 int page_size; in TEST() local [all …]
|
| /linux/tools/testing/selftests/lsm/ |
| A D | lsm_get_self_attr_test.c | 43 __u32 size = page_size; in TEST() 86 size = page_size; in TEST() 98 size = page_size; in TEST() 125 size = page_size; in TEST() 131 size = page_size; in TEST() 188 size = page_size; in TEST() 200 size = page_size; in TEST() 214 size = page_size; in TEST() 228 size = page_size; in TEST() 242 size = page_size; in TEST() [all …]
|
| A D | lsm_list_modules_test.c | 20 const long page_size = sysconf(_SC_PAGESIZE); in TEST() local 21 __u64 *syscall_lsms = calloc(page_size, 1); in TEST() 33 const long page_size = sysconf(_SC_PAGESIZE); in TEST() local 34 __u32 size = page_size; in TEST() 45 __u64 *syscall_lsms = calloc(page_size, 1); in TEST() 60 __u64 *syscall_lsms = calloc(page_size, 1); in TEST() 61 __u32 size = page_size; in TEST() 67 ASSERT_EQ(page_size, size); in TEST() 75 __u32 size = page_size; in TEST() 76 __u64 *syscall_lsms = calloc(page_size, 1); in TEST() [all …]
|
| A D | lsm_set_self_attr_test.c | 26 const long page_size = sysconf(_SC_PAGESIZE); in TEST() local 27 struct lsm_ctx *ctx = calloc(page_size, 1); in TEST() 28 __u32 size = page_size; in TEST() 42 const long page_size = sysconf(_SC_PAGESIZE); in TEST() local 43 struct lsm_ctx *ctx = calloc(page_size, 1); in TEST() 44 __u32 size = page_size; in TEST() 58 const long page_size = sysconf(_SC_PAGESIZE); in TEST() local 59 char *ctx = calloc(page_size, 1); in TEST() 60 __u32 size = page_size; in TEST()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| A D | mem.c | 45 rdma_umem_for_each_dma_block (umem, &biter, page_size) { in mlx5_ib_populate_pas() 63 unsigned long page_size; in __mlx5_umem_find_best_quantized_pgoff() local 66 page_size = ib_umem_find_best_pgoff(umem, pgsz_bitmap, pgoff_bitmask); in __mlx5_umem_find_best_quantized_pgoff() 67 if (!page_size) in __mlx5_umem_find_best_quantized_pgoff() 77 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff() 78 while (page_offset & ~(u64)(page_offset_mask * (page_size / scale))) { in __mlx5_umem_find_best_quantized_pgoff() 79 page_size /= 2; in __mlx5_umem_find_best_quantized_pgoff() 80 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff() 87 if (!(pgsz_bitmap & page_size)) in __mlx5_umem_find_best_quantized_pgoff() 91 (unsigned long)page_offset / (page_size / scale); in __mlx5_umem_find_best_quantized_pgoff() [all …]
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| A D | mmap.c | 12 long page_size = sysconf(_SC_PAGE_SIZE); in roundup_page() local 13 return (sz + page_size - 1) / page_size * page_size; in roundup_page() 57 munmap(tmp1, page_size); in test_mmap() 209 munmap(tmp0, 4 * page_size); in test_mmap() 214 err = munmap(tmp1 + page_size, page_size); in test_mmap() 216 munmap(tmp1, 4 * page_size); in test_mmap() 221 tmp2 = mmap(tmp1 + page_size, page_size, PROT_READ, in test_mmap() 224 munmap(tmp1, page_size); in test_mmap() 225 munmap(tmp1 + 2*page_size, 2 * page_size); in test_mmap() 250 munmap(tmp2, 4 * page_size); in test_mmap() [all …]
|
| A D | ringbuf.c | 92 int page_size = getpagesize(); in ringbuf_write_subtest() local 147 int page_size = getpagesize(); in ringbuf_subtest() local 157 skel->maps.ringbuf.max_entries = page_size; in ringbuf_subtest() 167 tmp_ptr = mremap(mmap_ptr, page_size, 2 * page_size, MREMAP_MAYMOVE); in ringbuf_subtest() 174 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, page_size); in ringbuf_subtest() 180 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, 2 * page_size); in ringbuf_subtest() 184 mmap_ptr = mmap(NULL, page_size, PROT_WRITE, MAP_SHARED, rb_fd, 3 * page_size); in ringbuf_subtest() 186 mmap_ptr = mmap(NULL, 2 * page_size, PROT_WRITE, MAP_SHARED, rb_fd, 2 * page_size); in ringbuf_subtest() 200 mmap_ptr = mmap(NULL, page_size, PROT_READ, MAP_SHARED, rb_fd, page_size); in ringbuf_subtest() 234 CHECK(skel->bss->ring_size != page_size, in ringbuf_subtest() [all …]
|
| /linux/drivers/pci/endpoint/ |
| A D | pci-epc-mem.c | 54 size_t page_size; in pci_epc_multi_mem_init() local 70 page_size = windows[i].page_size; in pci_epc_multi_mem_init() 71 if (page_size < PAGE_SIZE) in pci_epc_multi_mem_init() 72 page_size = PAGE_SIZE; in pci_epc_multi_mem_init() 73 page_shift = ilog2(page_size); in pci_epc_multi_mem_init() 94 mem->window.page_size = page_size; in pci_epc_multi_mem_init() 135 mem_window.page_size = page_size; in pci_epc_mem_init() 249 size_t page_size; in pci_epc_mem_free_addr() local 259 page_size = mem->window.page_size; in pci_epc_mem_free_addr() 260 page_shift = ilog2(page_size); in pci_epc_mem_free_addr() [all …]
|
| /linux/tools/testing/selftests/powerpc/primitives/ |
| A D | load_unaligned_zeropad.c | 38 static int page_size; variable 43 if (mprotect(mem_region + page_size, page_size, PROT_NONE)) { in protect_region() 53 if (mprotect(mem_region + page_size, page_size, PROT_READ|PROT_WRITE)) { in unprotect_region() 125 page_size = getpagesize(); in test_body() 126 mem_region = mmap(NULL, page_size * 2, PROT_READ|PROT_WRITE, in test_body() 131 for (i = 0; i < page_size; i++) in test_body() 134 memset(mem_region+page_size, 0, page_size); in test_body() 138 for (i = 0; i < page_size; i++) in test_body()
|
| /linux/arch/powerpc/mm/ |
| A D | init_64.c | 188 unsigned long page_size) in altmap_cross_boundary() argument 209 start = ALIGN_DOWN(start, page_size); in __vmemmap_populate() 213 for (; start < end; start += page_size) { in __vmemmap_populate() 223 if (vmemmap_populated(start, page_size)) in __vmemmap_populate() 252 int nr_pfns = page_size >> PAGE_SHIFT; in __vmemmap_populate() 263 start, start + page_size, p); in __vmemmap_populate() 327 start = ALIGN_DOWN(start, page_size); in __vmemmap_free() 335 for (; start < end; start += page_size) { in __vmemmap_free() 345 if (vmemmap_populated(start, page_size)) in __vmemmap_free() 360 if (page_size < PAGE_SIZE) { in __vmemmap_free() [all …]
|
| /linux/drivers/misc/ |
| A D | vmw_balloon.c | 675 ctl->page_size); in vmballoon_alloc_page_list() 686 ctl->page_size); in vmballoon_alloc_page_list() 717 page_size); in vmballoon_handle_one_result() 933 ctl->page_size); in vmballoon_release_refused_pages() 936 ctl->page_size); in vmballoon_release_refused_pages() 1099 .page_size = b->max_page_size, in vmballoon_inflate() 1134 ctl.page_size); in vmballoon_inflate() 1150 ctl.page_size--; in vmballoon_inflate() 1263 ctl.page_size); in vmballoon_deflate() 1268 ctl.page_size); in vmballoon_deflate() [all …]
|
| /linux/tools/testing/selftests/kvm/lib/s390x/ |
| A D | processor.c | 17 TEST_ASSERT(vm->page_size == PAGE_SIZE, "Unsupported page size: 0x%x", in virt_arch_pgd_alloc() 18 vm->page_size); in virt_arch_pgd_alloc() 26 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_arch_pgd_alloc() 43 memset(addr_gpa2hva(vm, taddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_alloc_region() 55 TEST_ASSERT((gva % vm->page_size) == 0, in virt_arch_pg_map() 58 gva, vm->page_size); in virt_arch_pg_map() 63 TEST_ASSERT((gpa % vm->page_size) == 0, in virt_arch_pg_map() 66 gva, vm->page_size); in virt_arch_pg_map() 70 gva, vm->max_gfn, vm->page_size); in virt_arch_pg_map() 95 vm->page_size); in addr_arch_gva2gpa() [all …]
|
| /linux/drivers/accel/habanalabs/common/mmu/ |
| A D | mmu.c | 182 if ((page_size % mmu_prop->page_size) == 0) { in hl_mmu_get_real_page_size() 188 page_size, mmu_prop->page_size >> 10); in hl_mmu_get_real_page_size() 200 else if ((page_size % prop->pmmu_huge.page_size) == 0) in hl_mmu_get_prop() 388 u32 page_size; in hl_mmu_map_contiguous() local 394 page_size = prop->dmmu.page_size; in hl_mmu_map_contiguous() 397 page_size = prop->pmmu.page_size; in hl_mmu_map_contiguous() 400 page_size = prop->pmmu_huge.page_size; in hl_mmu_map_contiguous() 448 u32 page_size; in hl_mmu_unmap_contiguous() local 454 page_size = prop->dmmu.page_size; in hl_mmu_unmap_contiguous() 457 page_size = prop->pmmu.page_size; in hl_mmu_unmap_contiguous() [all …]
|