| /linux/mm/ |
| A D | nommu.c | 457 if (region->vm_start < pregion->vm_start) in add_nommu_region() 459 else if (region->vm_start > pregion->vm_start) in add_nommu_region() 961 vma->vm_start = region->vm_start; in do_mmap_private() 987 region->vm_start = vma->vm_start = 0; in do_mmap_private() 1119 vma->vm_start = 0; in do_mmap() 1154 vma->vm_start = region->vm_start = addr; in do_mmap() 1181 result = vma->vm_start; in do_mmap() 1330 region->vm_start = new->vm_start = addr; in split_vma() 1347 vma->vm_region->vm_start = vma->vm_start = addr; in split_vma() 1389 vma->vm_start = to; in vmi_shrink_vma() [all …]
|
| A D | vma.c | 374 WARN_ON(vma->vm_start >= addr); in __split_vma() 390 new->vm_start = addr; in __split_vma() 422 vma->vm_start = addr; in __split_vma() 611 adjust->vm_start += adj_start; in commit_merge() 765 vmg->start = prev->vm_start; in vma_merge_existing_range() 785 vmg->start = prev->vm_start; in vma_merge_existing_range() 821 vmg->start = vma->vm_start; in vma_merge_existing_range() 944 vmg->start = prev->vm_start; in vma_merge_new_range() 1058 if (vma->vm_start < start) in vma_shrink() 1428 if (vma->vm_start < vmg->start) { in vma_modify() [all …]
|
| A D | vma.h | 101 return vma->vm_pgoff + PHYS_PFN(addr - vma->vm_start); in vma_pgoff_offset() 157 ((vmi->mas.index > vma->vm_start) || (vmi->mas.last < vma->vm_start))) in vma_iter_store_gfp() 160 __mas_set_range(&vmi->mas, vma->vm_start, vma->vm_end - 1); in vma_iter_store_gfp() 458 vmi->mas.index > vma->vm_start)) { in vma_iter_store() 460 vmi->mas.index, vma->vm_start, vma->vm_start, in vma_iter_store() 464 vmi->mas.last < vma->vm_start)) { in vma_iter_store() 466 vmi->mas.last, vma->vm_start, vma->vm_start, vma->vm_end, in vma_iter_store() 472 ((vmi->mas.index > vma->vm_start) || (vmi->mas.last < vma->vm_start))) in vma_iter_store() 475 __mas_set_range(&vmi->mas, vma->vm_start, vma->vm_end - 1); in vma_iter_store()
|
| A D | msync.c | 73 if (start < vma->vm_start) { in SYSCALL_DEFINE3() 76 start = vma->vm_start; in SYSCALL_DEFINE3() 88 fstart = (start - vma->vm_start) + in SYSCALL_DEFINE3()
|
| A D | mremap.c | 515 if (!for_stack && vma->vm_start != addr_to_align) in can_align_down() 519 if (for_stack && addr_masked >= vma->vm_start) in can_align_down() 687 if (vma->vm_start != old_addr) in move_vma() 713 new_pgoff = vma->vm_pgoff + ((old_addr - vma->vm_start) >> PAGE_SHIFT); in move_vma() 753 if (vma->vm_start < old_addr) in move_vma() 754 account_start = vma->vm_start; in move_vma() 783 if (new_vma != vma && vma->vm_start == old_addr && in move_vma() 856 pgoff = (addr - vma->vm_start) >> PAGE_SHIFT; in vma_to_resize() 950 ((addr - vma->vm_start) >> PAGE_SHIFT), in mremap_to() 974 if (get_unmapped_area(NULL, vma->vm_start, end - vma->vm_start, in vma_expandable() [all …]
|
| A D | mmap.c | 696 high_limit = tmp->vm_start; in unmapped_area_topdown() 1049 VMA_ITERATOR(vmi, mm, vma->vm_start); in expand_upwards() 1100 size = address - vma->vm_start; in expand_upwards() 1147 VMA_ITERATOR(vmi, mm, vma->vm_start); in expand_downwards() 1189 if (address < vma->vm_start) { in expand_downwards() 1213 vma->vm_start = address; in expand_downwards() 1258 if (vma && (vma->vm_start <= addr)) in find_extend_vma_locked() 1283 if (vma->vm_start <= addr) in find_extend_vma_locked() 1285 start = vma->vm_start; in find_extend_vma_locked() 1326 if (vma && vma->vm_start <= addr) in expand_stack() [all …]
|
| A D | mlock.c | 533 if (start > vma->vm_start) in apply_vma_lock_flags() 537 tmp = vma->vm_start; in apply_vma_lock_flags() 542 if (vma->vm_start != tmp) in apply_vma_lock_flags() 587 if (start > vma->vm_start) in count_mm_mlocked_page_nr() 588 count -= (start - vma->vm_start); in count_mm_mlocked_page_nr() 590 count += end - vma->vm_start; in count_mm_mlocked_page_nr() 593 count += vma->vm_end - vma->vm_start; in count_mm_mlocked_page_nr() 734 error = mlock_fixup(&vmi, vma, &prev, vma->vm_start, vma->vm_end, in apply_mlockall_flags()
|
| /linux/tools/testing/vma/ |
| A D | vma.c | 60 ret->vm_start = start; in alloc_vma() 249 ASSERT_EQ(vma->vm_start, 0); in test_simple_merge() 292 ASSERT_EQ(vma->vm_start, 0); in test_simple_modify() 338 ASSERT_EQ(vma->vm_start, 0); in test_simple_expand() 359 ASSERT_EQ(vma->vm_start, 0); in test_simple_shrink() 442 ASSERT_EQ(vma->vm_start, 0); in test_merge_new() 459 ASSERT_EQ(vma->vm_start, 0); in test_merge_new() 496 ASSERT_EQ(vma->vm_start, 0); in test_merge_new() 531 ASSERT_EQ(vma->vm_start, 0); in test_merge_new() 549 ASSERT_EQ(vma->vm_start, 0); in test_merge_new() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | iters_task_vma.c | 13 __u64 vm_start; member 34 vm_ranges[seen].vm_start = vma->vm_start; in iter_task_vma_for_each()
|
| A D | find_vma_fail1.c | 5 #define vm_flags vm_start 17 vma->vm_start = 0xffffffffff600000; in write_vma()
|
| A D | bpf_iter_vma_offset.c | 32 if (vma->vm_start <= address && vma->vm_end > address) { in get_vma_offset() 33 offset = address - vma->vm_start + (vma->vm_pgoff << page_shift); in get_vma_offset()
|
| /linux/mm/damon/tests/ |
| A D | vaddr-kunit.h | 28 mas_set_range(&mas, vmas[i].vm_start, vmas[i].vm_end - 1); in __link_vmas() 72 (struct vm_area_struct) {.vm_start = 10, .vm_end = 20}, in damon_test_three_regions_in_vmas() 73 (struct vm_area_struct) {.vm_start = 20, .vm_end = 25}, in damon_test_three_regions_in_vmas() 74 (struct vm_area_struct) {.vm_start = 200, .vm_end = 210}, in damon_test_three_regions_in_vmas() 75 (struct vm_area_struct) {.vm_start = 210, .vm_end = 220}, in damon_test_three_regions_in_vmas() 76 (struct vm_area_struct) {.vm_start = 300, .vm_end = 305}, in damon_test_three_regions_in_vmas() 77 (struct vm_area_struct) {.vm_start = 307, .vm_end = 330}, in damon_test_three_regions_in_vmas()
|
| /linux/include/trace/events/ |
| A D | mmap.h | 79 __field(unsigned long, vm_start) 86 __entry->vm_start = vma->vm_start; 92 (unsigned long) __entry->vm_start,
|
| /linux/fs/proc/ |
| A D | task_nommu.c | 35 size += region->vm_end - region->vm_start; in task_mem() 37 size = vma->vm_end - vma->vm_start; in task_mem() 89 vsize += vma->vm_end - vma->vm_start; in task_vsize() 109 size += region->vm_end - region->vm_start; in task_statm() 149 vma->vm_start, in nommu_vma_show() 184 *ppos = vma->vm_start; in proc_get_vma()
|
| /linux/arch/powerpc/include/asm/ |
| A D | video.h | 8 unsigned long vm_start, unsigned long vm_end, in pgprot_framebuffer() argument 11 return __phys_mem_access_prot(PHYS_PFN(offset), vm_end - vm_start, prot); in pgprot_framebuffer()
|
| /linux/scripts/coccinelle/api/ |
| A D | vma_pages.cocci | 22 * (vma->vm_end - vma->vm_start) >> PAGE_SHIFT 32 - ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) 44 (vma->vm_end@p - vma->vm_start) >> PAGE_SHIFT
|
| /linux/drivers/accel/habanalabs/common/ |
| A D | memory_mgr.c | 205 new_mmap_size = buf->real_mapped_size - (vma->vm_end - vma->vm_start); in hl_mmap_mem_buf_vm_close() 253 user_mem_size = vma->vm_end - vma->vm_start; 263 if (!access_ok(VERIFY_WRITE, (void __user *)(uintptr_t)vma->vm_start, 266 if (!access_ok((void __user *)(uintptr_t)vma->vm_start, 270 buf->behavior->topic, vma->vm_start);
|
| /linux/arch/parisc/mm/ |
| A D | fault.c | 130 if (tree->vm_start > addr) { 136 if (prev->vm_next->vm_start > addr) 262 vma->vm_start, vma->vm_end); in show_signal_msg() 296 if (!vma || address < vma->vm_start) { in do_page_fault() 373 address < vma->vm_start || address >= vma->vm_end) { in do_page_fault()
|
| /linux/arch/x86/um/ |
| A D | mem_32.c | 17 gate_vma.vm_start = FIXADDR_USER_START; in gate_vma_init() 49 return (addr >= vma->vm_start) && (addr < vma->vm_end); in in_gate_area()
|
| /linux/arch/arc/kernel/ |
| A D | arc_hostlink.c | 22 if (io_remap_pfn_range(vma, vma->vm_start, vma->vm_pgoff, in arc_hl_mmap() 23 vma->vm_end - vma->vm_start, in arc_hl_mmap()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| A D | i915_gem_mman.c | 36 return vma->vm_start == addr && in __vma_matches() 37 (vma->vm_end - vma->vm_start) == PAGE_ALIGN(size); in __vma_matches() 280 area->vm_start, area->vm_end - area->vm_start, in vm_fault_cpu() 303 unsigned long vm_start, vm_end, vma_size; /* user's memory parameters */ in set_address_limits() local 310 vm_start = area->vm_start >> PAGE_SHIFT; in set_address_limits() 319 start = vm_start; in set_address_limits() 324 start = max_t(long, start, vm_start); in set_address_limits() 332 *pfn += (*start_vaddr - area->vm_start) >> PAGE_SHIFT; in set_address_limits() 358 page_offset = (vmf->address - area->vm_start) >> PAGE_SHIFT; in vm_fault_gtt() 518 addr -= area->vm_start; in vm_access()
|
| /linux/drivers/soc/qcom/ |
| A D | rmtfs_mem.c | 136 if (vma->vm_end - vma->vm_start > rmtfs_mem->size) { in qcom_rmtfs_mem_mmap() 139 vma->vm_end, vma->vm_start, in qcom_rmtfs_mem_mmap() 140 (vma->vm_end - vma->vm_start), &rmtfs_mem->size); in qcom_rmtfs_mem_mmap() 146 vma->vm_start, in qcom_rmtfs_mem_mmap() 148 vma->vm_end - vma->vm_start, in qcom_rmtfs_mem_mmap()
|
| /linux/arch/powerpc/platforms/book3s/ |
| A D | vas-api.c | 438 fault = vmf_insert_pfn(vma, vma->vm_start, in vas_mmap_fault() 482 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) { in coproc_mmap() 484 (vma->vm_end - vma->vm_start), PAGE_SIZE); in coproc_mmap() 532 rc = remap_pfn_range(vma, vma->vm_start, pfn + vma->vm_pgoff, in coproc_mmap() 533 vma->vm_end - vma->vm_start, prot); in coproc_mmap() 536 vma->vm_start, rc); in coproc_mmap()
|
| /linux/drivers/xen/xenfs/ |
| A D | xenstored.c | 36 size_t size = vma->vm_end - vma->vm_start; in xsd_kva_mmap() 41 if (remap_pfn_range(vma, vma->vm_start, in xsd_kva_mmap()
|
| /linux/arch/powerpc/kernel/ |
| A D | proc_powerpc.c | 33 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) in page_map_mmap() 36 remap_pfn_range(vma, vma->vm_start, in page_map_mmap()
|