Lines Matching refs:vm_flags
18 vm_flags_t vm_flags; member
48 .vm_flags = vm_flags_, \
59 .vm_flags = (map_)->vm_flags, \
98 if ((vma->vm_flags ^ vmg->vm_flags) & ~VM_SOFTDIRTY) in is_mergeable_vma()
846 if (vmg->vm_flags & VM_SPECIAL || (!left_side && !right_side)) in vma_merge_existing_range()
976 khugepaged_enter_vma(vmg->target, vmg->vm_flags); in vma_merge_existing_range()
1058 if ((vmg->vm_flags & VM_SPECIAL) || (!prev && !next)) in vma_merge_new_range()
1096 khugepaged_enter_vma(vmg->target, vmg->vm_flags); in vma_merge_new_range()
1394 if (next->vm_flags & VM_LOCKED) in vms_gather_munmap_vmas()
1397 if (next->vm_flags & VM_ACCOUNT) in vms_gather_munmap_vmas()
1400 if (is_exec_mapping(next->vm_flags)) in vms_gather_munmap_vmas()
1402 else if (is_stack_mapping(next->vm_flags)) in vms_gather_munmap_vmas()
1404 else if (is_data_mapping(next->vm_flags)) in vms_gather_munmap_vmas()
1643 vm_flags_t vm_flags) in vma_modify_flags() argument
1647 vmg.vm_flags = vm_flags; in vma_modify_flags()
1686 vm_flags_t vm_flags, in vma_modify_flags_uffd() argument
1692 vmg.vm_flags = vm_flags; in vma_modify_flags_uffd()
1920 !((a->vm_flags ^ b->vm_flags) & ~(VM_ACCESS_FLAGS | VM_SOFTDIRTY)) && in anon_vma_compatible()
2008 return (vma->vm_flags & (VM_WRITE | VM_SHARED)) == in vma_is_shared_writable()
2015 if (vma->vm_flags & VM_PFNMAP) in vma_fs_can_writeback()
2062 pgprot_val(vm_pgprot_modify(vm_page_prot, vma->vm_flags))) in vma_wants_writenotify()
2281 static bool accountable_mapping(struct file *file, vm_flags_t vm_flags) in accountable_mapping() argument
2290 return (vm_flags & (VM_NORESERVE | VM_SHARED | VM_WRITE)) == VM_WRITE; in accountable_mapping()
2328 map->vm_flags = ksm_vma_flags(map->mm, map->file, map->vm_flags); in update_ksm_flags()
2373 if (!may_expand_vm(map->mm, map->vm_flags, map->pglen - vms->nr_pages)) in __mmap_prepare()
2377 if (accountable_mapping(map->file, map->vm_flags)) { in __mmap_prepare()
2387 map->vm_flags |= VM_ACCOUNT; in __mmap_prepare()
2431 VM_WARN_ON_ONCE(map->vm_flags != vma->vm_flags && in __mmap_new_file_vma()
2432 !(map->vm_flags & VM_MAYWRITE) && in __mmap_new_file_vma()
2433 (vma->vm_flags & VM_MAYWRITE)); in __mmap_new_file_vma()
2436 map->vm_flags = vma->vm_flags; in __mmap_new_file_vma()
2467 vm_flags_init(vma, map->vm_flags); in __mmap_new_vma()
2477 else if (map->vm_flags & VM_SHARED) in __mmap_new_vma()
2487 vm_flags_init(vma, map->vm_flags); in __mmap_new_vma()
2492 WARN_ON_ONCE(!arch_validate_flags(map->vm_flags)); in __mmap_new_vma()
2506 khugepaged_enter_vma(vma, map->vm_flags); in __mmap_new_vma()
2527 vm_flags_t vm_flags = vma->vm_flags; in __mmap_complete() local
2534 vm_stat_account(mm, vma->vm_flags, map->pglen); in __mmap_complete()
2535 if (vm_flags & VM_LOCKED) { in __mmap_complete()
2536 if ((vm_flags & VM_SPECIAL) || vma_is_dax(vma) || in __mmap_complete()
2580 .vm_flags = map->vm_flags, in call_mmap_prepare()
2592 map->vm_flags = desc.vm_flags; in call_mmap_prepare()
2639 unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, in __mmap_region() argument
2647 MMAP_STATE(map, mm, &vmi, addr, len, pgoff, vm_flags, file); in __mmap_region()
2714 unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, in mmap_region() argument
2723 if (map_deny_write_exec(vm_flags, vm_flags)) in mmap_region()
2727 if (!arch_validate_flags(vm_flags)) in mmap_region()
2731 if (file && is_shared_maywrite(vm_flags)) { in mmap_region()
2739 ret = __mmap_region(file, addr, len, vm_flags, pgoff, uf); in mmap_region()
2762 unsigned long addr, unsigned long len, vm_flags_t vm_flags) in do_brk_flags() argument
2770 vm_flags |= VM_DATA_DEFAULT_FLAGS | VM_ACCOUNT | mm->def_flags; in do_brk_flags()
2771 vm_flags = ksm_vma_flags(mm, NULL, vm_flags); in do_brk_flags()
2772 if (!may_expand_vm(mm, vm_flags, len >> PAGE_SHIFT)) in do_brk_flags()
2786 VMG_STATE(vmg, mm, vmi, addr, addr + len, vm_flags, PHYS_PFN(addr)); in do_brk_flags()
2807 vm_flags_init(vma, vm_flags); in do_brk_flags()
2808 vma->vm_page_prot = vm_get_page_prot(vm_flags); in do_brk_flags()
2819 if (vm_flags & VM_LOCKED) in do_brk_flags()
2870 if (tmp && (tmp->vm_flags & VM_STARTGAP_FLAGS)) { /* Avoid prev check if possible */ in unmapped_area()
2922 if (tmp && (tmp->vm_flags & VM_STARTGAP_FLAGS)) { /* Avoid prev check if possible */ in unmapped_area_topdown()
2952 if (!may_expand_vm(mm, vma->vm_flags, grow)) in acct_stack_growth()
2960 if (!mlock_future_ok(mm, vma->vm_flags, grow << PAGE_SHIFT)) in acct_stack_growth()
2964 new_start = (vma->vm_flags & VM_GROWSUP) ? vma->vm_start : in acct_stack_growth()
2992 if (!(vma->vm_flags & VM_GROWSUP)) in expand_upwards()
3012 if (!(next->vm_flags & VM_GROWSUP)) in expand_upwards()
3046 if (vma->vm_flags & VM_LOCKED) in expand_upwards()
3048 vm_stat_account(mm, vma->vm_flags, grow); in expand_upwards()
3077 if (!(vma->vm_flags & VM_GROWSDOWN)) in expand_downwards()
3090 if (!(prev->vm_flags & VM_GROWSDOWN) && in expand_downwards()
3125 if (vma->vm_flags & VM_LOCKED) in expand_downwards()
3127 vm_stat_account(mm, vma->vm_flags, grow); in expand_downwards()
3175 if ((vma->vm_flags & VM_ACCOUNT) && in insert_vm_struct()
3197 if (vma->vm_flags & VM_ACCOUNT) in insert_vm_struct()