| /arch/arm64/include/asm/ |
| A D | mman.h | 71 static inline bool arch_validate_flags(vm_flags_t vm_flags) in arch_validate_flags() argument 78 if ((vm_flags & VM_MTE) && !(vm_flags & VM_MTE_ALLOWED)) in arch_validate_flags() 82 if (system_supports_gcs() && (vm_flags & VM_SHADOW_STACK)) { in arch_validate_flags() 84 if (vm_flags & VM_EXEC) in arch_validate_flags() 88 VM_WARN_ON(vm_flags & VM_SHARED); in arch_validate_flags() 94 #define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags) argument
|
| /arch/arm64/mm/ |
| A D | mmap.c | 84 pgprot_t vm_get_page_prot(vm_flags_t vm_flags) in vm_get_page_prot() argument 89 if (system_supports_gcs() && (vm_flags & VM_SHADOW_STACK)) { in vm_get_page_prot() 92 prot = pgprot_val(protection_map[vm_flags & in vm_get_page_prot() 96 if (vm_flags & VM_ARM64_BTI) in vm_get_page_prot() 109 if (vm_flags & VM_MTE) in vm_get_page_prot() 114 if (vm_flags & VM_PKEY_BIT0) in vm_get_page_prot() 116 if (vm_flags & VM_PKEY_BIT1) in vm_get_page_prot() 118 if (vm_flags & VM_PKEY_BIT2) in vm_get_page_prot()
|
| A D | fault.c | 542 if (!(vma->vm_flags & VM_SHADOW_STACK)) in is_invalid_gcs_access() 558 vm_flags_t vm_flags; in do_page_fault() local 586 vm_flags = VM_EXEC; in do_page_fault() 594 vm_flags = VM_WRITE; in do_page_fault() 598 vm_flags = VM_WRITE; in do_page_fault() 602 vm_flags = VM_READ; in do_page_fault() 604 vm_flags |= VM_WRITE; in do_page_fault() 607 vm_flags |= VM_EXEC; in do_page_fault() 636 if (!(vma->vm_flags & vm_flags)) { in do_page_fault() 681 if (!(vma->vm_flags & vm_flags)) { in do_page_fault() [all …]
|
| /arch/powerpc/include/asm/book3s/64/ |
| A D | hash-pkey.h | 8 static inline u64 hash__vmflag_to_pte_pkey_bits(u64 vm_flags) in hash__vmflag_to_pte_pkey_bits() argument 10 return (((vm_flags & VM_PKEY_BIT0) ? H_PTE_PKEY_BIT0 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 11 ((vm_flags & VM_PKEY_BIT1) ? H_PTE_PKEY_BIT1 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 12 ((vm_flags & VM_PKEY_BIT2) ? H_PTE_PKEY_BIT2 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 13 ((vm_flags & VM_PKEY_BIT3) ? H_PTE_PKEY_BIT3 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 14 ((vm_flags & VM_PKEY_BIT4) ? H_PTE_PKEY_BIT4 : 0x0UL)); in hash__vmflag_to_pte_pkey_bits()
|
| A D | pkeys.h | 8 static inline u64 vmflag_to_pte_pkey_bits(vm_flags_t vm_flags) in vmflag_to_pte_pkey_bits() argument 15 return hash__vmflag_to_pte_pkey_bits(vm_flags); in vmflag_to_pte_pkey_bits()
|
| /arch/sparc/include/asm/ |
| A D | mman.h | 57 #define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags) argument 61 static inline bool arch_validate_flags(vm_flags_t vm_flags) in arch_validate_flags() argument 67 if (vm_flags & VM_SPARC_ADI) { in arch_validate_flags() 72 if (vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) in arch_validate_flags() 83 if (vm_flags & VM_MERGEABLE) in arch_validate_flags()
|
| /arch/x86/mm/ |
| A D | pgprot.c | 35 pgprot_t vm_get_page_prot(vm_flags_t vm_flags) in vm_get_page_prot() argument 37 unsigned long val = pgprot_val(protection_map[vm_flags & in vm_get_page_prot() 48 if (vm_flags & VM_PKEY_BIT0) in vm_get_page_prot() 50 if (vm_flags & VM_PKEY_BIT1) in vm_get_page_prot() 52 if (vm_flags & VM_PKEY_BIT2) in vm_get_page_prot() 54 if (vm_flags & VM_PKEY_BIT3) in vm_get_page_prot()
|
| /arch/x86/kernel/ |
| A D | sys_x86_64.c | 118 static inline unsigned long stack_guard_placement(vm_flags_t vm_flags) in stack_guard_placement() argument 120 if (vm_flags & VM_SHADOW_STACK) in stack_guard_placement() 128 unsigned long pgoff, unsigned long flags, vm_flags_t vm_flags) in arch_get_unmapped_area() argument 156 info.start_gap = stack_guard_placement(vm_flags); in arch_get_unmapped_area() 169 unsigned long flags, vm_flags_t vm_flags) in arch_get_unmapped_area_topdown() argument 209 info.start_gap = stack_guard_placement(vm_flags); in arch_get_unmapped_area_topdown()
|
| /arch/nios2/mm/ |
| A D | cacheflush.c | 90 if (!(vma->vm_flags & VM_MAYSHARE)) in flush_aliases() 138 if (vma == NULL || (vma->vm_flags & VM_EXEC)) in flush_cache_range() 159 if (vma->vm_flags & VM_EXEC) in flush_cache_page() 236 if (vma->vm_flags & VM_EXEC) in update_mmu_cache_range() 268 if (vma->vm_flags & VM_EXEC) in copy_from_user_page() 279 if (vma->vm_flags & VM_EXEC) in copy_to_user_page()
|
| A D | fault.c | 105 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 109 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 113 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
| /arch/hexagon/mm/ |
| A D | vm_fault.c | 70 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 74 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 78 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
| /arch/csky/abiv1/ |
| A D | cacheflush.c | 63 if (vma->vm_flags & VM_EXEC) in update_mmu_cache_range() 73 if (vma->vm_flags & VM_EXEC) in flush_cache_range()
|
| /arch/arm/mm/ |
| A D | fault.c | 271 vm_flags_t vm_flags = VM_ACCESS_FLAGS; in do_page_fault() local 293 vm_flags = VM_WRITE; in do_page_fault() 297 vm_flags = VM_EXEC; in do_page_fault() 321 if (!(vma->vm_flags & vm_flags)) { in do_page_fault() 360 if (!(vma->vm_flags & vm_flags)) { in do_page_fault()
|
| A D | flush.c | 94 if (vma->vm_flags & VM_EXEC) in flush_cache_range() 110 if (vma->vm_flags & VM_EXEC && icache_is_vivt_asid_tagged()) in flush_cache_pages() 165 if (vma->vm_flags & VM_EXEC) in flush_ptrace_access() 263 if (!(vma->vm_flags & VM_MAYSHARE)) in __flush_dcache_aliases()
|
| /arch/alpha/mm/ |
| A D | fault.c | 130 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 134 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) in do_page_fault() 137 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
| /arch/microblaze/mm/ |
| A D | fault.c | 157 if (unlikely(!(vma->vm_flags & VM_GROWSDOWN))) in do_page_fault() 204 if (unlikely(!(vma->vm_flags & VM_WRITE))) in do_page_fault() 212 if (unlikely(!(vma->vm_flags & (VM_READ | VM_EXEC)))) in do_page_fault()
|
| /arch/sparc/mm/ |
| A D | fault_64.c | 352 (vma->vm_flags & VM_WRITE) != 0) { in do_sparc64_fault() 373 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_sparc64_fault() 402 if ((fault_code & FAULT_CODE_ITLB) && !(vma->vm_flags & VM_EXEC)) { in do_sparc64_fault() 410 if (!(vma->vm_flags & VM_WRITE)) in do_sparc64_fault() 417 (vma->vm_flags & VM_EXEC) != 0 && in do_sparc64_fault() 425 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in do_sparc64_fault()
|
| A D | fault_32.c | 161 if (!(vma->vm_flags & VM_WRITE)) in do_sparc_fault() 165 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in do_sparc_fault() 320 if (!(vma->vm_flags & VM_WRITE)) in force_user_fault() 324 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in force_user_fault()
|
| /arch/openrisc/mm/ |
| A D | fault.c | 119 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 147 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 152 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in do_page_fault()
|
| /arch/xtensa/mm/ |
| A D | fault.c | 145 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 149 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 152 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) in do_page_fault()
|
| /arch/mips/mm/ |
| A D | mmap.c | 102 vm_flags_t vm_flags) in arch_get_unmapped_area() argument 114 unsigned long flags, vm_flags_t vm_flags) in arch_get_unmapped_area_topdown() argument
|
| /arch/um/kernel/ |
| A D | trap.c | 97 if (!vma || !(vma->vm_flags & VM_GROWSDOWN)) { in um_lock_mm_and_find_vma() 120 if (!(vma->vm_flags & VM_GROWSDOWN)) in um_lock_mm_and_find_vma() 168 if (!(vma->vm_flags & VM_WRITE)) in handle_page_fault() 173 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in handle_page_fault()
|
| /arch/parisc/mm/ |
| A D | fault.c | 297 if (!prev_vma || !(prev_vma->vm_flags & VM_GROWSUP)) in do_page_fault() 309 if ((vma->vm_flags & acc_type) != acc_type) in do_page_fault() 384 if ((vma->vm_flags & acc_type) != acc_type) { in do_page_fault() 506 && (vma->vm_flags & acc_type) == acc_type) in handle_nadtlb_fault()
|
| /arch/powerpc/mm/book3s64/ |
| A D | pgtable.c | 645 pgprot_t vm_get_page_prot(vm_flags_t vm_flags) in vm_get_page_prot() argument 650 if (!radix_enabled() && ((vm_flags & VM_ACCESS_FLAGS) == VM_EXEC)) in vm_get_page_prot() 651 vm_flags |= VM_READ; in vm_get_page_prot() 653 prot = pgprot_val(protection_map[vm_flags & (VM_ACCESS_FLAGS | VM_SHARED)]); in vm_get_page_prot() 655 if (vm_flags & VM_SAO) in vm_get_page_prot() 659 prot |= vmflag_to_pte_pkey_bits(vm_flags); in vm_get_page_prot()
|
| /arch/powerpc/mm/ |
| A D | copro_fault.c | 41 if (!(vma->vm_flags & VM_WRITE)) in copro_handle_mm_fault() 44 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in copro_handle_mm_fault()
|