| /mm/ |
| A D | early_ioremap.c | 37 pgprot_t prot) in early_memremap_pgprot_adjust() argument 39 return prot; in early_memremap_pgprot_adjust() 54 phys_addr_t phys, pgprot_t prot) in __late_set_fixmap() argument 100 __early_ioremap(resource_size_t phys_addr, unsigned long size, pgprot_t prot) in __early_ioremap() argument 148 __late_set_fixmap(idx, phys_addr, prot); in __early_ioremap() 150 __early_set_fixmap(idx, phys_addr, prot); in __early_ioremap() 220 pgprot_t prot = early_memremap_pgprot_adjust(phys_addr, size, in early_memremap() local 223 return (__force void *)__early_ioremap(phys_addr, size, prot); in early_memremap() 229 pgprot_t prot = early_memremap_pgprot_adjust(phys_addr, size, in early_memremap_ro() local 232 return (__force void *)__early_ioremap(phys_addr, size, prot); in early_memremap_ro()
|
| A D | nommu.c | 692 unsigned long prot, in validate_mmap_request() argument 786 if (prot & PROT_WRITE) in validate_mmap_request() 806 if (prot & PROT_EXEC) in validate_mmap_request() 808 } else if ((prot & PROT_READ) && !(prot & PROT_EXEC)) { in validate_mmap_request() 812 prot |= PROT_EXEC; in validate_mmap_request() 815 (prot & PROT_EXEC) && in validate_mmap_request() 828 if ((prot & PROT_READ) && in validate_mmap_request() 830 prot |= PROT_EXEC; in validate_mmap_request() 848 unsigned long prot, in determine_vm_flags() argument 1014 unsigned long prot, in do_mmap() argument [all …]
|
| A D | vmalloc.c | 95 phys_addr_t phys_addr, pgprot_t prot, in vmap_pte_range() argument 122 pte_t entry = pfn_pte(pfn, prot); in vmap_pte_range() 140 phys_addr_t phys_addr, pgprot_t prot, in vmap_try_huge_pmd() argument 146 if (!arch_vmap_pmd_supported(prot)) in vmap_try_huge_pmd() 196 if (!arch_vmap_pud_supported(prot)) in vmap_try_huge_pud() 247 if (!arch_vmap_p4d_supported(prot)) in vmap_try_huge_p4d() 612 pgprot_t prot, struct page **pages) in vmap_small_pages_range_noflush() argument 662 page_to_phys(pages[i]), prot, in __vmap_pages_range_noflush() 3532 pgprot_t prot; member 3563 struct vmap_pfn_data data = { .pfns = pfns, .prot = pgprot_nx(prot) }; in vmap_pfn() [all …]
|
| A D | mprotect.c | 862 unsigned long prot, int pkey) in do_mprotect_pkey() argument 869 (prot & PROT_READ); in do_mprotect_pkey() 875 prot &= ~(PROT_GROWSDOWN|PROT_GROWSUP); in do_mprotect_pkey() 887 if (!arch_validate_prot(prot, start)) in do_mprotect_pkey() 890 reqprot = prot; in do_mprotect_pkey() 946 prot |= PROT_EXEC; in do_mprotect_pkey() 996 prot = reqprot; in do_mprotect_pkey() 1009 unsigned long, prot) in SYSCALL_DEFINE3() 1011 return do_mprotect_pkey(start, len, prot, -1); in SYSCALL_DEFINE3() 1017 unsigned long, prot, int, pkey) in SYSCALL_DEFINE4() [all …]
|
| A D | ioremap.c | 16 pgprot_t prot) in generic_ioremap_prot() argument 43 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) { in generic_ioremap_prot() 53 pgprot_t prot) in ioremap_prot() argument 55 return generic_ioremap_prot(phys_addr, size, prot); in ioremap_prot()
|
| A D | mmap.c | 335 unsigned long len, unsigned long prot, in do_mmap() argument 358 prot |= PROT_EXEC; in do_mmap() 390 if (prot == PROT_EXEC) { in do_mmap() 448 if (prot & PROT_WRITE) { in do_mmap() 567 unsigned long prot, unsigned long flags, in ksys_mmap_pgoff() argument 612 unsigned long, prot, unsigned long, flags, in SYSCALL_DEFINE6() argument 622 unsigned long prot; member 1101 if (prot) in SYSCALL_DEFINE5() 1128 prot |= vma->vm_flags & VM_READ ? PROT_READ : 0; in SYSCALL_DEFINE5() 1144 ret = security_mmap_file(file, prot, flags); in SYSCALL_DEFINE5() [all …]
|
| A D | util.c | 566 unsigned long len, unsigned long prot, in vm_mmap_pgoff() argument 574 ret = security_mmap_file(file, prot, flag); in vm_mmap_pgoff() 576 ret = fsnotify_mmap_perm(file, prot, pgoff >> PAGE_SHIFT, len); in vm_mmap_pgoff() 580 ret = do_mmap(file, addr, len, prot, flag, 0, pgoff, &populate, in vm_mmap_pgoff() 608 unsigned long len, unsigned long prot, in vm_mmap() argument 616 return vm_mmap_pgoff(file, addr, len, prot, flag, offset >> PAGE_SHIFT); in vm_mmap()
|
| A D | memory.c | 2150 pgprot_t prot, bool mkwrite) in insert_page_into_pte_locked() argument 2172 pteval = mk_pte(page, prot); in insert_page_into_pte_locked() 2177 pteval = mk_pte(page, prot); in insert_page_into_pte_locked() 2684 unsigned long pfn, pgprot_t prot) in remap_pte_range() argument 2710 unsigned long pfn, pgprot_t prot) in remap_pmd_range() argument 2733 unsigned long pfn, pgprot_t prot) in remap_pud_range() argument 2755 unsigned long pfn, pgprot_t prot) in remap_p4d_range() argument 2851 unsigned long size, pgprot_t *prot) in pfnmap_track_ctx_alloc() argument 2855 if (pfnmap_track(pfn, size, prot)) in pfnmap_track_ctx_alloc() 6636 pgprot_t prot = __pgprot(0); in generic_access_phys() local [all …]
|
| A D | highmem.c | 549 void *__kmap_local_pfn_prot(unsigned long pfn, pgprot_t prot) in __kmap_local_pfn_prot() argument 565 pteval = pfn_pte(pfn, prot); in __kmap_local_pfn_prot() 575 void *__kmap_local_page_prot(struct page *page, pgprot_t prot) in __kmap_local_page_prot() argument 592 return __kmap_local_pfn_prot(page_to_pfn(page), prot); in __kmap_local_page_prot()
|
| A D | debug_vm_pgtable.c | 81 pgprot_t prot = vm_get_page_prot(idx); in pte_basic_tests() local 82 pte_t pte = pfn_pte(args->fixed_pte_pfn, prot); in pte_basic_tests() 169 pgprot_t prot = vm_get_page_prot(idx); in pmd_basic_tests() local 177 pmd = pfn_pmd(args->fixed_pmd_pfn, prot); in pmd_basic_tests() 288 pgprot_t prot = vm_get_page_prot(idx); in pud_basic_tests() local 296 pud = pfn_pud(args->fixed_pud_pfn, prot); in pud_basic_tests()
|
| A D | internal.h | 1362 pgprot_t prot, struct page **pages, unsigned int page_shift); 1371 pgprot_t prot, struct page **pages, unsigned int page_shift) in vmap_pages_range_noflush() argument 1378 unsigned long end, pgprot_t prot,
|
| A D | huge_memory.c | 1383 pmd_t *pmd, struct folio_or_pfn fop, pgprot_t prot, in insert_pmd() argument 1416 entry = pmd_mkhuge(pfn_pmd(fop.pfn, prot)); in insert_pmd() 1535 pud_t *pud, struct folio_or_pfn fop, pgprot_t prot, bool write) in insert_pud() argument 1562 entry = pud_mkhuge(pfn_pud(fop.pfn, prot)); in insert_pud()
|
| /mm/kmsan/ |
| A D | shadow.c | 217 pgprot_t prot, struct page **pages, in kmsan_vmap_pages_range_noflush() argument 243 prot = PAGE_KERNEL; in kmsan_vmap_pages_range_noflush() 248 mapped = __vmap_pages_range_noflush(shadow_start, shadow_end, prot, in kmsan_vmap_pages_range_noflush() 256 mapped = __vmap_pages_range_noflush(origin_start, origin_end, prot, in kmsan_vmap_pages_range_noflush()
|
| A D | hooks.c | 148 phys_addr_t phys_addr, pgprot_t prot, in kmsan_ioremap_page_range() argument 170 vmalloc_shadow(start + off + PAGE_SIZE), prot, &shadow, in kmsan_ioremap_page_range() 179 vmalloc_origin(start + off + PAGE_SIZE), prot, &origin, in kmsan_ioremap_page_range()
|