Lines Matching refs:prot
95 phys_addr_t phys_addr, pgprot_t prot, in vmap_pte_range() argument
122 pte_t entry = pfn_pte(pfn, prot); in vmap_pte_range()
130 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); in vmap_pte_range()
140 phys_addr_t phys_addr, pgprot_t prot, in vmap_try_huge_pmd() argument
146 if (!arch_vmap_pmd_supported(prot)) in vmap_try_huge_pmd()
161 return pmd_set_huge(pmd, phys_addr, prot); in vmap_try_huge_pmd()
165 phys_addr_t phys_addr, pgprot_t prot, in vmap_pmd_range() argument
177 if (vmap_try_huge_pmd(pmd, addr, next, phys_addr, prot, in vmap_pmd_range()
183 if (vmap_pte_range(pmd, addr, next, phys_addr, prot, max_page_shift, mask)) in vmap_pmd_range()
190 phys_addr_t phys_addr, pgprot_t prot, in vmap_try_huge_pud() argument
196 if (!arch_vmap_pud_supported(prot)) in vmap_try_huge_pud()
211 return pud_set_huge(pud, phys_addr, prot); in vmap_try_huge_pud()
215 phys_addr_t phys_addr, pgprot_t prot, in vmap_pud_range() argument
227 if (vmap_try_huge_pud(pud, addr, next, phys_addr, prot, in vmap_pud_range()
233 if (vmap_pmd_range(pud, addr, next, phys_addr, prot, in vmap_pud_range()
241 phys_addr_t phys_addr, pgprot_t prot, in vmap_try_huge_p4d() argument
247 if (!arch_vmap_p4d_supported(prot)) in vmap_try_huge_p4d()
262 return p4d_set_huge(p4d, phys_addr, prot); in vmap_try_huge_p4d()
266 phys_addr_t phys_addr, pgprot_t prot, in vmap_p4d_range() argument
278 if (vmap_try_huge_p4d(p4d, addr, next, phys_addr, prot, in vmap_p4d_range()
284 if (vmap_pud_range(p4d, addr, next, phys_addr, prot, in vmap_p4d_range()
292 phys_addr_t phys_addr, pgprot_t prot, in vmap_range_noflush() argument
308 err = vmap_p4d_range(pgd, addr, next, phys_addr, prot, in vmap_range_noflush()
321 phys_addr_t phys_addr, pgprot_t prot) in vmap_page_range() argument
325 err = vmap_range_noflush(addr, end, phys_addr, pgprot_nx(prot), in vmap_page_range()
329 err = kmsan_ioremap_page_range(addr, end, phys_addr, prot, in vmap_page_range()
335 phys_addr_t phys_addr, pgprot_t prot) in ioremap_page_range() argument
351 return vmap_page_range(addr, end, phys_addr, prot); in ioremap_page_range()
514 unsigned long end, pgprot_t prot, struct page **pages, int *nr, in vmap_pages_pte_range() argument
547 set_pte_at(&init_mm, addr, pte, mk_pte(page, prot)); in vmap_pages_pte_range()
558 unsigned long end, pgprot_t prot, struct page **pages, int *nr, in vmap_pages_pmd_range() argument
569 if (vmap_pages_pte_range(pmd, addr, next, prot, pages, nr, mask)) in vmap_pages_pmd_range()
576 unsigned long end, pgprot_t prot, struct page **pages, int *nr, in vmap_pages_pud_range() argument
587 if (vmap_pages_pmd_range(pud, addr, next, prot, pages, nr, mask)) in vmap_pages_pud_range()
594 unsigned long end, pgprot_t prot, struct page **pages, int *nr, in vmap_pages_p4d_range() argument
605 if (vmap_pages_pud_range(p4d, addr, next, prot, pages, nr, mask)) in vmap_pages_p4d_range()
612 pgprot_t prot, struct page **pages) in vmap_small_pages_range_noflush() argument
627 err = vmap_pages_p4d_range(pgd, addr, next, prot, pages, &nr, &mask); in vmap_small_pages_range_noflush()
648 pgprot_t prot, struct page **pages, unsigned int page_shift) in __vmap_pages_range_noflush() argument
656 return vmap_small_pages_range_noflush(addr, end, prot, pages); in __vmap_pages_range_noflush()
662 page_to_phys(pages[i]), prot, in __vmap_pages_range_noflush()
674 pgprot_t prot, struct page **pages, unsigned int page_shift) in vmap_pages_range_noflush() argument
676 int ret = kmsan_vmap_pages_range_noflush(addr, end, prot, pages, in vmap_pages_range_noflush()
681 return __vmap_pages_range_noflush(addr, end, prot, pages, page_shift); in vmap_pages_range_noflush()
697 pgprot_t prot, struct page **pages, unsigned int page_shift) in vmap_pages_range() argument
701 err = vmap_pages_range_noflush(addr, end, prot, pages, page_shift); in vmap_pages_range()
3488 unsigned long flags, pgprot_t prot) in vmap() argument
3515 if (vmap_pages_range(addr, addr + size, pgprot_nx(prot), in vmap()
3532 pgprot_t prot; member
3545 ptent = pte_mkspecial(pfn_pte(pfn, data->prot)); in vmap_pfn_apply()
3561 void *vmap_pfn(unsigned long *pfns, unsigned int count, pgprot_t prot) in vmap_pfn() argument
3563 struct vmap_pfn_data data = { .pfns = pfns, .prot = pgprot_nx(prot) }; in vmap_pfn()
3675 pgprot_t prot, unsigned int page_shift, in __vmalloc_area_node() argument
3762 ret = vmap_pages_range(addr, addr + size, prot, area->pages, in __vmalloc_area_node()
3817 pgprot_t prot, unsigned long vm_flags, int node, in __vmalloc_node_range_noprof() argument
3844 if (arch_vmap_pmd_supported(prot) && size >= PMD_SIZE) in __vmalloc_node_range_noprof()
3872 if (pgprot_val(prot) == pgprot_val(PAGE_KERNEL)) { in __vmalloc_node_range_noprof()
3878 prot = arch_vmap_pgprot_tagged(prot); in __vmalloc_node_range_noprof()
3893 ret = __vmalloc_area_node(area, gfp_mask, prot, shift, node); in __vmalloc_node_range_noprof()