Lines Matching refs:pages
514 unsigned long end, pgprot_t prot, struct page **pages, int *nr, in vmap_pages_pte_range() argument
532 struct page *page = pages[*nr]; in vmap_pages_pte_range()
558 unsigned long end, pgprot_t prot, struct page **pages, int *nr, in vmap_pages_pmd_range() argument
569 if (vmap_pages_pte_range(pmd, addr, next, prot, pages, nr, mask)) in vmap_pages_pmd_range()
576 unsigned long end, pgprot_t prot, struct page **pages, int *nr, in vmap_pages_pud_range() argument
587 if (vmap_pages_pmd_range(pud, addr, next, prot, pages, nr, mask)) in vmap_pages_pud_range()
594 unsigned long end, pgprot_t prot, struct page **pages, int *nr, in vmap_pages_p4d_range() argument
605 if (vmap_pages_pud_range(p4d, addr, next, prot, pages, nr, mask)) in vmap_pages_p4d_range()
612 pgprot_t prot, struct page **pages) in vmap_small_pages_range_noflush() argument
627 err = vmap_pages_p4d_range(pgd, addr, next, prot, pages, &nr, &mask); in vmap_small_pages_range_noflush()
648 pgprot_t prot, struct page **pages, unsigned int page_shift) in __vmap_pages_range_noflush() argument
656 return vmap_small_pages_range_noflush(addr, end, prot, pages); in __vmap_pages_range_noflush()
662 page_to_phys(pages[i]), prot, in __vmap_pages_range_noflush()
674 pgprot_t prot, struct page **pages, unsigned int page_shift) in vmap_pages_range_noflush() argument
676 int ret = kmsan_vmap_pages_range_noflush(addr, end, prot, pages, in vmap_pages_range_noflush()
681 return __vmap_pages_range_noflush(addr, end, prot, pages, page_shift); in vmap_pages_range_noflush()
697 pgprot_t prot, struct page **pages, unsigned int page_shift) in vmap_pages_range() argument
701 err = vmap_pages_range_noflush(addr, end, prot, pages, page_shift); in vmap_pages_range()
732 unsigned long end, struct page **pages) in vm_area_map_pages() argument
740 return vmap_pages_range(start, end, PAGE_KERNEL, pages, PAGE_SHIFT); in vm_area_map_pages()
3023 void *vm_map_ram(struct page **pages, unsigned int count, int node) in vm_map_ram() argument
3048 pages, PAGE_SHIFT) < 0) { in vm_map_ram()
3306 if (page_address(area->pages[i])) in set_area_direct_map()
3307 set_direct_map(area->pages[i]); in set_area_direct_map()
3325 unsigned long addr = (unsigned long)page_address(area->pages[i]); in vm_reset_perms()
3425 mod_memcg_page_state(vm->pages[0], MEMCG_VMALLOC, -vm->nr_pages); in vfree()
3427 struct page *page = vm->pages[i]; in vfree()
3439 kvfree(vm->pages); in vfree()
3487 void *vmap(struct page **pages, unsigned int count, in vmap() argument
3516 pages, PAGE_SHIFT) < 0) { in vmap()
3522 area->pages = pages; in vmap()
3586 unsigned int order, unsigned int nr_pages, struct page **pages) in vm_area_alloc_pages() argument
3618 pages + nr_allocated); in vm_area_alloc_pages()
3622 pages + nr_allocated); in vm_area_alloc_pages()
3665 pages[nr_allocated + i] = page + i; in vm_area_alloc_pages()
3695 area->pages = __vmalloc_node_noprof(array_size, 1, nested_gfp, node, in __vmalloc_area_node()
3698 area->pages = kmalloc_node_noprof(array_size, nested_gfp, node); in __vmalloc_area_node()
3701 if (!area->pages) { in __vmalloc_area_node()
3722 node, page_order, nr_small_pages, area->pages); in __vmalloc_area_node()
3727 mod_memcg_page_state(area->pages[0], MEMCG_VMALLOC, in __vmalloc_area_node()
3762 ret = vmap_pages_range(addr, addr + size, prot, area->pages, in __vmalloc_area_node()
5005 counters[page_to_nid(v->pages[nr])] += step; in show_numa_info()
5086 if (is_vmalloc_addr(v->pages)) in vmalloc_info_show()