Searched refs:alloc_page (Results 1 – 25 of 223) sorted by relevance
123456789
52 lock_op_page = alloc_page(GFP_NOIO); in ceph_cls_lock()111 unlock_op_page = alloc_page(GFP_NOIO); in ceph_cls_unlock()164 break_op_page = alloc_page(GFP_NOIO); in ceph_cls_break_lock()213 cookie_op_page = alloc_page(GFP_NOIO); in ceph_cls_set_cookie()352 get_info_op_page = alloc_page(GFP_NOIO); in ceph_cls_lock_info()356 reply_page = alloc_page(GFP_NOIO); in ceph_cls_lock_info()
80 page = alloc_page(gfp); in rds_page_remainder_alloc()122 page = alloc_page(gfp); in rds_page_remainder_alloc()
158 page = alloc_page(GFP_KERNEL); in gso_test_func()169 page = alloc_page(GFP_KERNEL); in gso_test_func()192 page = alloc_page(GFP_KERNEL); in gso_test_func()
118 pte_page = alloc_page(GFP_KERNEL); in __split_linear_mapping_pmd()158 pmd_page = alloc_page(GFP_KERNEL); in __split_linear_mapping_pud()207 pud_page = alloc_page(GFP_KERNEL); in __split_linear_mapping_p4d()
99 glob->dummy_read_page = alloc_page(__GFP_ZERO | GFP_DMA32 | in ttm_global_init()104 glob->dummy_read_page = alloc_page(__GFP_ZERO); in ttm_global_init()
236 src_page = alloc_page(GFP_KERNEL);239 dst_page = alloc_page(GFP_KERNEL);
33 um_vdso = alloc_page(GFP_KERNEL); in init_vdso()
49 selinux_state.status_page = alloc_page(GFP_KERNEL|__GFP_ZERO); in selinux_kernel_status_page()
25 vdso_page = alloc_page(GFP_KERNEL); in vdso_init()
32 p = alloc_page(GFP_KERNEL | __GFP_ZERO); in kvm_arch_ptp_init()
214 io->header_page = alloc_page(gfp_mask); in ppl_io_pool_alloc()802 page1 = alloc_page(GFP_KERNEL); in ppl_recover_entry()803 page2 = alloc_page(GFP_KERNEL); in ppl_recover_entry()971 page = alloc_page(GFP_KERNEL); in ppl_recover()1045 page = alloc_page(GFP_NOIO | __GFP_ZERO); in ppl_write_empty_header()1082 page = alloc_page(GFP_KERNEL); in ppl_load_distributed()1086 page2 = alloc_page(GFP_KERNEL); in ppl_load_distributed()
158 p = alloc_page(GFP_KERNEL); in do_mount_root()186 struct page *page = alloc_page(GFP_KERNEL); in mount_root_generic()
168 pd->p = alloc_page(GFP_DMA32); in psb_mmu_alloc_pd()171 pd->dummy_pt = alloc_page(GFP_DMA32); in psb_mmu_alloc_pd()174 pd->dummy_page = alloc_page(GFP_DMA32); in psb_mmu_alloc_pd()275 pt->p = alloc_page(GFP_DMA32); in psb_mmu_alloc_pt()
64 pub fn alloc_page(flags: Flags) -> Result<Self, AllocError> { in alloc_page() method
54 page = alloc_page(GFP | __GFP_HIGHMEM); in huge_get_pages()
113 mock->pages[i] = alloc_page(GFP_KERNEL); in mock_dmabuf()
276 cp = alloc_page(GFP_KERNEL); in init_vdso_image()297 dp = alloc_page(GFP_KERNEL); in init_vdso_image()
373 if (trans_pcie->alloc_page) { in iwl_pcie_rx_alloc_page()376 if (trans_pcie->alloc_page) { in iwl_pcie_rx_alloc_page()378 page = trans_pcie->alloc_page; in iwl_pcie_rx_alloc_page()381 trans_pcie->alloc_page = NULL; in iwl_pcie_rx_alloc_page()408 if (!trans_pcie->alloc_page) { in iwl_pcie_rx_alloc_page()410 trans_pcie->alloc_page = page; in iwl_pcie_rx_alloc_page()1247 if (trans_pcie->alloc_page) in iwl_pcie_rx_free()1248 __free_pages(trans_pcie->alloc_page, trans_pcie->rx_page_order); in iwl_pcie_rx_free()
215 struct page *page = alloc_page(GFP_KERNEL); in gve_rx_prefill_pages()541 bool alloc_page = false; in gve_rx_copy_to_pool() local553 alloc_page = !recycle; in gve_rx_copy_to_pool()556 if (alloc_page) { in gve_rx_copy_to_pool()566 page = alloc_page(GFP_ATOMIC); in gve_rx_copy_to_pool()
126 struct page *page = alloc_page(balloon_mapping_gfp_mask() | in balloon_page_alloc()
99 page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO); in kvm_coalesced_mmio_init()
149 vma_priv->pages[i] = alloc_page(GFP_KERNEL | __GFP_ZERO); in privcmd_buf_mmap()
238 sg_pages[i] = alloc_page(GFP_KERNEL); in efi_capsule_update()
42 page = alloc_page(GFP_KERNEL); in hv_call_deposit_pages()
161 page = alloc_page(GFP_USER); in isofs_lookup()
Completed in 54 milliseconds