Searched refs:page (Results 1 – 6 of 6) sorted by relevance
97 struct page *page; in kvm_coalesced_mmio_init() local99 page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO); in kvm_coalesced_mmio_init()100 if (!page) in kvm_coalesced_mmio_init()103 kvm->coalesced_mmio_ring = page_address(page); in kvm_coalesced_mmio_init()
351 void *page; in mmu_memory_cache_alloc_obj() local361 return page; in mmu_memory_cache_alloc_obj()2792 static void kvm_set_page_dirty(struct page *page) in kvm_set_page_dirty() argument2798 static void kvm_set_page_accessed(struct page *page) in kvm_set_page_accessed() argument2804 void kvm_release_page_clean(struct page *page) in kvm_release_page_clean() argument2814 void kvm_release_page_dirty(struct page *page) in kvm_release_page_dirty() argument2850 struct page *page; in hva_to_pfn_fast() local2894 struct page *page, *wpage; in hva_to_pfn_slow() local4046 struct page *page; in kvm_vcpu_fault() local4065 vmf->page = page; in kvm_vcpu_fault()[all …]
162 struct page *page; in hva_to_pfn_retry() local169 .refcounted_page = &page, in hva_to_pfn_retry()204 kvm_release_page_unused(page); in hva_to_pfn_retry()224 kvm_release_page_unused(page); in hva_to_pfn_retry()246 kvm_release_page_clean(page); in hva_to_pfn_retry()
368 struct page *page = folio_page(folio, 0); in kvm_gmem_free_folio() local369 kvm_pfn_t pfn = page_to_pfn(page); in kvm_gmem_free_folio()596 gfn_t gfn, kvm_pfn_t *pfn, struct page **page, in kvm_gmem_get_pfn() argument620 *page = folio_file_page(folio, index); in kvm_gmem_get_pfn()
53 struct page **refcounted_page;
262 struct page *kvm_dirty_ring_get_page(struct kvm_dirty_ring *ring, u32 offset) in kvm_dirty_ring_get_page()
Completed in 27 milliseconds