Lines Matching refs:gpa
179 gpa_t gpa, phys_addr_t hpa, in gstage_map_page() argument
219 return gstage_set_pte(kvm, level, pcache, gpa, &new_pte); in gstage_map_page()
346 int kvm_riscv_gstage_ioremap(struct kvm *kvm, gpa_t gpa, in kvm_riscv_gstage_ioremap() argument
359 end = (gpa + size + PAGE_SIZE - 1) & PAGE_MASK; in kvm_riscv_gstage_ioremap()
362 for (addr = gpa; addr < end; addr += PAGE_SIZE) { in kvm_riscv_gstage_ioremap()
386 void kvm_riscv_gstage_iounmap(struct kvm *kvm, gpa_t gpa, unsigned long size) in kvm_riscv_gstage_iounmap() argument
389 gstage_unmap_range(kvm, gpa, size, false); in kvm_riscv_gstage_iounmap()
431 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot() local
435 gstage_unmap_range(kvm, gpa, size, false); in kvm_arch_flush_shadow_memslot()
516 gpa_t gpa = base_gpa + (vm_start - hva); in kvm_arch_prepare_memory_region() local
528 ret = kvm_riscv_gstage_ioremap(kvm, gpa, pa, in kvm_arch_prepare_memory_region()
617 gpa_t gpa, unsigned long hva, bool is_write) in kvm_riscv_gstage_map() argument
623 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_riscv_gstage_map()
649 gfn = (gpa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT; in kvm_riscv_gstage_map()
693 ret = gstage_map_page(kvm, pcache, gpa, hfn << PAGE_SHIFT, in kvm_riscv_gstage_map()
696 ret = gstage_map_page(kvm, pcache, gpa, hfn << PAGE_SHIFT, in kvm_riscv_gstage_map()