Lines Matching refs:gpa_t
37 #define gstage_gpa_size ((gpa_t)(1ULL << gstage_gpa_bits))
42 static inline unsigned long gstage_pte_index(gpa_t addr, u32 level) in gstage_pte_index()
97 static bool gstage_get_leaf_entry(struct kvm *kvm, gpa_t addr, in gstage_get_leaf_entry()
126 static void gstage_remote_tlb_flush(struct kvm *kvm, u32 level, gpa_t addr) in gstage_remote_tlb_flush()
139 gpa_t addr, const pte_t *new_pte) in gstage_set_pte()
179 gpa_t gpa, phys_addr_t hpa, in gstage_map_page()
228 static void gstage_op_pte(struct kvm *kvm, gpa_t addr, in gstage_op_pte()
269 static void gstage_unmap_range(struct kvm *kvm, gpa_t start, in gstage_unmap_range()
270 gpa_t size, bool may_block) in gstage_unmap_range()
277 gpa_t addr = start, end = start + size; in gstage_unmap_range()
305 static void gstage_wp_range(struct kvm *kvm, gpa_t start, gpa_t end) in gstage_wp_range()
311 gpa_t addr = start; in gstage_wp_range()
346 int kvm_riscv_gstage_ioremap(struct kvm *kvm, gpa_t gpa, in kvm_riscv_gstage_ioremap()
386 void kvm_riscv_gstage_iounmap(struct kvm *kvm, gpa_t gpa, unsigned long size) in kvm_riscv_gstage_iounmap()
431 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot()
459 gpa_t base_gpa; in kvm_arch_prepare_memory_region()
516 gpa_t gpa = base_gpa + (vm_start - hva); in kvm_arch_prepare_memory_region()
617 gpa_t gpa, unsigned long hva, bool is_write) in kvm_riscv_gstage_map()