/linux-6.3-rc2/arch/powerpc/kvm/ |
A D | book3s_64_mmu_radix.c | 143 u64 pte, base, gpa; in kvmppc_mmu_walk_radix_tree() local 209 gpte->raddr = gpa; in kvmppc_mmu_walk_radix_tree() 1172 unsigned long gpa; in kvmppc_radix_flush_memslot() local 1188 gpa += PAGE_SIZE; in kvmppc_radix_flush_memslot() 1256 unsigned long gpa; member 1334 gpa = p->gpa; in debugfs_radix_read() 1339 gpa = 0; in debugfs_radix_read() 1377 gpa = (gpa & P4D_MASK) + P4D_SIZE; in debugfs_radix_read() 1384 gpa = (gpa & PUD_MASK) + PUD_SIZE; in debugfs_radix_read() 1396 gpa = (gpa & PMD_MASK) + PMD_SIZE; in debugfs_radix_read() [all …]
|
A D | book3s_hv_uvmem.c | 234 unsigned long gpa; member 568 gpa, 0, page_shift); in __kvmppc_svm_page_out() 644 PAGE_SHIFT, kvm, pvt->gpa, NULL)) in kvmppc_uvmem_drop_pages() 646 pvt->gpa, addr); in kvmppc_uvmem_drop_pages() 720 pvt->gpa = gpa; in kvmppc_uvmem_get_page() 769 dpage = kvmppc_uvmem_get_page(gpa, kvm); in kvmppc_svm_page_in() 780 gpa, 0, page_shift); in kvmppc_svm_page_in() 884 unsigned long gfn = gpa >> page_shift; in kvmppc_share_page() 942 unsigned long gfn = gpa >> page_shift; in kvmppc_h_svm_page_in() 1004 pvt->kvm, pvt->gpa, vmf->page)) in kvmppc_uvmem_migrate_to_ram() [all …]
|
/linux-6.3-rc2/tools/testing/selftests/kvm/ |
A D | max_guest_memory_test.c | 23 uint64_t gpa; in guest_code() local 25 for (gpa = start_gpa; gpa < end_gpa; gpa += stride) in guest_code() 26 *((volatile uint64_t *)gpa) = gpa; in guest_code() 96 uint64_t gpa, nr_bytes; in spawn_workers() local 110 for (i = 0, gpa = start_gpa; i < nr_vcpus; i++, gpa += nr_bytes) { in spawn_workers() 112 info[i].start_gpa = gpa; in spawn_workers() 113 info[i].end_gpa = gpa + nr_bytes; in spawn_workers() 232 gpa = 0; in main() 235 if (gpa + slot_size > max_gpa) in main() 246 __virt_pg_map(vm, gpa + i, gpa + i, PG_LEVEL_1G); in main() [all …]
|
A D | memslot_perf_test.c | 195 gpa -= MEM_GPA; in vm_gpa2hva() 197 gpage = gpa / guest_page_size; in vm_gpa2hva() 198 pgoffs = gpa % guest_page_size; in vm_gpa2hva() 331 uint64_t gpa; in prepare_vm() local 338 TEST_ASSERT(gpa == guest_addr, in prepare_vm() 632 uint64_t gpa, ctr; in test_memslot_do_unmap() local 640 hva = vm_gpa2hva(data, gpa, &npages); in test_memslot_do_unmap() 646 hva, gpa); in test_memslot_do_unmap() 648 gpa += npages * guest_page_size; in test_memslot_do_unmap() 657 uint64_t gpa; in test_memslot_map_unmap_check() local [all …]
|
A D | memslot_modification_stress_test.c | 69 uint64_t gpa; in add_remove_memslot() local 76 gpa = memstress_args.gpa - pages * vm->page_size; in add_remove_memslot() 80 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, gpa, in add_remove_memslot()
|
/linux-6.3-rc2/tools/testing/selftests/kvm/lib/ |
A D | memstress.c | 105 vcpu_args->gpa = args->gpa + (i * vcpu_memory_bytes); in memstress_setup_vcpus() 110 vcpu_args->gpa = args->gpa; in memstress_setup_vcpus() 116 i, vcpu_args->gpa, vcpu_args->gpa + in memstress_setup_vcpus() 193 args->gpa = (region_end_gfn - guest_num_pages - 1) * args->guest_page_size; in memstress_create_vm() 194 args->gpa = align_down(args->gpa, backing_src_pagesz); in memstress_create_vm() 197 args->gpa = align_down(args->gpa, 1 << 20); in memstress_create_vm() 201 args->gpa, args->gpa + args->size); in memstress_create_vm() 206 vm_paddr_t region_start = args->gpa + region_pages * args->guest_page_size * i; in memstress_create_vm() 214 virt_map(vm, guest_test_virt_mem, args->gpa, guest_num_pages); in memstress_create_vm()
|
/linux-6.3-rc2/arch/s390/kvm/ |
A D | gaccess.h | 152 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in write_guest_lc() local 154 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_lc() 178 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in read_guest_lc() local 180 return kvm_read_guest(vcpu->kvm, gpa, data, len); in read_guest_lc() 190 unsigned long *gpa, enum gacc_mode mode, 196 int check_gpa_range(struct kvm *kvm, unsigned long gpa, unsigned long length, 199 int access_guest_abs_with_key(struct kvm *kvm, gpa_t gpa, void *data, 371 int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in write_guest_abs() argument 374 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_abs() 394 int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in read_guest_abs() argument [all …]
|
A D | vsie.c | 733 gpa_t gpa; in pin_blocks() local 739 if (gpa) { in pin_blocks() 740 if (gpa < 2 * PAGE_SIZE) in pin_blocks() 754 vsie_page->sca_gpa = gpa; in pin_blocks() 761 if (gpa < 2 * PAGE_SIZE) { in pin_blocks() 777 if (gpa < 2 * PAGE_SIZE) { in pin_blocks() 790 vsie_page->gvrd_gpa = gpa; in pin_blocks() 796 if (gpa < 2 * PAGE_SIZE) { in pin_blocks() 816 if (!gpa || gpa < 2 * PAGE_SIZE) { in pin_blocks() 847 gpa_t gpa) in unpin_scb() argument [all …]
|
A D | gaccess.c | 792 *gpa = raddr.addr; in guest_translate() 962 unsigned long gpa; in guest_range_to_gpas() local 989 *gpas++ = gpa; in guest_range_to_gpas() 1021 gfn = gpa >> PAGE_SHIFT; in access_guest_page_with_key() 1033 hva += offset_in_page(gpa); in access_guest_page_with_key() 1060 gpa += fragment_len; in access_guest_abs_with_key() 1150 unsigned long gpa; in access_guest_real() local 1192 gfn_t gfn = gpa_to_gfn(gpa); in cmpxchg_guest_abs_with_key() 1198 if (!IS_ALIGNED(gpa, len)) in cmpxchg_guest_abs_with_key() 1212 hva += offset_in_page(gpa); in cmpxchg_guest_abs_with_key() [all …]
|
/linux-6.3-rc2/arch/x86/include/asm/uv/ |
A D | uv_hub.h | 460 uv_gpa_in_mmr_space(unsigned long gpa) in uv_gpa_in_mmr_space() argument 462 return (gpa >> 62) == 0x3UL; in uv_gpa_in_mmr_space() 474 gpa = ((gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift) | in uv_gpa_to_soc_phys_ram() 475 ((gpa >> uv_hub_info->n_lshift) << uv_hub_info->m_val); in uv_gpa_to_soc_phys_ram() 477 paddr = gpa & uv_hub_info->gpa_mask; in uv_gpa_to_soc_phys_ram() 489 return gpa >> n_lshift; in uv_gpa_to_gnode() 491 return uv_gam_range(gpa)->nasid >> 1; in uv_gpa_to_gnode() 495 static inline int uv_gpa_to_pnode(unsigned long gpa) in uv_gpa_to_pnode() argument 497 return uv_gpa_to_gnode(gpa) & uv_hub_info->pnode_mask; in uv_gpa_to_pnode() 506 return (gpa << m_shift) >> m_shift; in uv_gpa_to_offset() [all …]
|
/linux-6.3-rc2/virt/kvm/ |
A D | pfncache.c | 86 if ((gpc->gpa & ~PAGE_MASK) + len > PAGE_SIZE) in kvm_gpc_check() 222 gpc->khva = new_khva + (gpc->gpa & ~PAGE_MASK); in hva_to_pfn_retry() 239 static int __kvm_gpc_refresh(struct gfn_to_pfn_cache *gpc, gpa_t gpa, in __kvm_gpc_refresh() argument 243 unsigned long page_offset = gpa & ~PAGE_MASK; in __kvm_gpc_refresh() 276 if (gpc->gpa != gpa || gpc->generation != slots->generation || in __kvm_gpc_refresh() 278 gfn_t gfn = gpa_to_gfn(gpa); in __kvm_gpc_refresh() 280 gpc->gpa = gpa; in __kvm_gpc_refresh() 336 return __kvm_gpc_refresh(gpc, gpc->gpa, len); in kvm_gpc_refresh() 357 int kvm_gpc_activate(struct gfn_to_pfn_cache *gpc, gpa_t gpa, unsigned long len) in kvm_gpc_activate() argument 378 return __kvm_gpc_refresh(gpc, gpa, len); in kvm_gpc_activate()
|
/linux-6.3-rc2/arch/x86/kvm/ |
A D | cpuid.h | 45 static inline bool kvm_vcpu_is_legal_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) in kvm_vcpu_is_legal_gpa() argument 47 return !(gpa & vcpu->arch.reserved_gpa_bits); in kvm_vcpu_is_legal_gpa() 50 static inline bool kvm_vcpu_is_illegal_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) in kvm_vcpu_is_illegal_gpa() argument 52 return !kvm_vcpu_is_legal_gpa(vcpu, gpa); in kvm_vcpu_is_illegal_gpa() 56 gpa_t gpa, gpa_t alignment) in kvm_vcpu_is_legal_aligned_gpa() argument 58 return IS_ALIGNED(gpa, alignment) && kvm_vcpu_is_legal_gpa(vcpu, gpa); in kvm_vcpu_is_legal_aligned_gpa() 61 static inline bool page_address_valid(struct kvm_vcpu *vcpu, gpa_t gpa) in page_address_valid() argument 63 return kvm_vcpu_is_legal_aligned_gpa(vcpu, gpa, PAGE_SIZE); in page_address_valid()
|
/linux-6.3-rc2/arch/riscv/kvm/ |
A D | tlb.c | 21 gpa_t gpa, gpa_t gpsz, in kvm_riscv_local_hfence_gvma_vmid_gpa() argument 33 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa() 38 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa() 49 void kvm_riscv_local_hfence_gvma_gpa(gpa_t gpa, gpa_t gpsz, in kvm_riscv_local_hfence_gvma_gpa() argument 61 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa() 66 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa() 334 gpa_t gpa, gpa_t gpsz, in kvm_riscv_hfence_gvma_vmid_gpa() argument 341 data.addr = gpa; in kvm_riscv_hfence_gvma_vmid_gpa()
|
A D | mmu.c | 179 gpa_t gpa, phys_addr_t hpa, in gstage_map_page() argument 346 int kvm_riscv_gstage_ioremap(struct kvm *kvm, gpa_t gpa, in kvm_riscv_gstage_ioremap() argument 359 end = (gpa + size + PAGE_SIZE - 1) & PAGE_MASK; in kvm_riscv_gstage_ioremap() 362 for (addr = gpa; addr < end; addr += PAGE_SIZE) { in kvm_riscv_gstage_ioremap() 389 gstage_unmap_range(kvm, gpa, size, false); in kvm_riscv_gstage_iounmap() 431 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot() local 435 gstage_unmap_range(kvm, gpa, size, false); in kvm_arch_flush_shadow_memslot() 516 gpa_t gpa = base_gpa + (vm_start - hva); in kvm_arch_prepare_memory_region() local 528 ret = kvm_riscv_gstage_ioremap(kvm, gpa, pa, in kvm_arch_prepare_memory_region() 617 gpa_t gpa, unsigned long hva, bool is_write) in kvm_riscv_gstage_map() argument [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/i915/gvt/ |
A D | page_track.c | 159 int intel_vgpu_page_track_handler(struct intel_vgpu *vgpu, u64 gpa, in intel_vgpu_page_track_handler() argument 167 page_track = intel_vgpu_find_page_track(vgpu, gpa >> PAGE_SHIFT); in intel_vgpu_page_track_handler() 175 intel_gvt_page_track_remove(vgpu, gpa >> PAGE_SHIFT); in intel_vgpu_page_track_handler() 177 ret = page_track->handler(page_track, gpa, data, bytes); in intel_vgpu_page_track_handler() 179 gvt_err("guest page write error, gpa %llx\n", gpa); in intel_vgpu_page_track_handler()
|
A D | page_track.h | 35 u64 gpa, void *data, int bytes); 56 int intel_vgpu_page_track_handler(struct intel_vgpu *vgpu, u64 gpa,
|
/linux-6.3-rc2/arch/riscv/include/asm/ |
A D | kvm_host.h | 245 gpa_t gpa, gpa_t gpsz, 248 void kvm_riscv_local_hfence_gvma_gpa(gpa_t gpa, gpa_t gpsz, 274 gpa_t gpa, gpa_t gpsz, 292 int kvm_riscv_gstage_ioremap(struct kvm *kvm, gpa_t gpa, 295 void kvm_riscv_gstage_iounmap(struct kvm *kvm, gpa_t gpa, 299 gpa_t gpa, unsigned long hva, bool is_write);
|
/linux-6.3-rc2/arch/mips/kvm/ |
A D | mmu.c | 449 gpa_t gpa = range->start << PAGE_SHIFT; in kvm_set_spte_gfn() local 451 pte_t *gpa_pte = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in kvm_set_spte_gfn() 484 gpa_t gpa = range->start << PAGE_SHIFT; in kvm_test_age_gfn() local 485 pte_t *gpa_pte = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in kvm_test_age_gfn() 510 static int _kvm_mips_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, in _kvm_mips_map_page_fast() argument 515 gfn_t gfn = gpa >> PAGE_SHIFT; in _kvm_mips_map_page_fast() 524 ptep = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in _kvm_mips_map_page_fast() 586 static int kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, in kvm_mips_map_page() argument 592 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mips_map_page() 602 err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, in kvm_mips_map_page() [all …]
|
/linux-6.3-rc2/arch/x86/kvm/vmx/ |
A D | sgx.c | 74 gpa_t *gpa) in sgx_gva_to_gpa() argument 79 *gpa = kvm_mmu_gva_to_gpa_write(vcpu, gva, &ex); in sgx_gva_to_gpa() 81 *gpa = kvm_mmu_gva_to_gpa_read(vcpu, gva, &ex); in sgx_gva_to_gpa() 83 if (*gpa == INVALID_GPA) { in sgx_gva_to_gpa() 91 static int sgx_gpa_to_hva(struct kvm_vcpu *vcpu, gpa_t gpa, unsigned long *hva) in sgx_gpa_to_hva() argument 93 *hva = kvm_vcpu_gfn_to_hva(vcpu, PFN_DOWN(gpa)); in sgx_gpa_to_hva() 95 sgx_handle_emulation_failure(vcpu, gpa, 1); in sgx_gpa_to_hva() 99 *hva |= gpa & ~PAGE_MASK; in sgx_gpa_to_hva()
|
A D | vmx_ops.h | 18 void invept_error(unsigned long ext, u64 eptp, gpa_t gpa); 309 static inline void __invept(unsigned long ext, u64 eptp, gpa_t gpa) in __invept() argument 312 u64 eptp, gpa; in __invept() member 313 } operand = {eptp, gpa}; in __invept() 315 vmx_asm2(invept, "r"(ext), "m"(operand), ext, eptp, gpa); in __invept()
|
/linux-6.3-rc2/tools/testing/selftests/kvm/x86_64/ |
A D | smaller_maxphyaddr_emulation_test.c | 55 uint64_t gpa; in main() local 75 gpa = vm_phy_pages_alloc(vm, MEM_REGION_SIZE / PAGE_SIZE, in main() 77 TEST_ASSERT(gpa == MEM_REGION_GPA, "Failed vm_phy_pages_alloc\n"); in main()
|
/linux-6.3-rc2/tools/testing/selftests/kvm/lib/s390x/ |
A D | processor.c | 50 void virt_arch_pg_map(struct kvm_vm *vm, uint64_t gva, uint64_t gpa) in virt_arch_pg_map() argument 63 TEST_ASSERT((gpa % vm->page_size) == 0, in virt_arch_pg_map() 67 TEST_ASSERT((gpa >> vm->page_shift) <= vm->max_gfn, in virt_arch_pg_map() 85 "WARNING: PTE for gpa=0x%"PRIx64" already set!\n", gpa); in virt_arch_pg_map() 86 entry[idx] = gpa; in virt_arch_pg_map()
|
/linux-6.3-rc2/drivers/misc/sgi-gru/ |
A D | grufault.c | 253 int write, int atomic, unsigned long *gpa, int *pageshift) in gru_vtop() argument 279 *gpa = uv_soc_phys_ram_to_gpa(paddr); in gru_vtop() 316 unsigned long vaddr = 0, gpa; in gru_preload_tlb() local 332 ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); in gru_preload_tlb() 333 if (ret || tfh_write_only(tfh, gpa, GAA_RAM, vaddr, asid, write, in gru_preload_tlb() 339 vaddr, asid, write, pageshift, gpa); in gru_preload_tlb() 363 unsigned long gpa = 0, vaddr = 0; in gru_try_dropin() local 413 ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); in gru_try_dropin() 434 tfh_write_restart(tfh, gpa, GAA_RAM, vaddr, asid, write, in gru_try_dropin() 440 indexway, write, pageshift, gpa); in gru_try_dropin()
|
/linux-6.3-rc2/arch/arm64/include/asm/ |
A D | kvm_mmu.h | 248 gpa_t gpa, void *data, unsigned long len) 251 int ret = kvm_read_guest(kvm, gpa, data, len); 258 static inline int kvm_write_guest_lock(struct kvm *kvm, gpa_t gpa, 262 int ret = kvm_write_guest(kvm, gpa, data, len);
|
/linux-6.3-rc2/arch/x86/coco/tdx/ |
A D | tdx.c | 468 if (!mmio_write(size, ve->gpa, val)) in handle_mmio() 473 if (!mmio_write(size, ve->gpa, val)) in handle_mmio() 495 if (!mmio_read(size, ve->gpa, &val)) in handle_mmio() 646 ve->gpa = out.r9; in tdx_get_ve_info() 668 static inline bool is_private_gpa(u64 gpa) in is_private_gpa() argument 670 return gpa == cc_mkenc(gpa); in is_private_gpa() 691 if (is_private_gpa(ve->gpa)) in virt_exception_kernel()
|