Home
last modified time | relevance | path

Searched refs:gpa (Results 1 – 25 of 136) sorted by relevance

123456

/linux/tools/testing/selftests/kvm/x86_64/
A Dprivate_mem_conversions_test.c48 pattern, gpa + i, mem[i]); in memcmp_h()
148 memset((void *)gpa, p1, size); in guest_test_explicit_conversion()
172 memcmp_g(gpa, p2, size); in guest_test_explicit_conversion()
173 if (gpa > base_gpa) in guest_test_explicit_conversion()
176 memcmp_g(gpa + size, init_p, in guest_test_explicit_conversion()
203 memcmp_g(gpa, p4, size); in guest_test_explicit_conversion()
256 memcmp_g(gpa, init_p, size); in guest_test_punch_hole()
266 guest_punch_hole(gpa, size); in guest_test_punch_hole()
267 memcmp_g(gpa, 0, size); in guest_test_punch_hole()
340 uint64_t gpa = uc.args[1]; in __test_mem_conversions() local
[all …]
A Dsmaller_maxphyaddr_emulation_test.c52 uint64_t gpa; in main() local
69 gpa = vm_phy_pages_alloc(vm, MEM_REGION_SIZE / PAGE_SIZE, in main()
71 TEST_ASSERT(gpa == MEM_REGION_GPA, "Failed vm_phy_pages_alloc"); in main()
/linux/tools/testing/selftests/kvm/
A Dmax_guest_memory_test.c21 uint64_t gpa; in guest_code() local
24 for (gpa = start_gpa; gpa < end_gpa; gpa += stride) in guest_code()
25 *((volatile uint64_t *)gpa) = gpa; in guest_code()
91 uint64_t gpa, nr_bytes; in spawn_workers() local
105 for (i = 0, gpa = start_gpa; i < nr_vcpus; i++, gpa += nr_bytes) { in spawn_workers()
107 info[i].start_gpa = gpa; in spawn_workers()
108 info[i].end_gpa = gpa + nr_bytes; in spawn_workers()
227 gpa = 0; in main()
230 if (gpa + slot_size > max_gpa) in main()
241 __virt_pg_map(vm, gpa + i, gpa + i, PG_LEVEL_1G); in main()
[all …]
A Dmemslot_perf_test.c198 gpa -= MEM_GPA; in vm_gpa2hva()
200 gpage = gpa / guest_page_size; in vm_gpa2hva()
201 pgoffs = gpa % guest_page_size; in vm_gpa2hva()
334 uint64_t gpa; in prepare_vm() local
341 TEST_ASSERT(gpa == guest_addr, in prepare_vm()
640 uint64_t gpa, ctr; in test_memslot_do_unmap() local
648 hva = vm_gpa2hva(data, gpa, &npages); in test_memslot_do_unmap()
654 hva, gpa); in test_memslot_do_unmap()
656 gpa += npages * guest_page_size; in test_memslot_do_unmap()
665 uint64_t gpa; in test_memslot_map_unmap_check() local
[all …]
A Dmemslot_modification_stress_test.c60 uint64_t gpa; in add_remove_memslot() local
67 gpa = memstress_args.gpa - pages * vm->page_size; in add_remove_memslot()
71 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, gpa, in add_remove_memslot()
/linux/arch/powerpc/kvm/
A Dbook3s_64_mmu_radix.c149 u64 pte, base, gpa; in kvmppc_mmu_walk_radix_tree() local
215 gpte->raddr = gpa; in kvmppc_mmu_walk_radix_tree()
1183 unsigned long gpa; in kvmppc_radix_flush_memslot() local
1199 gpa += PAGE_SIZE; in kvmppc_radix_flush_memslot()
1267 unsigned long gpa; member
1345 gpa = p->gpa; in debugfs_radix_read()
1350 gpa = 0; in debugfs_radix_read()
1388 gpa = (gpa & P4D_MASK) + P4D_SIZE; in debugfs_radix_read()
1395 gpa = (gpa & PUD_MASK) + PUD_SIZE; in debugfs_radix_read()
1407 gpa = (gpa & PMD_MASK) + PMD_SIZE; in debugfs_radix_read()
[all …]
A Dbook3s_hv_uvmem.c234 unsigned long gpa; member
569 gpa, 0, page_shift); in __kvmppc_svm_page_out()
645 PAGE_SHIFT, kvm, pvt->gpa, NULL)) in kvmppc_uvmem_drop_pages()
647 pvt->gpa, addr); in kvmppc_uvmem_drop_pages()
721 pvt->gpa = gpa; in kvmppc_uvmem_get_page()
770 dpage = kvmppc_uvmem_get_page(gpa, kvm); in kvmppc_svm_page_in()
781 gpa, 0, page_shift); in kvmppc_svm_page_in()
885 unsigned long gfn = gpa >> page_shift; in kvmppc_share_page()
943 unsigned long gfn = gpa >> page_shift; in kvmppc_h_svm_page_in()
1005 pvt->kvm, pvt->gpa, vmf->page)) in kvmppc_uvmem_migrate_to_ram()
[all …]
/linux/virt/kvm/
A Dpfncache.c64 offset_in_page(gpa); in kvm_gpc_is_valid_len()
275 if (kvm_is_error_gpa(gpa)) { in __kvm_gpc_refresh()
278 gpc->gpa = INVALID_GPA; in __kvm_gpc_refresh()
287 page_offset = offset_in_page(gpa); in __kvm_gpc_refresh()
289 if (gpc->gpa != gpa || gpc->generation != slots->generation || in __kvm_gpc_refresh()
291 gfn_t gfn = gpa_to_gfn(gpa); in __kvm_gpc_refresh()
293 gpc->gpa = gpa; in __kvm_gpc_refresh()
384 gpc->gpa = INVALID_GPA; in kvm_gpc_init()
394 if (!kvm_gpc_is_valid_len(gpa, uhva, len)) in __kvm_gpc_activate()
416 return __kvm_gpc_refresh(gpc, gpa, uhva); in __kvm_gpc_activate()
[all …]
/linux/arch/s390/kvm/
A Dgaccess.h152 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in write_guest_lc() local
154 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_lc()
178 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in read_guest_lc() local
180 return kvm_read_guest(vcpu->kvm, gpa, data, len); in read_guest_lc()
190 unsigned long *gpa, enum gacc_mode mode,
196 int check_gpa_range(struct kvm *kvm, unsigned long gpa, unsigned long length,
199 int access_guest_abs_with_key(struct kvm *kvm, gpa_t gpa, void *data,
371 int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in write_guest_abs() argument
374 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_abs()
394 int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in read_guest_abs() argument
[all …]
A Dvsie.c740 gpa_t gpa; in pin_blocks() local
746 if (gpa) { in pin_blocks()
747 if (gpa < 2 * PAGE_SIZE) in pin_blocks()
761 vsie_page->sca_gpa = gpa; in pin_blocks()
768 if (gpa < 2 * PAGE_SIZE) { in pin_blocks()
784 if (gpa < 2 * PAGE_SIZE) { in pin_blocks()
797 vsie_page->gvrd_gpa = gpa; in pin_blocks()
803 if (gpa < 2 * PAGE_SIZE) { in pin_blocks()
823 if (!gpa || gpa < 2 * PAGE_SIZE) { in pin_blocks()
854 gpa_t gpa) in unpin_scb() argument
[all …]
A Dgaccess.c619 *gpa = raddr.addr; in guest_translate()
789 unsigned long gpa; in guest_range_to_gpas() local
816 *gpas++ = gpa; in guest_range_to_gpas()
850 gfn = gpa >> PAGE_SHIFT; in access_guest_page_with_key()
862 hva += offset_in_page(gpa); in access_guest_page_with_key()
889 gpa += fragment_len; in access_guest_abs_with_key()
979 unsigned long gpa; in access_guest_real() local
1023 gfn_t gfn = gpa_to_gfn(gpa); in cmpxchg_guest_abs_with_key()
1029 if (!IS_ALIGNED(gpa, len)) in cmpxchg_guest_abs_with_key()
1043 hva += offset_in_page(gpa); in cmpxchg_guest_abs_with_key()
[all …]
/linux/tools/testing/selftests/kvm/lib/
A Dmemstress.c108 vcpu_args->gpa = args->gpa + (i * vcpu_memory_bytes); in memstress_setup_vcpus()
113 vcpu_args->gpa = args->gpa; in memstress_setup_vcpus()
119 i, vcpu_args->gpa, vcpu_args->gpa + in memstress_setup_vcpus()
197 args->gpa = (region_end_gfn - guest_num_pages - 1) * args->guest_page_size; in memstress_create_vm()
198 args->gpa = align_down(args->gpa, backing_src_pagesz); in memstress_create_vm()
201 args->gpa = align_down(args->gpa, 1 << 20); in memstress_create_vm()
205 args->gpa, args->gpa + args->size); in memstress_create_vm()
210 vm_paddr_t region_start = args->gpa + region_pages * args->guest_page_size * i; in memstress_create_vm()
218 virt_map(vm, guest_test_virt_mem, args->gpa, guest_num_pages); in memstress_create_vm()
A Dkvm_util.c847 uint64_t gpa, uint64_t size, void *hva) in __vm_set_user_memory_region() argument
852 .guest_phys_addr = gpa, in __vm_set_user_memory_region()
874 uint64_t gpa, uint64_t size, void *hva, in __vm_set_user_memory_region2() argument
880 .guest_phys_addr = gpa, in __vm_set_user_memory_region2()
893 uint64_t gpa, uint64_t size, void *hva, in vm_set_user_memory_region2() argument
1209 uint64_t gpa, len; in vm_guest_mem_fallocate() local
1213 for (gpa = base; gpa < end; gpa += len) { in vm_guest_mem_fallocate()
1216 region = userspace_mem_region_find(vm, gpa, gpa); in vm_guest_mem_fallocate()
1545 gpa = vm_untag_gpa(vm, gpa); in addr_gpa2hva()
1547 region = userspace_mem_region_find(vm, gpa, gpa); in addr_gpa2hva()
[all …]
/linux/arch/x86/include/asm/uv/
A Duv_hub.h461 uv_gpa_in_mmr_space(unsigned long gpa) in uv_gpa_in_mmr_space() argument
463 return (gpa >> 62) == 0x3UL; in uv_gpa_in_mmr_space()
475 gpa = ((gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift) | in uv_gpa_to_soc_phys_ram()
476 ((gpa >> uv_hub_info->n_lshift) << uv_hub_info->m_val); in uv_gpa_to_soc_phys_ram()
478 paddr = gpa & uv_hub_info->gpa_mask; in uv_gpa_to_soc_phys_ram()
490 return gpa >> n_lshift; in uv_gpa_to_gnode()
492 return uv_gam_range(gpa)->nasid >> 1; in uv_gpa_to_gnode()
496 static inline int uv_gpa_to_pnode(unsigned long gpa) in uv_gpa_to_pnode() argument
498 return uv_gpa_to_gnode(gpa) & uv_hub_info->pnode_mask; in uv_gpa_to_pnode()
507 return (gpa << m_shift) >> m_shift; in uv_gpa_to_offset()
[all …]
/linux/arch/x86/kvm/mmu/
A Dpage_track.h30 void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, const u8 *new, int bytes);
41 static inline void __kvm_page_track_write(struct kvm *kvm, gpa_t gpa, in __kvm_page_track_write() argument
50 static inline void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, in kvm_page_track_write() argument
53 __kvm_page_track_write(vcpu->kvm, gpa, new, bytes); in kvm_page_track_write()
55 kvm_mmu_track_write(vcpu, gpa, new, bytes); in kvm_page_track_write()
/linux/arch/riscv/kvm/
A Dtlb.c21 gpa_t gpa, gpa_t gpsz, in kvm_riscv_local_hfence_gvma_vmid_gpa() argument
33 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
38 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
49 void kvm_riscv_local_hfence_gvma_gpa(gpa_t gpa, gpa_t gpsz, in kvm_riscv_local_hfence_gvma_gpa() argument
61 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
66 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
334 gpa_t gpa, gpa_t gpsz, in kvm_riscv_hfence_gvma_vmid_gpa() argument
341 data.addr = gpa; in kvm_riscv_hfence_gvma_vmid_gpa()
A Dmmu.c179 gpa_t gpa, phys_addr_t hpa, in gstage_map_page() argument
346 int kvm_riscv_gstage_ioremap(struct kvm *kvm, gpa_t gpa, in kvm_riscv_gstage_ioremap() argument
359 end = (gpa + size + PAGE_SIZE - 1) & PAGE_MASK; in kvm_riscv_gstage_ioremap()
362 for (addr = gpa; addr < end; addr += PAGE_SIZE) { in kvm_riscv_gstage_ioremap()
389 gstage_unmap_range(kvm, gpa, size, false); in kvm_riscv_gstage_iounmap()
425 gpa_t gpa = slot->base_gfn << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot() local
429 gstage_unmap_range(kvm, gpa, size, false); in kvm_arch_flush_shadow_memslot()
510 gpa_t gpa = base_gpa + (vm_start - hva); in kvm_arch_prepare_memory_region() local
522 ret = kvm_riscv_gstage_ioremap(kvm, gpa, pa, in kvm_arch_prepare_memory_region()
591 gpa_t gpa, unsigned long hva, bool is_write) in kvm_riscv_gstage_map() argument
[all …]
/linux/drivers/gpu/drm/i915/gvt/
A Dpage_track.c159 int intel_vgpu_page_track_handler(struct intel_vgpu *vgpu, u64 gpa, in intel_vgpu_page_track_handler() argument
165 page_track = intel_vgpu_find_page_track(vgpu, gpa >> PAGE_SHIFT); in intel_vgpu_page_track_handler()
171 intel_gvt_page_track_remove(vgpu, gpa >> PAGE_SHIFT); in intel_vgpu_page_track_handler()
173 ret = page_track->handler(page_track, gpa, data, bytes); in intel_vgpu_page_track_handler()
175 gvt_err("guest page write error, gpa %llx\n", gpa); in intel_vgpu_page_track_handler()
/linux/arch/loongarch/kvm/
A Dtlb.c24 void kvm_flush_tlb_gpa(struct kvm_vcpu *vcpu, unsigned long gpa) in kvm_flush_tlb_gpa() argument
27 gpa &= (PAGE_MASK << 1); in kvm_flush_tlb_gpa()
28 invtlb(INVTLB_GID_ADDR, read_csr_gstat() & CSR_GSTAT_GID, gpa); in kvm_flush_tlb_gpa()
A Dmmu.c528 gpa_t gpa = range->start << PAGE_SHIFT; in kvm_test_age_gfn() local
529 kvm_pte_t *ptep = kvm_populate_gpa(kvm, NULL, gpa, 0); in kvm_test_age_gfn()
557 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_map_page_fast()
565 ptep = kvm_populate_gpa(kvm, NULL, gpa, 0); in kvm_map_page_fast()
788 static int kvm_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, bool write) in kvm_map_page() argument
795 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_map_page()
802 err = kvm_map_page_fast(vcpu, gpa, write); in kvm_map_page()
890 ptep = kvm_populate_gpa(kvm, NULL, gpa, 0); in kvm_map_page()
902 ptep = kvm_populate_gpa(kvm, memcache, gpa, level); in kvm_map_page()
935 ret = kvm_map_page(vcpu, gpa, write); in kvm_handle_mm_fault()
[all …]
/linux/arch/x86/kvm/
A Dcpuid.h45 static inline bool kvm_vcpu_is_legal_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) in kvm_vcpu_is_legal_gpa() argument
47 return !(gpa & vcpu->arch.reserved_gpa_bits); in kvm_vcpu_is_legal_gpa()
51 gpa_t gpa, gpa_t alignment) in kvm_vcpu_is_legal_aligned_gpa() argument
53 return IS_ALIGNED(gpa, alignment) && kvm_vcpu_is_legal_gpa(vcpu, gpa); in kvm_vcpu_is_legal_aligned_gpa()
56 static inline bool page_address_valid(struct kvm_vcpu *vcpu, gpa_t gpa) in page_address_valid() argument
58 return kvm_vcpu_is_legal_aligned_gpa(vcpu, gpa, PAGE_SIZE); in page_address_valid()
/linux/tools/testing/selftests/kvm/include/
A Dkvm_util.h382 .address = gpa, in vm_set_memory_attributes()
407 vm_set_memory_attributes(vm, gpa, size, 0); in vm_mem_set_shared()
416 vm_guest_mem_fallocate(vm, gpa, size, true); in vm_guest_mem_punch_hole()
422 vm_guest_mem_fallocate(vm, gpa, size, false); in vm_guest_mem_allocate()
568 uint64_t gpa, uint64_t size, void *hva);
570 uint64_t gpa, uint64_t size, void *hva);
572 uint64_t gpa, uint64_t size, void *hva,
575 uint64_t gpa, uint64_t size, void *hva,
612 void *addr_gpa2hva(struct kvm_vm *vm, vm_paddr_t gpa);
615 void *addr_gpa2alias(struct kvm_vm *vm, vm_paddr_t gpa);
[all …]
/linux/arch/riscv/include/asm/
A Dkvm_host.h294 gpa_t gpa, gpa_t gpsz,
297 void kvm_riscv_local_hfence_gvma_gpa(gpa_t gpa, gpa_t gpsz,
323 gpa_t gpa, gpa_t gpsz,
341 int kvm_riscv_gstage_ioremap(struct kvm *kvm, gpa_t gpa,
344 void kvm_riscv_gstage_iounmap(struct kvm *kvm, gpa_t gpa,
348 gpa_t gpa, unsigned long hva, bool is_write);
/linux/arch/x86/kvm/vmx/
A Dsgx.c75 gpa_t *gpa) in sgx_gva_to_gpa() argument
80 *gpa = kvm_mmu_gva_to_gpa_write(vcpu, gva, &ex); in sgx_gva_to_gpa()
82 *gpa = kvm_mmu_gva_to_gpa_read(vcpu, gva, &ex); in sgx_gva_to_gpa()
84 if (*gpa == INVALID_GPA) { in sgx_gva_to_gpa()
92 static int sgx_gpa_to_hva(struct kvm_vcpu *vcpu, gpa_t gpa, unsigned long *hva) in sgx_gpa_to_hva() argument
94 *hva = kvm_vcpu_gfn_to_hva(vcpu, PFN_DOWN(gpa)); in sgx_gpa_to_hva()
96 sgx_handle_emulation_failure(vcpu, gpa, 1); in sgx_gpa_to_hva()
100 *hva |= gpa & ~PAGE_MASK; in sgx_gpa_to_hva()
/linux/arch/mips/kvm/
A Dmmu.c454 gpa_t gpa = range->start << PAGE_SHIFT; in kvm_test_age_gfn() local
455 pte_t *gpa_pte = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in kvm_test_age_gfn()
480 static int _kvm_mips_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, in _kvm_mips_map_page_fast() argument
485 gfn_t gfn = gpa >> PAGE_SHIFT; in _kvm_mips_map_page_fast()
494 ptep = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in _kvm_mips_map_page_fast()
556 static int kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, in kvm_mips_map_page() argument
562 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mips_map_page()
572 err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, in kvm_mips_map_page()
622 ptep = kvm_mips_pte_for_gpa(kvm, memcache, gpa); in kvm_mips_map_page()

Completed in 74 milliseconds

123456