Home
last modified time | relevance | path

Searched refs:eaddr (Results 1 – 25 of 39) sorted by relevance

12

/linux-6.3-rc2/arch/powerpc/kvm/
A Dbook3s_32_mmu.c75 static u32 find_sr(struct kvm_vcpu *vcpu, gva_t eaddr) in find_sr() argument
77 return kvmppc_get_sr(vcpu, (eaddr >> 28) & 0xf); in find_sr()
90 return (((u64)eaddr >> 12) & 0xffff) | (vsid << 16); in kvmppc_mmu_book3s_32_ea_to_vp()
94 u32 sre, gva_t eaddr, in kvmppc_mmu_book3s_32_get_pteg() argument
101 page = (eaddr & 0x0FFFFFFF) >> 12; in kvmppc_mmu_book3s_32_get_pteg()
152 data ? 'd' : 'i', i, eaddr, bat->bepi, in kvmppc_mmu_book3s_32_xlate_bat()
155 if ((eaddr & bat->bepi_mask) == bat->bepi) { in kvmppc_mmu_book3s_32_xlate_bat()
158 eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_book3s_32_xlate_bat()
160 pte->vpage = (((u64)eaddr >> 12) & 0xffff) | vsid; in kvmppc_mmu_book3s_32_xlate_bat()
194 sre = find_sr(vcpu, eaddr); in kvmppc_mmu_book3s_32_xlate_pte()
[all …]
A Dbook3s_64_mmu.c29 gva_t eaddr) in kvmppc_mmu_book3s_64_find_slbe() argument
32 u64 esid = GET_ESID(eaddr); in kvmppc_mmu_book3s_64_find_slbe()
33 u64 esid_1t = GET_ESID_1T(eaddr); in kvmppc_mmu_book3s_64_find_slbe()
49 eaddr, esid, esid_1t); in kvmppc_mmu_book3s_64_find_slbe()
75 eaddr &= kvmppc_slb_offset_mask(slb); in kvmppc_slb_calc_vpn()
77 return (eaddr >> VPN_SHIFT) | in kvmppc_slb_calc_vpn()
90 return kvmppc_slb_calc_vpn(slb, eaddr); in kvmppc_mmu_book3s_64_ea_to_vp()
117 struct kvmppc_slb *slbe, gva_t eaddr, in kvmppc_mmu_book3s_64_get_pteg() argument
128 vpn = kvmppc_slb_calc_vpn(slbe, eaddr); in kvmppc_mmu_book3s_64_get_pteg()
213 gpte->eaddr = eaddr; in kvmppc_mmu_book3s_64_xlate()
[all …]
A Dtrace_pr.h39 __field( unsigned long, eaddr )
49 __entry->eaddr = orig_pte->eaddr;
57 __entry->flag_w, __entry->flag_x, __entry->eaddr,
70 __field( ulong, eaddr )
79 __entry->eaddr = pte->pte.eaddr;
88 __entry->host_vpn, __entry->pfn, __entry->eaddr,
99 __field( ulong, eaddr )
108 __entry->eaddr = pte->pte.eaddr;
117 __entry->host_vpn, __entry->pfn, __entry->eaddr,
A Dbook3s_32_mmu_host.c59 asm volatile ("tlbie %0" : : "r" (pte->pte.eaddr) : "memory"); in kvmppc_mmu_invalidate_pte()
112 page = (eaddr & ~ESID_MASK) >> 12; in kvmppc_mmu_get_pteg()
138 u32 eaddr = orig_pte->eaddr; in kvmppc_mmu_map_page() local
158 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page()
161 kvmppc_mmu_map_segment(vcpu, eaddr); in kvmppc_mmu_map_page()
168 ((eaddr & ~ESID_MASK) >> VPN_SHIFT); in kvmppc_mmu_map_page()
176 pteg = kvmppc_mmu_get_pteg(vcpu, vsid, eaddr, primary); in kvmppc_mmu_map_page()
194 pteg0 = ((eaddr & 0x0fffffff) >> 22) | (vsid << 7) | PTE_V | in kvmppc_mmu_map_page()
243 orig_pte->eaddr, (ulong)pteg, vpn, in kvmppc_mmu_map_page()
302 int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr) in kvmppc_mmu_map_segment() argument
[all …]
A De500_mmu.c92 eaddr > vcpu_e500->tlb1_max_eaddr) in kvmppc_e500_tlb_index()
104 if (eaddr < get_tlb_eaddr(tlbe)) in kvmppc_e500_tlb_index()
107 if (eaddr > get_tlb_end(tlbe)) in kvmppc_e500_tlb_index()
127 gva_t eaddr, int as) in kvmppc_e500_deliver_tlb_miss() argument
155 gva_t eaddr; in kvmppc_recalc_tlb1map_range() local
169 eaddr = get_tlb_eaddr(tlbe); in kvmppc_recalc_tlb1map_range()
173 eaddr = get_tlb_end(tlbe); in kvmppc_recalc_tlb1map_range()
435 u64 eaddr = get_tlb_eaddr(gtlbe); in kvmppc_e500_emul_tlbwe() local
473 gva_t eaddr; in kvmppc_core_vcpu_translate() local
477 eaddr = tr->linear_address; in kvmppc_core_vcpu_translate()
[all …]
A Dbook3s_mmu_hpte.c26 static inline u64 kvmppc_mmu_hash_pte(u64 eaddr) in kvmppc_mmu_hash_pte() argument
28 return hash_64(eaddr >> PTE_SIZE, HPTEG_HASH_BITS_PTE); in kvmppc_mmu_hash_pte()
31 static inline u64 kvmppc_mmu_hash_pte_long(u64 eaddr) in kvmppc_mmu_hash_pte_long() argument
33 return hash_64((eaddr & 0x0ffff000) >> PTE_SIZE, in kvmppc_mmu_hash_pte_long()
66 index = kvmppc_mmu_hash_pte(pte->pte.eaddr); in kvmppc_mmu_hpte_cache_map()
70 index = kvmppc_mmu_hash_pte_long(pte->pte.eaddr); in kvmppc_mmu_hpte_cache_map()
163 if ((pte->pte.eaddr & ~0xfffUL) == guest_ea) in kvmppc_mmu_pte_flush_page()
183 if ((pte->pte.eaddr & 0x0ffff000UL) == guest_ea) in kvmppc_mmu_pte_flush_long()
A Dbook3s_64_mmu_radix.c35 gva_t eaddr, void *to, void *from, in __kvmhv_copy_tofrom_guest_radix() argument
48 if (eaddr & (0xFFFUL << 52)) in __kvmhv_copy_tofrom_guest_radix()
55 from = (void *) (eaddr | (quadrant << 62)); in __kvmhv_copy_tofrom_guest_radix()
57 to = (void *) (eaddr | (quadrant << 62)); in __kvmhv_copy_tofrom_guest_radix()
102 if (eaddr & (0x3FFUL << 52)) in kvmhv_copy_tofrom_guest_radix()
110 if (((eaddr >> 62) & 0x3) == 0x3) in kvmhv_copy_tofrom_guest_radix()
113 eaddr &= ~(0xFFFUL << 52); in kvmhv_copy_tofrom_guest_radix()
166 index = (eaddr >> offset) & ((1UL << bits) - 1); in kvmppc_mmu_walk_radix_tree()
201 gpa |= eaddr & ((1ul << offset) - 1); in kvmppc_mmu_walk_radix_tree()
208 gpte->eaddr = eaddr; in kvmppc_mmu_walk_radix_tree()
[all …]
A Dbook3s_64_mmu_host.c107 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page()
110 ret = kvmppc_mmu_map_segment(vcpu, orig_pte->eaddr); in kvmppc_mmu_map_page()
116 vsid, orig_pte->eaddr); in kvmppc_mmu_map_page()
122 vpn = hpt_vpn(orig_pte->eaddr, map->host_vsid, MMU_SEGSIZE_256M); in kvmppc_mmu_map_page()
219 vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_unmap_page()
312 int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr) in kvmppc_mmu_map_segment() argument
315 u64 esid = eaddr >> SID_SHIFT; in kvmppc_mmu_map_segment()
316 u64 slb_esid = (eaddr & ESID_MASK) | SLB_ESID_V; in kvmppc_mmu_map_segment()
323 slb_index = kvmppc_mmu_next_segment(vcpu, eaddr & ESID_MASK); in kvmppc_mmu_map_segment()
A De500_mmu_host.c105 static u32 get_host_mas0(unsigned long eaddr) in get_host_mas0() argument
115 asm volatile("tlbsx 0, %0" : : "b" (eaddr & ~CONFIG_PAGE_OFFSET)); in get_host_mas0()
586 void kvmppc_mmu_map(struct kvm_vcpu *vcpu, u64 eaddr, gpa_t gpaddr, in kvmppc_mmu_map() argument
606 &priv->ref, eaddr, &stlbe); in kvmppc_mmu_map()
613 kvmppc_e500_tlb1_map(vcpu_e500, eaddr, gfn, gtlbe, &stlbe, in kvmppc_mmu_map()
631 hva_t eaddr; in kvmppc_load_last_inst() local
708 eaddr = (unsigned long)kmap_atomic(page); in kvmppc_load_last_inst()
709 *instr = *(u32 *)(eaddr | (unsigned long)(addr & ~PAGE_MASK)); in kvmppc_load_last_inst()
710 kunmap_atomic((u32 *)eaddr); in kvmppc_load_last_inst()
A Dbooke.c1277 unsigned long eaddr = vcpu->arch.fault_dear; in kvmppc_handle_exit() local
1294 gtlb_index = kvmppc_mmu_dtlb_index(vcpu, eaddr); in kvmppc_handle_exit()
1325 vcpu->arch.vaddr_accessed = eaddr; in kvmppc_handle_exit()
1335 unsigned long eaddr = vcpu->arch.regs.nip; in kvmppc_handle_exit() local
1343 gtlb_index = kvmppc_mmu_itlb_index(vcpu, eaddr); in kvmppc_handle_exit()
1981 pte->eaddr = eaddr; in kvmppc_xlate()
1983 (eaddr & ~PAGE_MASK); in kvmppc_xlate()
1984 pte->vpage = eaddr >> PAGE_SHIFT; in kvmppc_xlate()
1996 gtlb_index = kvmppc_mmu_itlb_index(vcpu, eaddr); in kvmppc_xlate()
2011 pte->eaddr = eaddr; in kvmppc_xlate()
[all …]
A De500mc.c59 gva_t eaddr; in kvmppc_e500_tlbil_one() local
68 eaddr = get_tlb_eaddr(gtlbe); in kvmppc_e500_tlbil_one()
75 asm volatile("tlbsx 0, %[eaddr]\n" : : [eaddr] "r" (eaddr)); in kvmppc_e500_tlbil_one()
A De500.c238 u32 val, eaddr; in kvmppc_e500_tlbil_one() local
270 eaddr = get_tlb_eaddr(gtlbe); in kvmppc_e500_tlbil_one()
275 asm volatile("tlbsx 0, %[eaddr]" : : [eaddr] "r" (eaddr)); in kvmppc_e500_tlbil_one()
A Dbook3s.c454 int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr, enum xlate_instdata xlid, in kvmppc_xlate() argument
463 r = vcpu->arch.mmu.xlate(vcpu, eaddr, pte, data, iswrite); in kvmppc_xlate()
465 pte->eaddr = eaddr; in kvmppc_xlate()
466 pte->raddr = eaddr & KVM_PAM; in kvmppc_xlate()
467 pte->vpage = VSID_REAL | eaddr >> 12; in kvmppc_xlate()
476 ((eaddr & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS)) in kvmppc_xlate()
A Dbook3s_64_mmu_hv.c311 gva_t eaddr) in kvmppc_mmu_book3s_hv_find_slbe() argument
325 if (((vcpu->arch.slb[i].orige ^ eaddr) & mask) == 0) in kvmppc_mmu_book3s_hv_find_slbe()
340 static int kvmppc_mmu_book3s_64_hv_xlate(struct kvm_vcpu *vcpu, gva_t eaddr, in kvmppc_mmu_book3s_64_hv_xlate() argument
353 return kvmppc_mmu_radix_xlate(vcpu, eaddr, gpte, data, iswrite); in kvmppc_mmu_book3s_64_hv_xlate()
357 slbe = kvmppc_mmu_book3s_hv_find_slbe(vcpu, eaddr); in kvmppc_mmu_book3s_64_hv_xlate()
368 index = kvmppc_hv_find_lock_hpte(kvm, eaddr, slb_v, in kvmppc_mmu_book3s_64_hv_xlate()
383 gpte->eaddr = eaddr; in kvmppc_mmu_book3s_64_hv_xlate()
384 gpte->vpage = ((v & HPTE_V_AVPN) << 4) | ((eaddr >> 12) & 0xfff); in kvmppc_mmu_book3s_64_hv_xlate()
406 gpte->raddr = kvmppc_mmu_get_real_addr(v, gr, eaddr); in kvmppc_mmu_book3s_64_hv_xlate()
A Dbook3s_hv_rm_mmu.c1063 unsigned long eaddr, unsigned long slb_v, long mmio_update) in mmio_cache_search() argument
1073 if ((entry->eaddr >> pshift) == (eaddr >> pshift) && in mmio_cache_search()
1097 long kvmppc_hv_find_lock_hpte(struct kvm *kvm, gva_t eaddr, unsigned long slb_v, in kvmppc_hv_find_lock_hpte() argument
1126 hash = (vsid ^ ((eaddr & somask) >> pshift)) & kvmppc_hpt_mask(&kvm->arch.hpt); in kvmppc_hv_find_lock_hpte()
1128 avpn |= (eaddr & somask) >> 16; in kvmppc_hv_find_lock_hpte()
1285 cache_entry->eaddr = addr; in kvmppc_hpte_hv_fault()
/linux-6.3-rc2/drivers/edac/
A Digen6_edac.c237 return eaddr; in ehl_err_addr_to_sys_addr()
242 if (eaddr < igen6_tolud) in ehl_err_addr_to_imc_addr()
243 return eaddr; in ehl_err_addr_to_imc_addr()
248 if (eaddr < _4GB) in ehl_err_addr_to_imc_addr()
251 return eaddr; in ehl_err_addr_to_imc_addr()
308 if (eaddr >= ms_s_size) in tgl_err_addr_to_mem_addr()
309 return eaddr + ms_s_size; in tgl_err_addr_to_mem_addr()
333 return eaddr; in tgl_err_addr_to_imc_addr()
348 if (eaddr >= 2 * ms_s_size) in adl_err_addr_to_imc_addr()
349 return eaddr - ms_s_size; in adl_err_addr_to_imc_addr()
[all …]
/linux-6.3-rc2/arch/arm64/kernel/
A Dcompat_alignment.c118 unsigned long eaddr, newaddr; in do_alignment_ldmstm() local
125 newaddr = eaddr = regs->regs[rn]; in do_alignment_ldmstm()
131 eaddr = newaddr; in do_alignment_ldmstm()
134 eaddr += 4; in do_alignment_ldmstm()
140 if (get_user(val, (u32 __user *)eaddr)) in do_alignment_ldmstm()
155 if (put_user(val, (u32 __user *)eaddr)) in do_alignment_ldmstm()
158 eaddr += 4; in do_alignment_ldmstm()
/linux-6.3-rc2/arch/powerpc/include/asm/
A Dkvm_book3s.h155 extern int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr);
156 extern void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong eaddr, ulong seg_size);
160 extern long kvmppc_hv_find_lock_hpte(struct kvm *kvm, gva_t eaddr,
179 gva_t eaddr, void *to, void *from,
181 extern long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr,
183 extern long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr,
185 extern int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr,
188 extern int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr,
191 extern int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
226 extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr, bool data);
A Dkvm_ppc.h89 extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr,
91 extern int kvmppc_st(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr,
108 extern int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr);
109 extern int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr);
111 gva_t eaddr);
114 extern int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr,
302 int (*load_from_eaddr)(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr,
304 int (*store_to_eaddr)(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr,
A Dkvm_host.h378 ulong eaddr; member
395 int (*slbfee)(struct kvm_vcpu *vcpu, gva_t eaddr, ulong *ret_slb);
401 int (*xlate)(struct kvm_vcpu *vcpu, gva_t eaddr,
405 u64 (*ea_to_vp)(struct kvm_vcpu *vcpu, gva_t eaddr, bool data);
473 unsigned long eaddr; member
/linux-6.3-rc2/arch/arm/mm/
A Dalignment.c503 unsigned long eaddr, newaddr; in do_alignment_ldmstm() local
517 newaddr = eaddr = regs->uregs[rn]; in do_alignment_ldmstm()
523 eaddr = newaddr; in do_alignment_ldmstm()
526 eaddr += 4; in do_alignment_ldmstm()
540 if (addr != eaddr) { in do_alignment_ldmstm()
543 instruction_pointer(regs), instr, addr, eaddr); in do_alignment_ldmstm()
555 get32t_unaligned_check(val, eaddr); in do_alignment_ldmstm()
559 eaddr += 4; in do_alignment_ldmstm()
568 get32_unaligned_check(val, eaddr); in do_alignment_ldmstm()
571 put32_unaligned_check(regs->uregs[rd], eaddr); in do_alignment_ldmstm()
[all …]
/linux-6.3-rc2/fs/freevxfs/
A Dvxfs_olt.c58 char *oaddr, *eaddr; in vxfs_read_olt() local
81 eaddr = bp->b_data + (infp->vsi_oltsize * sbp->s_blocksize); in vxfs_read_olt()
83 while (oaddr < eaddr) { in vxfs_read_olt()
/linux-6.3-rc2/arch/powerpc/platforms/pseries/
A Dras.c572 unsigned long eaddr = 0, paddr = 0; in mce_handle_err_virtmode() local
642 eaddr = be64_to_cpu(mce_log->effective_address); in mce_handle_err_virtmode()
649 pfn = addr_to_pfn(regs, eaddr); in mce_handle_err_virtmode()
670 eaddr = be64_to_cpu(mce_log->effective_address); in mce_handle_err_virtmode()
687 eaddr = be64_to_cpu(mce_log->effective_address); in mce_handle_err_virtmode()
704 eaddr = be64_to_cpu(mce_log->effective_address); in mce_handle_err_virtmode()
725 eaddr = be64_to_cpu(mce_log->effective_address); in mce_handle_err_virtmode()
734 &mce_err, regs->nip, eaddr, paddr); in mce_handle_err_virtmode()
/linux-6.3-rc2/drivers/gpu/drm/amd/amdgpu/
A Damdgpu_vm.c1436 uint64_t eaddr; in amdgpu_vm_bo_map() local
1444 eaddr = saddr + size - 1; in amdgpu_vm_bo_map()
1445 if (saddr >= eaddr || in amdgpu_vm_bo_map()
1467 mapping->last = eaddr; in amdgpu_vm_bo_map()
1501 uint64_t eaddr; in amdgpu_vm_bo_replace_map() local
1510 eaddr = saddr + size - 1; in amdgpu_vm_bo_replace_map()
1511 if (saddr >= eaddr || in amdgpu_vm_bo_replace_map()
1531 mapping->last = eaddr; in amdgpu_vm_bo_replace_map()
1614 uint64_t eaddr; in amdgpu_vm_bo_clear_mappings() local
1670 if (tmp->last > eaddr) in amdgpu_vm_bo_clear_mappings()
[all …]
/linux-6.3-rc2/drivers/slimbus/
A Dcore.c176 struct slim_eaddr *eaddr, in slim_alloc_device() argument
186 sbdev->e_addr = *eaddr; in slim_alloc_device()
348 struct slim_eaddr *eaddr) in find_slim_device() argument
353 dev = device_find_child(ctrl->dev, eaddr, slim_match_dev); in find_slim_device()

Completed in 64 milliseconds

12