| /arch/powerpc/mm/ |
| A D | copro_fault.c | 83 u64 vsid, vsidkey; in copro_calculate_slb() local 93 vsid = get_user_vsid(&mm->context, ea, ssize); in copro_calculate_slb() 100 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); in copro_calculate_slb() 107 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); in copro_calculate_slb() 114 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); in copro_calculate_slb() 122 if (!vsid) in copro_calculate_slb() 125 vsid = (vsid << slb_vsid_shift(ssize)) | vsidkey; in copro_calculate_slb() 127 vsid |= mmu_psize_defs[psize].sllp | in copro_calculate_slb() 131 slb->vsid = vsid; in copro_calculate_slb()
|
| /arch/powerpc/include/asm/book3s/64/ |
| A D | mmu-hash.h | 159 void (*hugepage_invalidate)(unsigned long vsid, 421 unsigned long vsid, int ssize) in hpt_vpn() argument 437 unsigned long hash, vsid; in hpt_hash() local 446 vsid = vpn >> (SID_SHIFT_1T - VPN_SHIFT); in hpt_hash() 447 hash = vsid ^ (vsid << 25) ^ in hpt_hash() 483 unsigned long vsid, pmd_t *pmdp, in __hash_page_thp() argument 515 u64 vsid; member 757 unsigned long vsid; in vsid_scramble() local 762 vsid = protovsid * vsid_multiplier; in vsid_scramble() 763 vsid = (vsid >> vsid_bits) + (vsid & vsid_modulus); in vsid_scramble() [all …]
|
| A D | tlbflush-hash.h | 65 extern void flush_hash_hugepage(unsigned long vsid, unsigned long addr,
|
| /arch/powerpc/kvm/ |
| A D | book3s_32_mmu.c | 73 u64 *vsid); 83 u64 vsid; in kvmppc_mmu_book3s_32_ea_to_vp() local 156 u64 vsid; in kvmppc_mmu_book3s_32_xlate_bat() local 158 eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_book3s_32_xlate_bat() 159 vsid <<= 16; in kvmppc_mmu_book3s_32_xlate_bat() 349 u64 *vsid) in kvmppc_mmu_book3s_32_esid_to_vsid() argument 367 *vsid = VSID_REAL | esid; in kvmppc_mmu_book3s_32_esid_to_vsid() 370 *vsid = VSID_REAL_IR | gvsid; in kvmppc_mmu_book3s_32_esid_to_vsid() 377 *vsid = sr_vsid(sr); in kvmppc_mmu_book3s_32_esid_to_vsid() 379 *vsid = VSID_BAT | gvsid; in kvmppc_mmu_book3s_32_esid_to_vsid() [all …]
|
| A D | book3s_32_mmu_host.c | 106 static u32 *kvmppc_mmu_get_pteg(struct kvm_vcpu *vcpu, u32 vsid, u32 eaddr, in kvmppc_mmu_get_pteg() argument 114 hash = ((vsid ^ page) << 6); in kvmppc_mmu_get_pteg() 134 u64 vsid; in kvmppc_mmu_map_page() local 157 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page() 158 map = find_sid_vsid(vcpu, vsid); in kvmppc_mmu_map_page() 161 map = find_sid_vsid(vcpu, vsid); in kvmppc_mmu_map_page() 165 vsid = map->host_vsid; in kvmppc_mmu_map_page() 166 vpn = (vsid << (SID_SHIFT - VPN_SHIFT)) | in kvmppc_mmu_map_page() 175 pteg = kvmppc_mmu_get_pteg(vcpu, vsid, eaddr, primary); in kvmppc_mmu_map_page() 193 pteg0 = ((eaddr & 0x0fffffff) >> 22) | (vsid << 7) | PTE_V | in kvmppc_mmu_map_page()
|
| A D | book3s_64_mmu_host.c | 77 u64 vsid; in kvmppc_mmu_map_page() local 108 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page() 109 map = find_sid_vsid(vcpu, vsid); in kvmppc_mmu_map_page() 113 map = find_sid_vsid(vcpu, vsid); in kvmppc_mmu_map_page() 117 vsid, orig_pte->eaddr); in kvmppc_mmu_map_page() 142 if (vsid & VSID_64K) in kvmppc_mmu_map_page() 217 u64 vsid; in kvmppc_mmu_unmap_page() local 219 vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_unmap_page() 220 if (vsid & VSID_64K) in kvmppc_mmu_unmap_page() 349 svcpu->slb[slb_index].vsid = slb_vsid; in kvmppc_mmu_map_segment()
|
| A D | book3s_64_mmu.c | 51 if (vcpu->arch.slb[i].vsid) in kvmppc_mmu_book3s_64_find_slbe() 57 vcpu->arch.slb[i].vsid); in kvmppc_mmu_book3s_64_find_slbe() 78 ((slb->vsid) << (kvmppc_slb_sid_shift(slb) - VPN_SHIFT)); in kvmppc_slb_calc_vpn() 141 page, vcpu_book3s->sdr1, pteg, slbe->vsid); in kvmppc_mmu_book3s_64_get_pteg() 161 avpn |= slbe->vsid << (kvmppc_slb_sid_shift(slbe) - p); in kvmppc_mmu_book3s_64_get_avpn() 383 slbe->vsid = (rs & ~SLB_VSID_B) >> (kvmppc_slb_sid_shift(slbe) - 16); in kvmppc_mmu_book3s_64_slbmte() 573 u64 *vsid) in kvmppc_mmu_book3s_64_esid_to_vsid() argument 585 gvsid = slb->vsid; in kvmppc_mmu_book3s_64_esid_to_vsid() 631 *vsid = gvsid; in kvmppc_mmu_book3s_64_esid_to_vsid() 639 *vsid = VSID_REAL | esid; in kvmppc_mmu_book3s_64_esid_to_vsid()
|
| A D | book3s_64_mmu_hv.c | 1287 unsigned long offset, vsid; in resize_hpt_rehash_hpte() local 1291 vsid = avpn >> 5; in resize_hpt_rehash_hpte() 1294 offset |= ((vsid ^ pteg) & old_hash_mask) << pshift; in resize_hpt_rehash_hpte() 1296 hash = vsid ^ (offset >> pshift); in resize_hpt_rehash_hpte() 1298 unsigned long offset, vsid; in resize_hpt_rehash_hpte() local 1302 vsid = avpn >> 17; in resize_hpt_rehash_hpte() 1304 offset |= ((vsid ^ (vsid << 25) ^ pteg) & old_hash_mask) << pshift; in resize_hpt_rehash_hpte() 1306 hash = vsid ^ (vsid << 25) ^ (offset >> pshift); in resize_hpt_rehash_hpte()
|
| /arch/powerpc/mm/book3s64/ |
| A D | hash_native.c | 485 unsigned long vsid; in native_hpte_updateboltedpp() local 492 vsid = get_kernel_vsid(ea, ssize); in native_hpte_updateboltedpp() 493 vpn = hpt_vpn(ea, vsid, ssize); in native_hpte_updateboltedpp() 521 unsigned long vsid; in native_hpte_removebolted() local 528 vsid = get_kernel_vsid(ea, ssize); in native_hpte_removebolted() 529 vpn = hpt_vpn(ea, vsid, ssize); in native_hpte_removebolted() 616 vpn = hpt_vpn(addr, vsid, ssize); in native_hugepage_invalidate() 666 unsigned long vsid, seg_off; in hpte_decode() local 695 vsid = avpn >> 5; in hpte_decode() 706 vsid = avpn >> 17; in hpte_decode() [all …]
|
| A D | hash_tlb.c | 45 unsigned long vsid; in hpte_need_flush() local 90 vsid = get_user_vsid(&mm->context, addr, ssize); in hpte_need_flush() 92 vsid = get_kernel_vsid(addr, mmu_kernel_ssize); in hpte_need_flush() 95 WARN_ON(vsid == 0); in hpte_need_flush() 96 vpn = hpt_vpn(addr, vsid, ssize); in hpte_need_flush()
|
| A D | hash_hugepage.c | 21 int __hash_page_thp(unsigned long ea, unsigned long access, unsigned long vsid, in __hash_page_thp() argument 76 vpn = hpt_vpn(ea, vsid, ssize); in __hash_page_thp() 85 flush_hash_hugepage(vsid, ea, pmdp, MMU_PAGE_64K, in __hash_page_thp() 163 hash_failure_debug(ea, access, vsid, trap, ssize, in __hash_page_thp()
|
| A D | hash_64k.c | 38 int __hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid, in __hash_page_4K() argument 90 vpn = hpt_vpn(ea, vsid, ssize); in __hash_page_4K() 213 hash_failure_debug(ea, access, vsid, trap, ssize, in __hash_page_4K() 229 unsigned long vsid, pte_t *ptep, unsigned long trap, in __hash_page_64K() argument 275 vpn = hpt_vpn(ea, vsid, ssize); in __hash_page_64K() 328 hash_failure_debug(ea, access, vsid, trap, ssize, in __hash_page_64K()
|
| A D | hash_utils.c | 289 if (!vsid) in kernel_map_linear_page() 649 if (!vsid) in htab_bolt_mapping() 1667 unsigned long vsid, unsigned long trap, in hash_failure_debug() argument 1675 trap, vsid, ssize, psize, lpsize, pte); in hash_failure_debug() 1707 unsigned long vsid; in hash_page_mm() local 1731 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); in hash_page_mm() 1738 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); in hash_page_mm() 1754 if (!vsid) { in hash_page_mm() 2016 unsigned long vsid; in hash_preload() local 2038 if (!vsid) in hash_preload() [all …]
|
| A D | hash_4k.c | 21 int __hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid, in __hash_page_4K() argument 66 vpn = hpt_vpn(ea, vsid, ssize); in __hash_page_4K() 117 hash_failure_debug(ea, access, vsid, trap, ssize, in __hash_page_4K()
|
| A D | slb.c | 78 WRITE_ONCE(p->save_area[index].vsid, cpu_to_be64(mk_vsid_data(ea, ssize, flags))); in slb_shadow_update() 117 : "r" (be64_to_cpu(p->save_area[index].vsid)), in __slb_restore_bolted_realmode() 164 ksp_vsid_data = be64_to_cpu(p->save_area[KSTACK_INDEX].vsid); in __slb_flush_and_restore_bolted() 217 slb_ptr->vsid = v; in slb_save_contents() 235 v = slb_ptr->vsid; in slb_dump_contents() 674 unsigned long vsid; in slb_insert_entry() local 678 vsid = get_vsid(context, ea, ssize); in slb_insert_entry() 679 if (!vsid) in slb_insert_entry() 695 vsid_data = __mk_vsid_data(vsid, ssize, flags); in slb_insert_entry()
|
| A D | hugetlbpage.c | 20 int __hash_page_huge(unsigned long ea, unsigned long access, unsigned long vsid, in __hash_page_huge() argument 33 vpn = hpt_vpn(ea, vsid, ssize); in __hash_page_huge() 122 hash_failure_debug(ea, access, vsid, trap, ssize, in __hash_page_huge()
|
| A D | hash_pgtable.c | 316 unsigned long vsid; in hpte_do_hugepage_flush() local 331 vsid = get_user_vsid(&mm->context, addr, ssize); in hpte_do_hugepage_flush() 332 WARN_ON(vsid == 0); in hpte_do_hugepage_flush() 334 vsid = get_kernel_vsid(addr, mmu_kernel_ssize); in hpte_do_hugepage_flush() 341 return flush_hash_hugepage(vsid, addr, pmdp, psize, ssize, flags); in hpte_do_hugepage_flush()
|
| /arch/microblaze/include/asm/ |
| A D | mmu.h | 20 unsigned long vsid:24; /* Virtual segment identifier */ member 48 unsigned long vsid:24; /* Virtual Segment Identifier */ member
|
| /arch/powerpc/mm/ptdump/ |
| A D | hashpagetable.c | 210 unsigned long hash, vsid, vpn, hpte_group, want_v, hpte_v; in native_find() local 215 vsid = get_kernel_vsid(ea, ssize); in native_find() 216 vpn = hpt_vpn(ea, vsid, ssize); in native_find() 245 unsigned long vsid, vpn, hash, hpte_group, want_v; in pseries_find() local 251 vsid = get_kernel_vsid(ea, ssize); in pseries_find() 252 vpn = hpt_vpn(ea, vsid, ssize); in pseries_find()
|
| /arch/powerpc/include/asm/ |
| A D | copro.h | 13 u64 esid, vsid; member
|
| A D | lppaca.h | 142 __be64 vsid; member
|
| A D | kvm_book3s_asm.h | 147 u64 vsid;
|
| /arch/powerpc/platforms/pseries/ |
| A D | lpar.c | 1018 unsigned long lpar_rc, slot, vsid, flags; in pSeries_lpar_hpte_updateboltedpp() local 1020 vsid = get_kernel_vsid(ea, ssize); in pSeries_lpar_hpte_updateboltedpp() 1021 vpn = hpt_vpn(ea, vsid, ssize); in pSeries_lpar_hpte_updateboltedpp() 1272 vpn = hpt_vpn(addr, vsid, ssize); in pSeries_lpar_hugepage_invalidate() 1312 unsigned long slot, vsid; in pSeries_lpar_hpte_removebolted() local 1314 vsid = get_kernel_vsid(ea, ssize); in pSeries_lpar_hpte_removebolted() 1315 vpn = hpt_vpn(ea, vsid, ssize); in pSeries_lpar_hpte_removebolted() 1952 if (vsid >= vsid_modulus) in vsid_unscramble() 1962 max_mod_inv = 0x7fffffffffffffffull / vsid; in vsid_unscramble() 1964 return (vsid * modinv) % vsid_modulus; in vsid_unscramble() [all …]
|
| /arch/powerpc/xmon/ |
| A D | xmon.c | 2640 u64 esid, vsid; in dump_one_paca() local 2646 vsid = be64_to_cpu(p->slb_shadow_ptr->save_area[i].vsid); in dump_one_paca() 2648 if (esid || vsid) { in dump_one_paca() 2650 22, "slb_shadow", i, esid, vsid); in dump_one_paca() 3736 unsigned long esid,vsid; in dump_segments() local 3745 if (!esid && !vsid) in dump_segments() 3748 printf("%02d %016lx %016lx", i, esid, vsid); in dump_segments() 3755 llp = vsid & SLB_VSID_LLP; in dump_segments() 3756 if (vsid & SLB_VSID_B_1T) { in dump_segments() 3759 (vsid & ~SLB_VSID_B) >> SLB_VSID_SHIFT_1T, in dump_segments() [all …]
|
| /arch/powerpc/platforms/cell/ |
| A D | spu_base.c | 138 __func__, slbe, slb->vsid, slb->esid); in spu_load_slb() 144 out_be64(&priv2->slb_vsid_RW, slb->vsid); in spu_load_slb() 217 slb->vsid = (get_kernel_vsid(ea, MMU_SEGSIZE_256M) << SLB_VSID_SHIFT) | in __spu_kernel_slb()
|