/linux-6.3-rc2/arch/arm64/mm/ |
A D | hugetlbpage.c | 157 if (!pte_present(orig_pte) || !pte_cont(orig_pte)) in huge_ptep_get() 158 return orig_pte; in huge_ptep_get() 165 orig_pte = pte_mkdirty(orig_pte); in huge_ptep_get() 168 orig_pte = pte_mkyoung(orig_pte); in huge_ptep_get() 170 return orig_pte; in huge_ptep_get() 199 orig_pte = pte_mkdirty(orig_pte); in get_clear_contig() 202 orig_pte = pte_mkyoung(orig_pte); in get_clear_contig() 204 return orig_pte; in get_clear_contig() 217 return orig_pte; in get_clear_contig_flush() 431 if (!pte_cont(orig_pte)) in huge_ptep_get_and_clear() [all …]
|
/linux-6.3-rc2/arch/powerpc/kvm/ |
A D | book3s_32_mmu_host.c | 138 u32 eaddr = orig_pte->eaddr; in kvmppc_mmu_map_page() 148 hpaddr = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page() 151 orig_pte->raddr); in kvmppc_mmu_map_page() 198 if (orig_pte->may_write && writable) { in kvmppc_mmu_map_page() 200 mark_page_dirty(vcpu->kvm, orig_pte->raddr >> PAGE_SHIFT); in kvmppc_mmu_map_page() 205 if (orig_pte->may_execute) in kvmppc_mmu_map_page() 241 orig_pte->may_write ? 'w' : '-', in kvmppc_mmu_map_page() 242 orig_pte->may_execute ? 'x' : '-', in kvmppc_mmu_map_page() 243 orig_pte->eaddr, (ulong)pteg, vpn, in kvmppc_mmu_map_page() 244 orig_pte->vpage, hpaddr); in kvmppc_mmu_map_page() [all …]
|
A D | book3s_64_mmu_host.c | 89 unsigned long gfn = orig_pte->raddr >> PAGE_SHIFT; in kvmppc_mmu_map_page() 100 orig_pte->raddr); in kvmppc_mmu_map_page() 110 ret = kvmppc_mmu_map_segment(vcpu, orig_pte->eaddr); in kvmppc_mmu_map_page() 116 vsid, orig_pte->eaddr); in kvmppc_mmu_map_page() 122 vpn = hpt_vpn(orig_pte->eaddr, map->host_vsid, MMU_SEGSIZE_256M); in kvmppc_mmu_map_page() 125 if (!orig_pte->may_write || !writable) in kvmppc_mmu_map_page() 132 if (!orig_pte->may_execute) in kvmppc_mmu_map_page() 138 rflags = (rflags & ~HPTE_R_WIMG) | orig_pte->wimg; in kvmppc_mmu_map_page() 147 hpaddr |= orig_pte->raddr & (~0xfffULL & ~PAGE_MASK); in kvmppc_mmu_map_page() 183 vpn, hpaddr, orig_pte); in kvmppc_mmu_map_page() [all …]
|
A D | trace_pr.h | 33 struct kvmppc_pte *orig_pte), 34 TP_ARGS(rflags, hpteg, va, hpaddr, orig_pte), 49 __entry->eaddr = orig_pte->eaddr; 52 __entry->vpage = orig_pte->vpage;
|
A D | book3s_hv_rm_mmu.c | 492 u64 pte, orig_pte, pte_r; in kvmppc_do_h_remove() local 501 pte = orig_pte = be64_to_cpu(hpte[0]); in kvmppc_do_h_remove() 510 __unlock_hpte(hpte, orig_pte); in kvmppc_do_h_remove()
|
/linux-6.3-rc2/mm/ |
A D | madvise.c | 343 pte_t *orig_pte, *pte, ptent; in madvise_cold_or_pageout_pte_range() local 459 pte_unmap_unlock(orig_pte, ptl); in madvise_cold_or_pageout_pte_range() 463 orig_pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in madvise_cold_or_pageout_pte_range() 468 orig_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in madvise_cold_or_pageout_pte_range() 514 pte_unmap_unlock(orig_pte, ptl); in madvise_cold_or_pageout_pte_range() 615 pte_t *orig_pte, *pte, ptent; in madvise_free_pte_range() local 629 orig_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in madvise_free_pte_range() 674 pte_unmap_unlock(orig_pte, ptl); in madvise_free_pte_range() 678 orig_pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in madvise_free_pte_range() 683 orig_pte = pte = pte_offset_map_lock(mm, pmd, addr, &ptl); in madvise_free_pte_range() [all …]
|
A D | memory.c | 2842 entry = pte_mkyoung(vmf->orig_pte); in __wp_page_copy_user() 3019 entry = pte_mkyoung(vmf->orig_pte); in wp_page_reuse() 3120 if (pte_soft_dirty(vmf->orig_pte)) in wp_page_copy() 3122 if (pte_uffd_wp(vmf->orig_pte)) in wp_page_copy() 3931 if (pte_swp_soft_dirty(vmf->orig_pte)) in do_swap_page() 3933 if (pte_swp_uffd_wp(vmf->orig_pte)) in do_swap_page() 3935 vmf->orig_pte = pte; in do_swap_page() 4887 vmf->orig_pte = *vmf->pte; in handle_pte_fault() 4899 if (pte_none(vmf->orig_pte)) { in handle_pte_fault() 4912 if (!pte_present(vmf->orig_pte)) in handle_pte_fault() [all …]
|
A D | swap_state.c | 705 pte_t *pte, *orig_pte; in swap_ra_info() local 733 orig_pte = pte = pte_offset_map(vmf->pmd, faddr); in swap_ra_info() 761 pte_unmap(orig_pte); in swap_ra_info()
|
A D | ksm.c | 1045 pte_t *orig_pte) in write_protect_page() argument 1115 *orig_pte = *pvmw.pte; in write_protect_page() 1136 struct page *kpage, pte_t orig_pte) in replace_page() argument 1171 if (!pte_same(*ptep, orig_pte)) { in replace_page() 1234 pte_t orig_pte = __pte(0); in try_to_merge_one_page() local 1264 if (write_protect_page(vma, page, &orig_pte) == 0) { in try_to_merge_one_page() 1281 err = replace_page(vma, page, kpage, orig_pte); in try_to_merge_one_page()
|
A D | khugepaged.c | 935 vmf.orig_pte = *vmf.pte; in __collapse_huge_page_swapin() 936 if (!is_swap_pte(vmf.orig_pte)) { in __collapse_huge_page_swapin()
|
/linux-6.3-rc2/arch/sparc/mm/ |
A D | tlb.c | 206 pte_t orig_pte = __pte(pmd_val(orig)); in __set_pmd_acct() local 207 bool exec = pte_exec(orig_pte); in __set_pmd_acct()
|
/linux-6.3-rc2/arch/x86/kvm/mmu/ |
A D | paging_tmpl.h | 203 pt_element_t pte, orig_pte; in FNAME() local 213 pte = orig_pte = walker->ptes[level - 1]; in FNAME() 230 if (pte == orig_pte) in FNAME() 249 ret = __try_cmpxchg_user(ptep_user, &orig_pte, pte, fault); in FNAME()
|
/linux-6.3-rc2/fs/proc/ |
A D | task_mmu.c | 1471 pte_t *pte, *orig_pte; in pagemap_pmd_range() local 1551 orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); in pagemap_pmd_range() 1560 pte_unmap_unlock(orig_pte, ptl); in pagemap_pmd_range() 1871 pte_t *orig_pte; in gather_pte_stats() local 1890 orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in gather_pte_stats() 1898 pte_unmap_unlock(orig_pte, ptl); in gather_pte_stats()
|
/linux-6.3-rc2/arch/arm64/kvm/hyp/ |
A D | pgtable.c | 1084 kvm_pte_t attr_clr, kvm_pte_t *orig_pte, in stage2_update_leaf_attrs() argument 1103 if (orig_pte) in stage2_update_leaf_attrs() 1104 *orig_pte = data.pte; in stage2_update_leaf_attrs()
|
/linux-6.3-rc2/include/linux/ |
A D | pgtable.h | 773 pte_t orig_pte) in arch_unmap_one() argument
|
A D | mm.h | 509 pte_t orig_pte; /* Value of PTE at the time of fault */ member
|