/xen-4.10.0-shim-comet/xen/arch/x86/ |
A D | domain_page.c | 80 if ( mfn_x(mfn) <= PFN_DOWN(__pa(HYPERVISOR_VIRT_END - 1)) ) in map_domain_page() 81 return mfn_to_virt(mfn_x(mfn)); in map_domain_page() 86 return mfn_to_virt(mfn_x(mfn)); in map_domain_page() 91 return mfn_to_virt(mfn_x(mfn)); in map_domain_page() 97 hashent = &vcache->hash[MAPHASH_HASHFN(mfn_x(mfn))]; in map_domain_page() 98 if ( hashent->mfn == mfn_x(mfn) ) in map_domain_page() 104 ASSERT(l1e_get_pfn(MAPCACHE_L1ENT(idx)) == mfn_x(mfn)); in map_domain_page() 139 i = MAPHASH_HASHFN(mfn_x(mfn)); in map_domain_page() 153 } while ( i != MAPHASH_HASHFN(mfn_x(mfn)) ); in map_domain_page() 314 if ( mfn_x(mfn) <= PFN_DOWN(__pa(HYPERVISOR_VIRT_END - 1)) ) in map_domain_page_global() [all …]
|
A D | debug.c | 69 vaddr, dp->domain_id, mfn_x(mfn)); in dbg_hvm_va2mfn() 113 l4_table_offset(vaddr), l4e, mfn_x(mfn)); in dbg_pv_va2mfn() 125 l3_table_offset(vaddr), l3e, mfn_x(mfn)); in dbg_pv_va2mfn() 139 l2t, l2_table_offset(vaddr), l2e, mfn_x(mfn)); in dbg_pv_va2mfn() 151 l1e, mfn_x(mfn)); in dbg_pv_va2mfn()
|
A D | mm.c | 134 #define mfn_to_page(mfn) __mfn_to_page(mfn_x(mfn)) 506 v->arch.cr3 = mfn_x(mfn) << PAGE_SHIFT; in make_cr3() 1751 unsigned long pfn = mfn_x(page_to_mfn(page)); in free_l2_table() 1780 unsigned long pfn = mfn_x(page_to_mfn(page)); in free_l3_table() 2342 mfn_x(page_to_mfn(page)), in alloc_page_type() 2550 mfn_x(page_to_mfn(page))); in __get_page_type() 2822 mfn_x(mfn)); in new_guest_cr3() 2860 mfn_x(mfn)); in new_guest_cr3() 2873 if ( likely(mfn_x(old_base_mfn) != 0) ) in new_guest_cr3() 3645 mfn = mfn_x(page_to_mfn(page)); in do_mmu_update() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/shadow/ |
A D | multi.c | 133 d->domain_id, mfn_x(gmfn), shadow_type, mfn_x(smfn)); in set_shadow_status() 162 d->domain_id, mfn_x(gmfn), shadow_type, mfn_x(smfn)); in delete_shadow_status() 1461 mfn_x(gmfn), shadow_type, mfn_x(smfn)); in sh_make_shadow() 2449 if ( mfn_x(smfn2) != mfn_x(map_mfn) ) in sh_map_and_validate() 4317 || (mfn_x(shadow_l1e_get_mfn(sl1e)) != mfn_x(gmfn)) ) in sh_rm_write_access_from_sl1p() 4386 || (mfn_x(shadow_l1e_get_mfn(sl1e)) != mfn_x(gmfn)) ) in sh_guess_wrmap() 4979 if ( mfn_x(gmfn) != mfn_x(mfn) ) in sh_audit_l2_table() 5034 if ( mfn_x(gmfn) != mfn_x(mfn) ) in sh_audit_l3_table() 5037 gfn_x(gfn), mfn_x(gmfn), mfn_x(mfn)); in sh_audit_l3_table() 5079 if ( mfn_x(gmfn) != mfn_x(mfn) ) in sh_audit_l4_table() [all …]
|
A D | common.c | 496 if ( mfn_x(oos[idx]) != mfn_x(gmfn) ) in oos_audit_hash_is_present() 499 if ( mfn_x(oos[idx]) == mfn_x(gmfn) ) in oos_audit_hash_is_present() 567 if ( mfn_x(oos[idx]) != mfn_x(gmfn) ) in oos_fixup_add() 569 if ( mfn_x(oos[idx]) == mfn_x(gmfn) ) in oos_fixup_add() 575 && (mfn_x(oos_fixup[idx].smfn[i]) == mfn_x(smfn)) in oos_fixup_add() 748 if ( mfn_x(oos[idx]) != mfn_x(gmfn) ) in oos_hash_remove() 750 if ( mfn_x(oos[idx]) == mfn_x(gmfn) ) in oos_hash_remove() 773 if ( mfn_x(oos[idx]) != mfn_x(gmfn) ) in oos_snapshot_lookup() 775 if ( mfn_x(oos[idx]) == mfn_x(gmfn) ) in oos_snapshot_lookup() 800 if ( mfn_x(oos[idx]) != mfn_x(gmfn) ) in sh_resync() [all …]
|
A D | private.h | 318 return page_is_out_of_sync(mfn_to_page(mfn_x(gmfn))); in mfn_is_out_of_sync() 329 return page_oos_may_write(mfn_to_page(mfn_x(gmfn))); in mfn_oos_may_write() 470 #define mfn_to_page(_m) __mfn_to_page(mfn_x(_m)) 535 __backpointer(sp), mfn_x(smfn)); in sh_get_ref() 575 mfn_x(smfn), sp->u.sh.count, sp->u.sh.type); in sh_put_ref()
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/ |
A D | p2m.c | 52 #define mfn_to_page(_m) __mfn_to_page(mfn_x(_m)) 922 gfn_x(gfn), mfn_x(mfn)); in guest_physmap_add_entry() 1121 gfn_l, order, rc, mfn_x(mfn)); in set_typed_p2m_entry() 1231 if ( mfn_x(mfn) != mfn_x(actual_mfn) ) in clear_mmio_p2m_entry() 1234 mfn_x(mfn), gfn_l); in clear_mmio_p2m_entry() 1657 set_gpfn_from_mfn(mfn_x(mfn), gfn_l); in p2m_mem_paging_prep() 2233 mfn = _mfn(mfn_x(mfn) & mask); in p2m_altp2m_lazy_copy() 2423 mfn = _mfn(mfn_x(mfn) & mask); in p2m_change_altp2m_gfn() 2564 mfn = mfn_x(page_to_mfn(page)); in audit_p2m() 2594 if ( mfn_x(p2mfn) != mfn ) in audit_p2m() [all …]
|
A D | mem_sharing.c | 156 #define mfn_to_page(_m) __mfn_to_page(mfn_x(_m)) 475 mfn_x(mfn), get_gpfn_from_mfn(mfn_x(mfn))); in audit() 483 mfn_x(mfn)); in audit() 508 if ( mfn_x(o_mfn) != mfn_x(mfn) ) in audit() 512 g->domain, g->gfn, mfn_x(mfn), mfn_x(o_mfn)); in audit() 530 mfn_x(mfn), nr_gfns, in audit() 701 if ( mem_sharing_lookup(mfn_x(mfn)) == NULL ) in __grab_shared_page() 723 mfn_x(page_to_mfn(page)), in debug_mfn() 882 set_gpfn_from_mfn(mfn_x(mfn), SHARED_M2P_ENTRY); in nominate_page() 912 if( mfn_x(smfn) == mfn_x(cmfn) ) in share_pages() [all …]
|
A D | p2m-pod.c | 34 #define mfn_to_page(_m) __mfn_to_page(mfn_x(_m)) 73 if ( mfn_x(mfn) & ((1UL << order) - 1) ) in p2m_pod_cache_add() 84 p = mfn_to_page(_mfn(mfn_x(mfn) + i)); in p2m_pod_cache_add() 164 mfn = mfn_x(page_to_mfn(p)); in p2m_pod_cache_get() 436 bmfn = mfn_x(page_to_mfn(p)); in p2m_pod_offline_or_broken_hit() 439 mfn = mfn_x(page_to_mfn(q)); in p2m_pod_offline_or_broken_hit() 457 mfn = mfn_x(page_to_mfn(q)); in p2m_pod_offline_or_broken_hit() 762 if ( !superpage_aligned(mfn_x(mfn)) ) in p2m_pod_zero_check_superpage() 843 t.mfn = mfn_x(mfn); in p2m_pod_zero_check_superpage() 1010 t.mfn = mfn_x(mfns[i]); in p2m_pod_zero_check() [all …]
|
A D | p2m-pt.c | 44 #define mfn_to_page(_m) __mfn_to_page(mfn_x(_m)) 114 if ( !rangeset_contains_singleton(mmio_ro_ranges, mfn_x(mfn)) ) in p2m_type_to_flags() 508 unsigned long old_mfn = mfn_x(INVALID_MFN); in p2m_pt_set_entry() 521 t.mfn = mfn_x(mfn); in p2m_pt_set_entry() 577 ? p2m_l3e_from_pfn(mfn_x(mfn), in p2m_pt_set_entry() 616 entry_content = p2m_l1e_from_pfn(mfn_x(mfn), in p2m_pt_set_entry() 669 ? p2m_l2e_from_pfn(mfn_x(mfn), in p2m_pt_set_entry() 687 (iommu_old_flags != iommu_pte_flags || old_mfn != mfn_x(mfn)) ) in p2m_pt_set_entry() 699 rc = iommu_map_page(p2m->domain, gfn + i, mfn_x(mfn) + i, in p2m_pt_set_entry()
|
A D | guest_walk.c | 472 put_page(mfn_to_page(mfn_x(gw->l3mfn))); in guest_walk_tables() 479 put_page(mfn_to_page(mfn_x(gw->l2mfn))); in guest_walk_tables() 485 put_page(mfn_to_page(mfn_x(gw->l1mfn))); in guest_walk_tables()
|
A D | p2m-ept.c | 52 unsigned long oldmfn = mfn_x(INVALID_MFN); in atomic_write_ept_entry() 93 if ( unlikely(oldmfn != mfn_x(INVALID_MFN)) ) in atomic_write_ept_entry() 237 ept_entry->mfn = mfn_x(mfn); in ept_set_middle_entry() 685 unsigned long fn_mask = !mfn_eq(mfn, INVALID_MFN) ? (gfn | mfn_x(mfn)) : gfn; in ept_set_entry() 814 new_entry.mfn = mfn_x(mfn); in ept_set_entry() 876 rc = iommu_map_page(d, gfn + i, mfn_x(mfn) + i, iommu_flags); in ept_set_entry() 1010 unsigned long split_mfn = mfn_x(mfn) + in ept_get_entry() 1393 if ( d->arch.altp2m_eptp[i] == mfn_x(INVALID_MFN) ) in p2m_find_altp2m_by_eptp()
|
/xen-4.10.0-shim-comet/xen/include/xen/ |
A D | domain_page.h | 56 #define map_domain_page(mfn) __mfn_to_virt(mfn_x(mfn)) 63 return mfn_to_virt(mfn_x(mfn)); in map_domain_page_global()
|
A D | mm.h | 69 #define mfn_x macro 72 #undef mfn_x 77 return _mfn(mfn_x(mfn) + i); in mfn_add() 82 return _mfn(max(mfn_x(x), mfn_x(y))); in mfn_max() 87 return _mfn(min(mfn_x(x), mfn_x(y))); in mfn_min() 92 return mfn_x(x) == mfn_x(y); in mfn_eq()
|
/xen-4.10.0-shim-comet/xen/arch/x86/hvm/ |
A D | grant_table.c | 63 if ( !p2m_is_grant(type) || mfn_x(old_mfn) != frame ) in replace_grant_p2m_mapping() 68 type, mfn_x(old_mfn), frame); in replace_grant_p2m_mapping()
|
A D | mtrr.c | 772 if ( rangeset_overlaps_range(mmio_ro_ranges, mfn_x(mfn), in epte_get_entry_emt() 773 mfn_x(mfn) | ((1UL << order) - 1)) ) in epte_get_entry_emt() 775 if ( !order || rangeset_contains_range(mmio_ro_ranges, mfn_x(mfn), in epte_get_entry_emt() 776 mfn_x(mfn) | ((1UL << order) - 1)) ) in epte_get_entry_emt() 787 if ( (mfn_x(mfn) ^ d->arch.hvm_domain.vmx.apic_access_mfn) >> order ) in epte_get_entry_emt() 821 hmtrr_mtype = get_mtrr_type(&mtrr_state, mfn_x(mfn) << PAGE_SHIFT, order); in epte_get_entry_emt()
|
/xen-4.10.0-shim-comet/xen/arch/x86/pv/ |
A D | grant_table.c | 32 #define mfn_to_page(mfn) __mfn_to_page(mfn_x(mfn)) 116 if ( UPDATE_ENTRY(l1, pl1e, ol1e, nl1e, mfn_x(gl1mfn), curr, 0) ) in create_grant_pv_mapping() 173 okay = UPDATE_ENTRY(l1, pl1e, ol1e, l1e_empty(), mfn_x(gl1mfn), curr, 0); in steal_linear_address() 301 if ( UPDATE_ENTRY(l1, pl1e, ol1e, nl1e, mfn_x(gl1mfn), curr, 0) ) in replace_grant_pv_mapping()
|
A D | descriptor-tables.c | 30 #define mfn_to_page(mfn) __mfn_to_page(mfn_x(mfn)) 79 frames[i] = mfn_x(page_to_mfn(page)); in pv_set_gdt() 152 mfn = mfn_x(page_to_mfn(page)); in do_update_descriptor()
|
/xen-4.10.0-shim-comet/xen/include/asm-arm/ |
A D | mm.h | 144 (_mfn >= mfn_x(xenheap_mfn_start) && \ 145 _mfn < mfn_x(xenheap_mfn_end)); \ 218 unsigned long __m_f_n = mfn_x(mfn); \ 236 #define mfn_to_maddr(mfn) pfn_to_paddr(mfn_x(mfn)) 311 pdx += pfn_to_pdx(mfn_x(xenheap_mfn_start)); in virt_to_page()
|
/xen-4.10.0-shim-comet/xen/arch/x86/guest/ |
A D | xen.c | 104 xatp.gpfn = mfn_x(mfn); in map_shared_info() 109 set_fixmap(FIX_XEN_SHARED_INFO, mfn_x(mfn) << PAGE_SHIFT); in map_shared_info() 303 return rangeset_remove_range(mem, mfn_x(mfn), mfn_x(mfn)); in hypervisor_free_unused_page()
|
/xen-4.10.0-shim-comet/xen/include/asm-x86/ |
A D | page.h | 124 #define l1e_from_mfn(m, f) l1e_from_pfn(mfn_x(m), f) 125 #define l2e_from_mfn(m, f) l2e_from_pfn(mfn_x(m), f) 126 #define l3e_from_mfn(m, f) l3e_from_pfn(mfn_x(m), f) 127 #define l4e_from_mfn(m, f) l4e_from_pfn(mfn_x(m), f) 223 #define pagetable_from_mfn(mfn) ((pagetable_t) { mfn_x(mfn) }) 255 #define mfn_to_maddr(mfn) __pfn_to_paddr(mfn_x(mfn)) 262 #define mfn_valid(mfn) __mfn_valid(mfn_x(mfn))
|
A D | guest_pt.h | 437 gw->l4e.l4, mfn_x(gw->l4mfn)); in print_gw() 439 gw->l3e.l3, mfn_x(gw->l3mfn)); in print_gw() 445 gw->l2e.l2, mfn_x(gw->l2mfn)); in print_gw() 448 gw->el1e, mfn_x(gw->l1mfn)); in print_gw() 451 gw->l1e.l1, mfn_x(gw->l1mfn)); in print_gw()
|
/xen-4.10.0-shim-comet/xen/arch/arm/ |
A D | p2m.c | 43 #define mfn_to_page(mfn) __mfn_to_page(mfn_x(mfn)) 529 e.p2m.base = mfn_x(mfn); in mfn_to_p2m_entry() 730 pte.p2m.base = mfn_x(mfn_add(mfn, i << level_order)); in p2m_split_superpage() 997 mask = !mfn_eq(smfn, INVALID_MFN) ? mfn_x(smfn) : 0; in p2m_set_entry() 1096 if ( !(nr && iomem_access_permitted(d, mfn_x(mfn), mfn_x(mfn) + nr - 1)) ) in map_dev_mmio_region() 1103 mfn_x(mfn), mfn_x(mfn) + nr - 1, d->domain_id); in map_dev_mmio_region() 1391 flush_page_to_ram(mfn_x(mfn), false); in p2m_cache_flush() 1441 v, mfn_x(maddr_to_mfn(maddr))); in get_page_from_gva() 1451 v, mfn_x(maddr_to_mfn(maddr))); in get_page_from_gva()
|
A D | mm.c | 51 #define mfn_to_virt(mfn) __mfn_to_virt(mfn_x(mfn)) 326 e.pt.base = mfn_x(mfn); in mfn_to_xen_entry() 399 unsigned long slot_mfn = mfn_x(mfn) & ~LPAE_ENTRY_MASK; in map_domain_page() 452 + ((mfn_x(mfn) & LPAE_ENTRY_MASK) << THIRD_SHIFT)); in map_domain_page() 832 if ( base_mfn < mfn_x(xenheap_mfn_start) ) in setup_xenheap_mappings() 834 base_mfn, mfn_x(xenheap_mfn_start)); in setup_xenheap_mappings() 922 create_mappings(second, 0, mfn_x(base_mfn), frametable_size >> PAGE_SHIFT, in setup_frametable_mappings() 925 create_mappings(xen_second, FRAMETABLE_VIRT_START, mfn_x(base_mfn), in setup_frametable_mappings() 1019 __func__, addr, mfn_x(mfn)); in create_xen_entries()
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/hap/ |
A D | hap.c | 47 #define mfn_to_page(_m) __mfn_to_page(mfn_x(_m)) 314 d->domain_id, mfn_x(page_to_mfn(pg)), in hap_free_p2m_page() 497 d->arch.altp2m_eptp[i] = mfn_x(INVALID_MFN); in hap_enable() 569 if ( mfn_valid(mfn) && (mfn_x(mfn) != 0) ) in hap_teardown() 738 flush_nestedp2m = !( mfn_x(omfn) == mfn_x(nmfn) in hap_write_p2m_entry()
|