/xen-4.10.0-shim-comet/xen/arch/x86/x86_64/ |
A D | traps.c | 196 l4t = map_domain_page(_mfn(mfn)); in show_page_walk() 200 pfn = mfn_valid(_mfn(mfn)) && machine_to_phys_mapping_valid ? in show_page_walk() 205 !mfn_valid(_mfn(mfn)) ) in show_page_walk() 208 l3t = map_domain_page(_mfn(mfn)); in show_page_walk() 212 pfn = mfn_valid(_mfn(mfn)) && machine_to_phys_mapping_valid ? in show_page_walk() 219 !mfn_valid(_mfn(mfn)) ) in show_page_walk() 222 l2t = map_domain_page(_mfn(mfn)); in show_page_walk() 226 pfn = mfn_valid(_mfn(mfn)) && machine_to_phys_mapping_valid ? in show_page_walk() 233 !mfn_valid(_mfn(mfn)) ) in show_page_walk() 236 l1t = map_domain_page(_mfn(mfn)); in show_page_walk() [all …]
|
A D | mm.c | 58 l4t = map_domain_page(_mfn(mfn)); in do_page_walk() 76 l2t = map_domain_page(_mfn(mfn)); in do_page_walk() 88 l1t = map_domain_page(_mfn(mfn)); in do_page_walk() 96 return map_domain_page(_mfn(mfn)) + (addr & ~PAGE_MASK); in do_page_walk() 367 if ( mfn_valid(_mfn(i + n * PDX_GROUP_COUNT)) ) in setup_compat_m2p_table() 437 if ( mfn_valid(_mfn(i + n * PDX_GROUP_COUNT)) ) in setup_m2p_table() 587 if ( mfn_valid(_mfn(MFN(i) + n * PDX_GROUP_COUNT)) ) in paging_init() 653 if ( mfn_valid(_mfn(MFN(i) + n * PDX_GROUP_COUNT)) ) in paging_init() 1194 pl4e = map_domain_page(_mfn(mfn)); in handle_memadd_fault() 1203 pl3e = map_domain_page(_mfn(mfn)); in handle_memadd_fault() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/pv/ |
A D | descriptor-tables.c | 32 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) 50 put_page_and_type(mfn_to_page(_mfn(pfn))); in pv_destroy_gdt() 99 put_page_and_type(mfn_to_page(_mfn(frames[i]))); in pv_set_gdt() 167 paging_mark_dirty(currd, _mfn(mfn)); in do_update_descriptor() 170 gdt_pent = map_domain_page(_mfn(mfn)); in do_update_descriptor()
|
A D | ro-page-fault.c | 40 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) 119 page = mfn_to_page(_mfn(mfn)); in ptwr_emulated_update() 123 ASSERT(mfn_valid(_mfn(mfn))); in ptwr_emulated_update() 163 pl1e = map_domain_page(_mfn(mfn)); in ptwr_emulated_update() 172 &t, l1e_get_intpte(nl1e), _mfn(mfn)); in ptwr_emulated_update()
|
A D | grant_table.c | 34 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) 81 gl1mfn = _mfn(addr >> PAGE_SHIFT); in create_grant_pv_mapping() 236 gl1mfn = _mfn(addr >> PAGE_SHIFT); in replace_grant_pv_mapping()
|
A D | mm.h | 48 rv = paging_write_guest_entry(v, p, new, _mfn(mfn)); in update_intpte() 62 rv = paging_cmpxchg_guest_entry(v, p, &t, _new, _mfn(mfn)); in update_intpte()
|
/xen-4.10.0-shim-comet/xen/include/xen/ |
A D | domain_page.h | 47 #define __map_domain_page(pg) map_domain_page(_mfn(__page_to_mfn(pg))) 51 return map_domain_page_global(_mfn(__page_to_mfn(pg))); in __map_domain_page_global()
|
A D | mm.h | 59 #define INVALID_MFN _mfn(~0UL) 68 #define _mfn macro 71 #undef _mfn 77 return _mfn(mfn_x(mfn) + i); in mfn_add() 82 return _mfn(max(mfn_x(x), mfn_x(y))); in mfn_max() 87 return _mfn(min(mfn_x(x), mfn_x(y))); in mfn_min()
|
/xen-4.10.0-shim-comet/xen/arch/x86/hvm/ |
A D | grant_table.c | 44 _mfn(frame), PAGE_ORDER_4K, p2mt); in create_grant_p2m_mapping() 71 if ( guest_physmap_remove_page(d, _gfn(gfn), _mfn(frame), PAGE_ORDER_4K) ) in replace_grant_p2m_mapping()
|
/xen-4.10.0-shim-comet/xen/arch/arm/ |
A D | mm.c | 49 #define virt_to_mfn(va) _mfn(__virt_to_mfn(va)) 241 mapping = map_domain_page(_mfn(pte.walk.base)); in dump_pt_walk() 372 pte = mfn_to_xen_entry(_mfn(base_mfn), MT_NORMAL); in create_mappings() 499 void *v = map_domain_page(_mfn(mfn)); in flush_page_to_ram() 810 xenheap_mfn_start = _mfn(base_mfn); in setup_xenheap_mappings() 811 xenheap_mfn_end = _mfn(base_mfn + nr_mfns); in setup_xenheap_mappings() 827 xenheap_mfn_start = _mfn(base_mfn); in setup_xenheap_mappings() 878 pte = mfn_to_xen_entry(_mfn(mfn), MT_NORMAL); in setup_xenheap_mappings() 947 mfn_t mfn = _mfn(PFN_DOWN(pa)); in ioremap_attr() 1291 mfn = _mfn(page_to_mfn(page)); in xenmem_add_to_physmap_one() [all …]
|
A D | p2m.c | 45 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) 275 mfn = _mfn(entry->p2m.base); in p2m_next_level() 359 mfn = _mfn(entry.p2m.base); in p2m_get_entry() 631 mfn_t mfn = _mfn(pte.p2m.base); in p2m_put_l3_page() 665 table = map_domain_page(_mfn(entry.p2m.base)); in p2m_free_entry() 680 mfn = _mfn(entry.p2m.base); in p2m_free_entry() 699 mfn_t mfn = _mfn(entry->p2m.base); in p2m_split_superpage()
|
/xen-4.10.0-shim-comet/xen/include/asm-arm/ |
A D | mm.h | 143 unsigned long _mfn = (mfn); \ 144 (_mfn >= mfn_x(xenheap_mfn_start) && \ 145 _mfn < mfn_x(xenheap_mfn_end)); \ 150 (mfn_valid(_mfn(mfn)) && is_xen_heap_page(__mfn_to_page(mfn))) 237 #define maddr_to_mfn(ma) _mfn(paddr_to_pfn(ma))
|
/xen-4.10.0-shim-comet/xen/common/ |
A D | tmem_xen.c | 71 return map_domain_page(_mfn(*pcli_mfn)); in cli_get_page() 80 paging_mark_dirty(current->domain, _mfn(cli_mfn)); in cli_put_page() 98 tmem_va = map_domain_page(_mfn(tmem_mfn)); in tmem_copy_from_client() 168 tmem_va = map_domain_page(_mfn(tmem_mfn)); in tmem_copy_to_client()
|
A D | page_alloc.c | 376 return _mfn(pg); in alloc_boot_pages() 383 return _mfn(pg); in alloc_boot_pages() 732 mfn_t mfn = _mfn(page_to_mfn(pg)); in poison_one_page() 747 mfn_t mfn = _mfn(page_to_mfn(pg)); in check_one_page() 1417 if ( !mfn_valid(_mfn(page_to_mfn(predecessor))) || in free_heap_pages() 1440 if ( !mfn_valid(_mfn(page_to_mfn(successor))) || in free_heap_pages() 1528 if ( !mfn_valid(_mfn(mfn)) ) in offline_page() 1637 if ( !mfn_valid(_mfn(mfn)) ) in online_page() 1688 if ( !mfn_valid(_mfn(mfn)) || !page_is_ram_type(mfn, RAM_TYPE_CONVENTIONAL) ) in query_page_offline() 1868 if ( !mfn_valid(_mfn(mfn)) || !page_state_is(pg, free) ) in smp_scrub_heap_pages() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/ |
A D | mm.c | 342 if ( !mfn_valid(_mfn(pfn)) ) in arch_init_memory() 905 if ( !mfn_valid(_mfn(mfn)) || in get_page_from_l1e() 911 if ( mfn_valid(_mfn(mfn)) ) in get_page_from_l1e() 1209 page = mfn_to_page(_mfn(pfn)); in put_page_from_l1e() 1450 pl2e = map_domain_page(_mfn(pfn)); in alloc_l2_table() 1493 pl3e = map_domain_page(_mfn(pfn)); in alloc_l3_table() 1756 pl2e = map_domain_page(_mfn(pfn)); in free_l2_table() 3652 map_mfn = _mfn(mfn); in do_mmu_update() 4122 mfn = _mfn(idx); in xenmem_add_to_physmap_one() 4140 if ( mfn_valid(_mfn(prev_mfn)) ) in xenmem_add_to_physmap_one() [all …]
|
A D | x86_emulate.c | 33 memset(((stb).ptr = map_domain_page(_mfn(this_cpu(stubs.mfn)))) + \
|
A D | tboot.c | 162 pt_vaddr = (struct dma_pte *)map_domain_page(_mfn(paddr_to_pfn(pt_maddr))); in update_iommu_mac() 189 if ( !mfn_valid(_mfn(mfn)) ) in update_pagetable_mac() 195 void *pg = map_domain_page(_mfn(mfn)); in update_pagetable_mac() 281 if ( !mfn_valid(_mfn(mfn)) ) in tboot_gen_xenheap_integrity()
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/ |
A D | p2m-ept.c | 73 if ( !mfn_valid(_mfn(new.mfn)) ) in atomic_write_ept_entry() 266 ept_entry_t *epte = map_domain_page(_mfn(ept_entry->mfn)); in ept_free_entry() 294 table = map_domain_page(_mfn(new_ept.mfn)); in ept_split_super_page() 382 *table = map_domain_page(_mfn(mfn)); in ept_next_level() 521 epte = map_domain_page(_mfn(mfn)); in resolve_misconfig() 623 ept_invalidate_emt(_mfn(e.mfn), e.recalc, level); in resolve_misconfig() 1003 mfn = _mfn(ept_entry->mfn); in ept_get_entry() 1013 mfn = _mfn(split_mfn); in ept_get_entry() 1063 next = map_domain_page(_mfn(ept_entry->mfn)); in ept_walk_table() 1084 if ( ept_invalidate_emt(_mfn(mfn), 1, p2m->ept.wl) ) in ept_change_entry_type_global() [all …]
|
A D | p2m-pt.c | 46 #define page_to_mfn(_pg) _mfn(__page_to_mfn(_pg)) 430 _mfn(mfn), level); in do_recalc() 565 _mfn(old_mfn)); in p2m_pt_set_entry() 613 _mfn(old_mfn)); in p2m_pt_set_entry() 657 _mfn(old_mfn)); in p2m_pt_set_entry() 817 mfn = _mfn(l3e_get_pfn(*l3e) + in p2m_pt_get_entry() 858 mfn = _mfn(l2e_get_pfn(*l2e) + l1_table_offset(addr)); in p2m_pt_get_entry() 1020 ASSERT(mfn_valid(_mfn(mfn))); in p2m_pt_audit_p2m() 1057 ASSERT(mfn_valid(_mfn(mfn))); in p2m_pt_audit_p2m() 1090 ASSERT(mfn_valid(_mfn(mfn))); in p2m_pt_audit_p2m()
|
/xen-4.10.0-shim-comet/xen/drivers/passthrough/vtd/x86/ |
A D | vtd.c | 43 return map_domain_page(_mfn(paddr_to_pfn(maddr))); in map_vtd_domain_page() 132 (!mfn_valid(_mfn(pfn)) || in vtd_set_hwdom_mapping()
|
/xen-4.10.0-shim-comet/xen/drivers/passthrough/amd/ |
A D | iommu_guest.c | 202 ASSERT(mfn_valid(_mfn(mfn))); in guest_iommu_add_ppr_log() 204 log_base = map_domain_page(_mfn(mfn)); in guest_iommu_add_ppr_log() 251 ASSERT(mfn_valid(_mfn(mfn))); in guest_iommu_add_event_log() 253 log_base = map_domain_page(_mfn(mfn)); in guest_iommu_add_event_log() 423 ASSERT(mfn_valid(_mfn(dte_mfn))); in do_invalidate_dte() 426 dte_base = map_domain_page(_mfn(dte_mfn)); in do_invalidate_dte() 444 ASSERT(mfn_valid(_mfn(gcr3_mfn))); in do_invalidate_dte() 505 ASSERT(mfn_valid(_mfn(cmd_mfn))); in guest_iommu_process_command() 507 cmd_base = map_domain_page(_mfn(cmd_mfn)); in guest_iommu_process_command()
|
A D | iommu_map.c | 42 table = map_domain_page(_mfn(l1_mfn)); in clear_iommu_pte_present() 115 table = map_domain_page(_mfn(pt_mfn)); in set_iommu_pte_present() 349 table = map_domain_page(_mfn(pt_mfn)); in iommu_update_pde_count() 354 ntable = map_domain_page(_mfn(paddr_to_pfn(ntable_maddr))); in iommu_update_pde_count() 400 table = map_domain_page(_mfn(pt_mfn)); in iommu_merge_pages() 412 ntable = map_domain_page(_mfn(ntable_mfn)); in iommu_merge_pages() 467 next_table_vaddr = map_domain_page(_mfn(next_table_mfn)); in iommu_pde_from_gfn()
|
/xen-4.10.0-shim-comet/xen/include/asm-x86/ |
A D | page.h | 75 #define l1e_get_mfn(x) _mfn(l1e_get_pfn(x)) 76 #define l2e_get_mfn(x) _mfn(l2e_get_pfn(x)) 77 #define l3e_get_mfn(x) _mfn(l3e_get_pfn(x)) 78 #define l4e_get_mfn(x) _mfn(l4e_get_pfn(x)) 220 #define pagetable_get_mfn(x) _mfn(((x).pfn)) 256 #define maddr_to_mfn(ma) _mfn(__paddr_to_pfn(ma))
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/hap/ |
A D | nested_hap.c | 75 #define page_to_mfn(_pg) _mfn(__page_to_mfn(_pg)) 117 mfn = _mfn((L0_gpa >> PAGE_SHIFT) & mask); in nestedhap_fix_p2m()
|
/xen-4.10.0-shim-comet/tools/libxc/ |
A D | xc_sr_common_x86_pv.c | 52 uint64_t mfn_to_cr3(struct xc_sr_context *ctx, xen_pfn_t _mfn) in mfn_to_cr3() argument 54 uint64_t mfn = _mfn; in mfn_to_cr3()
|