/xen-4.10.0-shim-comet/xen/arch/x86/pv/ |
A D | descriptor-tables.c | 31 #undef page_to_mfn 32 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro 79 frames[i] = mfn_x(page_to_mfn(page)); in pv_set_gdt() 152 mfn = mfn_x(page_to_mfn(page)); in do_update_descriptor()
|
A D | mm.c | 32 #undef page_to_mfn 33 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro
|
A D | domain.c | 17 #undef page_to_mfn 18 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro 37 mfn = page_to_mfn(pg); in setup_compat_l4()
|
A D | iret.c | 30 #undef page_to_mfn 31 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro
|
A D | traps.c | 35 #undef page_to_mfn 36 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro
|
A D | grant_table.c | 33 #undef page_to_mfn 34 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro
|
A D | ro-page-fault.c | 39 #undef page_to_mfn 40 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro
|
A D | callback.c | 37 #undef page_to_mfn 38 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro
|
A D | emul-gate-op.c | 47 #undef page_to_mfn 48 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro
|
/xen-4.10.0-shim-comet/xen/common/ |
A D | tmem_xen.c | 69 *pcli_mfn = page_to_mfn(page); in cli_get_page() 97 tmem_mfn = page_to_mfn(pfp); in tmem_copy_from_client() 167 tmem_mfn = page_to_mfn(pfp); in tmem_copy_to_client()
|
A D | page_alloc.c | 400 (flsl(page_to_mfn(pg)) ? : 1)) 732 mfn_t mfn = _mfn(page_to_mfn(pg)); in poison_one_page() 747 mfn_t mfn = _mfn(page_to_mfn(pg)); in check_one_page() 1347 unsigned long mask, mfn = page_to_mfn(pg); in free_heap_pages() 1412 if ( (page_to_mfn(pg) & mask) ) in free_heap_pages() 1417 if ( !mfn_valid(_mfn(page_to_mfn(predecessor))) || in free_heap_pages() 1440 if ( !mfn_valid(_mfn(page_to_mfn(successor))) || in free_heap_pages() 1738 unsigned long s = page_to_mfn(pg + i); in init_heap_pages() 1739 unsigned long e = page_to_mfn(pg + nr_pages - 1) + 1; in init_heap_pages() 1745 n = init_node_heap(nid, page_to_mfn(pg+i), nr_pages - i, in init_heap_pages() [all …]
|
A D | vmap.c | 39 map_pages_to_xen(va, page_to_mfn(pg), 1, PAGE_HYPERVISOR); in vm_init_type() 110 if ( !map_pages_to_xen(va, page_to_mfn(pg), 1, PAGE_HYPERVISOR) ) in vm_alloc() 261 mfn[i] = _mfn(page_to_mfn(pg)); in vmalloc_type()
|
A D | kimage.c | 29 #undef page_to_mfn 30 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro 85 clear_domain_page(page_to_mfn(page)); in kimage_alloc_zeroed_page() 414 clear_domain_page(page_to_mfn(page)); in kimage_alloc_crash_control_page() 922 copy_domain_page(page_to_mfn(xen_page), mfn); in kimage_build_ind()
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/hap/ |
A D | hap.c | 48 #undef page_to_mfn 49 #define page_to_mfn(_pg) _mfn(__page_to_mfn(_pg)) macro 261 clear_domain_page(page_to_mfn(pg)); in hap_alloc() 314 d->domain_id, mfn_x(page_to_mfn(pg)), in hap_free_p2m_page() 323 hap_free(d, page_to_mfn(pg)); in hap_free_p2m_page() 407 m4mfn = page_to_mfn(pg); in hap_make_monitor_table()
|
A D | nested_hap.c | 74 #undef page_to_mfn 75 #define page_to_mfn(_pg) _mfn(__page_to_mfn(_pg)) macro
|
A D | guest_walk.c | 86 top_mfn = _mfn(page_to_mfn(top_page)); in hap_p2m_ga_to_gfn()
|
/xen-4.10.0-shim-comet/xen/arch/x86/ |
A D | mm.c | 135 #undef page_to_mfn 136 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro 1445 unsigned long pfn = mfn_x(page_to_mfn(page)); in alloc_l2_table() 1751 unsigned long pfn = mfn_x(page_to_mfn(page)); in free_l2_table() 1780 unsigned long pfn = mfn_x(page_to_mfn(page)); in free_l3_table() 1816 unsigned long pfn = mfn_x(page_to_mfn(page)); in free_l4_table() 2342 mfn_x(page_to_mfn(page)), in alloc_page_type() 2408 type, mfn_x(page_to_mfn(page))); in free_page_type() 2550 mfn_x(page_to_mfn(page))); in __get_page_type() 3447 copy_domain_page(page_to_mfn(dst_page), page_to_mfn(src_page)); in do_mmuext_op() [all …]
|
A D | smpboot.c | 53 #undef page_to_mfn 54 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) macro 623 if ( map_pages_to_xen(stub_va, mfn_x(page_to_mfn(pg)), 1, in alloc_stub_page() 631 *mfn = mfn_x(page_to_mfn(pg)); in alloc_stub_page()
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/ |
A D | p2m-pod.c | 35 #undef page_to_mfn 36 #define page_to_mfn(_pg) _mfn(__page_to_mfn(_pg)) macro 70 mfn = page_to_mfn(page); in p2m_pod_cache_add() 104 clear_domain_page(mfn_add(page_to_mfn(page), i)); in p2m_pod_cache_add() 164 mfn = mfn_x(page_to_mfn(p)); in p2m_pod_cache_get() 436 bmfn = mfn_x(page_to_mfn(p)); in p2m_pod_offline_or_broken_hit() 439 mfn = mfn_x(page_to_mfn(q)); in p2m_pod_offline_or_broken_hit() 457 mfn = mfn_x(page_to_mfn(q)); in p2m_pod_offline_or_broken_hit() 1210 mfn = page_to_mfn(p); in p2m_pod_demand_populate()
|
A D | mem_sharing.c | 157 #undef page_to_mfn 158 #define page_to_mfn(_pg) _mfn(__page_to_mfn(_pg)) macro 443 mfn = page_to_mfn(pg); in audit() 723 mfn_x(page_to_mfn(page)), in debug_mfn() 1187 copy_domain_page(page_to_mfn(page), page_to_mfn(old_page)); in __mem_sharing_unshare_page() 1189 BUG_ON(set_shared_p2m_entry(d, gfn, page_to_mfn(page))); in __mem_sharing_unshare_page() 1203 set_gpfn_from_mfn(mfn_x(page_to_mfn(page)), gfn); in __mem_sharing_unshare_page() 1207 paging_mark_dirty(d, page_to_mfn(page)); in __mem_sharing_unshare_page()
|
/xen-4.10.0-shim-comet/xen/include/asm-arm/ |
A D | mm.h | 141 #define is_xen_heap_page(page) is_xen_heap_mfn(page_to_mfn(page)) 297 #define page_to_mfn(pg) __page_to_mfn(pg) macro 317 return mfn_to_virt(page_to_mfn(pg)); in page_to_virt()
|
/xen-4.10.0-shim-comet/xen/drivers/passthrough/x86/ |
A D | iommu.c | 61 unsigned long mfn = page_to_mfn(page); in arch_iommu_populate_page_table()
|
/xen-4.10.0-shim-comet/xen/drivers/passthrough/amd/ |
A D | iommu_map.c | 454 next_table_mfn = page_to_mfn(table); in iommu_pde_from_gfn() 496 next_table_mfn = page_to_mfn(table); in iommu_pde_from_gfn() 523 next_table_mfn = page_to_mfn(table); in iommu_pde_from_gfn() 580 old_root_mfn = page_to_mfn(old_root); in update_paging_mode()
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/shadow/ |
A D | common.c | 1209 smfn = page_to_mfn(sp); in _shadow_prealloc() 1278 smfn = page_to_mfn(sp); in shadow_blow_tables() 1402 clear_domain_page(page_to_mfn(sp)); in shadow_alloc() 1419 return page_to_mfn(sp); in shadow_alloc() 1450 == mfn_x(page_to_mfn(sp)) ) in shadow_free() 1522 d->domain_id, mfn_x(page_to_mfn(pg)), in shadow_free_p2m_page() 1534 shadow_free(d, page_to_mfn(pg)); in shadow_free_p2m_page() 1672 mfn = page_to_mfn(page); in emulate_gva_to_mfn() 1911 mfn_x(page_to_mfn(sp)), in sh_hash_audit_bucket() 2014 return page_to_mfn(sp); in shadow_hash_lookup() [all …]
|
A D | private.h | 471 #undef page_to_mfn 472 #define page_to_mfn(_pg) _mfn(__page_to_mfn(_pg)) macro 478 #define pagetable_from_page(pg) pagetable_from_mfn(page_to_mfn(pg))
|