Lines Matching refs:mfn

49     unsigned long mfn = pagetable_get_pfn(v->arch.guest_table);  in do_page_walk()  local
58 l4t = map_domain_page(_mfn(mfn)); in do_page_walk()
67 mfn = l3e_get_pfn(l3e); in do_page_walk()
68 if ( !(l3e_get_flags(l3e) & _PAGE_PRESENT) || !mfn_valid(_mfn(mfn)) ) in do_page_walk()
72 mfn += PFN_DOWN(addr & ((1UL << L3_PAGETABLE_SHIFT) - 1)); in do_page_walk()
76 l2t = map_domain_page(_mfn(mfn)); in do_page_walk()
79 mfn = l2e_get_pfn(l2e); in do_page_walk()
80 if ( !(l2e_get_flags(l2e) & _PAGE_PRESENT) || !mfn_valid(_mfn(mfn)) ) in do_page_walk()
84 mfn += PFN_DOWN(addr & ((1UL << L2_PAGETABLE_SHIFT) - 1)); in do_page_walk()
88 l1t = map_domain_page(_mfn(mfn)); in do_page_walk()
91 mfn = l1e_get_pfn(l1e); in do_page_walk()
92 if ( !(l1e_get_flags(l1e) & _PAGE_PRESENT) || !mfn_valid(_mfn(mfn)) ) in do_page_walk()
96 return map_domain_page(_mfn(mfn)) + (addr & ~PAGE_MASK); in do_page_walk()
116 unsigned mfn; in alloc_hotadd_mfn() local
121 mfn = info->cur; in alloc_hotadd_mfn()
123 return mfn; in alloc_hotadd_mfn()
320 unsigned long i, va, smap, emap, rwva, epfn = info->epfn, mfn; in setup_compat_m2p_table() local
372 mfn = alloc_hotadd_mfn(info); in setup_compat_m2p_table()
373 err = map_pages_to_xen(rwva, mfn, 1UL << PAGETABLE_ORDER, in setup_compat_m2p_table()
381 l2e_from_pfn(mfn, _PAGE_PSE|_PAGE_PRESENT)); in setup_compat_m2p_table()
441 unsigned long mfn = alloc_hotadd_mfn(info); in setup_m2p_table() local
445 mfn, 1UL << PAGETABLE_ORDER, in setup_m2p_table()
476 l2e_write(l2_ro_mpt, l2e_from_pfn(mfn, in setup_m2p_table()
773 unsigned long mfn; in setup_frametable_chunk() local
781 mfn = alloc_hotadd_mfn(info); in setup_frametable_chunk()
782 err = map_pages_to_xen(s, mfn, 1UL << PAGETABLE_ORDER, in setup_frametable_chunk()
913 xen_pfn_t mfn, last_mfn; in subarch_memory_op() local
933 mfn = last_mfn; in subarch_memory_op()
938 mfn = l2e_get_pfn(l2e); in subarch_memory_op()
940 mfn = last_mfn; in subarch_memory_op()
944 mfn = l3e_get_pfn(l3e) in subarch_memory_op()
947 ASSERT(mfn); in subarch_memory_op()
948 if ( copy_to_guest_offset(xmml.extent_start, i, &mfn, 1) ) in subarch_memory_op()
950 last_mfn = mfn; in subarch_memory_op()
972 mfn = l2e_get_pfn(l2e); in subarch_memory_op()
974 mfn = last_mfn; in subarch_memory_op()
975 ASSERT(mfn); in subarch_memory_op()
976 if ( copy_to_guest_offset(xmml.extent_start, i, &mfn, 1) ) in subarch_memory_op()
978 last_mfn = mfn; in subarch_memory_op()
1182 unsigned long mfn, idle_index; in handle_memadd_fault() local
1192 mfn = (read_cr3()) >> PAGE_SHIFT; in handle_memadd_fault()
1194 pl4e = map_domain_page(_mfn(mfn)); in handle_memadd_fault()
1201 mfn = l4e_get_pfn(l4e); in handle_memadd_fault()
1203 pl3e = map_domain_page(_mfn(mfn)); in handle_memadd_fault()
1210 mfn = l3e_get_pfn(l3e); in handle_memadd_fault()
1211 pl2e = map_domain_page(_mfn(mfn)); in handle_memadd_fault()