/xen-4.10.0-shim-comet/xen/common/ |
A D | page_alloc.c | 151 #define p2m_pod_offline_or_broken_replace(pg) BUG_ON(pg != NULL) argument 359 if ( pg >= r->e || pg < r->s ) in alloc_boot_pages() 372 pg = r->s; in alloc_boot_pages() 381 r->e = pg; in alloc_boot_pages() 399 #define page_to_zone(pg) (is_xen_heap_page(pg) ? MEMZONE_XEN : \ argument 923 return pg; in alloc_heap_pages() 931 if ( !pg ) in alloc_heap_pages() 1031 return pg; in alloc_heap_pages() 1302 st.pg = pg; in scrub_free_pages() 1378 pg[i].u.free.need_tlbflush = (page_get_owner(&pg[i]) != NULL); in free_heap_pages() [all …]
|
A D | vmap.c | 37 struct page_info *pg = alloc_domheap_page(NULL, 0); in vm_init_type() local 65 struct page_info *pg; in vm_alloc() local 100 pg = alloc_domheap_page(NULL, 0); in vm_alloc() 101 if ( !pg ) in vm_alloc() 120 free_domheap_page(pg); in vm_alloc() 246 struct page_info *pg; in vmalloc_type() local 258 pg = alloc_domheap_page(NULL, 0); in vmalloc_type() 259 if ( pg == NULL ) in vmalloc_type() 261 mfn[i] = _mfn(page_to_mfn(pg)); in vmalloc_type() 305 struct page_info *pg; in vfree() local [all …]
|
A D | grant_table.c | 1000 if ( !pg ) in map_grant_ref() 1003 if ( pg ) in map_grant_ref() 1015 if ( pg ) in map_grant_ref() 1017 put_page(pg); in map_grant_ref() 1161 put_page_type(pg); in map_grant_ref() 1164 put_page(pg); in map_grant_ref() 1404 struct page_info *pg; in unmap_common_complete() local 1435 put_page(pg); in unmap_common_complete() 1454 put_page(pg); in unmap_common_complete() 1650 put_page(pg); in gnttab_unpopulate_status_frames() [all …]
|
/xen-4.10.0-shim-comet/xen/include/xen/ |
A D | domain_page.h | 47 #define __map_domain_page(pg) map_domain_page(_mfn(__page_to_mfn(pg))) argument 49 static inline void *__map_domain_page_global(const struct page_info *pg) in __map_domain_page_global() argument 51 return map_domain_page_global(_mfn(__page_to_mfn(pg))); in __map_domain_page_global() 57 #define __map_domain_page(pg) page_to_virt(pg) argument 66 static inline void *__map_domain_page_global(const struct page_info *pg) in __map_domain_page_global() argument 68 return page_to_virt(pg); in __map_domain_page_global()
|
A D | pdx.h | 21 #define page_to_pdx(pg) ((pg) - frame_table) argument
|
A D | mm.h | 189 void free_domheap_pages(struct page_info *pg, unsigned int order); 205 struct page_info *pg, 520 struct page_info *pg; in page_list_remove_head() local 525 pg = page_list_first(head); in page_list_remove_head() 526 list_del(&pg->list); in page_list_remove_head() 527 return pg; in page_list_remove_head() 573 #define arch_free_heap_page(d, pg) \ argument 574 page_list_del(pg, is_xen_heap_page(pg) ? \
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/hap/ |
A D | hap.c | 251 struct page_info *pg; in hap_alloc() local 256 if ( unlikely(!pg) ) in hap_alloc() 263 return pg; in hap_alloc() 278 struct page_info *pg; in hap_alloc_p2m_page() local 283 pg = hap_alloc(d); in hap_alloc_p2m_page() 289 ASSERT(!page_get_owner(pg) && !(pg->count_info & PGC_count_mask)); in hap_alloc_p2m_page() 299 return pg; in hap_alloc_p2m_page() 316 pg->count_info, pg->u.inuse.type_info); in hap_free_p2m_page() 343 struct page_info *pg; in hap_set_allocation() local 376 ASSERT(pg); in hap_set_allocation() [all …]
|
/xen-4.10.0-shim-comet/stubdom/vtpmmgr/ |
A D | vtpm_disk.c | 150 struct mem_vtpm_page *pg = find_mem_vtpm_page(group, vtpm); in vtpm_sync() local 151 if (!pg) in vtpm_sync() 153 pg->disk_loc.value = 0; in vtpm_sync() 191 struct mem_vtpm_page *pg = find_mem_vtpm_page(group, vtpm); in delete_vtpm() local 200 pg->vtpms[vtidx] = last; in delete_vtpm() 201 pg->disk_loc.value = 0; in delete_vtpm() 224 struct mem_vtpm_page *pg = &group->data[j]; in find_vtpm() local 225 for (k = 0; k < pg->size; k++) { in find_vtpm() 226 struct mem_vtpm *vt = pg->vtpms[k]; in find_vtpm()
|
/xen-4.10.0-shim-comet/xen/include/asm-x86/ |
A D | mm.h | 259 #define page_state_is(pg, st) (((pg)->count_info&PGC_state) == PGC_state_##st) argument 326 static inline void *__page_to_virt(const struct page_info *pg) in __page_to_virt() argument 328 ASSERT((unsigned long)pg - FRAMETABLE_VIRT_START < FRAMETABLE_SIZE); in __page_to_virt() 336 ((unsigned long)pg - FRAMETABLE_VIRT_START) / in __page_to_virt() 337 (sizeof(*pg) / (sizeof(*pg) & -sizeof(*pg))) * in __page_to_virt() 338 (PAGE_SIZE / (sizeof(*pg) & -sizeof(*pg)))); in __page_to_virt() 612 #define arch_free_heap_page(d, pg) \ argument 613 page_list_del2(pg, is_xen_heap_page(pg) ? \
|
A D | page.h | 224 #define pagetable_from_page(pg) pagetable_from_pfn(__page_to_mfn(pg)) argument 244 #define __page_to_mfn(pg) pdx_to_pfn((unsigned long)((pg) - frame_table)) argument 248 #define __page_to_maddr(pg) ((paddr_t)__page_to_mfn(pg) << PAGE_SHIFT) argument 268 #define page_to_mfn(pg) __page_to_mfn(pg) argument 270 #define page_to_maddr(pg) __page_to_maddr(pg) argument 272 #define page_to_virt(pg) __page_to_virt(pg) argument
|
/xen-4.10.0-shim-comet/xen/include/asm-arm/ |
A D | mm.h | 121 #define page_state_is(pg, st) (((pg)->count_info&PGC_state) == PGC_state_##st) argument 224 #define __page_to_mfn(pg) pdx_to_pfn((unsigned long)((pg) - frame_table) + frametable_base_pdx) argument 228 #define page_to_maddr(pg) ((paddr_t)__page_to_mfn(pg) << PAGE_SHIFT) argument 297 #define page_to_mfn(pg) __page_to_mfn(pg) argument 315 static inline void *page_to_virt(const struct page_info *pg) in page_to_virt() argument 317 return mfn_to_virt(page_to_mfn(pg)); in page_to_virt()
|
/xen-4.10.0-shim-comet/xen/arch/x86/pv/ |
A D | domain.c | 18 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) argument 29 struct page_info *pg; in setup_compat_l4() local 33 pg = alloc_domheap_page(v->domain, MEMF_no_owner); in setup_compat_l4() 34 if ( pg == NULL ) in setup_compat_l4() 37 mfn = page_to_mfn(pg); in setup_compat_l4() 44 pg->u.inuse.type_info = PGT_l4_page_table | PGT_validated | 1; in setup_compat_l4() 46 v->arch.guest_table = pagetable_from_page(pg); in setup_compat_l4()
|
A D | mm.c | 33 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) argument
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/ |
A D | mem_sharing.c | 121 rc = page_lock(pg); in mem_sharing_page_lock() 138 page_unlock(pg); in mem_sharing_page_unlock() 239 rmap_count(struct page_info *pg) in rmap_count() argument 436 struct page_info *pg; in audit() local 442 pg = pg_shared_info->pg; in audit() 443 mfn = page_to_mfn(pg); in audit() 480 if ( (!pg->sharing) || !rmap_has_entries(pg) ) in audit() 694 pg = mfn_to_page(mfn); in __grab_shared_page() 707 return pg; in __grab_shared_page() 793 if ( !pg ) in nominate_page() [all …]
|
/xen-4.10.0-shim-comet/xen/include/asm-x86/hvm/svm/ |
A D | amd-iommu-proto.h | 170 struct page_info *pg; in alloc_amd_iommu_pgtable() local 173 pg = alloc_domheap_page(NULL, 0); in alloc_amd_iommu_pgtable() 174 if ( pg == NULL ) in alloc_amd_iommu_pgtable() 176 vaddr = __map_domain_page(pg); in alloc_amd_iommu_pgtable() 179 return pg; in alloc_amd_iommu_pgtable() 182 static inline void free_amd_iommu_pgtable(struct page_info *pg) in free_amd_iommu_pgtable() argument 184 if ( pg != 0 ) in free_amd_iommu_pgtable() 185 free_domheap_page(pg); in free_amd_iommu_pgtable()
|
/xen-4.10.0-shim-comet/xen/arch/x86/mm/shadow/ |
A D | common.c | 439 struct page_info *pg; in sh_oos_audit() local 902 struct page_info *pg; in sh_unsync() local 908 pg = mfn_to_page(gmfn); in sh_unsync() 913 if ( pg->shadow_flags & in sh_unsync() 1483 struct page_info *pg; in shadow_alloc_p2m_page() local 1506 ASSERT(!page_get_owner(pg) && !(pg->count_info & PGC_count_mask)); in shadow_alloc_p2m_page() 1510 return pg; in shadow_alloc_p2m_page() 1524 pg->count_info, pg->u.inuse.type_info); in shadow_free_p2m_page() 3122 if ( pg == NULL ) in shadow_enable() 3132 pg->count_info = 1; in shadow_enable() [all …]
|
/xen-4.10.0-shim-comet/xen/drivers/passthrough/amd/ |
A D | pci_amd_iommu.c | 399 PFN_ORDER(pg) = level; in deallocate_next_page_table() 401 page_list_add_tail(pg, &iommu_pt_cleanup_list); in deallocate_next_page_table() 405 static void deallocate_page_table(struct page_info *pg) in deallocate_page_table() argument 409 unsigned int index, level = PFN_ORDER(pg), next_level; in deallocate_page_table() 411 PFN_ORDER(pg) = 0; in deallocate_page_table() 415 free_amd_iommu_pgtable(pg); in deallocate_page_table() 419 table_vaddr = __map_domain_page(pg); in deallocate_page_table() 438 free_amd_iommu_pgtable(pg); in deallocate_page_table() 527 static void amd_dump_p2m_table_level(struct page_info* pg, int level, in amd_dump_p2m_table_level() argument 539 table_vaddr = __map_domain_page(pg); in amd_dump_p2m_table_level() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/ |
A D | tboot.c | 195 void *pg = map_domain_page(_mfn(mfn)); in update_pagetable_mac() local 197 vmac_update(pg, PAGE_SIZE, ctx); in update_pagetable_mac() 198 unmap_domain_page(pg); in update_pagetable_mac() 222 void *pg = __map_domain_page(page); in tboot_gen_domain_integrity() local 223 vmac_update(pg, PAGE_SIZE, &ctx); in tboot_gen_domain_integrity() 224 unmap_domain_page(pg); in tboot_gen_domain_integrity() 293 void *pg; in tboot_gen_xenheap_integrity() local 298 pg = mfn_to_virt(mfn); in tboot_gen_xenheap_integrity() 299 vmac_update((uint8_t *)pg, PAGE_SIZE, &ctx); in tboot_gen_xenheap_integrity()
|
A D | smpboot.c | 54 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) argument 604 struct page_info *pg; in alloc_stub_page() local 609 pg = mfn_to_page(_mfn(*mfn)); in alloc_stub_page() 615 pg = alloc_domheap_page(NULL, memflags); in alloc_stub_page() 616 if ( !pg ) in alloc_stub_page() 619 unmap_domain_page(memset(__map_domain_page(pg), 0xcc, PAGE_SIZE)); in alloc_stub_page() 623 if ( map_pages_to_xen(stub_va, mfn_x(page_to_mfn(pg)), 1, in alloc_stub_page() 627 free_domheap_page(pg); in alloc_stub_page() 631 *mfn = mfn_x(page_to_mfn(pg)); in alloc_stub_page()
|
A D | mm.c | 136 #define page_to_mfn(pg) _mfn(__page_to_mfn(pg)) argument 643 typeof(pg->linear_pt_count) nc = read_atomic(&pg->linear_pt_count), oc; in inc_linear_entries() 674 typeof(pg->linear_pt_count) nc = read_atomic(&pg->linear_pt_count), oc; in inc_linear_uses() 1282 put_page(pg); in put_page_from_l2e() 1291 struct page_info *pg; in put_page_from_l3e() local 1327 put_page(pg); in put_page_from_l3e() 1357 put_page(pg); in put_page_from_l4e() 5251 struct page_info *pg; in create_perdomain_mapping() local 5263 if ( !pg ) in create_perdomain_mapping() 5284 if ( !pg ) in create_perdomain_mapping() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/hvm/ |
A D | stdvga.c | 87 struct page_info *pg = s->vram_page[(a >> 12) & 0x3f]; in vram_getb() local 88 uint8_t *p = __map_domain_page(pg); in vram_getb() 94 struct page_info *pg = s->vram_page[(a >> 10) & 0x3f]; in vram_getl() local 95 uint32_t *p = __map_domain_page(pg); in vram_getl() 578 struct page_info *pg; in stdvga_init() local 589 pg = alloc_domheap_page(d, MEMF_no_owner); in stdvga_init() 590 if ( pg == NULL ) in stdvga_init() 592 s->vram_page[i] = pg; in stdvga_init() 593 clear_domain_page(_mfn(page_to_mfn(pg))); in stdvga_init()
|
A D | ioreq.c | 1267 buffered_iopage_t *pg; in hvm_send_buffered_ioreq() local 1279 pg = iorp->va; in hvm_send_buffered_ioreq() 1281 if ( !pg ) in hvm_send_buffered_ioreq() 1317 if ( (pg->ptrs.write_pointer - pg->ptrs.read_pointer) >= in hvm_send_buffered_ioreq() 1325 pg->buf_ioreq[pg->ptrs.write_pointer % IOREQ_BUFFER_SLOT_NUM] = bp; in hvm_send_buffered_ioreq() 1330 pg->buf_ioreq[(pg->ptrs.write_pointer+1) % IOREQ_BUFFER_SLOT_NUM] = bp; in hvm_send_buffered_ioreq() 1335 pg->ptrs.write_pointer += qw ? 2 : 1; in hvm_send_buffered_ioreq() 1339 pg->ptrs.read_pointer >= IOREQ_BUFFER_SLOT_NUM ) in hvm_send_buffered_ioreq() 1341 union bufioreq_pointers old = pg->ptrs, new; in hvm_send_buffered_ioreq() 1346 cmpxchg(&pg->ptrs.full, old.full, new.full); in hvm_send_buffered_ioreq()
|
/xen-4.10.0-shim-comet/xen/drivers/acpi/apei/ |
A D | apei-io.c | 79 int i, pg; in apei_range_map() local 82 pg = ((((paddr + size -1) & PAGE_MASK) in apei_range_map() 84 if (apei_range_nr + pg > FIX_APEI_RANGE_MAX) in apei_range_map() 87 start_nr = apei_range_nr + pg -1; in apei_range_map() 88 for (i = 0; i < pg; i++) { in apei_range_map()
|
/xen-4.10.0-shim-comet/xen/arch/arm/ |
A D | guest_walk.c | 130 if ( pte.pg.page ) /* Small page. */ in guest_walk_sd() 133 *ipa = ((paddr_t)pte.pg.base << L2DESC_SMALL_PAGE_SHIFT) | (gva & mask); in guest_walk_sd() 136 if ( !pte.pg.xn ) in guest_walk_sd() 150 if ( !pte.pg.ro ) in guest_walk_sd()
|
/xen-4.10.0-shim-comet/xen/drivers/passthrough/ |
A D | iommu.c | 308 struct page_info *pg; in iommu_free_pagetables() local 311 pg = page_list_remove_head(&iommu_pt_cleanup_list); in iommu_free_pagetables() 313 if ( !pg ) in iommu_free_pagetables() 315 iommu_get_ops()->free_page_table(pg); in iommu_free_pagetables()
|