| /hypervisor/dm/ |
| A D | mmio_dev.c | 23 if (mem_aligned_check(res->user_vm_pa, PAGE_SIZE) && in assign_mmio_dev() 24 mem_aligned_check(res->host_pa, PAGE_SIZE) && in assign_mmio_dev() 25 mem_aligned_check(res->size, PAGE_SIZE)) { in assign_mmio_dev() 50 if (mem_aligned_check(gpa, PAGE_SIZE) && in deassign_mmio_dev() 51 mem_aligned_check(res->size, PAGE_SIZE)) { in deassign_mmio_dev()
|
| /hypervisor/include/arch/x86/asm/ |
| A D | page.h | 36 #define PAGE_SIZE (1U << PAGE_SHIFT) macro 95 uint8_t contents[PAGE_SIZE]; /**< A 4-KByte page in the memory. */ 96 } __aligned(PAGE_SIZE);
|
| A D | per_cpu.h | 25 uint8_t vmxon_region[PAGE_SIZE]; 75 } __aligned(PAGE_SIZE); /* per_cpu_region size aligned with PAGE_SIZE */
|
| A D | mmu.h | 105 return (((addr + (uint64_t)PAGE_SIZE) - 1UL) & PAGE_MASK); in round_page_up()
|
| A D | apicreg.h | 80 } __aligned(PAGE_SIZE);
|
| /hypervisor/include/arch/x86/asm/guest/ |
| A D | vm.h | 50 } __aligned(PAGE_SIZE); 106 uint8_t io_bitmap[PAGE_SIZE*2]; 136 } __aligned(PAGE_SIZE); 185 } __aligned(PAGE_SIZE);
|
| A D | vcpu.h | 235 uint8_t vmcs[PAGE_SIZE]; 241 uint8_t msr_bitmap[PAGE_SIZE]; 303 } __aligned(PAGE_SIZE); 326 } __aligned(PAGE_SIZE);
|
| A D | nested.h | 333 uint8_t vmcs02[PAGE_SIZE]; /* VMCS to run L2 and as Link Pointer in VMCS01 */ 339 } __aligned(PAGE_SIZE); 349 } __aligned(PAGE_SIZE);
|
| A D | vlapic.h | 86 } __aligned(PAGE_SIZE);
|
| /hypervisor/debug/ |
| A D | uart16550.c | 124 static uint8_t uart_pde_page[PAGE_SIZE]__aligned(PAGE_SIZE); 125 static uint8_t uart_pdpte_page[PAGE_SIZE]__aligned(PAGE_SIZE);
|
| A D | shell_priv.h | 13 #define SHELL_STRING_MAX_LEN (PAGE_SIZE << 2U)
|
| /hypervisor/arch/x86/guest/ |
| A D | vmx_io.c | 119 gpa & PAGE_MASK, PAGE_SIZE, EPT_EXE, 0UL); in ept_violation_vmexit_handler() 122 gpa & PAGE_MASK, PAGE_SIZE, EPT_EXE, 0UL); in ept_violation_vmexit_handler()
|
| A D | ept.c | 69 return CONFIG_MAX_VM_NUM * (get_ept_page_num()) * PAGE_SIZE; in get_total_ept_4k_pages_size() 111 offset += get_ept_page_num() * PAGE_SIZE; in reserve_buffer_for_ept_pages() 239 (void)memset(vm->arch_vm.nworld_eptp, 0U, PAGE_SIZE); in destroy_ept()
|
| A D | vm.c | 53 static struct acrn_vm vm_array[CONFIG_MAX_VM_NUM] __aligned(PAGE_SIZE); 284 (void)memset(vm->arch_vm.io_bitmap, 0x00U, PAGE_SIZE * 2U); in setup_io_bitmap() 287 (void)memset(vm->arch_vm.io_bitmap, 0xFFU, PAGE_SIZE * 2U); in setup_io_bitmap() 591 ept_del_mr(vm, pml4_page, plat_dmar_info.drhd_units[i].reg_base_addr, PAGE_SIZE); in prepare_service_vm_memmap() 856 (void)memset(vm->arch_vm.nworld_eptp, 0U, PAGE_SIZE); in create_vm()
|
| A D | nested.c | 380 static uint64_t vmcs_shadowing_bitmap[PAGE_SIZE / sizeof(uint64_t)] __aligned(PAGE_SIZE); 392 memset((void *)vmcs_shadowing_bitmap, 0xFFU, PAGE_SIZE); in setup_vmcs_shadowing_bitmap() 635 return (mem_aligned_check(vmptr_gpa, PAGE_SIZE) && ((vmptr_gpa >> cpu_info->phys_bits) == 0UL)); in validate_vmptr_gpa() 724 (void)memset(vvmcs->vmcs02, 0U, PAGE_SIZE); in reset_vvmcs()
|
| A D | vept.c | 34 return (get_e820_ram_size() * 8UL) / PAGE_SIZE; in calc_sept_size() 39 return calc_sept_size() / PAGE_SIZE; in calc_sept_page_num()
|
| A D | hyperv.c | 138 (void)memset(page_hva, 0U, PAGE_SIZE); in hyperv_setup_hypercall_page()
|
| /hypervisor/arch/x86/ |
| A D | mmu.c | 48 static uint8_t sanitized_page[PAGE_SIZE] __aligned(PAGE_SIZE); 338 for (linear_addr = addr; linear_addr < (addr + size); linear_addr += PAGE_SIZE) { in flush_tlb_range()
|
| A D | page.c | 56 (void)memset(page, 0U, PAGE_SIZE); in alloc_page()
|
| A D | vtd.c | 159 static struct page root_tables[CONFIG_MAX_IOMMU_NUM] __aligned(PAGE_SIZE); 165 static struct context_table ctx_tables[CONFIG_MAX_IOMMU_NUM] __aligned(PAGE_SIZE); 174 static struct page qi_queues[CONFIG_MAX_IOMMU_NUM] __aligned(PAGE_SIZE); 180 static struct intr_remap_table ir_tables[CONFIG_MAX_IOMMU_NUM] __aligned(PAGE_SIZE); 213 set_paging_supervisor(drhd_rt->drhd->reg_base_addr, PAGE_SIZE); in register_hrhd_units()
|
| A D | e820.c | 135 hv_e820[e820_idx].length = efi_mmap_entry[i].num_pages * PAGE_SIZE; in init_e820_from_efi_mmap()
|
| A D | cpu.c | 45 struct per_cpu_region per_cpu_data[MAX_PCPU_NUM] __aligned(PAGE_SIZE);
|
| /hypervisor/common/ |
| A D | hypercall.c | 671 if ((region->size & (PAGE_SIZE - 1UL)) == 0UL) { in set_vm_memory_region() 751 if ((!mem_aligned_check(wp->gpa, PAGE_SIZE)) || in write_protect_page() 752 (!ept_is_valid_mr(vm, wp->gpa, PAGE_SIZE))) { in write_protect_page() 765 if (((hpa <= base_paddr) && ((hpa + PAGE_SIZE) > base_paddr)) || in write_protect_page() 774 wp->gpa, PAGE_SIZE, prot_set, prot_clr); in write_protect_page()
|
| /hypervisor/dm/vpci/ |
| A D | vmsix.c | 59 ASSERT(vdev->msix.table_count <= (PAGE_SIZE/ MSIX_TABLE_ENTRY_SIZE), ""); in read_vmsix_cap_reg()
|
| /hypervisor/boot/guest/ |
| A D | bzimage_loader.c | 225 efi_mmap_desc[desc_idx].num_pages = vm->e820_entries[i].length / PAGE_SIZE; in create_service_vm_efi_mmap_desc() 233 efi_mmap_desc[i].phys_addr, efi_mmap_desc[i].num_pages * PAGE_SIZE, efi_mmap_desc[i].type); in create_service_vm_efi_mmap_desc()
|