/lk-master/arch/arm64/ |
A D | mmu.c | 34 return (vaddr >= aspace->base && vaddr <= aspace->base + aspace->size - 1); in is_valid_vaddr() 116 vaddr_rem = vaddr; in arch_mmu_query() 209 if (!vaddr) in alloc_page_table() 213 free(vaddr); in alloc_page_table() 234 free(vaddr); in free_page_table() 241 void *vaddr; in arm64_mmu_get_page_table() local 262 return vaddr; in arm64_mmu_get_page_table() 346 vaddr += chunk_size; in arm64_mmu_unmap_pt() 360 vaddr_t vaddr = vaddr_in; in arm64_mmu_map_pt() local 416 vaddr += chunk_size; in arm64_mmu_map_pt() [all …]
|
A D | start.S | 25 vaddr .req x23 label 123 subs size, tmp, vaddr 129 subs size, tmp, vaddr 133 subs size, tmp, vaddr 137 subs size, tmp, vaddr 148 orr tmp, vaddr, paddr 154 eor vaddr, vaddr, #(~0 << MMU_KERNEL_SIZE_SHIFT) 177 orr tmp2, vaddr, paddr 201 add vaddr, vaddr, tmp 208 eor vaddr, vaddr, #(~0 << MMU_KERNEL_SIZE_SHIFT) [all …]
|
/lk-master/arch/x86/64/ |
A D | mmu.c | 46 static bool x86_mmu_check_vaddr(vaddr_t vaddr) { in x86_mmu_check_vaddr() argument 47 uint64_t addr = (uint64_t)vaddr; in x86_mmu_check_vaddr() 211 pdpe = get_pdp_entry_from_pdp_table(vaddr, pml4e); in x86_mmu_get_mapping() 219 pde = get_pd_entry_from_pd_table(vaddr, pdpe); in x86_mmu_get_mapping() 236 pte = get_pt_entry_from_pt_table(vaddr, pde); in x86_mmu_get_mapping() 266 (!x86_mmu_check_vaddr(vaddr)) || in x86_mmu_check_mapping() 547 if (!(x86_mmu_check_vaddr(vaddr))) in x86_mmu_unmap() 553 next_aligned_v_addr = vaddr; in x86_mmu_unmap() 569 if (!(x86_mmu_check_vaddr(vaddr))) in arch_mmu_unmap() 663 if (!x86_mmu_check_vaddr(vaddr)) in arch_mmu_map() [all …]
|
/lk-master/arch/or1k/ |
A D | mmu.c | 28 void or1k_invalidate_tlb(vaddr_t vaddr, uint count) { in or1k_invalidate_tlb() argument 67 vaddr += PAGE_SIZE; in or1k_invalidate_tlb() 72 uint index = vaddr / SECTION_SIZE; in arch_mmu_query() 82 index = (vaddr % SECTION_SIZE) / PAGE_SIZE; in arch_mmu_query() 108 if (!IS_PAGE_ALIGNED(vaddr)) in arch_mmu_unmap() 113 uint index = vaddr / SECTION_SIZE; in arch_mmu_unmap() 116 vaddr += PAGE_SIZE; in arch_mmu_unmap() 126 vaddr += SECTION_SIZE; in arch_mmu_unmap() 155 l1_index = vaddr / SECTION_SIZE; in arch_mmu_map() 160 vaddr += SECTION_SIZE; in arch_mmu_map() [all …]
|
/lk-master/arch/x86/32/ |
A D | mmu.c | 157 pdt = get_pdp_entry_from_pdp_table(vaddr, pdpt); in x86_mmu_get_mapping() 164 pt = get_pd_entry_from_pd_table(vaddr, pdt); in x86_mmu_get_mapping() 175 pt = get_pd_entry_from_pd_table(vaddr, pdt); in x86_mmu_get_mapping() 195 pte = get_pt_entry_from_page_table(vaddr, pt); in x86_mmu_get_mapping() 224 (!IS_ALIGNED(vaddr, PAGE_SIZE)) || in x86_mmu_check_mapping() 333 pt = get_pd_entry_from_pd_table(vaddr, pdt); in x86_mmu_add_mapping() 349 pt = get_pd_entry_from_pd_table(vaddr, init_table); in x86_mmu_add_mapping() 445 if (!IS_ALIGNED(vaddr, PAGE_SIZE)) in x86_mmu_unmap() 451 next_aligned_v_addr = vaddr; in x86_mmu_unmap() 469 if (!IS_ALIGNED(vaddr, PAGE_SIZE)) in arch_mmu_unmap() [all …]
|
/lk-master/arch/arm/arm/ |
A D | mmu.c | 143 DEBUG_ASSERT(IS_SECTION_ALIGNED(vaddr)); in arm_mmu_map_section() 147 index = vaddr / SECTION_SIZE; in arm_mmu_map_section() 168 DEBUG_ASSERT(IS_SECTION_ALIGNED(vaddr)); in arm_mmu_unmap_section() 246 uint index = vaddr / MB; in arch_mmu_query() 498 DEBUG_ASSERT(IS_PAGE_ALIGNED(vaddr)); in arch_mmu_map() 520 vaddr += SECTION_SIZE; in arch_mmu_map() 562 vaddr += PAGE_SIZE; in arch_mmu_map() 587 DEBUG_ASSERT(IS_PAGE_ALIGNED(vaddr)); in arch_mmu_unmap() 588 if (!IS_PAGE_ALIGNED(vaddr)) in arch_mmu_unmap() 612 vaddr += SECTION_SIZE; in arch_mmu_unmap() [all …]
|
/lk-master/kernel/vm/ |
A D | vmm.c | 44 return (vaddr >= aspace->base && vaddr <= aspace->base + aspace->size - 1); in is_inside_aspace() 56 if (vaddr + size - 1 < vaddr) in is_region_inside_aspace() 271 if (vaddr == (vaddr_t)-1) { in alloc_region() 279 r->base = (vaddr_t)vaddr; in alloc_region() 346 vaddr_t vaddr = 0; in vmm_alloc_physical() local 354 vaddr = (vaddr_t)*ptr; in vmm_alloc_physical() 398 vaddr_t vaddr = 0; in vmm_alloc_contiguous() local 407 vaddr = (vaddr_t)*ptr; in vmm_alloc_contiguous() 472 vaddr_t vaddr = 0; in vmm_alloc() local 481 vaddr = (vaddr_t)*ptr; in vmm_alloc() [all …]
|
/lk-master/arch/riscv/ |
A D | mmu.cpp | 315 LTRACEF("vaddr %#lx\n", vaddr); in riscv_pt_walk() 322 uint index = vaddr_to_index(vaddr, level); in riscv_pt_walk() 327 level, index, ptep, *ptep, vaddr); in riscv_pt_walk() 340 index = vaddr_to_index(vaddr, level); in riscv_pt_walk() 345 auto ret = callback(level, index, pte, &vaddr); in riscv_pt_walk() 381 index = vaddr_to_index(vaddr, level); in riscv_pt_walk() 416 level, index, pte, *vaddr, paddr, count, flags); in arch_mmu_map() 449 *vaddr += PAGE_SIZE; in arch_mmu_map() 494 *paddr = pa | (*vaddr & page_mask); in arch_mmu_query() 549 *vaddr += PAGE_SIZE; in arch_mmu_unmap() [all …]
|
/lk-master/arch/arm/arm/include/arch/ |
A D | aspace.h | 28 static inline bool arch_mmu_is_valid_vaddr(struct arch_aspace *aspace, vaddr_t vaddr) { in arch_mmu_is_valid_vaddr() argument 29 return (vaddr >= aspace->base && vaddr <= aspace->base + aspace->size - 1); in arch_mmu_is_valid_vaddr()
|
/lk-master/arch/x86/include/arch/x86/ |
A D | mmu.h | 132 vaddr_t vaddr, arch_flags_t in_flags, 137 vaddr_t vaddr, arch_flags_t in_flags, 142 status_t x86_mmu_get_mapping(map_addr_t init_table, vaddr_t vaddr, uint32_t *ret_level, 147 vaddr_t vaddr, arch_flags_t flags); 148 status_t x86_mmu_unmap(map_addr_t init_table, vaddr_t vaddr, uint count);
|
/lk-master/arch/include/arch/ |
A D | mmu.h | 42 int arch_mmu_map(arch_aspace_t *aspace, vaddr_t vaddr, paddr_t paddr, uint count, uint flags) __NON… 43 int arch_mmu_unmap(arch_aspace_t *aspace, vaddr_t vaddr, uint count) __NONNULL((1)); 44 status_t arch_mmu_query(arch_aspace_t *aspace, vaddr_t vaddr, paddr_t *paddr, uint *flags) __NONNUL…
|
/lk-master/external/platform/lpc15xx/lpcopen/lpc_chip_15xx/inc/usbd/ |
A D | usbd_core.h | 276 uint32_t (* virt_to_phys)(void* vaddr); 504 uint32_t (* virt_to_phys)(void* vaddr);
|
/lk-master/arch/arm64/include/arch/arm64/ |
A D | mmu.h | 324 int arm64_mmu_map(vaddr_t vaddr, paddr_t paddr, size_t size, pte_t attrs, 328 int arm64_mmu_unmap(vaddr_t vaddr, size_t size,
|
/lk-master/kernel/include/kernel/ |
A D | vm.h | 227 status_t vmm_reserve_space(vmm_aspace_t *aspace, const char *name, size_t size, vaddr_t vaddr)
|