/lk-master/kernel/vm/ |
A D | vm.c | 42 paddr_t pa; in mark_pages_in_use() local 49 pmm_alloc_range(pa, 1, &list); in mark_pages_in_use() 95 void *paddr_to_kvaddr(paddr_t pa) { in paddr_to_kvaddr() argument 100 pa >= map->phys && in paddr_to_kvaddr() 101 pa <= map->phys + map->size - 1) { in paddr_to_kvaddr() 102 return (void *)(map->virt + (pa - map->phys)); in paddr_to_kvaddr() 114 paddr_t pa; in vaddr_to_paddr() local 115 status_t rc = arch_mmu_query(&aspace->arch_aspace, (vaddr_t)ptr, &pa, NULL); in vaddr_to_paddr() 119 return pa; in vaddr_to_paddr() 159 paddr_t pa; in cmd_vm() local [all …]
|
A D | pmm.c | 254 paddr_t pa; in pmm_alloc_kpages() local 255 size_t alloc_count = pmm_alloc_contiguous(count, PAGE_SIZE_SHIFT, &pa, list); in pmm_alloc_kpages() 259 return paddr_to_kvaddr(pa); in pmm_alloc_kpages() 283 size_t pmm_alloc_contiguous(uint count, uint8_t alignment_log2, paddr_t *pa, struct list_node *list… in pmm_alloc_contiguous() argument 345 if (pa) in pmm_alloc_contiguous() 346 *pa = a->base + start * PAGE_SIZE; in pmm_alloc_contiguous() 477 paddr_t pa; in cmd_pmm() local 478 size_t ret = pmm_alloc_contiguous(argv[2].u, argv[3].u, &pa, &list); in cmd_pmm() 479 printf("pmm_alloc_contiguous returns %zu, address 0x%lx\n", ret, pa); in cmd_pmm() 480 printf("address %% align = 0x%lx\n", pa % argv[3].u); in cmd_pmm()
|
A D | vmm.c | 414 paddr_t pa = 0; in vmm_alloc_contiguous() local 416 size_t count = pmm_alloc_contiguous(size / PAGE_SIZE, align_pow2, &pa, &page_list); in vmm_alloc_contiguous() 438 arch_mmu_map(&aspace->arch_aspace, r->base, pa, size / PAGE_SIZE, arch_mmu_flags); in vmm_alloc_contiguous() 521 paddr_t pa = vm_page_to_paddr(p); in vmm_alloc() local 522 DEBUG_ASSERT(IS_PAGE_ALIGNED(pa)); in vmm_alloc() 524 arch_mmu_map(&aspace->arch_aspace, va, pa, 1, arch_mmu_flags); in vmm_alloc()
|
/lk-master/dev/cache/pl310/ |
A D | pl310.c | 154 addr_t pa = 0; \ 159 if (unlikely(pa / PAGE_SIZE != last_pa_page)) { \ 161 if (unlikely(arm_vtop(start, &pa))) \ 163 last_pa_page = pa / PAGE_SIZE; \ 165 PL310_REG(reg) = pa; \ 167 pa += CACHE_LINE; \
|
/lk-master/app/lkboot/ |
A D | dcc.c | 212 paddr_t pa; in lkboot_dcc_init() local 217 pa = vaddr_to_paddr(htod_buffer); in lkboot_dcc_init() 218 DEBUG_ASSERT(pa); in lkboot_dcc_init() 220 buffer_desc.htod_buffer_phys = pa; in lkboot_dcc_init() 223 pa = vaddr_to_paddr(dtoh_buffer); in lkboot_dcc_init() 224 DEBUG_ASSERT(pa); in lkboot_dcc_init() 226 buffer_desc.dtoh_buffer_phys = pa; in lkboot_dcc_init()
|
A D | commands.c | 351 paddr_t pa = vaddr_to_paddr(buf); in lkb_handle_command() local 352 if (pa == 0) { in lkb_handle_command() 370 zynq_program_fpga(pa, len); in lkb_handle_command()
|
/lk-master/lib/libc/ |
A D | qsort.c | 83 char *pa, *pb, *pc, *pd, *pl, *pm, *pn; in qsort() local 110 pa = pb = (char *)a + es; in qsort() 117 swap(pa, pb); in qsort() 118 pa += es; in qsort() 146 r = min(pa - (char *)a, pb - pa); in qsort() 150 if ((r = pb - pa) > (int)es) in qsort()
|
/lk-master/lib/acpi_lite/ |
A D | acpi_lite.cpp | 29 static const void* phys_to_ptr(uintptr_t pa) { in phys_to_ptr() argument 30 void *ptr = paddr_to_kvaddr(pa); in phys_to_ptr() 132 paddr_t pa; in acpi_get_table_at_index() local 134 pa = acpi.sdt->addr64[index]; in acpi_get_table_at_index() 136 pa = acpi.sdt->addr32[index]; in acpi_get_table_at_index() 139 return static_cast<const acpi_sdt_header*>(phys_to_ptr(pa)); in acpi_get_table_at_index()
|
/lk-master/app/zynq-common/ |
A D | init.c | 123 paddr_t pa = vaddr_to_paddr((void *)fpga_ptr); in zynq_common_target_init() local 124 if (pa != 0) { in zynq_common_target_init() 126 … printf("loading fpga image at %p (phys 0x%lx), len %zx\n", fpga_ptr, pa, fpga_len); in zynq_common_target_init() 128 err = zynq_program_fpga(pa, fpga_len); in zynq_common_target_init()
|
/lk-master/dev/virtio/block/ |
A D | virtio-block.c | 227 paddr_t pa = vaddr_to_paddr((void *)va); in virtio_block_read_write() local 228 desc->addr = (uint64_t)pa; in virtio_block_read_write() 239 paddr_t next_pa = PAGE_ALIGN(pa + 1); in virtio_block_read_write() 240 desc->len = MIN(next_pa - pa, len); in virtio_block_read_write() 249 pa = vaddr_to_paddr((void *)va); in virtio_block_read_write() 250 LTRACEF("va now 0x%lx, pa 0x%lx, next_pa 0x%lx, remaining len %zu\n", va, pa, next_pa, len); in virtio_block_read_write() 253 if (next_pa == pa) { in virtio_block_read_write() 264 next_desc->addr = (uint64_t)pa; in virtio_block_read_write()
|
/lk-master/arch/test/ |
A D | mmu.c | 54 paddr_t pa; in map_user_pages() local 56 err = arch_mmu_query(&as, va, &pa, &flags); in map_user_pages() 58 EXPECT_EQ(vm_page_to_paddr(p), pa, "pa"); in map_user_pages()
|
/lk-master/dev/virtio/ |
A D | virtio.c | 328 paddr_t pa; in virtio_alloc_ring() local 329 pa = vaddr_to_paddr(vptr); in virtio_alloc_ring() 330 if (pa == 0) { in virtio_alloc_ring() 334 LTRACEF("virtio_ring at pa 0x%lx\n", pa); in virtio_alloc_ring() 344 paddr_t pa = (paddr_t)vptr; in virtio_alloc_ring() local 363 dev->mmio_config->queue_pfn = pa / PAGE_SIZE; in virtio_alloc_ring()
|
/lk-master/arch/riscv/ |
A D | mmu.cpp | 152 volatile riscv_pte_t *alloc_ptable(arch_aspace_t *aspace, addr_t *pa) { in alloc_ptable() argument 160 *pa = vm_page_to_paddr(p); in alloc_ptable() 161 riscv_pte_t *pte = (riscv_pte_t *)paddr_to_kvaddr(*pa); in alloc_ptable() 171 LTRACEF_LEVEL(3, "returning pa %#lx, va %p\n", *pa, pte); in alloc_ptable() 490 paddr_t pa = RISCV_PTE_PPN(pte); in arch_mmu_query() local 494 *paddr = pa | (*vaddr & page_mask); in arch_mmu_query() 495 … LTRACEF_LEVEL(3, "raw pa %#lx, page_mask %#lx, final pa %#lx\n", pa, page_mask, *paddr); in arch_mmu_query()
|
/lk-master/app/tests/ |
A D | mem_tests.c | 183 paddr_t pa; in mem_test() local 184 pa = vaddr_to_paddr(ptr); in mem_test() 185 printf("physical address 0x%lx\n", pa); in mem_test()
|
/lk-master/arch/arm/arm/ |
A D | arch.c | 285 status_t arm_vtop(addr_t va, addr_t *pa) { in arm_vtop() argument 298 if (pa) { in arm_vtop() 299 *pa = (par & 0xfffff000) | (va & 0xfff); in arm_vtop()
|
A D | mmu.c | 369 paddr_t pa; in get_l2_table() local 395 ret = arm_vtop((vaddr_t)l2_va, &pa); in get_l2_table() 397 ASSERT(paddr_to_kvaddr(pa)); in get_l2_table() 400 DEBUG_ASSERT(IS_PAGE_ALIGNED(pa)); in get_l2_table() 402 *ppa = pa + (PAGE_SIZE / L1E_PER_PAGE) * (l1_index & (L1E_PER_PAGE-1)); in get_l2_table() 404 LTRACEF("allocated pagetable at %p, pa 0x%lx, pa 0x%lx\n", l2_va, pa, *ppa); in get_l2_table()
|
/lk-master/kernel/include/kernel/ |
A D | vm.h | 157 size_t pmm_alloc_contiguous(uint count, uint8_t align_log2, paddr_t *pa, struct list_node *list); 170 void *paddr_to_kvaddr(paddr_t pa);
|
/lk-master/dev/virtio/gpu/ |
A D | virtio-gpu.c | 214 paddr_t pa; in attach_backing() local 216 pa = vaddr_to_paddr(ptr); in attach_backing() 218 pa = (paddr_t)ptr; in attach_backing() 220 req.mem.addr = pa; in attach_backing()
|
/lk-master/arch/arm/include/arch/arm/ |
A D | mmu.h | 202 status_t arm_vtop(addr_t va, addr_t *pa);
|