Lines Matching refs:base
127 void riscv_tlb_flush_vma_range(vaddr_t base, size_t count) { in riscv_tlb_flush_vma_range() argument
133 sbi_rfence_vma(&hart_mask, base, count * PAGE_SIZE); in riscv_tlb_flush_vma_range()
138 asm volatile("sfence.vma %0, zero" :: "r"(base)); in riscv_tlb_flush_vma_range()
139 base += PAGE_SIZE; in riscv_tlb_flush_vma_range()
198 status_t arch_mmu_init_aspace(arch_aspace_t *aspace, vaddr_t base, size_t size, uint flags) { in arch_mmu_init_aspace() argument
199 LTRACEF("aspace %p, base %#lx, size %#zx, flags %#x\n", aspace, base, size, flags); in arch_mmu_init_aspace()
206 DEBUG_ASSERT(base + size - 1 > base); in arch_mmu_init_aspace()
213 DEBUG_ASSERT(base == KERNEL_ASPACE_BASE); in arch_mmu_init_aspace()
217 aspace->base = base; in arch_mmu_init_aspace()
227 DEBUG_ASSERT(base == USER_ASPACE_BASE); in arch_mmu_init_aspace()
230 aspace->base = base; in arch_mmu_init_aspace()
405 if (_vaddr < aspace->base || _vaddr > aspace->base + aspace->size - 1) { in arch_mmu_map()
476 if (_vaddr < aspace->base || _vaddr > aspace->base + aspace->size - 1) { in arch_mmu_query()
527 if (_vaddr < aspace->base || _vaddr > aspace->base + aspace->size - 1) { in arch_mmu_unmap()