Home
last modified time | relevance | path

Searched refs:size (Results 1 – 22 of 22) sorted by relevance

/arch/arm64/
A Dmmu.c216 if (size == PAGE_SIZE) { in alloc_page_table()
228 void *vaddr = memalign(size, size); in alloc_page_table()
248 if (size >= PAGE_SIZE) { in free_page_table()
331 while (size) { in arm64_mmu_unmap_pt()
368 size -= chunk_size; in arm64_mmu_unmap_pt()
383 size_t size = size_in; in arm64_mmu_map_pt() local
399 while (size) { in arm64_mmu_map_pt()
462 if (vaddr_rel > vaddr_rel_max - size || size > vaddr_rel_max) { in arm64_mmu_map()
488 if (vaddr_rel > vaddr_rel_max - size || size > vaddr_rel_max) { in arm64_mmu_unmap()
591 aspace->size = size; in arch_mmu_init_aspace()
[all …]
A Dstart.S27 size .req x26 label
116 mov mapping_size, size
148 subs size, tmp, vaddr
154 subs size, tmp, vaddr
158 subs size, tmp, vaddr
162 subs size, tmp, vaddr
164 cmp mapping_size, size
166 mov size, mapping_size
174 orr tmp, tmp, size
209 lsr tmp, size, idx_shift
[all …]
A Drules.mk98 ARCH_LDFLAGS += -z max-page-size=4096
/arch/arm64/include/arch/arm64/
A Dmmu.h71 #error User address space size must be larger than page size
84 #error Kernel address space size must be larger than page size
108 #define MMU_TCR_IPS(size) BM(32, 3, (size)) argument
115 #define MMU_TCR_T1SZ(size) BM(16, 6, (size)) argument
121 #define MMU_TCR_T0SZ(size) BM( 0, 6, (size)) argument
271 int arm64_mmu_map(vaddr_t vaddr, paddr_t paddr, size_t size, pte_t attrs,
275 int arm64_mmu_unmap(vaddr_t vaddr, size_t size,
/arch/arm/arm/include/arch/
A Daspace.h22 size_t size; member
29 return (vaddr >= aspace->base && vaddr <= aspace->base + aspace->size - 1); in arch_mmu_is_valid_vaddr()
/arch/x86/32/
A Dmmu.c346 if (IS_ALIGNED(range->size, PAGE_SIZE)) in x86_mmu_map_range()
347 no_of_pages = (range->size) >> PAGE_DIV_SHIFT; in x86_mmu_map_range()
349 no_of_pages = ((range->size) >> PAGE_DIV_SHIFT) + 1; in x86_mmu_map_range()
409 range.size = count * PAGE_SIZE; in arch_mmu_map()
460 LTRACEF("aspace %p, base %#lx, size %#zx, flags %#x\n", aspace, base, size, flags); in arch_mmu_init_aspace()
463 DEBUG_ASSERT(size > PAGE_SIZE); in arch_mmu_init_aspace()
464 DEBUG_ASSERT(base + size - 1 > base); in arch_mmu_init_aspace()
471 DEBUG_ASSERT(size == KERNEL_ASPACE_SIZE); in arch_mmu_init_aspace()
474 aspace->size = size; in arch_mmu_init_aspace()
479 DEBUG_ASSERT(size == USER_ASPACE_SIZE); in arch_mmu_init_aspace()
[all …]
/arch/arm/arm/
A Dmmu.c178 while (map->size > 0) { in arm_mmu_init()
181 size_t size = map->size; in arm_mmu_init() local
183 DEBUG_ASSERT(IS_SECTION_ALIGNED(size)); in arm_mmu_init()
185 while (size > 0) { in arm_mmu_init()
188 size -= MB; in arm_mmu_init()
451 base, end, align, size, flags); in arch_mmu_pick_spot()
462 vaddr_t spot_end = spot + size - 1; in arch_mmu_pick_spot()
674 DEBUG_ASSERT(size > PAGE_SIZE); in arch_mmu_init_aspace()
675 DEBUG_ASSERT(base + size - 1 > base); in arch_mmu_init_aspace()
681 aspace->size = size; in arch_mmu_init_aspace()
[all …]
/arch/or1k/
A Dmmu.c204 status_t arch_mmu_init_aspace(arch_aspace_t *aspace, vaddr_t base, size_t size, uint flags) { in arch_mmu_init_aspace() argument
205 LTRACEF("aspace %p, base %#lx, size %#zx, flags %#x\n", aspace, base, size, flags); in arch_mmu_init_aspace()
210 DEBUG_ASSERT(size > PAGE_SIZE); in arch_mmu_init_aspace()
211 DEBUG_ASSERT(base + size - 1 > base); in arch_mmu_init_aspace()
218 DEBUG_ASSERT(size == KERNEL_ASPACE_SIZE); in arch_mmu_init_aspace()
221 aspace->size = size; in arch_mmu_init_aspace()
/arch/riscv/
A Dmmu.cpp57 .size = RISCV_MMU_PHYSMAP_SIZE,
199 LTRACEF("aspace %p, base %#lx, size %#zx, flags %#x\n", aspace, base, size, flags); in arch_mmu_init_aspace()
205 DEBUG_ASSERT(size > PAGE_SIZE); in arch_mmu_init_aspace()
206 DEBUG_ASSERT(base + size - 1 > base); in arch_mmu_init_aspace()
214 DEBUG_ASSERT(size == KERNEL_ASPACE_SIZE); in arch_mmu_init_aspace()
218 aspace->size = size; in arch_mmu_init_aspace()
228 DEBUG_ASSERT(size == USER_ASPACE_SIZE); in arch_mmu_init_aspace()
231 aspace->size = size; in arch_mmu_init_aspace()
405 if (_vaddr < aspace->base || _vaddr > aspace->base + aspace->size - 1) { in arch_mmu_map()
476 if (_vaddr < aspace->base || _vaddr > aspace->base + aspace->size - 1) { in arch_mmu_query()
[all …]
A Dsbi.c110 void sbi_rfence_vma(const unsigned long *hart_mask, vaddr_t vma, size_t size) { in sbi_rfence_vma() argument
113 sbi_call(SBI_EXT_RFENCE_SIG, 1, *hart_mask, 0, vma, size); in sbi_rfence_vma()
A Drules.mk229 ARCH_LDFLAGS += -z max-page-size=8
/arch/include/arch/
A Dmmu.h45 status_t arch_mmu_init_aspace(arch_aspace_t *aspace, vaddr_t base, size_t size, uint flags) __NONNU…
56 vaddr_t align, size_t size, uint arch_mmu_flags) __NONNULL((1));
/arch/x86/64/
A Dmmu.c548 range->start_vaddr, range->start_paddr, range->size, flags); in x86_mmu_map_range()
556 if (IS_ALIGNED(range->size, PAGE_SIZE)) in x86_mmu_map_range()
557 no_of_pages = (range->size) >> PAGE_DIV_SHIFT; in x86_mmu_map_range()
559 no_of_pages = ((range->size) >> PAGE_DIV_SHIFT) + 1; in x86_mmu_map_range()
623 range.size = count * PAGE_SIZE; in arch_mmu_map()
691 DEBUG_ASSERT(size > PAGE_SIZE); in arch_mmu_init_aspace()
692 DEBUG_ASSERT(base + size - 1 > base); in arch_mmu_init_aspace()
698 DEBUG_ASSERT(size == KERNEL_ASPACE_SIZE); in arch_mmu_init_aspace()
701 aspace->size = size; in arch_mmu_init_aspace()
706 DEBUG_ASSERT(size == USER_ASPACE_SIZE); in arch_mmu_init_aspace()
[all …]
/arch/arm64/include/arch/
A Daspace.h25 size_t size; member
/arch/x86/include/arch/
A Daspace.h25 size_t size; member
/arch/riscv/include/arch/
A Daspace.h30 size_t size; member
/arch/x86/
A Darch.c32 … .size = PHYSMAP_SIZE, /* x86-64 maps first 64GB by default, 1GB on x86-32, 16MB in legacy mode */
42 .size = 1*GB,
A Drules.mk122 ARCH_LDFLAGS += -z max-page-size=4096
/arch/x86/include/arch/x86/
A Dmmu.h107 uint32_t size; member
/arch/riscv/include/arch/riscv/
A Dsbi.h77 void sbi_rfence_vma(const unsigned long *hart_mask, vaddr_t vma, size_t size);
/arch/arm/
A Drules.mk343 ARCH_LDFLAGS += -z max-page-size=4096
/arch/mips/
A Dmips.ld159 we can shorten the on-disk segment size. */

Completed in 30 milliseconds