Home
last modified time | relevance | path

Searched refs:SMALL_PAGE_SIZE (Results 1 – 25 of 45) sorted by relevance

12

/optee_os-3.20.0/core/arch/arm/kernel/
A Dkern.ld.S124 . = ALIGN(SMALL_PAGE_SIZE);
186 . = ALIGN(SMALL_PAGE_SIZE);
228 . = ALIGN(SMALL_PAGE_SIZE);
304 . = ALIGN(SMALL_PAGE_SIZE);
314 ASSERT(!(__flatmap_init_rx_start & (SMALL_PAGE_SIZE - 1)),
333 . = ALIGN(SMALL_PAGE_SIZE);
369 . = ALIGN(SMALL_PAGE_SIZE);
375 . = ALIGN(SMALL_PAGE_SIZE);
506 __asan_map_start = (__asan_shadow_start / SMALL_PAGE_SIZE) *
507 SMALL_PAGE_SIZE;
[all …]
A Dboot.c366 stats.npages_all * SMALL_PAGE_SIZE / 1024); in print_pager_pool_size()
425 const unsigned int num_pages = tee_mm_get_bytes(mm) / SMALL_PAGE_SIZE; in ro_paged_alloc()
447 size_t hash_size = (pageable_size / SMALL_PAGE_SIZE) * in init_runtime()
456 assert(pageable_size % SMALL_PAGE_SIZE == 0); in init_runtime()
513 for (n = 0; (n * SMALL_PAGE_SIZE) < pageable_size; n++) { in init_runtime()
515 const uint8_t *page = paged_store + n * SMALL_PAGE_SIZE; in init_runtime()
519 res = hash_sha256_check(hash, page, SMALL_PAGE_SIZE); in init_runtime()
573 tee_pager_add_pages(pageable_start, init_size / SMALL_PAGE_SIZE, false); in init_runtime()
575 (pageable_size - init_size) / SMALL_PAGE_SIZE, in init_runtime()
579 SMALL_PAGE_SIZE, true); in init_runtime()
[all …]
A Dsecure_partition.c311 len = reg->page_count * SMALL_PAGE_SIZE; in sp_unmap_ffa_regions()
464 *num_pgs = total_size / SMALL_PAGE_SIZE; in sp_init_info()
581 res = vm_map(&ctx->uctx, &va, pages_cnt * SMALL_PAGE_SIZE, in handle_fdt_dev_regions()
598 vm_unmap(&ctx->uctx, va, pages_cnt * SMALL_PAGE_SIZE); in handle_fdt_dev_regions()
713 if (MUL_OVERFLOW(pages_cnt, SMALL_PAGE_SIZE, &size)) in handle_fdt_mem_regions()
854 page_count = ROUNDUP_DIV(log_size, SMALL_PAGE_SIZE); in handle_tpm_event_log()
1112 if (!tee_pbuf_is_sec(sp_pkg_pa, SMALL_PAGE_SIZE)) in process_sp_pkg()
1115 mm = tee_mm_alloc(&tee_mm_sec_ddr, SMALL_PAGE_SIZE); in process_sp_pkg()
1162 num_pages = ROUNDUP_DIV(sp_pkg_size, SMALL_PAGE_SIZE); in process_sp_pkg()
1198 SMALL_PAGE_SIZE)); in process_sp_pkg()
[all …]
A Dstmm_sp.c70 static const unsigned int stmm_stack_size = 4 * SMALL_PAGE_SIZE;
71 static const unsigned int stmm_heap_size = 398 * SMALL_PAGE_SIZE;
72 static const unsigned int stmm_sec_buf_size = 4 * SMALL_PAGE_SIZE;
73 static const unsigned int stmm_ns_comm_buf_size = 4 * SMALL_PAGE_SIZE;
178 size_t num_pgs = ROUNDUP(sz, SMALL_PAGE_SIZE) / SMALL_PAGE_SIZE; in alloc_and_map_sp_fobj()
188 res = vm_map(&spc->uctx, va, num_pgs * SMALL_PAGE_SIZE, in alloc_and_map_sp_fobj()
244 SMALL_PAGE_SIZE); in load_stmm()
513 res = vm_get_prot(&spc->uctx, va, SMALL_PAGE_SIZE, &attrs); in sp_svc_get_mem_attr()
538 if (!va || !nr_pages || MUL_OVERFLOW(nr_pages, SMALL_PAGE_SIZE, &sz)) in sp_svc_set_mem_attr()
/optee_os-3.20.0/core/arch/riscv/kernel/
A Dkern.ld.S83 ASSERT(!(TEE_TEXT_VA_START & (SMALL_PAGE_SIZE - 1)),
93 __flatmap_rx_start = LD_ROUNDDOWN(__text_start, SMALL_PAGE_SIZE);
110 . = ALIGN(SMALL_PAGE_SIZE);
144 . = ALIGN(SMALL_PAGE_SIZE);
278 __asan_map_start = (__asan_shadow_start / SMALL_PAGE_SIZE) *
279 SMALL_PAGE_SIZE;
280 __asan_map_end = ((__asan_shadow_end - 1) / SMALL_PAGE_SIZE) *
281 SMALL_PAGE_SIZE + SMALL_PAGE_SIZE;
/optee_os-3.20.0/core/arch/arm/mm/
A Dsp_mem.c60 m->mobj.size = pages * SMALL_PAGE_SIZE; in sp_mem_new_mobj()
61 m->mobj.phys_granule = SMALL_PAGE_SIZE; in sp_mem_new_mobj()
72 return ROUNDUP(ms->mobj.size, SMALL_PAGE_SIZE) / SMALL_PAGE_SIZE; in get_page_count()
89 if (!tee_pbuf_is_sec(pa, num_pages * SMALL_PAGE_SIZE)) in sp_mem_add_pages()
93 num_pages * SMALL_PAGE_SIZE)) in sp_mem_add_pages()
99 ms->pages[n + *idx] = pa + n * SMALL_PAGE_SIZE; in sp_mem_add_pages()
138 p = ms->pages[offset / SMALL_PAGE_SIZE] + in get_pa()
141 case SMALL_PAGE_SIZE: in get_pa()
142 p = ms->pages[offset / SMALL_PAGE_SIZE]; in get_pa()
267 (new_reg->page_count * SMALL_PAGE_SIZE); in sp_mem_is_shared()
[all …]
A Dmobj_dyn_shm.c82 p = mobj_reg_shm->pages[full_offset / SMALL_PAGE_SIZE] + in mobj_reg_shm_get_pa()
85 case SMALL_PAGE_SIZE: in mobj_reg_shm_get_pa()
86 p = mobj_reg_shm->pages[full_offset / SMALL_PAGE_SIZE]; in mobj_reg_shm_get_pa()
207 sz = ROUNDUP(mobj->size + r->page_offset, SMALL_PAGE_SIZE); in mobj_reg_shm_inc_map()
215 sz / SMALL_PAGE_SIZE, in mobj_reg_shm_inc_map()
304 if (!num_pages || page_offset >= SMALL_PAGE_SIZE) in mobj_reg_shm_alloc()
315 mobj_reg_shm->mobj.size = num_pages * SMALL_PAGE_SIZE - page_offset; in mobj_reg_shm_alloc()
316 mobj_reg_shm->mobj.phys_granule = SMALL_PAGE_SIZE; in mobj_reg_shm_alloc()
330 SMALL_PAGE_SIZE)) in mobj_reg_shm_alloc()
A Dtee_pager.c415 tlbi_mva_range(smem, nbytes, SMALL_PAGE_SIZE); in tee_pager_set_alias_area()
455 .idx = (va & mask) / SMALL_PAGE_SIZE, in region_va2tblidx()
578 pager_alias_next_free += SMALL_PAGE_SIZE; in pager_add_alias_page()
742 size_t s = fobj->num_pages * SMALL_PAGE_SIZE; in pager_add_um_region()
1049 size_t s = ROUNDUP(size, SMALL_PAGE_SIZE); in tee_pager_rem_um_region()
1267 icache_inv_user_range(va, SMALL_PAGE_SIZE); in tee_pager_unhide_page()
1449 dcache_clean_range_pou(va, SMALL_PAGE_SIZE); in pager_deploy_page()
1451 icache_inv_user_range(va, SMALL_PAGE_SIZE); in pager_deploy_page()
1453 icache_inv_range(va, SMALL_PAGE_SIZE); in pager_deploy_page()
1576 asan_tag_access(va, va + SMALL_PAGE_SIZE); in pager_get_page()
[all …]
A Dmobj_ffa.c78 mf->mobj.size = num_pages * SMALL_PAGE_SIZE; in ffa_new()
79 mf->mobj.phys_granule = SMALL_PAGE_SIZE; in ffa_new()
120 return ROUNDUP(mf->mobj.size, SMALL_PAGE_SIZE) / SMALL_PAGE_SIZE; in get_page_count()
218 if (!core_pbuf_is(CORE_MEM_NON_SEC, pa, num_pages * SMALL_PAGE_SIZE)) in mobj_ffa_add_pages_at()
222 mf->pages[n + *idx] = pa + n * SMALL_PAGE_SIZE; in mobj_ffa_add_pages_at()
354 if (internal_offs >= SMALL_PAGE_SIZE) in mobj_ffa_get_by_cookie()
445 p = mf->pages[full_offset / SMALL_PAGE_SIZE] + in ffa_get_pa()
448 case SMALL_PAGE_SIZE: in ffa_get_pa()
449 p = mf->pages[full_offset / SMALL_PAGE_SIZE]; in ffa_get_pa()
551 sz = ROUNDUP(mobj->size + mf->page_offset, SMALL_PAGE_SIZE); in ffa_inc_map()
[all …]
A Dpgt_cache.c58 #define PGT_PARENT_SIZE (4 * SMALL_PAGE_SIZE)
245 idx = (b - p->vabase) / SMALL_PAGE_SIZE; in pgt_clear_range()
246 n = (e - b) / SMALL_PAGE_SIZE; in pgt_clear_range()
388 COMPILE_TIME_ASSERT(PGT_SIZE * PGT_NUM_PGT_PER_PAGE == SMALL_PAGE_SIZE); in pgt_init()
391 uint8_t *tbl = tee_pager_alloc(SMALL_PAGE_SIZE); in pgt_init()
478 tee_pager_release_phys((void *)va, SMALL_PAGE_SIZE); in push_to_free_list()
750 idx = (b - p->vabase) / SMALL_PAGE_SIZE; in clear_ctx_range_from_list()
751 n = (e - b) / SMALL_PAGE_SIZE; in clear_ctx_range_from_list()
A Dcore_mmu.c36 assert(granule == CORE_MMU_PGDIR_SIZE || granule == SMALL_PAGE_SIZE); in tlbi_mva_range()
51 assert(granule == CORE_MMU_PGDIR_SIZE || granule == SMALL_PAGE_SIZE); in tlbi_mva_range_asid()
/optee_os-3.20.0/core/mm/
A Dfobj.c95 memset(va, 0, SMALL_PAGE_SIZE); in rwp_load_page()
100 NULL, 0, src, SMALL_PAGE_SIZE, va, in rwp_load_page()
125 NULL, 0, va, SMALL_PAGE_SIZE, dst, in rwp_save_page()
138 return rwp_store_base + idx * SMALL_PAGE_SIZE; in idx_to_store()
153 if (MUL_OVERFLOW(num_pages, SMALL_PAGE_SIZE, &size)) in rwp_paged_iv_alloc()
266 if (MUL_OVERFLOW(num_pages, SMALL_PAGE_SIZE, &size)) in rwp_unpaged_iv_alloc()
378 SMALL_PAGE_SIZE) / SMALL_PAGE_SIZE; in rwp_init()
474 memcpy(va, src, SMALL_PAGE_SIZE); in rop_load_page_helper()
578 pg_idx = r / SMALL_PAGE_SIZE; in init_rels()
584 rrp->relocs[reln] = r - pg_idx * SMALL_PAGE_SIZE; in init_rels()
[all …]
A Dmobj.c67 if (granule != SMALL_PAGE_SIZE && in mobj_phys_get_pa()
378 if (granule != SMALL_PAGE_SIZE && in mobj_shm_get_pa()
457 m->mobj.phys_granule = SMALL_PAGE_SIZE; in mobj_shm_alloc()
559 m->fobj = fobj_rw_paged_alloc(ROUNDUP(size, SMALL_PAGE_SIZE) / in mobj_seccpy_shm_alloc()
560 SMALL_PAGE_SIZE); in mobj_seccpy_shm_alloc()
606 m->mobj.size = fobj->num_pages * SMALL_PAGE_SIZE; in mobj_with_fobj_alloc()
607 m->mobj.phys_granule = SMALL_PAGE_SIZE; in mobj_with_fobj_alloc()
674 p = f->fobj->ops->get_pa(f->fobj, offs / SMALL_PAGE_SIZE) + in mobj_with_fobj_get_pa()
675 offs % SMALL_PAGE_SIZE; in mobj_with_fobj_get_pa()
678 if (granule != SMALL_PAGE_SIZE && in mobj_with_fobj_get_pa()
A Dcore_mmu.c908 mmap[pos].region_size = SMALL_PAGE_SIZE; in add_pager_vaspace()
972 map->region_size = SMALL_PAGE_SIZE; in assign_mem_granularity()
977 map->region_size = SMALL_PAGE_SIZE; in assign_mem_granularity()
1173 vaddr_t end = ROUNDUP(id_map_end, SMALL_PAGE_SIZE); in mem_map_add_id_map()
1192 .region_size = SMALL_PAGE_SIZE, in mem_map_add_id_map()
1343 SMALL_PAGE_SIZE); in core_init_mmu_map()
1357 .region_size = SMALL_PAGE_SIZE, in core_init_mmu_map()
1619 r.size = ROUNDUP(r.size, SMALL_PAGE_SIZE); in set_pg_region()
1770 vaddr += SMALL_PAGE_SIZE; in core_mmu_map_pages()
1843 paddr += SMALL_PAGE_SIZE; in core_mmu_map_contiguous_pages()
[all …]
/optee_os-3.20.0/core/arch/arm/plat-aspeed/
A Dplatform_ast2600.c46 register_phys_mem(MEM_AREA_IO_NSEC, CONSOLE_UART_BASE, SMALL_PAGE_SIZE);
49 register_phys_mem(MEM_AREA_IO_SEC, AHBC_BASE, SMALL_PAGE_SIZE);
50 register_phys_mem(MEM_AREA_IO_NSEC, SCU_BASE, SMALL_PAGE_SIZE);
93 MEM_AREA_IO_SEC, SMALL_PAGE_SIZE); in plat_primary_init_early()
/optee_os-3.20.0/core/kernel/
A Dmsg_param.c82 va = mobj_get_va(mobj, 0, SMALL_PAGE_SIZE); in msg_param_extract_pages()
101 va = mobj_get_va(mobj, 0, SMALL_PAGE_SIZE); in msg_param_extract_pages()
128 num_pages = (size_plus_offs - 1) / SMALL_PAGE_SIZE + 1; in msg_param_mobj_from_noncontig()
/optee_os-3.20.0/core/arch/arm/include/mm/
A Dgeneric_ram_layout.h162 SMALL_PAGE_SIZE)
167 SMALL_PAGE_SIZE) - TA_RAM_START)
/optee_os-3.20.0/core/arch/riscv/include/mm/
A Dgeneric_ram_layout.h163 SMALL_PAGE_SIZE)
168 SMALL_PAGE_SIZE) - TA_RAM_START)
/optee_os-3.20.0/core/drivers/
A Dversal_sha3_384.c29 len = MIN(src_len, SMALL_PAGE_SIZE); in input_plaintext()
31 versal_mbox_alloc(len, src + i * SMALL_PAGE_SIZE, &p); in input_plaintext()
/optee_os-3.20.0/core/drivers/crypto/aspeed/
A Dcrypto_ast2600.c28 scu_virt = core_mmu_get_va(SCU_BASE, MEM_AREA_IO_NSEC, SMALL_PAGE_SIZE); in crypto_ast2600_init()
/optee_os-3.20.0/core/drivers/crypto/caam/utils/
A Dutils_mem.c253 if (buf->length > SMALL_PAGE_SIZE) { in caam_mem_get_pa_area()
254 nb_pa_area = buf->length / SMALL_PAGE_SIZE + 1; in caam_mem_get_pa_area()
255 if (buf->length % SMALL_PAGE_SIZE) in caam_mem_get_pa_area()
289 MIN(SMALL_PAGE_SIZE - (va & SMALL_PAGE_MASK), len); in caam_mem_get_pa_area()
/optee_os-3.20.0/core/arch/arm/plat-rcar/
A Dmain.c43 register_phys_mem_pgdir(MEM_AREA_IO_SEC, PRR_BASE, SMALL_PAGE_SIZE);
/optee_os-3.20.0/ldelf/
A Dta_elf.c501 elf->max_offs = SMALL_PAGE_SIZE; in init_elf()
517 if (sz > SMALL_PAGE_SIZE) in init_elf()
525 return ROUNDUP(v, SMALL_PAGE_SIZE); in roundup()
530 return ROUNDDOWN(v, SMALL_PAGE_SIZE); in rounddown()
769 return min * SMALL_PAGE_SIZE; in get_pad_begin()
774 return (min + rnd) * SMALL_PAGE_SIZE; in get_pad_begin()
837 SMALL_PAGE_SIZE; in populate_segments()
848 vaddr += SMALL_PAGE_SIZE; in populate_segments()
849 filesz -= SMALL_PAGE_SIZE; in populate_segments()
850 memsz -= SMALL_PAGE_SIZE; in populate_segments()
[all …]
A Dsys.h26 #define SMALL_PAGE_SIZE 0x00001000 macro
/optee_os-3.20.0/core/arch/arm/plat-marvell/
A Dmain.c67 SMALL_PAGE_SIZE);

Completed in 62 milliseconds

12