Searched refs:map_size (Results 1 – 9 of 9) sorted by relevance
| /components/lwp/ |
| A D | lwp_user_mm.c | 204 map_size, ret); in _lwp_map_user() 240 if (!map_size) in lwp_map_user() 245 map_size += (offset + ARCH_PAGE_SIZE - 1); in lwp_map_user() 246 map_size &= ~ARCH_PAGE_MASK; in lwp_map_user() 305 map_size, ret); in _lwp_map_user_varea() 315 if (!map_size) in _map_user_varea_ext() 320 map_size += (offset + ARCH_PAGE_SIZE - 1); in _map_user_varea_ext() 321 map_size &= ~ARCH_PAGE_MASK; in _map_user_varea_ext() 344 if (!map_size) in lwp_map_user_phy() 358 map_size &= ~ARCH_PAGE_MASK; in lwp_map_user_phy() [all …]
|
| A D | lwp_elf.c | 74 rt_ubase_t map_size; /* Total size required for memory mapping */ member 172 rt_ubase_t map_size, size_t prot, size_t flags, rt_ubase_t offset) in file_mmap() argument 577 if (map_size != 0) in elf_map() 579 size = map_size; in elf_map() 620 rt_uint32_t map_size, rt_ubase_t *load_base) in elf_file_mmap() argument 670 map_size = 0; in elf_file_mmap() 722 LOG_D("%s : total_mapping_size 0x%x", __func__, load_info->interp_info.map_size); in load_elf_interp() 725 load_info->interp_info.map_size, &load_base); in load_elf_interp() 818 load_info->exec_info.map_size = 0; /* Total size of the mapped segments. */ in elf_load_segment() 828 LOG_D("%s : map_size : 0x%x", __func__, load_info->exec_info.map_size); in elf_load_segment() [all …]
|
| A D | lwp_user_mm.h | 152 void *lwp_map_user(struct rt_lwp *lwp, void *map_va, size_t map_size, rt_bool_t text); 154 rt_varea_t lwp_map_user_varea(struct rt_lwp *lwp, void *map_va, size_t map_size); 157 rt_varea_t lwp_map_user_varea_ext(struct rt_lwp *lwp, void *map_va, size_t map_size, size_t flags); 159 void *lwp_map_user_phy(struct rt_lwp *lwp, void *map_va, void *map_pa, size_t map_size, rt_bool_t c…
|
| /components/dfs/dfs_v2/src/ |
| A D | dfs_file_mmap.c | 55 static void *_do_mmap(struct rt_lwp *lwp, void *map_vaddr, size_t map_size, size_t attr, in _do_mmap() argument 62 ret = rt_aspace_map(lwp->aspace, &vaddr, map_size, in _do_mmap() 68 map_size, ret); in _do_mmap() 97 size_t map_size = mmap2->length; in _map_data_to_uspace() local 102 if (map_size) in _map_data_to_uspace() 105 map_size += (offset + ARCH_PAGE_SIZE - 1); in _map_data_to_uspace() 106 map_size &= ~ARCH_PAGE_MASK; in _map_data_to_uspace() 113 … map_vaddr = _do_mmap(lwp, map_vaddr, map_size, k_attr, k_flags, mmap2->pgoffset, data, code); in _map_data_to_uspace()
|
| /components/mm/ |
| A D | mm_aspace.c | 328 map_off + (map_size >> MM_PAGE_SHIFT) == n_off; in _contiguous_offset() 331 static rt_bool_t _compatible(rt_varea_t neighbour, rt_size_t map_size, in _compatible() argument 336 _contiguous_offset(neighbour, map_size, prop)); in _compatible() 424 size_t operand_size = hint->map_size; in _try_expand_and_merge_okay() 469 varea = _varea_create(hint->prefer, hint->map_size); in _insert_new_varea() 520 err = _unmap_range_locked(aspace, alloc_va, hint->map_size); in _varea_install() 592 .map_size = length, in _mm_aspace_map() 759 if (!aspace || !hint || !hint->limit_range_size || !hint->map_size) in _mm_aspace_map_phy() 764 else if (_not_align(hint->prefer, hint->map_size, ARCH_PAGE_MASK)) in _mm_aspace_map_phy() 847 varea->size = hint->map_size; in rt_aspace_map_phy_static() [all …]
|
| A D | mm_kmem.c | 64 .map_size = RT_ALIGN(length + lo_off, ARCH_PAGE_SIZE)}; in rt_kmem_map_phy()
|
| A D | ioremap.c | 48 .map_size = RT_ALIGN(size + lo_off, RT_PAGE_AFFINITY_BLOCK_SIZE), in _ioremap_type()
|
| A D | mm_aspace.h | 82 const rt_size_t map_size; member
|
| A D | mm_memblock.c | 389 .map_size = reg.end - reg.start, in rt_memblock_setup_memory_environment()
|
Completed in 16 milliseconds