Lines Matching refs:v_addr
276 static void _kenrel_unmap_4K(unsigned long *lv0_tbl, void *v_addr) in _kenrel_unmap_4K() argument
278 size_t loop_va = (size_t)v_addr & ~ARCH_PAGE_MASK; in _kenrel_unmap_4K()
311 static int _kenrel_map_4K(unsigned long *lv0_tbl, void *v_addr, void *p_addr, in _kenrel_map_4K() argument
314 size_t loop_va = (size_t)v_addr & ~ARCH_PAGE_MASK; in _kenrel_map_4K()
358 void *rt_hw_mmu_map(rt_aspace_t aspace, void *v_addr, void *p_addr, size_t size, in rt_hw_mmu_map() argument
362 void *unmap_va = v_addr; in rt_hw_mmu_map()
368 ret = _kenrel_map_4K(aspace->page_table, v_addr, p_addr, attr); in rt_hw_mmu_map()
372 while (unmap_va != v_addr) in rt_hw_mmu_map()
382 v_addr += ARCH_PAGE_SIZE; in rt_hw_mmu_map()
388 return v_addr; in rt_hw_mmu_map()
394 void rt_hw_mmu_unmap(rt_aspace_t aspace, void *v_addr, size_t size) in rt_hw_mmu_unmap() argument
407 _kenrel_unmap_4K(aspace->page_table, v_addr); in rt_hw_mmu_unmap()
410 v_addr += ARCH_PAGE_SIZE; in rt_hw_mmu_unmap()
427 void *rt_hw_mmu_v2p(rt_aspace_t aspace, void* v_addr) in rt_hw_mmu_v2p() argument
434 l1_off = (size_t)v_addr >> ARCH_SECTION_SHIFT; in rt_hw_mmu_v2p()
448 l2_off = (((size_t)v_addr & ARCH_SECTION_MASK) >> ARCH_PAGE_SHIFT); in rt_hw_mmu_v2p()
458 pa += ((size_t)v_addr & ARCH_PAGE_MASK); in rt_hw_mmu_v2p()
471 pa += ((size_t)v_addr & ARCH_SECTION_MASK); in rt_hw_mmu_v2p()