Lines Matching refs:end
94 static void vmemmap_mark_sub_pmd_used(unsigned long start, unsigned long end) in vmemmap_mark_sub_pmd_used() argument
105 static void vmemmap_use_sub_pmd(unsigned long start, unsigned long end) in vmemmap_use_sub_pmd() argument
112 unused_sub_pmd_start = end; in vmemmap_use_sub_pmd()
118 vmemmap_mark_sub_pmd_used(start, end); in vmemmap_use_sub_pmd()
121 static void vmemmap_use_new_sub_pmd(unsigned long start, unsigned long end) in vmemmap_use_new_sub_pmd() argument
128 vmemmap_mark_sub_pmd_used(start, end); in vmemmap_use_new_sub_pmd()
138 if (!IS_ALIGNED(end, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
139 unused_sub_pmd_start = end; in vmemmap_use_new_sub_pmd()
143 static bool vmemmap_unuse_sub_pmd(unsigned long start, unsigned long end) in vmemmap_unuse_sub_pmd() argument
148 memset((void *)start, PAGE_UNUSED, end - start); in vmemmap_unuse_sub_pmd()
154 unsigned long end, bool add, bool direct) in modify_pte_table() argument
165 for (; addr < end; addr += PAGE_SIZE, pte++) { in modify_pte_table()
211 unsigned long end, bool add, bool direct) in modify_pmd_table() argument
223 for (; addr < end; addr = next, pmd++) { in modify_pmd_table()
224 next = pmd_addr_end(addr, end); in modify_pmd_table()
293 const unsigned long end = start + PUD_SIZE; in try_free_pmd_table() local
298 if (end > VMALLOC_START) in try_free_pmd_table()
301 if (start < KASAN_SHADOW_END && KASAN_SHADOW_START > end) in try_free_pmd_table()
312 static int modify_pud_table(p4d_t *p4d, unsigned long addr, unsigned long end, in modify_pud_table() argument
324 for (; addr < end; addr = next, pud++) { in modify_pud_table()
325 next = pud_addr_end(addr, end); in modify_pud_table()
368 const unsigned long end = start + P4D_SIZE; in try_free_pud_table() local
373 if (end > VMALLOC_START) in try_free_pud_table()
376 if (start < KASAN_SHADOW_END && KASAN_SHADOW_START > end) in try_free_pud_table()
389 static int modify_p4d_table(pgd_t *pgd, unsigned long addr, unsigned long end, in modify_p4d_table() argument
398 for (; addr < end; addr = next, p4d++) { in modify_p4d_table()
399 next = p4d_addr_end(addr, end); in modify_p4d_table()
422 const unsigned long end = start + PGDIR_SIZE; in try_free_p4d_table() local
427 if (end > VMALLOC_START) in try_free_p4d_table()
430 if (start < KASAN_SHADOW_END && KASAN_SHADOW_START > end) in try_free_p4d_table()
443 static int modify_pagetable(unsigned long start, unsigned long end, bool add, in modify_pagetable() argument
451 if (WARN_ON_ONCE(!PAGE_ALIGNED(start | end))) in modify_pagetable()
453 for (addr = start; addr < end; addr = next) { in modify_pagetable()
454 next = pgd_addr_end(addr, end); in modify_pagetable()
475 flush_tlb_kernel_range(start, end); in modify_pagetable()
479 static int add_pagetable(unsigned long start, unsigned long end, bool direct) in add_pagetable() argument
481 return modify_pagetable(start, end, true, direct); in add_pagetable()
484 static int remove_pagetable(unsigned long start, unsigned long end, bool direct) in remove_pagetable() argument
486 return modify_pagetable(start, end, false, direct); in remove_pagetable()
508 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node, in vmemmap_populate() argument
515 ret = add_pagetable(start, end, false); in vmemmap_populate()
517 remove_pagetable(start, end, false); in vmemmap_populate()
522 void vmemmap_free(unsigned long start, unsigned long end, in vmemmap_free() argument
526 remove_pagetable(start, end, false); in vmemmap_free()
542 mhp_range.end = VMEM_MAX_PHYS - 1; in arch_get_mappable_range()
552 start + size > range.end + 1 || in vmem_add_mapping()
571 phys_addr_t base, end; in vmem_map_init() local
574 for_each_mem_range(i, &base, &end) in vmem_map_init()
575 vmem_add_range(base, end - base); in vmem_map_init()