Lines Matching refs:end

161 static int vmemmap_remap_range(unsigned long start, unsigned long end,  in vmemmap_remap_range()  argument
166 VM_BUG_ON(!PAGE_ALIGNED(start | end)); in vmemmap_remap_range()
169 ret = walk_kernel_page_table_range(start, end, &vmemmap_remap_ops, in vmemmap_remap_range()
176 flush_tlb_kernel_range(start, end); in vmemmap_remap_range()
289 static int vmemmap_remap_split(unsigned long start, unsigned long end, in vmemmap_remap_split() argument
300 return vmemmap_remap_range(reuse, end, &walk); in vmemmap_remap_split()
318 static int vmemmap_remap_free(unsigned long start, unsigned long end, in vmemmap_remap_free() argument
364 ret = vmemmap_remap_range(reuse, end, &walk); in vmemmap_remap_free()
366 end = reuse + walk.nr_walked * PAGE_SIZE; in vmemmap_remap_free()
380 vmemmap_remap_range(reuse, end, &walk); in vmemmap_remap_free()
386 static int alloc_vmemmap_page_list(unsigned long start, unsigned long end, in alloc_vmemmap_page_list() argument
390 unsigned long nr_pages = (end - start) >> PAGE_SHIFT; in alloc_vmemmap_page_list()
423 static int vmemmap_remap_alloc(unsigned long start, unsigned long end, in vmemmap_remap_alloc() argument
437 if (alloc_vmemmap_page_list(start, end, &vmemmap_pages)) in vmemmap_remap_alloc()
440 return vmemmap_remap_range(reuse, end, &walk); in vmemmap_remap_alloc()
792 unsigned long start, end; in hugetlb_vmemmap_init_early() local
819 end = start + nr_pages * sizeof(struct page); in hugetlb_vmemmap_init_early()
821 if (vmemmap_populate_hvo(start, end, nid, in hugetlb_vmemmap_init_early()
844 unsigned long phys, nr_pages, start, end; in hugetlb_vmemmap_init_late() local
874 end = start + nr_pages * sizeof(struct page); in hugetlb_vmemmap_init_late()
876 vmemmap_undo_hvo(start, end, nid, in hugetlb_vmemmap_init_late()
878 nr_mmap = end - start - HUGETLB_VMEMMAP_RESERVE_SIZE; in hugetlb_vmemmap_init_late()