Lines Matching refs:end
21 unsigned long end, in migrate_vma_collect_skip() argument
27 for (addr = start; addr < end; addr += PAGE_SIZE) { in migrate_vma_collect_skip()
36 unsigned long end, in migrate_vma_collect_hole() argument
45 return migrate_vma_collect_skip(start, end, walk); in migrate_vma_collect_hole()
47 for (addr = start; addr < end; addr += PAGE_SIZE) { in migrate_vma_collect_hole()
59 unsigned long end, in migrate_vma_collect_pmd() argument
73 return migrate_vma_collect_hole(start, end, -1, walk); in migrate_vma_collect_pmd()
95 return migrate_vma_collect_skip(start, end, in migrate_vma_collect_pmd()
98 return migrate_vma_collect_skip(start, end, in migrate_vma_collect_pmd()
105 return migrate_vma_collect_skip(start, end, in migrate_vma_collect_pmd()
115 for (; addr < end; addr += PAGE_SIZE, ptep++) { in migrate_vma_collect_pmd()
288 flush_tlb_range(walk->vma, start, end); in migrate_vma_collect_pmd()
320 migrate->vma->vm_mm, migrate->start, migrate->end, in migrate_vma_collect()
324 walk_page_range(migrate->vma->vm_mm, migrate->start, migrate->end, in migrate_vma_collect()
328 migrate->end = migrate->start + (migrate->npages << PAGE_SHIFT); in migrate_vma_collect()
538 long nr_pages = (args->end - args->start) >> PAGE_SHIFT; in migrate_vma_setup()
541 args->end &= PAGE_MASK; in migrate_vma_setup()
550 if (args->end <= args->vma->vm_start || args->end > args->vma->vm_end) in migrate_vma_setup()
740 migrate->vma->vm_mm, addr, migrate->end, in __migrate_device_pages()