Lines Matching refs:start

320 		struct vm_area_struct *vma, unsigned long start,  in allow_mlock_munlock()  argument
340 if (!folio_within_range(folio, vma, start, end)) in allow_mlock_munlock()
360 unsigned long start = addr; in mlock_pte_range() local
393 if (!allow_mlock_munlock(folio, vma, start, end, step)) in mlock_pte_range()
424 unsigned long start, unsigned long end, vm_flags_t newflags) in mlock_vma_pages_range() argument
448 walk_page_range(vma->vm_mm, start, end, &mlock_walk_ops, NULL); in mlock_vma_pages_range()
467 struct vm_area_struct **prev, unsigned long start, in mlock_fixup() argument
481 vma = vma_modify_flags(vmi, *prev, vma, start, end, newflags); in mlock_fixup()
490 nr_pages = (end - start) >> PAGE_SHIFT; in mlock_fixup()
507 mlock_vma_pages_range(vma, start, end, newflags); in mlock_fixup()
514 static int apply_vma_lock_flags(unsigned long start, size_t len, in apply_vma_lock_flags() argument
519 VMA_ITERATOR(vmi, current->mm, start); in apply_vma_lock_flags()
521 VM_BUG_ON(offset_in_page(start)); in apply_vma_lock_flags()
523 end = start + len; in apply_vma_lock_flags()
524 if (end < start) in apply_vma_lock_flags()
526 if (end == start) in apply_vma_lock_flags()
533 if (start > vma->vm_start) in apply_vma_lock_flags()
536 nstart = start; in apply_vma_lock_flags()
572 unsigned long start, size_t len) in count_mm_mlocked_page_nr() argument
577 VMA_ITERATOR(vmi, mm, start); in count_mm_mlocked_page_nr()
580 if (unlikely(ULONG_MAX - len < start)) in count_mm_mlocked_page_nr()
583 end = start + len; in count_mm_mlocked_page_nr()
587 if (start > vma->vm_start) in count_mm_mlocked_page_nr()
588 count -= (start - vma->vm_start); in count_mm_mlocked_page_nr()
612 static __must_check int do_mlock(unsigned long start, size_t len, vm_flags_t flags) in do_mlock() argument
618 start = untagged_addr(start); in do_mlock()
623 len = PAGE_ALIGN(len + (offset_in_page(start))); in do_mlock()
624 start &= PAGE_MASK; in do_mlock()
642 start, len); in do_mlock()
647 error = apply_vma_lock_flags(start, len, flags); in do_mlock()
653 error = __mm_populate(start, len, 0); in do_mlock()
659 SYSCALL_DEFINE2(mlock, unsigned long, start, size_t, len) in SYSCALL_DEFINE2() argument
661 return do_mlock(start, len, VM_LOCKED); in SYSCALL_DEFINE2()
664 SYSCALL_DEFINE3(mlock2, unsigned long, start, size_t, len, int, flags) in SYSCALL_DEFINE3() argument
674 return do_mlock(start, len, vm_flags); in SYSCALL_DEFINE3()
677 SYSCALL_DEFINE2(munlock, unsigned long, start, size_t, len) in SYSCALL_DEFINE2() argument
681 start = untagged_addr(start); in SYSCALL_DEFINE2()
683 len = PAGE_ALIGN(len + (offset_in_page(start))); in SYSCALL_DEFINE2()
684 start &= PAGE_MASK; in SYSCALL_DEFINE2()
688 ret = apply_vma_lock_flags(start, len, 0); in SYSCALL_DEFINE2()