Lines Matching refs:dst_start

100 					      unsigned long dst_start,  in uffd_mfill_lock()  argument
105 dst_vma = uffd_lock_vma(dst_mm, dst_start); in uffd_mfill_lock()
106 if (IS_ERR(dst_vma) || validate_dst_vma(dst_vma, dst_start + len)) in uffd_mfill_lock()
121 unsigned long dst_start, in uffd_mfill_lock() argument
127 dst_vma = find_vma_and_prepare_anon(dst_mm, dst_start); in uffd_mfill_lock()
131 if (validate_dst_vma(dst_vma, dst_start + len)) in uffd_mfill_lock()
491 unsigned long dst_start, in mfill_atomic_hugetlb() argument
520 dst_addr = dst_start; in mfill_atomic_hugetlb()
529 if (dst_start & (vma_hpagesize - 1) || len & (vma_hpagesize - 1)) in mfill_atomic_hugetlb()
538 dst_vma = uffd_mfill_lock(dst_mm, dst_start, len); in mfill_atomic_hugetlb()
564 VM_WARN_ON_ONCE(dst_addr >= dst_start + len); in mfill_atomic_hugetlb()
647 unsigned long dst_start,
698 unsigned long dst_start, in mfill_atomic() argument
714 VM_WARN_ON_ONCE(dst_start & ~PAGE_MASK); in mfill_atomic()
719 VM_WARN_ON_ONCE(dst_start + len <= dst_start); in mfill_atomic()
722 dst_addr = dst_start; in mfill_atomic()
730 dst_vma = uffd_mfill_lock(dst_mm, dst_start, len); in mfill_atomic()
766 return mfill_atomic_hugetlb(ctx, dst_vma, dst_start, in mfill_atomic()
778 VM_WARN_ON_ONCE(dst_addr >= dst_start + len); in mfill_atomic()
861 ssize_t mfill_atomic_copy(struct userfaultfd_ctx *ctx, unsigned long dst_start, in mfill_atomic_copy() argument
865 return mfill_atomic(ctx, dst_start, src_start, len, in mfill_atomic_copy()
1529 unsigned long dst_start, in find_vmas_mm_locked() argument
1537 vma = find_vma_and_prepare_anon(mm, dst_start); in find_vmas_mm_locked()
1556 unsigned long dst_start, in uffd_move_lock() argument
1564 vma = uffd_lock_vma(mm, dst_start); in uffd_move_lock()
1598 err = find_vmas_mm_locked(mm, dst_start, src_start, dst_vmap, src_vmap); in uffd_move_lock()
1632 unsigned long dst_start, in uffd_move_lock() argument
1640 err = find_vmas_mm_locked(mm, dst_start, src_start, dst_vmap, src_vmap); in uffd_move_lock()
1730 ssize_t move_pages(struct userfaultfd_ctx *ctx, unsigned long dst_start, in move_pages() argument
1742 VM_WARN_ON_ONCE(dst_start & ~PAGE_MASK); in move_pages()
1747 VM_WARN_ON_ONCE(dst_start + len < dst_start); in move_pages()
1749 err = uffd_move_lock(mm, dst_start, src_start, &dst_vma, &src_vma); in move_pages()
1771 if (dst_start + len > dst_vma->vm_end) in move_pages()
1778 for (src_addr = src_start, dst_addr = dst_start; in move_pages()