Searched refs:src_vma (Results 1 – 4 of 4) sorted by relevance
| /mm/ |
| A D | userfaultfd.c | 1031 struct vm_area_struct *src_vma, in move_present_pte() argument 1148 struct vm_area_struct *src_vma, in move_zeropage_pte() argument 1181 struct vm_area_struct *src_vma, in move_pages_pte() argument 1508 if (!(src_vma->vm_flags & VM_WRITE)) in validate_move_areas() 1624 vma_end_read(src_vma); in uffd_move_unlock() 1625 if (src_vma != dst_vma) in uffd_move_unlock() 1649 mmap_assert_locked(src_vma->vm_mm); in uffd_move_unlock() 1764 if (src_vma->vm_flags & VM_SHARED) in move_pages() 1843 dst_pmdval, dst_vma, src_vma, in move_pages() 1864 dst_vma, src_vma, in move_pages() [all …]
|
| A D | memory.c | 965 if (src_vma->vm_flags & VM_SHARED) in __copy_present_ptes() 1004 if (!(src_vma->vm_flags & VM_SHARED)) in copy_present_ptes() 1006 if (vma_soft_dirty_enabled(src_vma)) in copy_present_ptes() 1013 nr, dst_vma, src_vma))) { in copy_present_ptes() 1152 dst_vma, src_vma, in copy_pte_range() 1252 addr, dst_vma, src_vma); in copy_pmd_range() 1349 if (src_vma->anon_vma) in vma_needs_copy() 1366 unsigned long end = src_vma->vm_end; in copy_page_range() 1374 if (!vma_needs_copy(dst_vma, src_vma)) in copy_page_range() 1377 if (is_vm_hugetlb_page(src_vma)) in copy_page_range() [all …]
|
| A D | hugetlb.c | 5553 struct vm_area_struct *src_vma) in copy_hugetlb_page_range() argument 5558 bool cow = is_cow_mapping(src_vma->vm_flags); in copy_hugetlb_page_range() 5559 struct hstate *h = hstate_vma(src_vma); in copy_hugetlb_page_range() 5568 src_vma->vm_start, in copy_hugetlb_page_range() 5569 src_vma->vm_end); in copy_hugetlb_page_range() 5571 vma_assert_write_locked(src_vma); in copy_hugetlb_page_range() 5580 hugetlb_vma_lock_read(src_vma); in copy_hugetlb_page_range() 5584 for (addr = src_vma->vm_start; addr < src_vma->vm_end; addr += sz) { in copy_hugetlb_page_range() 5586 src_pte = hugetlb_walk(src_vma, addr, sz); in copy_hugetlb_page_range() 5636 if (userfaultfd_wp(src_vma) && uffd_wp) in copy_hugetlb_page_range() [all …]
|
| A D | huge_memory.c | 1668 struct vm_area_struct *dst_vma, struct vm_area_struct *src_vma) in copy_huge_pmd() argument 1691 VM_WARN_ON_ONCE(is_cow_mapping(src_vma->vm_flags) && pmd_write(pmd)); in copy_huge_pmd() 1760 if (unlikely(folio_try_dup_anon_rmap_pmd(src_folio, src_page, dst_vma, src_vma))) { in copy_huge_pmd() 1766 __split_huge_pmd(src_vma, src_pmd, addr, false); in copy_huge_pmd() 2538 struct vm_area_struct *dst_vma, struct vm_area_struct *src_vma, in move_pages_huge_pmd() argument 2554 vma_assert_locked(src_vma); in move_pages_huge_pmd() 2588 flush_cache_range(src_vma, src_addr, src_addr + HPAGE_PMD_SIZE); in move_pages_huge_pmd() 2630 src_pmdval = pmdp_huge_clear_flush(src_vma, src_addr, src_pmd); in move_pages_huge_pmd() 2645 src_pmdval = pmdp_huge_clear_flush(src_vma, src_addr, src_pmd); in move_pages_huge_pmd()
|
Completed in 43 milliseconds