| /mm/ |
| A D | maccess.c | 22 __get_kernel_nofault(dst, src, type, err_label); \ 24 dst += sizeof(type); \ 34 align = (unsigned long)dst | (unsigned long)src; in copy_from_kernel_nofault() 58 instrument_write(dst, sizeof(type)); \ 59 dst += sizeof(type); \ 97 __get_kernel_nofault(dst, src, u8, Efault); in strncpy_from_kernel_nofault() 98 dst++; in strncpy_from_kernel_nofault() 103 dst[-1] = '\0'; in strncpy_from_kernel_nofault() 107 dst[0] = '\0'; in strncpy_from_kernel_nofault() 153 if (access_ok(dst, size)) { in copy_to_user_nofault() [all …]
|
| A D | migrate.c | 710 folio_ref_add(dst, folio_nr_pages(dst)); in migrate_huge_page_move_mapping() 1061 VM_BUG_ON_FOLIO(!folio_test_locked(dst), dst); in move_to_new_folio() 1150 folio_put(dst); in migrate_folio_undo_dst() 1172 struct folio *dst; in migrate_folio_unmap() local 1190 if (!dst) in migrate_folio_unmap() 1192 *dstp = dst; in migrate_folio_unmap() 1376 folio_put(dst); in migrate_folio_move() 1444 if (!dst) in unmap_and_move_huge_page() 1534 folio_put(dst); in unmap_and_move_huge_page() 1722 dst = dst2; in migrate_folios_move() [all …]
|
| A D | migrate_device.c | 839 if (dst) { in __migrate_device_finalize() 841 folio_unlock(dst); in __migrate_device_finalize() 842 folio_put(dst); in __migrate_device_finalize() 850 if (dst) { in __migrate_device_finalize() 852 folio_unlock(dst); in __migrate_device_finalize() 853 folio_put(dst); in __migrate_device_finalize() 855 dst = src; in __migrate_device_finalize() 859 folio_add_lru(dst); in __migrate_device_finalize() 865 if (dst != src) { in __migrate_device_finalize() 867 folio_unlock(dst); in __migrate_device_finalize() [all …]
|
| A D | list_lru.c | 448 struct list_lru_one *dst; in memcg_reparent_list_lru_one() local 451 dst = list_lru_from_memcg_idx(lru, nid, dst_idx); in memcg_reparent_list_lru_one() 452 spin_lock_nested(&dst->lock, SINGLE_DEPTH_NESTING); in memcg_reparent_list_lru_one() 454 list_splice_init(&src->list, &dst->list); in memcg_reparent_list_lru_one() 457 dst->nr_items += src->nr_items; in memcg_reparent_list_lru_one() 463 spin_unlock(&dst->lock); in memcg_reparent_list_lru_one()
|
| A D | compaction.c | 1800 struct folio *dst; in compaction_alloc_noprof() local 1834 dst = (struct folio *)freepage; in compaction_alloc_noprof() 1836 post_alloc_hook(&dst->page, order, __GFP_MOVABLE); in compaction_alloc_noprof() 1837 set_page_refcounted(&dst->page); in compaction_alloc_noprof() 1839 prep_compound_page(&dst->page, order); in compaction_alloc_noprof() 1842 return page_rmappable_folio(&dst->page); in compaction_alloc_noprof() 1855 static void compaction_free(struct folio *dst, unsigned long data) in compaction_free() argument 1858 int order = folio_order(dst); in compaction_free() 1859 struct page *page = &dst->page; in compaction_free() 1861 if (folio_put_testzero(dst)) { in compaction_free() [all …]
|
| A D | numa_memblks.c | 180 static void __init numa_move_tail_memblk(struct numa_meminfo *dst, int idx, in numa_move_tail_memblk() argument 183 dst->blk[dst->nr_blks++] = src->blk[idx]; in numa_move_tail_memblk()
|
| A D | hugetlb.c | 4013 mutex_lock(&dst->resize_lock); in demote_free_hugetlb_folios() 4035 prep_compound_page(page, dst->order); in demote_free_hugetlb_folios() 4038 init_new_hugetlb_folio(dst, new_folio); in demote_free_hugetlb_folios() 4046 prep_and_add_allocated_folios(dst, &dst_list); in demote_free_hugetlb_folios() 4048 mutex_unlock(&dst->resize_lock); in demote_free_hugetlb_folios() 4058 struct hstate *dst; in demote_pool_huge_page() local 4108 dst->max_huge_pages += nr_demoted << (huge_page_order(src) - huge_page_order(dst)); in demote_pool_huge_page() 5610 dst_ptl = huge_pte_lock(h, dst, dst_pte); in copy_hugetlb_page_range() 5648 set_huge_pte_at(dst, addr, dst_pte, in copy_hugetlb_page_range() 5689 dst_ptl = huge_pte_lock(h, dst, dst_pte); in copy_hugetlb_page_range() [all …]
|
| A D | util.c | 720 void folio_copy(struct folio *dst, struct folio *src) in folio_copy() argument 726 copy_highpage(folio_page(dst, i), folio_page(src, i)); in folio_copy() 734 int folio_mc_copy(struct folio *dst, struct folio *src) in folio_mc_copy() argument 740 if (copy_mc_highpage(folio_page(dst, i), folio_page(src, i))) in folio_mc_copy()
|
| A D | rmap.c | 280 int anon_vma_clone(struct vm_area_struct *dst, struct vm_area_struct *src) in anon_vma_clone() argument 298 anon_vma_chain_link(dst, avc, anon_vma); in anon_vma_clone() 307 if (!dst->anon_vma && src->anon_vma && in anon_vma_clone() 310 dst->anon_vma = anon_vma; in anon_vma_clone() 312 if (dst->anon_vma) in anon_vma_clone() 313 dst->anon_vma->num_active_vmas++; in anon_vma_clone() 324 dst->anon_vma = NULL; in anon_vma_clone() 325 unlink_anon_vmas(dst); in anon_vma_clone()
|
| A D | khugepaged.c | 1852 struct page *dst; in collapse_file() local 2102 dst = folio_page(new_folio, 0); in collapse_file() 2107 clear_highpage(dst); in collapse_file() 2109 dst++; in collapse_file() 2113 if (copy_mc_highpage(dst, folio_page(folio, i)) > 0) { in collapse_file() 2118 dst++; in collapse_file() 2122 clear_highpage(dst); in collapse_file() 2124 dst++; in collapse_file()
|
| A D | zswap.c | 950 u8 *dst; in zswap_compress() local 953 dst = acomp_ctx->buffer; in zswap_compress() 962 sg_init_one(&output, dst, PAGE_SIZE * 2); in zswap_compress() 988 zpool_obj_write(zpool, handle, dst, dlen); in zswap_compress()
|
| A D | zsmalloc.c | 1243 void *dst = kmap_local_zpdesc(zpdesc); in zs_obj_write() local 1245 memcpy(dst + off, handle_mem, mem_len); in zs_obj_write() 1246 kunmap_local(dst); in zs_obj_write() 1459 static void zs_object_copy(struct size_class *class, unsigned long dst, in zs_object_copy() argument 1472 obj_to_location(dst, &d_zpdesc, &d_objidx); in zs_object_copy()
|
| A D | memory.c | 3243 if (copy_mc_user_highpage(dst, src, addr, vma)) in __wp_page_copy_user() 3254 kaddr = kmap_local_page(dst); in __wp_page_copy_user() 3325 flush_dcache_page(dst); in __wp_page_copy_user() 7065 unsigned long addr = ALIGN_DOWN(addr_hint, folio_size(dst)); in copy_user_gigantic_page() 7071 dst_page = folio_page(dst, i); in copy_user_gigantic_page() 7083 struct folio *dst; member 7091 struct page *dst = folio_page(copy_arg->dst, idx); in copy_subpage() local 7094 if (copy_mc_user_highpage(dst, src, addr, copy_arg->vma)) in copy_subpage() 7099 int copy_user_large_folio(struct folio *dst, struct folio *src, in copy_user_large_folio() argument 7102 unsigned int nr_pages = folio_nr_pages(dst); in copy_user_large_folio() [all …]
|
| A D | vma.c | 611 static int dup_anon_vma(struct vm_area_struct *dst, in dup_anon_vma() argument 628 if (src->anon_vma && !dst->anon_vma) { in dup_anon_vma() 631 vma_assert_write_locked(dst); in dup_anon_vma() 632 dst->anon_vma = src->anon_vma; in dup_anon_vma() 633 ret = anon_vma_clone(dst, src); in dup_anon_vma() 637 *dup = dst; in dup_anon_vma()
|
| A D | secretmem.c | 148 struct folio *dst, struct folio *src, enum migrate_mode mode) in secretmem_migrate_folio() argument
|
| A D | vmscan.c | 1011 struct folio *dst; in alloc_demote_folio() local 1029 dst = alloc_migration_target(src, (unsigned long)mtc); in alloc_demote_folio() 1030 if (dst) in alloc_demote_folio() 1031 return dst; in alloc_demote_folio() 1738 struct lruvec *lruvec, struct list_head *dst, in isolate_lru_folios() argument 1800 move_to = dst; in isolate_lru_folios()
|
| A D | mempolicy.c | 2728 int vma_dup_policy(struct vm_area_struct *src, struct vm_area_struct *dst) in vma_dup_policy() argument 2734 dst->vm_policy = pol; in vma_dup_policy()
|
| /mm/damon/ |
| A D | core.c | 762 switch (dst->metric) { in damos_commit_quota_goal_union() 765 dst->nid = src->nid; in damos_commit_quota_goal_union() 775 dst->metric = src->metric; in damos_commit_quota_goal() 825 dst->ms = src->ms; in damos_commit_quota() 826 dst->sz = src->sz; in damos_commit_quota() 851 switch (dst->type) { in damos_commit_filter_arg() 872 dst->type = src->type; in damos_commit_filter() 1025 dst->action = src->action; in damos_commit() 1131 put_pid(dst->pid); in damon_commit_target() 1134 dst->pid = src->pid; in damon_commit_target() [all …]
|
| /mm/kmsan/ |
| A D | kmsan_test.c | 425 return memcpy((void *)dst, (const void *)src, size); in memcpy_noinline() 433 volatile long long dst = 0; in test_init_memcpy() local 440 kmsan_check_memory((void *)&dst, sizeof(dst)); in test_init_memcpy() 452 volatile int dst = 0; in test_memcpy_aligned_to_aligned() local 458 kmsan_check_memory((void *)&dst, sizeof(dst)); in test_memcpy_aligned_to_aligned() 474 volatile char dst[8] = { 0 }; in test_memcpy_aligned_to_unaligned() local 482 kmsan_check_memory((void *)dst, 4); in test_memcpy_aligned_to_unaligned() 515 volatile char dst[8] = { 0 }; in test_memcpy_initialized_gap() local 531 kmsan_check_memory((void *)&dst[0], 4); in test_memcpy_initialized_gap() 534 kmsan_check_memory((void *)&dst[2], 4); in test_memcpy_initialized_gap() [all …]
|
| A D | instrumentation.c | 158 void *__msan_memmove(void *dst, const void *src, uintptr_t n) in __msan_memmove() argument 165 result = __memmove(dst, src, n); in __msan_memmove() 173 kmsan_internal_memmove_metadata(dst, (void *)src, n); in __msan_memmove() 182 void *__msan_memcpy(void *dst, const void *src, uintptr_t n); 183 void *__msan_memcpy(void *dst, const void *src, uintptr_t n) in __msan_memcpy() argument 190 result = __memcpy(dst, src, n); in __msan_memcpy() 200 kmsan_internal_memmove_metadata(dst, (void *)src, n); in __msan_memcpy() 209 void *__msan_memset(void *dst, int c, uintptr_t n); 210 void *__msan_memset(void *dst, int c, uintptr_t n) in __msan_memset() argument 217 result = __memset(dst, c, n); in __msan_memset() [all …]
|
| A D | core.c | 83 void kmsan_internal_memmove_metadata(void *dst, void *src, size_t n) in kmsan_internal_memmove_metadata() argument 93 shadow_dst = kmsan_get_metadata(dst, KMSAN_META_SHADOW); in kmsan_internal_memmove_metadata() 96 KMSAN_WARN_ON(!kmsan_metadata_is_contiguous(dst, n)); in kmsan_internal_memmove_metadata() 103 kmsan_internal_unpoison_memory(dst, n, /*checked*/ false); in kmsan_internal_memmove_metadata() 108 origin_dst = kmsan_get_metadata(dst, KMSAN_META_ORIGIN); in kmsan_internal_memmove_metadata() 112 backwards = dst > src; in kmsan_internal_memmove_metadata() 116 dst_off = (u64)dst % KMSAN_ORIGIN_SIZE; in kmsan_internal_memmove_metadata()
|
| A D | shadow.c | 151 void kmsan_copy_page_meta(struct page *dst, struct page *src) in kmsan_copy_page_meta() argument 155 if (!dst || !page_has_metadata(dst)) in kmsan_copy_page_meta() 158 kmsan_internal_unpoison_memory(page_address(dst), PAGE_SIZE, in kmsan_copy_page_meta() 164 __memcpy(shadow_ptr_for(dst), shadow_ptr_for(src), PAGE_SIZE); in kmsan_copy_page_meta() 165 __memcpy(origin_ptr_for(dst), origin_ptr_for(src), PAGE_SIZE); in kmsan_copy_page_meta()
|
| A D | kmsan.h | 154 void kmsan_internal_memmove_metadata(void *dst, void *src, size_t n);
|
| /mm/kfence/ |
| A D | kfence_test.c | 391 char dst; in test_use_after_free_read_nofault() local 398 ret = copy_from_kernel_nofault(&dst, addr, 1); in test_use_after_free_read_nofault()
|