| /mm/ |
| A D | mmu_notifier.c | 269 .mm = mm, in mn_itree_release() 659 mmgrab(mm); in __mmu_notifier_register() 660 subscription->mm = mm; in __mmu_notifier_register() 835 mmdrop(mm); in mmu_notifier_unregister() 843 struct mm_struct *mm = subscription->mm; in mmu_notifier_free_rcu() local 847 mmdrop(mm); in mmu_notifier_free_rcu() 874 struct mm_struct *mm = subscription->mm; in mmu_notifier_put() local 895 interval_sub->mm = mm; in __mmu_interval_notifier_insert() 913 mmgrab(mm); in __mmu_interval_notifier_insert() 1040 struct mm_struct *mm = interval_sub->mm; in mmu_interval_notifier_remove() local [all …]
|
| A D | mmap.c | 111 return mlock_future_ok(current->mm, current->mm->def_flags, len) in check_brk_limits() 118 struct mm_struct *mm = current->mm; in SYSCALL_DEFINE1() local 150 mm->end_data, mm->start_data)) in SYSCALL_DEFINE1() 156 mm->brk = brk; in SYSCALL_DEFINE1() 172 mm->brk = brk; in SYSCALL_DEFINE1() 197 mm->brk = brk; in SYSCALL_DEFINE1() 340 struct mm_struct *mm = current->mm; in do_mmap() local 692 struct mm_struct *mm = current->mm; in generic_get_unmapped_area() local 742 struct mm_struct *mm = current->mm; in generic_get_unmapped_area_topdown() local 1091 struct mm_struct *mm = current->mm; in SYSCALL_DEFINE5() local [all …]
|
| A D | debug.c | 202 mm, mm->task_size, in dump_mm() 203 mm->mmap_base, mm->mmap_legacy_base, in dump_mm() 204 mm->pgd, atomic_read(&mm->mm_users), in dump_mm() 208 mm->hiwater_rss, mm->hiwater_vm, mm->total_vm, mm->locked_vm, in dump_mm() 210 mm->data_vm, mm->exec_vm, mm->stack_vm, in dump_mm() 211 mm->start_code, mm->end_code, mm->start_data, mm->end_data, in dump_mm() 212 mm->start_brk, mm->brk, mm->start_stack, in dump_mm() 213 mm->arg_start, mm->arg_end, mm->env_start, mm->env_end, in dump_mm() 214 mm->binfmt, mm->flags, in dump_mm() 226 mm->numa_next_scan, mm->numa_scan_offset, mm->numa_scan_seq, in dump_mm() [all …]
|
| A D | mmap_lock.c | 147 vma = vma_start_read(mm, vma); in lock_vma_under_rcu() 201 mmap_read_unlock(mm); in lock_next_vma_under_mmap_lock() 304 mmap_read_unlock(mm); in upgrade_mmap_lock_carefully() 340 vma = find_vma(mm, addr); in lock_mm_and_find_vma() 349 mmap_read_unlock(mm); in lock_mm_and_find_vma() 366 vma = find_vma(mm, addr); in lock_mm_and_find_vma() 379 mmap_write_downgrade(mm); in lock_mm_and_find_vma() 383 mmap_write_unlock(mm); in lock_mm_and_find_vma() 399 mmap_read_lock(mm); in lock_mm_and_find_vma() 400 vma = vma_lookup(mm, addr); in lock_mm_and_find_vma() [all …]
|
| A D | oom_kill.c | 231 points = get_mm_rss(p->mm) + get_mm_counter(p->mm, MM_SWAPENTS) + in oom_badness() 405 task->tgid, task->mm->total_vm, get_mm_rss(task->mm), in dump_task() 763 struct mm_struct *mm = tsk->mm; local 876 struct mm_struct *mm = task->mm; in task_will_free_mem() local 885 if (!mm) in task_will_free_mem() 940 mm = victim->mm; in __oom_kill_process() 941 mmgrab(mm); in __oom_kill_process() 999 mmdrop(mm); in __oom_kill_process() 1231 mm = p->mm; in SYSCALL_DEFINE2() 1232 mmgrab(mm); in SYSCALL_DEFINE2() [all …]
|
| A D | khugepaged.c | 447 VM_BUG_ON_MM(hpage_collapse_test_exit(mm), mm); in __khugepaged_enter() 467 mmgrab(mm); in __khugepaged_enter() 502 mmdrop(mm); in __khugepaged_exit() 1136 mmap_read_lock(mm); in collapse_huge_page() 1450 struct mm_struct *mm = slot->mm; in collect_mm_slot() local 1467 mmdrop(mm); in collect_mm_slot() 1708 flush_tlb_mm(mm); in collapse_pte_mapped_thp() 1752 mm = vma->vm_mm; in retract_page_tables() 2413 mm = slot->mm; in khugepaged_scan_mm_slot() 2778 mmgrab(mm); in madvise_collapse() [all …]
|
| A D | mremap.c | 81 pgd = pgd_offset(mm, addr); in get_old_pud() 117 pgd = pgd_offset(mm, addr); in alloc_new_pud() 1014 struct mm_struct *mm = current->mm; in vrm_stat_account() local 1075 struct mm_struct *mm = current->mm; in unmap_source_vma() local 1266 struct mm_struct *mm = current->mm; in move_vma() local 1327 struct mm_struct *mm = current->mm; in shrink_vma() local 1363 struct mm_struct *mm = current->mm; in mremap_to() local 1452 struct mm_struct *mm = current->mm; in expand_vma_in_place() local 1609 struct mm_struct *mm = current->mm; in notify_uffd() local 1650 struct mm_struct *mm = current->mm; in check_prep_vma() local [all …]
|
| A D | vma.c | 11 struct mm_struct *mm; member 42 .mm = mm_, \ 55 .mm = (map_)->mm, \ 355 mm->map_count++; in vma_complete() 386 mm->map_count--; in vma_complete() 1268 mm = current->mm; in vms_complete_munmap_vmas() 2526 struct mm_struct *mm = map->mm; in __mmap_complete() local 2574 .mm = map->mm, in call_mmap_prepare() 2642 struct mm_struct *mm = current->mm; in __mmap_region() local 2764 struct mm_struct *mm = current->mm; in do_brk_flags() local [all …]
|
| A D | madvise.c | 286 struct mm_struct *mm = madv_behavior->mm; in madvise_willneed() local 363 struct mm_struct *mm = tlb->mm; in madvise_cold_or_pageout_pte_range() local 661 struct mm_struct *mm = tlb->mm; in madvise_free_pte_range() local 805 struct mm_struct *mm = madv_behavior->mm; in madvise_free_single_vma() local 909 struct mm_struct *mm = madv_behavior->mm; in madvise_dontneed_free() local 970 struct mm_struct *mm = madv_behavior->mm; in madvise_populate() local 1016 struct mm_struct *mm = madv_behavior->mm; in madvise_remove() local 1967 .mm = mm, in do_madvise() 1998 .mm = mm, in vector_madvise() 2140 .mm = mm, in madvise_set_anon_name() [all …]
|
| A D | nommu.c | 381 struct mm_struct *mm = current->mm; in SYSCALL_DEFINE1() local 383 if (brk < mm->start_brk || brk > mm->context.end_brk) in SYSCALL_DEFINE1() 384 return mm->brk; in SYSCALL_DEFINE1() 387 return mm->brk; in SYSCALL_DEFINE1() 393 mm->brk = brk; in SYSCALL_DEFINE1() 1489 struct mm_struct *mm = current->mm; in vm_munmap() local 1512 if (!mm) in exit_mmap() 1703 if (!mm) in access_process_vm() 1708 mmput(mm); in access_process_vm() 1778 if (!mm) { in copy_remote_vm_str() [all …]
|
| A D | ksm.c | 720 struct mm_struct *mm = rmap_item->mm; in break_cow() local 739 struct mm_struct *mm = rmap_item->mm; in get_mergeable_page() local 1187 mm = mm_slot->slot.mm; in unmerge_and_remove_all_rmap_items() 1222 mmdrop(mm); in unmerge_and_remove_all_rmap_items() 1520 struct mm_struct *mm = rmap_item->mm; in try_to_merge_with_zero_page() local 1560 struct mm_struct *mm = rmap_item->mm; in try_to_merge_with_ksm_page() local 2384 rmap_item->mm = mm_slot->slot.mm; in get_next_rmap_item() 2532 mm = slot->mm; in scan_get_next_rmap_item() 2626 mmdrop(mm); in scan_get_next_rmap_item() 2916 mmgrab(mm); in __ksm_enter() [all …]
|
| A D | vma_exec.c | 32 struct mm_struct *mm = vma->vm_mm; in relocate_vma_down() local 38 VMA_ITERATOR(vmi, mm, new_start); in relocate_vma_down() 69 tlb_gather_mmu(&tlb, mm); in relocate_vma_down() 111 struct vm_area_struct *vma = vm_area_alloc(mm); in create_init_stack_vma() 118 if (mmap_write_lock_killable(mm)) { in create_init_stack_vma() 127 err = ksm_execve(mm); in create_init_stack_vma() 143 err = insert_vm_struct(mm, vma); in create_init_stack_vma() 147 mm->stack_vm = mm->total_vm = 1; in create_init_stack_vma() 148 mmap_write_unlock(mm); in create_init_stack_vma() 154 ksm_exit(mm); in create_init_stack_vma() [all …]
|
| A D | util.c | 553 if (pages == 0 || !mm) in account_locked_vm() 556 mmap_write_lock(mm); in account_locked_vm() 559 mmap_write_unlock(mm); in account_locked_vm() 570 struct mm_struct *mm = current->mm; in vm_mmap_pgoff() local 582 mmap_write_unlock(mm); in vm_mmap_pgoff() 958 if (mm) { in __vm_enough_memory() 991 if (!mm) in get_cmdline() 993 if (!mm->arg_end) in get_cmdline() 998 arg_end = mm->arg_end; in get_cmdline() 1000 env_end = mm->env_end; in get_cmdline() [all …]
|
| A D | userfaultfd.c | 47 mmap_assert_locked(mm); in find_vma_and_prepare_anon() 48 vma = vma_lookup(mm, addr); in find_vma_and_prepare_anon() 86 mmap_read_lock(mm); in uffd_lock_vma() 95 mmap_read_unlock(mm); in uffd_lock_vma() 1597 mmap_read_lock(mm); in uffd_move_lock() 1639 mmap_read_lock(mm); in uffd_move_lock() 1733 struct mm_struct *mm = ctx->mm; in move_pages() local 2028 struct mm_struct *mm = ctx->mm; in userfaultfd_release_new() local 2033 mmap_write_lock(mm); in userfaultfd_release_new() 2058 mmap_write_lock(mm); in userfaultfd_release_all() [all …]
|
| A D | pagewalk.c | 422 mmap_assert_locked(mm); in process_mm_walk_lock() 464 .mm = mm, in walk_page_range_mm() 471 if (!walk.mm) in walk_page_range_mm() 612 .mm = mm, in walk_kernel_page_table_range() 631 mmap_assert_locked(mm); in walk_kernel_page_table_range() 657 .mm = mm, in walk_page_range_debug() 664 if (mm == &init_mm) in walk_page_range_debug() 691 .mm = vma->vm_mm, in walk_page_range_vma() 713 .mm = vma->vm_mm, in walk_page_vma() 718 if (!walk.mm) in walk_page_vma() [all …]
|
| A D | page_table_check.c | 150 if (&init_mm == mm) in __page_table_check_pte_clear() 161 if (&init_mm == mm) in __page_table_check_pmd_clear() 172 if (&init_mm == mm) in __page_table_check_pud_clear() 201 if (&init_mm == mm) in __page_table_check_ptes_set() 207 __page_table_check_pte_clear(mm, ptep_get(ptep + i)); in __page_table_check_ptes_set() 227 if (&init_mm == mm) in __page_table_check_pmds_set() 233 __page_table_check_pmd_clear(mm, *(pmdp + i)); in __page_table_check_pmds_set() 245 if (&init_mm == mm) in __page_table_check_puds_set() 249 __page_table_check_pud_clear(mm, *(pudp + i)); in __page_table_check_puds_set() 259 if (&init_mm == mm) in __page_table_check_pte_clear_range() [all …]
|
| A D | pgtable-generic.c | 97 struct mm_struct *mm = (vma)->vm_mm; in ptep_clear_flush() local 100 if (pte_accessible(mm, pte)) in ptep_clear_flush() 167 assert_spin_locked(pmd_lockptr(mm, pmdp)); in pgtable_trans_huge_deposit() 170 if (!pmd_huge_pte(mm, pmdp)) in pgtable_trans_huge_deposit() 174 pmd_huge_pte(mm, pmdp) = pgtable; in pgtable_trans_huge_deposit() 184 assert_spin_locked(pmd_lockptr(mm, pmdp)); in pgtable_trans_huge_withdraw() 187 pgtable = pmd_huge_pte(mm, pmdp); in pgtable_trans_huge_withdraw() 190 if (pmd_huge_pte(mm, pmdp)) in pgtable_trans_huge_withdraw() 315 *ptlp = pte_lockptr(mm, &pmdval); in pte_offset_map_ro_nolock() 328 *ptlp = pte_lockptr(mm, pmdvalp); in pte_offset_map_rw_nolock() [all …]
|
| A D | debug_vm_pgtable.c | 46 struct mm_struct *mm; member 317 if (mm_pmd_folded(args->mm)) in pud_basic_tests() 488 if (mm_pmd_folded(args->mm)) in pud_clear_tests() 502 if (mm_pmd_folded(args->mm)) in pud_populate_tests() 994 mm_dec_nr_ptes(args->mm); in destroy_args() 999 mm_dec_nr_pmds(args->mm); in destroy_args() 1004 mm_dec_nr_puds(args->mm); in destroy_args() 1014 if (args->mm) in destroy_args() 1015 mmdrop(args->mm); in destroy_args() 1144 args->mm = mm_alloc(); in init_args() [all …]
|
| A D | gup.c | 1616 mmap_read_lock(mm); in fixup_user_fault() 1630 mmap_read_lock(mm); in fixup_user_fault() 1799 mmap_read_unlock(mm); in __get_user_pages_locked() 1846 mmap_assert_locked(mm); in populate_vma_page_range() 1947 struct mm_struct *mm = current->mm; in __mm_populate() local 1962 mmap_read_lock(mm); in __mm_populate() 1995 mmap_read_unlock(mm); in __mm_populate() 2050 mmap_read_unlock(mm); in __get_user_pages_locked() 2140 struct mm_struct *mm = current->mm; in fault_in_safe_writeable() local 2146 mmap_read_lock(mm); in fault_in_safe_writeable() [all …]
|
| A D | rmap.c | 188 struct mm_struct *mm = vma->vm_mm; in __anon_vma_prepare() local 192 mmap_assert_locked(mm); in __anon_vma_prepare() 211 spin_lock(&mm->page_table_lock); in __anon_vma_prepare() 219 spin_unlock(&mm->page_table_lock); in __anon_vma_prepare() 681 if (!pte_accessible(mm, pteval)) in set_tlb_ubc_flush_pending() 749 flush_tlb_mm(mm); in flush_tlb_batched_pending() 822 pgd = pgd_offset(mm, address); in mm_find_pmd() 2066 update_hiwater_rss(mm); in try_to_unmap_one() 2174 if (list_empty(&mm->mmlist)) in try_to_unmap_one() 2455 update_hiwater_rss(mm); in try_to_migrate_one() [all …]
|
| A D | mseal.c | 38 static bool range_contains_unmapped(struct mm_struct *mm, in range_contains_unmapped() argument 43 VMA_ITERATOR(vmi, current->mm, start); in range_contains_unmapped() 55 static int mseal_apply(struct mm_struct *mm, in mseal_apply() argument 60 VMA_ITERATOR(vmi, mm, start); in mseal_apply() 144 struct mm_struct *mm = current->mm; in do_mseal() local 166 if (mmap_write_lock_killable(mm)) in do_mseal() 169 if (range_contains_unmapped(mm, start, end)) { in do_mseal() 180 ret = mseal_apply(mm, start, end); in do_mseal() 183 mmap_write_unlock(mm); in do_mseal()
|
| A D | memory.c | 458 pte_free(mm, new); in __pte_alloc() 1495 struct mm_struct *mm = tlb->mm; in zap_present_folio_ptes() local 1546 struct mm_struct *mm = tlb->mm; in zap_present_ptes() local 1698 struct mm_struct *mm = tlb->mm; in zap_pte_range() local 6699 if (!vma_lookup(mm, addr) && !expand_stack(mm, addr)) in __access_remote_vm() 6794 if (!mm) in access_process_vm() 6799 mmput(mm); in access_process_vm() 6911 if (!mm) { in copy_remote_vm_str() 6918 mmput(mm); in copy_remote_vm_str() 6930 struct mm_struct *mm = current->mm; in print_vma_addr() local [all …]
|
| A D | msync.c | 35 struct mm_struct *mm = current->mm; in SYSCALL_DEFINE3() local 62 mmap_read_lock(mm); in SYSCALL_DEFINE3() 63 vma = find_vma(mm, start); in SYSCALL_DEFINE3() 95 mmap_read_unlock(mm); in SYSCALL_DEFINE3() 100 mmap_read_lock(mm); in SYSCALL_DEFINE3() 101 vma = find_vma(mm, start); in SYSCALL_DEFINE3() 107 vma = find_vma(mm, vma->vm_end); in SYSCALL_DEFINE3() 111 mmap_read_unlock(mm); in SYSCALL_DEFINE3()
|
| A D | pt_reclaim.c | 14 bool try_get_and_clear_pmd(struct mm_struct *mm, pmd_t *pmd, pmd_t *pmdval) in try_get_and_clear_pmd() argument 16 spinlock_t *pml = pmd_lockptr(mm, pmd); in try_get_and_clear_pmd() 28 void free_pte(struct mm_struct *mm, unsigned long addr, struct mmu_gather *tlb, in free_pte() argument 32 mm_dec_nr_ptes(mm); in free_pte() 35 void try_to_free_pte(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, in try_to_free_pte() argument 43 pml = pmd_lock(mm, pmd); in try_to_free_pte() 44 start_pte = pte_offset_map_rw_nolock(mm, pmd, addr, &pmdval, &ptl); in try_to_free_pte() 63 free_pte(mm, addr, tlb, pmdval); in try_to_free_pte()
|
| /mm/damon/ |
| A D | vaddr.c | 53 return mm; in damon_get_mm() 186 if (!mm) in damon_va_three_regions() 193 mmput(mm); in damon_va_three_regions() 427 if (!mm) in damon_va_prepare_access_checks() 431 mmput(mm); in damon_va_prepare_access_checks() 572 if (!mm) { in __damon_va_check_access() 607 if (mm) in damon_va_check_accesses() 833 if (!mm) in damos_madvise() 837 mmput(mm); in damos_madvise() 874 if (!mm) in damos_va_migrate() [all …]
|