| /mm/ |
| A D | init-mm.c | 32 struct mm_struct init_mm = { variable 37 .write_protect_seq = SEQCNT_ZERO(init_mm.write_protect_seq), 38 MMAP_LOCK_INITIALIZER(init_mm) 40 .arg_lock = __SPIN_LOCK_UNLOCKED(init_mm.arg_lock), 41 .mmlist = LIST_HEAD_INIT(init_mm.mmlist), 44 .mm_lock_seq = SEQCNT_ZERO(init_mm.mm_lock_seq), 48 INIT_MM_CONTEXT(init_mm) 54 init_mm.start_code = (unsigned long)start_code; in setup_initial_init_mm() 55 init_mm.end_code = (unsigned long)end_code; in setup_initial_init_mm() 56 init_mm.end_data = (unsigned long)end_data; in setup_initial_init_mm() [all …]
|
| A D | hugetlb_vmemmap.c | 59 pgtable = pte_alloc_one_kernel(&init_mm); in vmemmap_split_pmd() 71 set_pte_at(&init_mm, addr, pte, entry); in vmemmap_split_pmd() 74 spin_lock(&init_mm.page_table_lock); in vmemmap_split_pmd() 86 pmd_populate_kernel(&init_mm, pmd, pgtable); in vmemmap_split_pmd() 90 pte_free_kernel(&init_mm, pgtable); in vmemmap_split_pmd() 92 spin_unlock(&init_mm.page_table_lock); in vmemmap_split_pmd() 108 spin_lock(&init_mm.page_table_lock); in vmemmap_pmd_entry() 131 spin_unlock(&init_mm.page_table_lock); in vmemmap_pmd_entry() 168 mmap_read_lock(&init_mm); in vmemmap_remap_range() 171 mmap_read_unlock(&init_mm); in vmemmap_remap_range() [all …]
|
| A D | page_table_check.c | 150 if (&init_mm == mm) in __page_table_check_pte_clear() 161 if (&init_mm == mm) in __page_table_check_pmd_clear() 172 if (&init_mm == mm) in __page_table_check_pud_clear() 201 if (&init_mm == mm) in __page_table_check_ptes_set() 227 if (&init_mm == mm) in __page_table_check_pmds_set() 245 if (&init_mm == mm) in __page_table_check_puds_set() 259 if (&init_mm == mm) in __page_table_check_pte_clear_range()
|
| A D | sparse-vmemmap.c | 182 set_pte_at(&init_mm, addr, pte, entry); in vmemmap_pte_populate() 206 pmd_populate_kernel(&init_mm, pmd, p); in vmemmap_pmd_populate() 219 pud_populate(&init_mm, pud, p); in vmemmap_pud_populate() 232 p4d_populate(&init_mm, p4d, p); in vmemmap_p4d_populate() 244 pgd_populate(&init_mm, pgd, p); in vmemmap_pgd_populate() 339 pte_clear(&init_mm, maddr, pte); in vmemmap_undo_hvo() 349 pte_clear(&init_mm, maddr, pte); in vmemmap_undo_hvo() 377 ptep_set_wrprotect(&init_mm, maddr, pte); in vmemmap_wrprotect_hvo()
|
| A D | highmem.c | 229 pte_clear(&init_mm, PKMAP_ADDR(i), &pkmap_page_table[i]); in flush_all_zero_pkmaps() 290 set_pte_at(&init_mm, vaddr, in map_new_virtual() 566 arch_kmap_local_set_pte(&init_mm, vaddr, kmap_pte, pteval); in __kmap_local_pfn_prot() 626 pte_clear(&init_mm, addr, kmap_pte); in kunmap_local_indexed() 677 pte_clear(&init_mm, addr, kmap_pte); in __kmap_local_sched_out() 706 set_pte_at(&init_mm, addr, kmap_pte, pteval); in __kmap_local_sched_in()
|
| A D | pagewalk.c | 76 if (walk->mm == &init_mm || addr >= TASK_SIZE) in walk_pte_range() 82 if (walk->mm != &init_mm && addr < TASK_SIZE) in walk_pte_range() 609 struct mm_struct *mm = &init_mm; in walk_kernel_page_table_range() 664 if (mm == &init_mm) in walk_page_range_debug()
|
| A D | vmalloc.c | 125 set_huge_pte_at(&init_mm, addr, pte, entry, size); in vmap_pte_range() 130 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); in vmap_pte_range() 171 pmd = pmd_alloc_track(&init_mm, pud, addr, mask); in vmap_pmd_range() 221 pud = pud_alloc_track(&init_mm, p4d, addr, mask); in vmap_pud_range() 272 p4d = p4d_alloc_track(&init_mm, pgd, addr, mask); in vmap_p4d_range() 377 ptent = ptep_get_and_clear(&init_mm, addr, pte); in vunmap_pte_range() 547 set_pte_at(&init_mm, addr, pte, mk_pte(page, prot)); in vmap_pages_pte_range() 564 pmd = pmd_alloc_track(&init_mm, pud, addr, mask); in vmap_pages_pmd_range() 582 pud = pud_alloc_track(&init_mm, p4d, addr, mask); in vmap_pages_pud_range() 600 p4d = p4d_alloc_track(&init_mm, pgd, addr, mask); in vmap_pages_p4d_range() [all …]
|
| A D | percpu.c | 3137 pgd_populate(&init_mm, pgd, p4d); in pcpu_populate_pte() 3143 p4d_populate(&init_mm, p4d, pud); in pcpu_populate_pte() 3149 pud_populate(&init_mm, pud, pmd); in pcpu_populate_pte() 3157 pmd_populate_kernel(&init_mm, pmd, new); in pcpu_populate_pte()
|
| A D | memory.c | 464 pte_t *new = pte_alloc_one_kernel(&init_mm); in __pte_alloc_kernel() 468 spin_lock(&init_mm.page_table_lock); in __pte_alloc_kernel() 471 pmd_populate_kernel(&init_mm, pmd, new); in __pte_alloc_kernel() 474 spin_unlock(&init_mm.page_table_lock); in __pte_alloc_kernel() 476 pte_free_kernel(&init_mm, new); in __pte_alloc_kernel() 2997 mapped_pte = pte = (mm == &init_mm) ? in apply_to_pte_range() 3003 mapped_pte = pte = (mm == &init_mm) ? in apply_to_pte_range() 3025 if (mm != &init_mm) in apply_to_pte_range()
|
| A D | swapfile.c | 2309 prev_mm = &init_mm; in try_to_unuse() 2313 p = &init_mm.mmlist; in try_to_unuse() 2316 (p = p->next) != &init_mm.mmlist) { in try_to_unuse() 2407 list_for_each_safe(p, next, &init_mm.mmlist) in drain_mmlist()
|
| A D | rmap.c | 2175 list_add(&mm->mmlist, &init_mm.mmlist); in try_to_unmap_one()
|
| /mm/kasan/ |
| A D | init.c | 119 pmd_populate_kernel(&init_mm, pmd, in zero_pmd_populate() 155 pud_populate(&init_mm, pud, in zero_pud_populate() 173 pud_populate(&init_mm, pud, p); in zero_pud_populate() 194 p4d_populate(&init_mm, p4d, in zero_p4d_populate() 197 pud_populate(&init_mm, pud, in zero_p4d_populate() 215 p4d_populate(&init_mm, p4d, p); in zero_p4d_populate() 254 pgd_populate(&init_mm, pgd, in kasan_populate_early_shadow() 257 p4d_populate(&init_mm, p4d, in kasan_populate_early_shadow() 260 pud_populate(&init_mm, pud, in kasan_populate_early_shadow() 276 pgd_populate(&init_mm, pgd, in kasan_populate_early_shadow() [all …]
|
| A D | shadow.c | 316 spin_lock(&init_mm.page_table_lock); in kasan_populate_vmalloc_pte() 318 set_pte_at(&init_mm, addr, ptep, pte); in kasan_populate_vmalloc_pte() 321 spin_unlock(&init_mm.page_table_lock); in kasan_populate_vmalloc_pte() 373 ret = apply_to_page_range(&init_mm, start, nr_pages * PAGE_SIZE, in __kasan_populate_vmalloc() 468 spin_lock(&init_mm.page_table_lock); in kasan_depopulate_vmalloc_pte() 471 pte_clear(&init_mm, addr, ptep); in kasan_depopulate_vmalloc_pte() 474 spin_unlock(&init_mm.page_table_lock); in kasan_depopulate_vmalloc_pte() 593 apply_to_existing_page_range(&init_mm, in kasan_release_vmalloc()
|