Home
last modified time | relevance | path

Searched refs:pgd (Results 1 – 25 of 338) sorted by relevance

12345678910>>...14

/linux/include/asm-generic/
A Dpgtable-nop4d.h9 typedef struct { pgd_t pgd; } p4d_t; member
21 static inline int pgd_none(pgd_t pgd) { return 0; } in pgd_none() argument
22 static inline int pgd_bad(pgd_t pgd) { return 0; } in pgd_bad() argument
23 static inline int pgd_present(pgd_t pgd) { return 1; } in pgd_present() argument
24 static inline void pgd_clear(pgd_t *pgd) { } in pgd_clear() argument
25 #define p4d_ERROR(p4d) (pgd_ERROR((p4d).pgd))
27 #define pgd_populate(mm, pgd, p4d) do { } while (0) argument
37 return (p4d_t *)pgd; in p4d_offset()
40 #define p4d_val(x) (pgd_val((x).pgd))
43 #define pgd_page(pgd) (p4d_page((p4d_t){ pgd })) argument
[all …]
/linux/arch/x86/mm/
A Dkasan_init_64.c130 if (pgd_none(*pgd)) { in kasan_populate_pgd()
145 pgd_t *pgd; in kasan_populate_shadow() local
150 pgd = pgd_offset_k(addr); in kasan_populate_shadow()
171 pgd_t *pgd; in clear_pgds() local
182 pgd_clear(pgd); in clear_pgds()
197 return (p4d_t *)pgd; in early_p4d_offset()
212 if (pgd_none(*pgd)) { in kasan_early_p4d_populate()
215 set_pgd(pgd, pgd_entry); in kasan_early_p4d_populate()
238 pgd += pgd_index(addr); in kasan_map_early_shadow()
267 pgd_t *pgd; in kasan_shallow_populate_pgds() local
[all …]
A Dpgtable.c142 pgd_list_add(pgd); in pgd_ctor()
152 pgd_list_del(pgd); in pgd_dtor()
276 pgd_t pgd = *pgdp; in mop_up_one_pmd() local
435 pgd_t *pgd; in pgd_alloc() local
441 if (pgd == NULL) in pgd_alloc()
444 mm->pgd = pgd; in pgd_alloc()
464 pgd_ctor(mm, pgd); in pgd_alloc()
473 return pgd; in pgd_alloc()
482 _pgd_free(pgd); in pgd_alloc()
490 pgd_dtor(pgd); in pgd_free()
[all …]
A Dpti.c136 return pgd; in __pti_set_user_pgtbl()
142 kernel_to_user_pgdp(pgdp)->pgd = pgd.pgd; in __pti_set_user_pgtbl()
157 if ((pgd.pgd & (_PAGE_USER|_PAGE_PRESENT)) == (_PAGE_USER|_PAGE_PRESENT) && in __pti_set_user_pgtbl()
159 pgd.pgd |= _PAGE_NX; in __pti_set_user_pgtbl()
162 return pgd; in __pti_set_user_pgtbl()
181 if (pgd_none(*pgd)) { in pti_user_pagetable_walk_p4d()
190 return p4d_offset(pgd, address); in pti_user_pagetable_walk_p4d()
320 pgd_t *pgd; in pti_clone_pgtable() local
328 pgd = pgd_offset_k(addr); in pti_clone_pgtable()
329 if (WARN_ON(pgd_none(*pgd))) in pti_clone_pgtable()
[all …]
A Dinit_64.c150 pgd_t *pgd; in sync_global_pgds_l5() local
161 if (pgd_none(*pgd)) in sync_global_pgds_l5()
191 pgd_t *pgd; in sync_global_pgds_l4() local
251 if (pgd_none(*pgd)) { in fill_p4d()
327 pgd_t *pgd; in set_pte_vaddr() local
333 if (pgd_none(*pgd)) { in set_pte_vaddr()
345 pgd_t *pgd; in populate_extra_pmd() local
369 pgd_t *pgd; in __init_extra_mapping() local
380 if (pgd_none(*pgd)) { in __init_extra_mapping()
1234 pgd_t *pgd; in remove_pagetable() local
[all …]
A Dident_map.c50 static void free_p4d(struct x86_mapping_info *info, pgd_t *pgd) in free_p4d() argument
52 p4d_t *p4d = p4d_offset(pgd, 0); in free_p4d()
66 void kernel_ident_mapping_free(struct x86_mapping_info *info, pgd_t *pgd) in kernel_ident_mapping_free() argument
71 if (!pgd_present(pgd[i])) in kernel_ident_mapping_free()
74 free_p4d(info, &pgd[i]); in kernel_ident_mapping_free()
77 info->free_pgt_page(pgd, info->context); in kernel_ident_mapping_free()
199 pgd_t *pgd = pgd_page + pgd_index(addr); in kernel_ident_mapping_init() local
206 if (pgd_present(*pgd)) { in kernel_ident_mapping_init()
207 p4d = p4d_offset(pgd, 0); in kernel_ident_mapping_init()
221 set_pgd(pgd, __pgd(__pa(p4d) | info->kernpg_flag)); in kernel_ident_mapping_init()
[all …]
/linux/arch/powerpc/include/asm/book3s/64/
A Dpgalloc.h42 free_page((unsigned long)pgd); in radix__pgd_free()
44 free_pages((unsigned long)pgd, 4); in radix__pgd_free()
50 pgd_t *pgd; in pgd_alloc() local
55 pgd = kmem_cache_alloc(PGT_CACHE(PGD_INDEX_SIZE), in pgd_alloc()
57 if (unlikely(!pgd)) in pgd_alloc()
58 return pgd; in pgd_alloc()
65 kmemleak_no_scan(pgd); in pgd_alloc()
76 memset(pgd, 0, PGD_TABLE_SIZE); in pgd_alloc()
78 return pgd; in pgd_alloc()
84 return radix__pgd_free(mm, pgd); in pgd_free()
[all …]
/linux/arch/arm/mm/
A Dpgd.c21 #define __pgd_free(pgd) kfree(pgd) argument
24 #define __pgd_free(pgd) free_pages((unsigned long)pgd, 2) argument
144 pgd_t *pgd; in pgd_free() local
153 pgd = pgd_base + pgd_index(0); in pgd_free()
154 if (pgd_none_or_clear_bad(pgd)) in pgd_free()
157 p4d = p4d_offset(pgd, 0); in pgd_free()
181 pgd_clear(pgd); in pgd_free()
188 for (pgd = pgd_base; pgd < pgd_base + PTRS_PER_PGD; pgd++) { in pgd_free()
189 if (pgd_none_or_clear_bad(pgd)) in pgd_free()
193 p4d = p4d_offset(pgd, 0); in pgd_free()
[all …]
/linux/arch/riscv/include/asm/
A Dpgtable-64.h346 WRITE_ONCE(*pgdp, pgd); in set_pgd()
351 static inline int pgd_none(pgd_t pgd) in pgd_none() argument
354 return (pgd_val(pgd) == 0); in pgd_none()
359 static inline int pgd_present(pgd_t pgd) in pgd_present() argument
362 return (pgd_val(pgd) & _PAGE_PRESENT); in pgd_present()
367 static inline int pgd_bad(pgd_t pgd) in pgd_bad() argument
370 return !pgd_present(pgd); in pgd_bad()
378 set_pgd(pgd, __pgd(0)); in pgd_clear()
388 #define pgd_page_vaddr(pgd) ((unsigned long)pgd_pgtable(pgd)) argument
394 #define pgd_page(pgd) pgd_page(pgd) argument
[all …]
A Dpgalloc.h76 set_pgd(pgd, __pgd((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE)); in pgd_populate()
86 set_pgd_safe(pgd, in pgd_populate_safe()
153 static inline void sync_kernel_mappings(pgd_t *pgd) in sync_kernel_mappings() argument
155 memcpy(pgd + USER_PTRS_PER_PGD, in sync_kernel_mappings()
156 init_mm.pgd + USER_PTRS_PER_PGD, in sync_kernel_mappings()
162 pgd_t *pgd; in pgd_alloc() local
164 pgd = (pgd_t *)__get_free_page(GFP_KERNEL); in pgd_alloc()
165 if (likely(pgd != NULL)) { in pgd_alloc()
166 memset(pgd, 0, USER_PTRS_PER_PGD * sizeof(pgd_t)); in pgd_alloc()
168 sync_kernel_mappings(pgd); in pgd_alloc()
[all …]
/linux/arch/x86/kernel/acpi/
A Dmadt_wakeup.c89 pgd += pgd_index(vaddr); in init_transition_pgtable()
90 if (!pgd_present(*pgd)) { in init_transition_pgtable()
96 p4d = p4d_offset(pgd, vaddr); in init_transition_pgtable()
133 pgd_t *pgd; in acpi_mp_setup_reset() local
135 pgd = alloc_pgt_page(NULL); in acpi_mp_setup_reset()
136 if (!pgd) in acpi_mp_setup_reset()
145 kernel_ident_mapping_free(&info, pgd); in acpi_mp_setup_reset()
153 kernel_ident_mapping_free(&info, pgd); in acpi_mp_setup_reset()
157 if (init_transition_pgtable(pgd)) { in acpi_mp_setup_reset()
158 kernel_ident_mapping_free(&info, pgd); in acpi_mp_setup_reset()
[all …]
/linux/arch/sh/mm/
A Dhugetlbpage.c27 pgd_t *pgd; in huge_pte_alloc() local
33 pgd = pgd_offset(mm, addr); in huge_pte_alloc()
34 if (pgd) { in huge_pte_alloc()
35 p4d = p4d_alloc(mm, pgd, addr); in huge_pte_alloc()
52 pgd_t *pgd; in huge_pte_offset() local
58 pgd = pgd_offset(mm, addr); in huge_pte_offset()
59 if (pgd) { in huge_pte_offset()
60 p4d = p4d_offset(pgd, addr); in huge_pte_offset()
A Dfault.c39 pgd_t *pgd; in show_pte() local
42 pgd = mm->pgd; in show_pte()
44 pgd = get_TTB(); in show_pte()
46 if (unlikely(!pgd)) in show_pte()
47 pgd = swapper_pg_dir; in show_pte()
50 pr_alert("pgd = %p\n", pgd); in show_pte()
51 pgd += pgd_index(addr); in show_pte()
53 (u64)pgd_val(*pgd)); in show_pte()
61 if (pgd_none(*pgd)) in show_pte()
64 if (pgd_bad(*pgd)) { in show_pte()
[all …]
A Dpgtable.c14 pgd_t *pgd = x; in pgd_ctor() local
16 memset(pgd, 0, USER_PTRS_PER_PGD * sizeof(pgd_t)); in pgd_ctor()
17 memcpy(pgd + USER_PTRS_PER_PGD, in pgd_ctor()
39 void pgd_free(struct mm_struct *mm, pgd_t *pgd) in pgd_free() argument
41 kmem_cache_free(pgd_cachep, pgd); in pgd_free()
/linux/mm/kasan/
A Dinit.c266 pgd_populate(&init_mm, pgd, in kasan_populate_early_shadow()
268 p4d = p4d_offset(pgd, addr); in kasan_populate_early_shadow()
280 if (pgd_none(*pgd)) { in kasan_populate_early_shadow()
355 pgd_clear(pgd); in kasan_free_p4d()
462 pgd_t *pgd; in kasan_remove_zero_shadow() local
476 pgd = pgd_offset_k(addr); in kasan_remove_zero_shadow()
477 if (!pgd_present(*pgd)) in kasan_remove_zero_shadow()
480 if (kasan_p4d_table(*pgd)) { in kasan_remove_zero_shadow()
483 pgd_clear(pgd); in kasan_remove_zero_shadow()
488 p4d = p4d_offset(pgd, addr); in kasan_remove_zero_shadow()
[all …]
/linux/arch/x86/power/
A Dhibernate_32.c30 static pmd_t *resume_one_md_table_init(pgd_t *pgd) in resume_one_md_table_init() argument
41 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in resume_one_md_table_init()
42 p4d = p4d_offset(pgd, 0); in resume_one_md_table_init()
47 p4d = p4d_offset(pgd, 0); in resume_one_md_table_init()
84 pgd_t *pgd; in resume_physical_mapping_init() local
90 pgd = pgd_base + pgd_idx; in resume_physical_mapping_init()
93 for (; pgd_idx < PTRS_PER_PGD; pgd++, pgd_idx++) { in resume_physical_mapping_init()
94 pmd = resume_one_md_table_init(pgd); in resume_physical_mapping_init()
147 pgd_t *pgd; in set_up_temporary_text_mapping() local
151 pgd = pgd_base + pgd_index(restore_jump_address); in set_up_temporary_text_mapping()
[all …]
A Dhibernate_64.c28 static int set_up_temporary_text_mapping(pgd_t *pgd) in set_up_temporary_text_mapping() argument
77 set_pgd(pgd + pgd_index(restore_jump_address), new_pgd); in set_up_temporary_text_mapping()
81 set_pgd(pgd + pgd_index(restore_jump_address), new_pgd); in set_up_temporary_text_mapping()
100 pgd_t *pgd; in set_up_temporary_mappings() local
104 pgd = (pgd_t *)get_safe_page(GFP_ATOMIC); in set_up_temporary_mappings()
105 if (!pgd) in set_up_temporary_mappings()
109 result = set_up_temporary_text_mapping(pgd); in set_up_temporary_mappings()
118 result = kernel_ident_mapping_init(&info, pgd, mstart, mend); in set_up_temporary_mappings()
123 temp_pgt = __pa(pgd); in set_up_temporary_mappings()
/linux/arch/mips/mm/
A Dhugetlbpage.c27 pgd_t *pgd; in huge_pte_alloc() local
32 pgd = pgd_offset(mm, addr); in huge_pte_alloc()
33 p4d = p4d_alloc(mm, pgd, addr); in huge_pte_alloc()
44 pgd_t *pgd; in huge_pte_offset() local
49 pgd = pgd_offset(mm, addr); in huge_pte_offset()
50 if (pgd_present(*pgd)) { in huge_pte_offset()
51 p4d = p4d_offset(pgd, addr); in huge_pte_offset()
/linux/arch/x86/kernel/
A Dmachine_kexec_32.c45 free_pages((unsigned long)image->arch.pgd, PGD_ALLOCATION_ORDER); in machine_kexec_free_page_tables()
46 image->arch.pgd = NULL; in machine_kexec_free_page_tables()
69 if (!image->arch.pgd || in machine_kexec_alloc_page_tables()
80 pgd_t *pgd, pmd_t *pmd, pte_t *pte, in machine_kexec_page_table_set_one() argument
86 pgd += pgd_index(vaddr); in machine_kexec_page_table_set_one()
88 if (!(pgd_val(*pgd) & _PAGE_PRESENT)) in machine_kexec_page_table_set_one()
89 set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT)); in machine_kexec_page_table_set_one()
91 p4d = p4d_offset(pgd, vaddr); in machine_kexec_page_table_set_one()
110 image->arch.pgd, pmd, image->arch.pte0, in machine_kexec_prepare_page_tables()
116 image->arch.pgd, pmd, image->arch.pte1, in machine_kexec_prepare_page_tables()
[all …]
/linux/arch/loongarch/mm/
A Dhugetlbpage.c19 pgd_t *pgd; in huge_pte_alloc() local
24 pgd = pgd_offset(mm, addr); in huge_pte_alloc()
25 p4d = p4d_alloc(mm, pgd, addr); in huge_pte_alloc()
36 pgd_t *pgd; in huge_pte_offset() local
41 pgd = pgd_offset(mm, addr); in huge_pte_offset()
42 if (pgd_present(pgdp_get(pgd))) { in huge_pte_offset()
43 p4d = p4d_offset(pgd, addr); in huge_pte_offset()
/linux/arch/parisc/include/asm/
A Dpgalloc.h21 pgd_t *pgd; in pgd_alloc() local
23 pgd = (pgd_t *) __get_free_pages(GFP_KERNEL, PGD_TABLE_ORDER); in pgd_alloc()
24 if (unlikely(pgd == NULL)) in pgd_alloc()
27 memset(pgd, 0, PAGE_SIZE << PGD_TABLE_ORDER); in pgd_alloc()
29 return pgd; in pgd_alloc()
32 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) in pgd_free() argument
34 free_pages((unsigned long)pgd, PGD_TABLE_ORDER); in pgd_free()
/linux/arch/hexagon/include/asm/
A Dpgalloc.h23 pgd_t *pgd; in pgd_alloc() local
25 pgd = (pgd_t *)__get_free_page(GFP_KERNEL | __GFP_ZERO); in pgd_alloc()
35 memcpy(pgd, swapper_pg_dir, PTRS_PER_PGD*sizeof(pgd_t)); in pgd_alloc()
39 mm->context.ptbase = __pa(pgd); in pgd_alloc()
41 return pgd; in pgd_alloc()
83 pmdindex = (pgd_t *)pmd - mm->pgd; in pmd_populate_kernel()
84 ppmd = (pmd_t *)current->active_mm->pgd + pmdindex; in pmd_populate_kernel()
/linux/arch/um/kernel/
A Dmem.c115 pgd_t *pgd; in fixrange_init() local
125 pgd = pgd_base + i; in fixrange_init()
127 for ( ; (i < PTRS_PER_PGD) && (vaddr < end); pgd++, i++) { in fixrange_init()
128 p4d = p4d_offset(pgd, vaddr); in fixrange_init()
205 pgd_t *pgd = (pgd_t *)__get_free_page(GFP_KERNEL); in pgd_alloc() local
207 if (pgd) { in pgd_alloc()
208 memset(pgd, 0, USER_PTRS_PER_PGD * sizeof(pgd_t)); in pgd_alloc()
209 memcpy(pgd + USER_PTRS_PER_PGD, in pgd_alloc()
213 return pgd; in pgd_alloc()
/linux/arch/x86/xen/
A Dmmu_pv.c383 return pte_mfn_to_pfn(pgd.pgd); in xen_pgd_val()
397 pgd = pte_pfn_to_mfn(pgd); in xen_make_pgd()
642 if (pgd_none(pgd[i])) in __xen_pgd_walk()
1128 pgd_t *pgd; in xen_cleanmfnmap() local
1134 pgd = pgd_offset_k(vaddr); in xen_cleanmfnmap()
1135 p4d = p4d_offset(pgd, 0); in xen_cleanmfnmap()
1403 pgd_t *pgd = mm->pgd; in xen_pgd_alloc() local
1741 l3 = m2v(pgd[pgd_index(__START_KERNEL_map)].pgd); in xen_setup_kernel_pagetable()
1849 pgd_t pgd; in xen_early_virt_to_phys() local
1857 if (!pgd_present(pgd)) in xen_early_virt_to_phys()
[all …]
/linux/arch/x86/platform/efi/
A Defi_64.c70 pgd_t *pgd, *efi_pgd; in efi_alloc_page_tables() local
80 pgd = efi_pgd + pgd_index(EFI_VA_END); in efi_alloc_page_tables()
81 p4d = p4d_alloc(&init_mm, pgd, EFI_VA_END); in efi_alloc_page_tables()
89 efi_mm.pgd = efi_pgd; in efi_alloc_page_tables()
97 free_page((unsigned long)pgd_page_vaddr(*pgd)); in efi_alloc_page_tables()
113 pgd_t *efi_pgd = efi_mm.pgd; in efi_sync_low_kernel_mappings()
183 pgd_t *pgd = efi_mm.pgd; in efi_setup_page_tables() local
218 if (sev_es_efi_map_ghcbs(pgd)) { in efi_setup_page_tables()
243 if (kernel_unmap_pages_in_pgd(pgd, text, npages)) { in efi_setup_page_tables()
274 pgd_t *pgd = efi_mm.pgd; in __map_region() local
[all …]

Completed in 37 milliseconds

12345678910>>...14