| /arch/mips/include/asm/ |
| A D | switch_to.h | 62 next->thread.emulated_fp = 0; \ 88 # define __sanitize_fcr31(next) \ argument 95 next->thread.fpu.fcr31 &= ~fcr31; \ 96 force_fcr31_sig(fcr31, pc, next); \ 100 # define __sanitize_fcr31(next) do { (void) (next); } while (0) argument 113 if (tsk_used_math(next)) \ 114 __sanitize_fcr31(next); \ 117 __restore_dsp(next); \ 130 cop2_restore(next); \ 138 __restore_watch(next); \ [all …]
|
| /arch/riscv/mm/ |
| A D | kasan_init.c | 54 unsigned long next; in kasan_populate_pmd() local 85 unsigned long next; in kasan_populate_pud() local 116 unsigned long next; in kasan_populate_p4d() local 146 unsigned long next; in kasan_populate_pgd() local 169 unsigned long next; in kasan_early_clear_pud() local 194 unsigned long next; in kasan_early_clear_p4d() local 219 unsigned long next; in kasan_early_clear_pgd() local 240 unsigned long next; in kasan_early_populate_pud() local 269 unsigned long next; in kasan_early_populate_p4d() local 306 unsigned long next; in kasan_early_populate_pgd() local [all …]
|
| A D | pageattr.c | 100 unsigned long next; in __split_linear_mapping_pmd() local 105 next = pmd_addr_end(vaddr, end); in __split_linear_mapping_pmd() 107 if (next - vaddr >= PMD_SIZE && in __split_linear_mapping_pmd() 139 unsigned long next; in __split_linear_mapping_pud() local 145 next = pud_addr_end(vaddr, end); in __split_linear_mapping_pud() 147 if (next - vaddr >= PUD_SIZE && in __split_linear_mapping_pud() 184 unsigned long next; in __split_linear_mapping_p4d() local 190 next = p4d_addr_end(vaddr, end); in __split_linear_mapping_p4d() 196 if (next - vaddr >= P4D_SIZE && in __split_linear_mapping_p4d() 242 unsigned long next; in __split_linear_mapping_pgd() local [all …]
|
| /arch/arm64/kvm/hyp/nvhe/ |
| A D | list_debug.c | 31 struct list_head *next) in __list_add_valid_or_report() argument 33 if (NVHE_CHECK_DATA_CORRUPTION(next->prev != prev) || in __list_add_valid_or_report() 34 NVHE_CHECK_DATA_CORRUPTION(prev->next != next) || in __list_add_valid_or_report() 35 NVHE_CHECK_DATA_CORRUPTION(new == prev || new == next)) in __list_add_valid_or_report() 44 struct list_head *prev, *next; in __list_del_entry_valid_or_report() local 47 next = entry->next; in __list_del_entry_valid_or_report() 49 if (NVHE_CHECK_DATA_CORRUPTION(next == LIST_POISON1) || in __list_del_entry_valid_or_report() 51 NVHE_CHECK_DATA_CORRUPTION(prev->next != entry) || in __list_del_entry_valid_or_report() 52 NVHE_CHECK_DATA_CORRUPTION(next->prev != entry)) in __list_del_entry_valid_or_report()
|
| /arch/parisc/include/asm/ |
| A D | mmu_context.h | 51 struct mm_struct *next, struct task_struct *tsk) in switch_mm_irqs_off() argument 53 if (prev != next) { in switch_mm_irqs_off() 57 spinlock_t *pgd_lock = &next->page_table_lock; in switch_mm_irqs_off() 60 mtctl(__pa(next->pgd), 25); in switch_mm_irqs_off() 61 load_context(next->context); in switch_mm_irqs_off() 66 struct mm_struct *next, struct task_struct *tsk) in switch_mm() argument 70 if (prev == next) in switch_mm() 74 switch_mm_irqs_off(prev, next, tsk); in switch_mm() 92 if (next->context.space_id == 0) in activate_mm() 93 next->context.space_id = alloc_sid(); in activate_mm() [all …]
|
| /arch/um/kernel/ |
| A D | tlb.c | 97 unsigned long next; in update_pmd_range() local 102 next = pmd_addr_end(addr, end); in update_pmd_range() 106 next - addr); in update_pmd_range() 120 unsigned long next; in update_pud_range() local 125 next = pud_addr_end(addr, end); in update_pud_range() 129 next - addr); in update_pud_range() 143 unsigned long next; in update_p4d_range() local 148 next = p4d_addr_end(addr, end); in update_p4d_range() 152 next - addr); in update_p4d_range() 186 next - addr); in um_tlb_sync() [all …]
|
| /arch/nios2/mm/ |
| A D | mmu_context.c | 80 void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 89 if (unlikely(CTX_VERSION(next->context) != in switch_mm() 91 next->context = get_new_context(); in switch_mm() 94 pgd_current = next->pgd; in switch_mm() 97 set_context(next->context); in switch_mm() 106 void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 108 next->context = get_new_context(); in activate_mm() 109 set_context(next->context); in activate_mm() 110 pgd_current = next->pgd; in activate_mm()
|
| /arch/arm64/kernel/ |
| A D | process.c | 595 gcs_set_el0_mode(next); in gcs_thread_switch() 639 struct task_struct *next) in cntkctl_thread_switch() argument 645 update_cntkctl_el1(next); in cntkctl_thread_switch() 706 struct task_struct *next) in __switch_to() argument 710 fpsimd_thread_switch(next); in __switch_to() 711 tls_thread_switch(next); in __switch_to() 713 contextidr_thread_switch(next); in __switch_to() 714 entry_task_switch(next); in __switch_to() 715 ssbs_thread_switch(next); in __switch_to() 719 gcs_thread_switch(next); in __switch_to() [all …]
|
| /arch/powerpc/mm/ |
| A D | mmu_context.c | 43 void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, in switch_mm_irqs_off() argument 50 if (!cpumask_test_cpu(cpu, mm_cpumask(next))) { in switch_mm_irqs_off() 51 VM_WARN_ON_ONCE(next == &init_mm); in switch_mm_irqs_off() 52 cpumask_set_cpu(cpu, mm_cpumask(next)); in switch_mm_irqs_off() 53 inc_mm_active_cpus(next); in switch_mm_irqs_off() 84 switch_mm_pgdir(tsk, next); in switch_mm_irqs_off() 87 if (prev == next) in switch_mm_irqs_off() 98 membarrier_arch_switch_mm(prev, next, tsk); in switch_mm_irqs_off() 104 switch_mmu_context(prev, next, tsk); in switch_mm_irqs_off()
|
| /arch/hexagon/include/asm/ |
| A D | mmu_context.h | 29 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 38 if (next->context.generation < prev->context.generation) { in switch_mm() 40 next->pgd[l1] = init_mm.pgd[l1]; in switch_mm() 42 next->context.generation = prev->context.generation; in switch_mm() 45 __vmnewmap((void *)next->context.ptbase); in switch_mm() 52 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 57 switch_mm(prev, next, current_thread_info()->task); in activate_mm()
|
| /arch/loongarch/include/asm/ |
| A D | mmu_context.h | 88 static inline void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, in switch_mm_irqs_off() argument 95 if (!asid_valid(next, cpu)) in switch_mm_irqs_off() 96 get_new_mmu_context(next, cpu, &need_flush); in switch_mm_irqs_off() 98 if (next != &init_mm) in switch_mm_irqs_off() 99 atomic_update_pgd_asid(cpu_asid(cpu, next), (unsigned long)next->pgd); in switch_mm_irqs_off() 101 atomic_update_pgd_asid(cpu_asid(cpu, next), (unsigned long)invalid_pg_dir); in switch_mm_irqs_off() 110 cpumask_set_cpu(cpu, mm_cpumask(next)); in switch_mm_irqs_off() 115 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 121 switch_mm_irqs_off(prev, next, tsk); in switch_mm() 133 #define activate_mm(prev, next) switch_mm(prev, next, current) argument
|
| A D | switch_to.h | 26 struct task_struct *next, struct thread_info *next_ti, 35 #define switch_to(prev, next, last) \ argument 39 hw_breakpoint_thread_switch(next); \ 40 (last) = __switch_to(prev, next, task_thread_info(next), \
|
| /arch/x86/mm/ |
| A D | kasan_init_64.c | 76 unsigned long next; in kasan_populate_pud() local 96 next = pmd_addr_end(addr, end); in kasan_populate_pud() 106 unsigned long next; in kasan_populate_p4d() local 116 next = pud_addr_end(addr, end); in kasan_populate_p4d() 127 unsigned long next; in kasan_populate_pgd() local 136 next = p4d_addr_end(addr, end); in kasan_populate_pgd() 145 unsigned long next; in kasan_populate_shadow() local 209 unsigned long next; in kasan_early_p4d_populate() local 235 unsigned long next; in kasan_map_early_shadow() local 249 unsigned long next; in kasan_shallow_populate_p4ds() local [all …]
|
| A D | ident_map.c | 97 unsigned long next; in ident_pud_init() local 99 for (; addr < end; addr = next) { in ident_pud_init() 104 next = pud_addr_end(addr, end); in ident_pud_init() 132 ident_pmd_init(info, pmd, addr, next); in ident_pud_init() 138 ident_pmd_init(info, pmd, addr, next); in ident_pud_init() 148 unsigned long next; in ident_p4d_init() local 151 for (; addr < end; addr = next) { in ident_p4d_init() 155 next = p4d_addr_end(addr, end); in ident_p4d_init() 183 unsigned long next; in kernel_ident_mapping_init() local 193 for (; addr < end; addr = next) { in kernel_ident_mapping_init() [all …]
|
| /arch/arm/mm/ |
| A D | idmap.c | 28 unsigned long next; in idmap_add_pmd() local 49 next = pmd_addr_end(addr, end); in idmap_add_pmd() 52 } while (pmd++, addr = next, addr != end); in idmap_add_pmd() 73 unsigned long next; in idmap_add_pud() local 76 next = pud_addr_end(addr, end); in idmap_add_pud() 77 idmap_add_pmd(pud, addr, next, prot); in idmap_add_pud() 78 } while (pud++, addr = next, addr != end); in idmap_add_pud() 85 unsigned long next; in identity_mapping_add() local 105 next = pgd_addr_end(addr, end); in identity_mapping_add() 106 idmap_add_pud(pgd, addr, next, prot); in identity_mapping_add() [all …]
|
| /arch/csky/include/asm/ |
| A D | switch_to.h | 10 struct task_struct *next) in __switch_to_fpu() argument 13 restore_from_user_fp(&next->thread.user_fp); in __switch_to_fpu() 17 struct task_struct *next) in __switch_to_fpu() argument 27 #define switch_to(prev, next, last) \ argument 30 struct task_struct *__next = (next); \ 32 ((last) = __switch_to((prev), (next))); \
|
| A D | mmu_context.h | 24 switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 29 if (prev != next) in switch_mm() 30 check_and_switch_context(next, cpu); in switch_mm() 32 setup_pgd(next->pgd, next->context.asid.counter); in switch_mm() 34 flush_icache_deferred(next); in switch_mm()
|
| /arch/alpha/kernel/ |
| A D | err_common.c | 159 struct el_subpacket *next = NULL; in el_process_subpacket_reg() local 163 if (h) next = h->handler(header); in el_process_subpacket_reg() 165 return next; in el_process_subpacket_reg() 197 struct el_subpacket *next = NULL; in el_process_subpacket() local 216 return next; in el_process_subpacket() 286 for (; a->next != NULL; a = a->next) { in cdl_register_subpacket_annotation() 294 a->next = new; in cdl_register_subpacket_annotation() 296 new->next = NULL; in cdl_register_subpacket_annotation() 308 for (; h->next != NULL; h = h->next) { in cdl_register_subpacket_handler() 315 h->next = new; in cdl_register_subpacket_handler() [all …]
|
| /arch/arm64/boot/dts/amd/ |
| A D | elba-16core.dtsi | 46 next-level-cache = <&l2_0>; 54 next-level-cache = <&l2_0>; 62 next-level-cache = <&l2_0>; 70 next-level-cache = <&l2_0>; 85 next-level-cache = <&l2_1>; 93 next-level-cache = <&l2_1>; 101 next-level-cache = <&l2_1>; 109 next-level-cache = <&l2_1>; 124 next-level-cache = <&l2_2>; 132 next-level-cache = <&l2_2>; [all …]
|
| /arch/arm64/mm/ |
| A D | mmu.c | 206 unsigned long next; in alloc_init_cont_pte() local 241 phys += next - addr; in alloc_init_cont_pte() 256 unsigned long next; in init_pmd() local 291 unsigned long next; in alloc_init_cont_pmd() local 340 unsigned long next; in alloc_init_pud() local 398 unsigned long next; in alloc_init_p4d() local 899 unsigned long next; in unmap_hotplug_pmd_range() local 932 unsigned long next; in unmap_hotplug_pud_range() local 965 unsigned long next; in unmap_hotplug_p4d_range() local 1306 next = addr; in pud_free_pmd_page() [all …]
|
| A D | trans_pgd.c | 90 unsigned long next; in copy_pmd() local 105 next = pmd_addr_end(addr, end); in copy_pmd() 109 if (copy_pte(info, dst_pmdp, src_pmdp, addr, next)) in copy_pmd() 126 unsigned long next; in copy_pud() local 141 next = pud_addr_end(addr, end); in copy_pud() 162 unsigned long next; in copy_p4d() local 175 next = p4d_addr_end(addr, end); in copy_p4d() 178 if (copy_pud(info, dst_p4dp, src_p4dp, addr, next)) in copy_p4d() 188 unsigned long next; in copy_page_tables() local 194 next = pgd_addr_end(addr, end); in copy_page_tables() [all …]
|
| /arch/s390/include/asm/ |
| A D | mmu_context.h | 71 static inline void switch_mm_irqs_off(struct mm_struct *prev, struct mm_struct *next, in switch_mm_irqs_off() argument 76 if (next == &init_mm) in switch_mm_irqs_off() 79 get_lowcore()->user_asce.val = next->context.asce; in switch_mm_irqs_off() 80 cpumask_set_cpu(cpu, &next->context.cpu_attach_mask); in switch_mm_irqs_off() 84 if (prev != next) in switch_mm_irqs_off() 89 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 95 switch_mm_irqs_off(prev, next, tsk); in switch_mm() 125 struct mm_struct *next) in activate_mm() argument 127 switch_mm_irqs_off(prev, next, current); in activate_mm() 128 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); in activate_mm()
|
| /arch/arm/include/asm/ |
| A D | mmu_context.h | 108 #define activate_mm(prev,next) switch_mm(prev, next, NULL) argument 117 switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 129 !cpumask_empty(mm_cpumask(next)) && in switch_mm() 130 !cpumask_test_cpu(cpu, mm_cpumask(next))) in switch_mm() 133 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { in switch_mm() 134 check_and_switch_context(next, tsk); in switch_mm()
|
| /arch/sh/include/asm/ |
| A D | switch_to_32.h | 65 struct task_struct *next); 70 #define switch_to(prev, next, last) \ argument 82 if (is_dsp_enabled(next)) \ 83 __restore_dsp(next); \ 88 __ts5 = (u32 *)next; \ 89 __ts6 = (u32 *)&next->thread.sp; \ 90 __ts7 = next->thread.pc; \
|
| /arch/s390/mm/ |
| A D | pgalloc.c | 227 return (next - 1) < (end - 1) ? next : end; \ 249 unsigned long *pte, next; in base_page_walk() local 256 next = base_page_addr_end(addr, end); in base_page_walk() 258 } while (pte++, addr = next, addr < end); in base_page_walk() 265 unsigned long *ste, next, *table; in base_segment_walk() local 271 next = base_segment_addr_end(addr, end); in base_segment_walk() 287 } while (ste++, addr = next, addr < end); in base_segment_walk() 294 unsigned long *rtte, next, *table; in base_region3_walk() local 300 next = base_region3_addr_end(addr, end); in base_region3_walk() 322 unsigned long *rste, next, *table; in base_region2_walk() local [all …]
|