| /arch/x86/include/asm/ |
| A D | smap.h | 19 ALTERNATIVE "", "clac", X86_FEATURE_SMAP 22 ALTERNATIVE "", "stac", X86_FEATURE_SMAP 43 ALTERNATIVE(ANNOTATE_IGNORE_ALTERNATIVE in smap_save() 54 ALTERNATIVE(ANNOTATE_IGNORE_ALTERNATIVE in smap_restore() 62 ALTERNATIVE("", "clac", X86_FEATURE_SMAP) 64 ALTERNATIVE("", "stac", X86_FEATURE_SMAP) 67 ALTERNATIVE("", ANNOTATE_IGNORE_ALTERNATIVE "clac", X86_FEATURE_SMAP) 69 ALTERNATIVE("", ANNOTATE_IGNORE_ALTERNATIVE "stac", X86_FEATURE_SMAP)
|
| A D | alternative.h | 215 #define ALTERNATIVE(oldinstr, newinstr, ft_flags) \ macro 221 ALTERNATIVE(ALTERNATIVE(oldinstr, newinstr1, ft_flags1), newinstr2, ft_flags2) 229 ALTERNATIVE(ALTERNATIVE_2(oldinstr, newinstr1, ft_flags1, newinstr2, ft_flags2), \ 245 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) : : : "memory") 259 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 264 asm_inline volatile(ALTERNATIVE(oldinstr, newinstr, ft_flags) \ 278 asm_inline volatile(ALTERNATIVE("call %c[old]", "call %c[new]", ft_flags) \ 372 .macro ALTERNATIVE oldinstr, newinstr, ft_flags
|
| A D | barrier.h | 15 #define mb() asm volatile(ALTERNATIVE("lock addl $0,-4(%%esp)", "mfence", \ 17 #define rmb() asm volatile(ALTERNATIVE("lock addl $0,-4(%%esp)", "lfence", \ 19 #define wmb() asm volatile(ALTERNATIVE("lock addl $0,-4(%%esp)", "sfence", \
|
| A D | nospec-branch.h | 299 ALTERNATIVE "", \ 313 ALTERNATIVE "", "verw x86_verw_sel(%rip)", \feature 320 ALTERNATIVE "", "verw %cs:x86_verw_sel", \feature 332 ALTERNATIVE "", "call clear_bhb_loop", X86_FEATURE_CLEAR_BHB_LOOP 336 ALTERNATIVE "", "call clear_bhb_loop", X86_FEATURE_CLEAR_BHB_VMEXIT 401 ALTERNATIVE("", \ 525 asm volatile(ALTERNATIVE("", "wrmsr", %c[feature]) 535 asm_inline volatile(ALTERNATIVE("", "call write_ibpb", X86_FEATURE_IBPB)
|
| A D | arch_hweight.h | 19 asm_inline (ALTERNATIVE("call __sw_hweight32", in __arch_hweight32() 48 asm_inline (ALTERNATIVE("call __sw_hweight64", in __arch_hweight64()
|
| A D | uaccess_64.h | 29 asm_inline (ALTERNATIVE("", "and " __percpu_arg([mask]) ", %[addr]", in __untagged_addr() 128 ALTERNATIVE("rep movsb", in copy_user_generic() 190 ALTERNATIVE("rep stosb", in __clear_user()
|
| /arch/parisc/include/asm/ |
| A D | cache.h | 53 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 56 ALTERNATIVE(ALT_COND_NO_SMP, INSN_PxTLB) \ 57 ALTERNATIVE(ALT_COND_NO_SPLIT_TLB, INSN_NOP) \ 61 ALTERNATIVE(ALT_COND_NO_DCACHE, INSN_NOP) \ 62 ALTERNATIVE(ALT_COND_NO_IOC_FDC, INSN_NOP) \ 65 ALTERNATIVE(ALT_COND_NO_DCACHE, INSN_NOP) \ 66 ALTERNATIVE(ALT_COND_NO_IOC_FDC, INSN_NOP) :::"memory")
|
| A D | alternative.h | 36 #define ALTERNATIVE(cond, replacement) "!0:" \ macro 47 #define ALTERNATIVE(from, to, cond, replacement)\ macro
|
| /arch/riscv/include/asm/ |
| A D | errata_list.h | 35 ALTERNATIVE(__stringify(RISCV_PTR do_trap_insn_fault), \ 41 ALTERNATIVE(__stringify(RISCV_PTR do_page_fault), \ 48 asm(ALTERNATIVE("sfence.vma x0, %0", "sfence.vma", SIFIVE_VENDOR_ID, \ 53 asm(ALTERNATIVE("sfence.vma %0", "sfence.vma", SIFIVE_VENDOR_ID, \ 58 asm(ALTERNATIVE("sfence.vma %0, %1", "sfence.vma", SIFIVE_VENDOR_ID, \ 87 asm volatile(ALTERNATIVE( \ 108 asm volatile(ALTERNATIVE( \
|
| A D | arch_hweight.h | 23 asm goto(ALTERNATIVE("j %l[legacy]", "nop", 0, in __arch_hweight32() 54 asm goto(ALTERNATIVE("j %l[legacy]", "nop", 0, in __arch_hweight64()
|
| A D | cpufeature-macros.h | 21 asm goto(ALTERNATIVE("j %l[l_no]", "nop", %[vendor], %[ext], 1) in __riscv_has_extension_likely() 35 asm goto(ALTERNATIVE("nop", "j %l[l_yes]", %[vendor], %[ext], 1) in __riscv_has_extension_unlikely()
|
| A D | bitops.h | 50 asm goto(ALTERNATIVE("j %l[legacy]", "nop", 0, in variable__ffs() 79 asm goto(ALTERNATIVE("j %l[legacy]", "nop", 0, in variable__fls() 108 asm goto(ALTERNATIVE("j %l[legacy]", "nop", 0, in variable_ffs() 140 asm goto(ALTERNATIVE("j %l[legacy]", "nop", 0, in variable_fls()
|
| /arch/x86/entry/ |
| A D | calling.h | 169 ALTERNATIVE "", "SET_NOFLUSH_BIT \reg", X86_FEATURE_PCID 175 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 188 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 218 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 224 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 232 ALTERNATIVE "jmp .Ldone_\@", "", X86_FEATURE_PTI 251 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 261 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 308 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_KERNEL_IBRS 337 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_KERNEL_IBRS [all …]
|
| A D | entry_64.S | 130 ALTERNATIVE "testb %al, %al; jz swapgs_restore_regs_and_return_to_usermode", \ 255 ALTERNATIVE "jmp swapgs_restore_regs_and_return_to_usermode", \ 299 ALTERNATIVE "call error_entry; movq %rax, %rsp", \ 563 ALTERNATIVE "", "jmp xenpv_restore_regs_and_return_to_usermode", X86_FEATURE_XENPV 566 ALTERNATIVE "", "jmp .Lpti_restore_regs_and_return_to_usermode", X86_FEATURE_PTI 752 2: ALTERNATIVE "", "mfence", X86_BUG_SWAPGS_FENCE 765 ALTERNATIVE "", "ZAP_GS", X86_BUG_NULL_SEG 899 ALTERNATIVE "jmp .Lparanoid_entry_checkgs", "", X86_FEATURE_FSGSBASE 985 ALTERNATIVE "jmp .Lparanoid_exit_checkgs", "", X86_FEATURE_FSGSBASE 1421 ALTERNATIVE "jmp nmi_no_fsgsbase", "", X86_FEATURE_FSGSBASE
|
| /arch/s390/include/asm/ |
| A D | alternative.h | 131 #define ALTERNATIVE(oldinstr, altinstr, feature) \ macro 164 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) : : : "memory") 172 asm_inline volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 177 asm_inline volatile(ALTERNATIVE(oldinstr, altinstr, feature) \ 209 .macro ALTERNATIVE oldinstr, newinstr, feature
|
| A D | lowcore.h | 225 ALTERNATIVE(" lghi %[lc],0", in get_lowcore() 243 ALTERNATIVE "lghi \reg,0", \ 249 ALTERNATIVE "stmg \start, \end, \savearea", \
|
| /arch/arm64/include/asm/ |
| A D | arch_timer.h | 69 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0", in arch_timer_read_cntpct_el0() 81 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0", in arch_timer_read_cntvct_el0() 183 asm volatile(ALTERNATIVE("isb\n mrs %0, cntpct_el0", in __arch_counter_get_cntpct() 204 asm volatile(ALTERNATIVE("isb\n mrs %0, cntvct_el0", in __arch_counter_get_cntvct()
|
| A D | alternative-macros.h | 219 #define ALTERNATIVE(oldinstr, newinstr, ...) \ macro 234 ALTERNATIVE("b %l[l_no]", "nop", %[cpucap]) 255 ALTERNATIVE("nop", "b %l[l_yes]", %[cpucap]) in alternative_has_cap_unlikely()
|
| A D | io.h | 57 asm volatile(ALTERNATIVE("ldrb %w0, [%1]", in __raw_readb() 69 asm volatile(ALTERNATIVE("ldrh %w0, [%1]", in __raw_readw() 80 asm volatile(ALTERNATIVE("ldr %w0, [%1]", in __raw_readl() 91 asm volatile(ALTERNATIVE("ldr %0, [%1]", in __raw_readq()
|
| A D | mte-kasan.h | 56 asm volatile(ALTERNATIVE("nop", SET_PSTATE_TCO(0), in mte_disable_tco() 62 asm volatile(ALTERNATIVE("nop", SET_PSTATE_TCO(1), in mte_enable_tco()
|
| /arch/s390/kernel/ |
| A D | entry.S | 37 ALTERNATIVE "nop", ".insn s,0xb2010000,\address", ALT_FACILITY(193) 41 ALTERNATIVE "nop", ".insn s,0xb2000000,\address", ALT_FACILITY(193) 52 ALTERNATIVE "brcl 0,0", __stringify(mvc __PT_LAST_BREAK(8,\reg),__LC_LAST_BREAK(\lowcore)),\ 97 ALTERNATIVE "nop", ".insn rrf,0xb2e80000,0,0,12,0", ALT_SPEC(82) 101 ALTERNATIVE "nop", ".insn rrf,0xb2e80000,0,0,13,0", ALT_SPEC(82) 105 ALTERNATIVE "TSTMSK \tif_ptr,\tif_mask; jz .+8; .insn rrf,0xb2e80000,0,0,13,0", \ 111 ALTERNATIVE "jz .+8; .insn rrf,0xb2e80000,0,0,12,0", \ 167 ALTERNATIVE "nop", "lpp _LPP_OFFSET(%r13)", ALT_FACILITY(40)
|
| /arch/parisc/kernel/ |
| A D | pacache.S | 244 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_ICACHE, INSN_NOP) 305 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 548 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 549 ALTERNATIVE(1b, 1b+4, ALT_COND_NO_SMP, INSN_PxTLB) 677 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 743 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 775 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 792 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) 824 89: ALTERNATIVE(88b, 89b, ALT_COND_NO_DCACHE, INSN_NOP) 847 ALTERNATIVE(0b, 0b+4, ALT_COND_NO_SMP, INSN_PxTLB) [all …]
|
| /arch/riscv/lib/ |
| A D | csum.c | 51 asm goto(ALTERNATIVE("j %l[no_zbb]", "nop", 0, in csum_ipv6_magic() 163 asm goto(ALTERNATIVE("j %l[no_zbb]", "nop", 0, in do_csum_with_alignment() 245 asm goto(ALTERNATIVE("j %l[no_zbb]", "nop", 0, in do_csum_no_alignment()
|
| /arch/s390/kernel/vdso64/ |
| A D | vgetrandom-chacha.S | 50 ALTERNATIVE __stringify(VL BEPERM,16,,%r1), "brcl 0,0", ALT_FACILITY(148) 133 ALTERNATIVE \ 172 ALTERNATIVE "nopr", "br %r14", ALT_FACILITY(148)
|
| /arch/loongarch/include/asm/ |
| A D | alternative.h | 71 #define ALTERNATIVE(oldinstr, newinstr, feature) \ macro 104 (asm volatile (ALTERNATIVE(oldinstr, newinstr, feature) : : : "memory"))
|