| /arch/arc/include/asm/ |
| A D | atomic64-arcv2.h | 49 #define ATOMIC64_OP(op, op1, op2) \ argument 58 " " #op2 " %H0, %H0, %H2 \n" \ 75 " " #op2 " %H0, %H0, %H2 \n" \ 97 " " #op2 " %H1, %H0, %H3 \n" \ 115 #define ATOMIC64_OPS(op, op1, op2) \ argument 116 ATOMIC64_OP(op, op1, op2) \ 117 ATOMIC64_OP_RETURN(op, op1, op2) \ 118 ATOMIC64_FETCH_OP(op, op1, op2) 124 #define ATOMIC64_OPS(op, op1, op2) \ argument 125 ATOMIC64_OP(op, op1, op2) \ [all …]
|
| /arch/arm/include/asm/ |
| A D | atomic.h | 311 #define ATOMIC64_OP(op, op1, op2) \ argument 321 " " #op2 " %R0, %R0, %R4\n" \ 342 " " #op2 " %R0, %R0, %R4\n" \ 365 " " #op2 " %R1, %R0, %R5\n" \ 376 #define ATOMIC64_OPS(op, op1, op2) \ argument 377 ATOMIC64_OP(op, op1, op2) \ 378 ATOMIC64_OP_RETURN(op, op1, op2) \ 379 ATOMIC64_FETCH_OP(op, op1, op2) 390 #define ATOMIC64_OPS(op, op1, op2) \ in ATOMIC64_OPS() argument 391 ATOMIC64_OP(op, op1, op2) \ in ATOMIC64_OPS() [all …]
|
| /arch/sh/kernel/ |
| A D | kprobes.c | 144 struct kprobe *op1, *op2; in prepare_singlestep() local 149 op2 = this_cpu_ptr(&saved_next_opcode2); in prepare_singlestep() 173 op2->addr = in prepare_singlestep() 175 op2->opcode = *(op2->addr); in prepare_singlestep() 176 arch_arm_kprobe(op2); in prepare_singlestep() 183 op2->addr = in prepare_singlestep() 185 op2->opcode = *(op2->addr); in prepare_singlestep() 186 arch_arm_kprobe(op2); in prepare_singlestep()
|
| /arch/sparc/kernel/ |
| A D | uprobes.c | 62 u32 op2 = (insn >> 22) & 0x7; in arch_uprobe_copy_ixol() local 65 (op2 == 1 || op2 == 2 || op2 == 3 || op2 == 5 || op2 == 6) && in arch_uprobe_copy_ixol()
|
| /arch/x86/crypto/ |
| A D | cast6-avx-x86_64-asm_64.S | 85 #define lookup_32bit(src, dst, op1, op2, op3, interleave_op, il_reg) \ argument 95 op2 (RID2,RID1,4), dst ## d; \ 115 #define F_tail(a, x, gi1, gi2, op1, op2, op3) \ argument 116 lookup_32bit(##gi1, RFS1, op1, op2, op3, shr_next, ##gi1); \ 117 lookup_32bit(##gi2, RFS3, op1, op2, op3, shr_next, ##gi2); \ 119 lookup_32bit(##gi1, RFS2, op1, op2, op3, dummy, none); \ 122 lookup_32bit(##gi2, RFS1, op1, op2, op3, dummy, none); \ 129 #define F_2(a1, b1, a2, b2, op0, op1, op2, op3) \ argument 133 F_tail(b1, RX, RGI1, RGI2, op1, op2, op3); \ 134 F_tail(b2, RTMP, RGI3, RGI4, op1, op2, op3); \
|
| A D | cast5-avx-x86_64-asm_64.S | 85 #define lookup_32bit(src, dst, op1, op2, op3, interleave_op, il_reg) \ argument 95 op2 (RID2,RID1,4), dst ## d; \ 115 #define F_tail(a, x, gi1, gi2, op1, op2, op3) \ argument 116 lookup_32bit(##gi1, RFS1, op1, op2, op3, shr_next, ##gi1); \ 117 lookup_32bit(##gi2, RFS3, op1, op2, op3, shr_next, ##gi2); \ 119 lookup_32bit(##gi1, RFS2, op1, op2, op3, dummy, none); \ 122 lookup_32bit(##gi2, RFS1, op1, op2, op3, dummy, none); \ 129 #define F_2(a1, b1, a2, b2, op0, op1, op2, op3) \ argument 133 F_tail(b1, RX, RGI1, RGI2, op1, op2, op3); \ 134 F_tail(b2, RTMP, RGI3, RGI4, op1, op2, op3); \
|
| A D | polyval-clmulni_glue.c | 48 asmlinkage void clmul_polyval_mul(u8 *op1, const u8 *op2); 63 static void internal_polyval_mul(u8 *op1, const u8 *op2) in internal_polyval_mul() argument 66 clmul_polyval_mul(op1, op2); in internal_polyval_mul()
|
| A D | twofish-x86_64-asm_64-3way.S | 78 #define do16bit_ror(rot, op1, op2, T0, T1, tmp1, tmp2, ab, dst) \ argument 83 op2##l T1(CTX, tmp1, 4), dst ## d;
|
| /arch/arm64/tools/ |
| A D | gen-sysreg.awk | 159 op2 = $7 165 define("REG_" reg, "S" op0 "_" op1 "_C" crn "_C" crm "_" op2) 166 define("SYS_" reg, "sys_reg(" op0 ", " op1 ", " crn ", " crm ", " op2 ")") 172 define("SYS_" reg "_Op2", op2) 200 op2 = null
|
| /arch/x86/kvm/svm/ |
| A D | svm_ops.h | 29 #define svm_asm2(insn, op1, op2, clobber...) \ argument 33 :: op1, op2 : clobber : fault); \
|
| /arch/arm64/crypto/ |
| A D | polyval-ce-glue.c | 43 asmlinkage void pmull_polyval_mul(u8 *op1, const u8 *op2); 53 static void internal_polyval_mul(u8 *op1, const u8 *op2) in internal_polyval_mul() argument 56 pmull_polyval_mul(op1, op2); in internal_polyval_mul()
|
| /arch/powerpc/math-emu/ |
| A D | math.c | 28 #define FLOATFUNC(x) static inline int x(void *op1, void *op2, void *op3, \ 228 void *op0 = NULL, *op1 = NULL, *op2 = NULL, *op3 = NULL; in do_mathemu() local 334 op2 = (void *)¤t->thread.TS_FPR((insn >> 11) & 0x1f); in do_mathemu() 340 op2 = (void *)¤t->thread.TS_FPR((insn >> 6) & 0x1f); in do_mathemu() 346 op2 = (void *)¤t->thread.TS_FPR((insn >> 11) & 0x1f); in do_mathemu() 400 op2 = (void *)¤t->thread.TS_FPR((insn >> 16) & 0x1f); in do_mathemu() 407 op2 = (void *)(long)((insn >> 18) & 0x7); in do_mathemu() 435 eflag = func(op0, op1, op2, op3); in do_mathemu()
|
| /arch/arm64/include/uapi/asm/ |
| A D | kvm.h | 247 #define __ARM64_SYS_REG(op0,op1,crn,crm,op2) \ argument 253 ARM64_SYS_REG_SHIFT_MASK(op2, OP2)) 545 #define KVM_ARM_FEATURE_ID_RANGE_IDX(op0, op1, crn, crm, op2) \ argument 549 (__op1 << 6 | ((crm) & 7) << 3 | (op2)); \
|
| /arch/arm64/include/asm/ |
| A D | esr.h | 234 #define ESR_ELx_SYS64_ISS_SYS_VAL(op0, op1, op2, crn, crm) \ argument 237 ((op2) << ESR_ELx_SYS64_ISS_OP2_SHIFT) | \ 350 #define ESR_ELx_CP15_32_ISS_SYS_VAL(op1, op2, crn, crm) \ argument 352 ((op2) << ESR_ELx_CP15_32_ISS_OP2_SHIFT) | \
|
| A D | sysreg.h | 40 #define sys_reg(op0, op1, crn, crm, op2) \ argument 43 ((op2) << Op2_shift)) 92 #define pstate_field(op1, op2) ((op1) << Op1_shift | (op2) << Op2_shift) argument 116 #define __SYS_BARRIER_INSN(op0, op1, CRn, CRm, op2, Rt) \ argument 118 sys_insn((op0), (op1), (CRn), (CRm), (op2)) | \ 438 #define SYS_AM_EL0(crm, op2) sys_reg(3, 3, 13, (crm), (op2)) argument
|
| /arch/s390/include/asm/ |
| A D | percpu.h | 66 #define arch_this_cpu_add(pcp, val, op1, op2, szcast) \ argument 76 op2 " %[ptr__],%[val__]\n" \
|
| /arch/x86/kvm/vmx/ |
| A D | vmx_ops.h | 208 #define vmx_asm2(insn, op1, op2, error_args...) \ argument 214 : : op1, op2 : "cc" : error, fault); \
|
| /arch/s390/net/ |
| A D | bpf_jit_comp.c | 218 #define _EMIT6(op1, op2) \ argument 222 *(u16 *) (jit->prg_buf + jit->prg + 4) = (op2); \ 227 #define _EMIT6_DISP(op1, op2, disp) \ argument 230 _EMIT6((op1) | __disp, op2); \ 233 #define _EMIT6_DISP_LH(op1, op2, disp) \ argument 238 _EMIT6((op1) | __disp_l, (op2) | __disp_h >> 4); \ 241 #define EMIT6_DISP_LH(op1, op2, b1, b2, b3, disp) \ argument 244 reg_high(b3) << 8, op2, disp); \ 254 (op2) | (mask) << 12); \ 263 (rel & 0xffff), (op2) | ((imm) & 0xff) << 8); \ [all …]
|
| /arch/arm64/kvm/hyp/nvhe/ |
| A D | sys_regs.c | 352 #define ID_UNALLOCATED(crm, op2) { \ argument 353 Op0(3), Op1(0), CRn(0), CRm(crm), Op2(op2), \
|
| /arch/arm/include/asm/hardware/ |
| A D | cp14.h | 17 #define MRC14(op1, crn, crm, op2) \ argument 20 asm volatile("mrc p14, "#op1", %0, "#crn", "#crm", "#op2 : "=r" (val)); \ 24 #define MCR14(val, op1, crn, crm, op2) \ argument 26 asm volatile("mcr p14, "#op1", %0, "#crn", "#crm", "#op2 : : "r" (val));\
|
| /arch/s390/kvm/ |
| A D | priv.c | 88 u64 op2; in handle_set_clock() local 95 op2 = kvm_s390_get_base_disp_s(vcpu, &ar); in handle_set_clock() 96 if (op2 & 7) /* Operand must be on a doubleword boundary */ in handle_set_clock() 98 rc = read_guest(vcpu, op2, ar, >od.tod, sizeof(gtod.tod)); in handle_set_clock()
|
| /arch/x86/kernel/ |
| A D | alternative.c | 953 u8 op1, op2; in apply_retpolines() local 961 op2 = insn.opcode.bytes[1]; in apply_retpolines() 980 if (op2 >= 0x80 && op2 <= 0x8f) in apply_retpolines()
|
| /arch/mips/kernel/ |
| A D | traps.c | 727 int op2 = opcode & CSR_OPCODE2_MASK; in simulate_loongson3_cpucfg() local 730 if (op == LWC2 && op2 == CSR_OPCODE2 && csr_func == CSR_FUNC_CPUCFG) { in simulate_loongson3_cpucfg()
|
| /arch/m68k/fpsp040/ |
| A D | bugfix.S | 195 bne op2sgl |not opclass 0, check op2
|
| /arch/arm64/kvm/ |
| A D | emulate-nested.c | 2086 u8 op0, op1, crn, crm, op2; in encoding_next() local 2092 op2 = sys_reg_Op2(encoding); in encoding_next() 2094 if (op2 < Op2_mask) in encoding_next() 2095 return sys_reg(op0, op1, crn, crm, op2 + 1); in encoding_next()
|