/linux-6.3-rc2/include/asm-generic/ |
A D | xor.h | 48 p3 += 8; in xor_8regs_3() 71 p3 += 8; in xor_8regs_4() 96 p3 += 8; in xor_8regs_5() 182 p3 += 8; in xor_32regs_3() 238 p3 += 8; in xor_32regs_4() 304 p3 += 8; in xor_32regs_5() 362 p3 += 8; in xor_8regs_p_3() 397 p3 += 8; in xor_8regs_p_4() 436 p3 += 8; in xor_8regs_p_5() 542 p3 += 8; in xor_32regs_p_3() [all …]
|
/linux-6.3-rc2/arch/arm/include/asm/ |
A D | xor.h | 70 const unsigned long * __restrict p3) in xor_arm4regs_3() argument 85 XOR_BLOCK_4(p3); in xor_arm4regs_3() 93 const unsigned long * __restrict p3, in xor_arm4regs_4() argument 105 XOR_BLOCK_2(p3); in xor_arm4regs_4() 114 const unsigned long * __restrict p3, in xor_arm4regs_5() argument 127 XOR_BLOCK_2(p3); in xor_arm4regs_5() 171 const unsigned long * __restrict p3) in xor_neon_3() argument 174 xor_arm4regs_3(bytes, p1, p2, p3); in xor_neon_3() 185 const unsigned long * __restrict p3, in xor_neon_4() argument 189 xor_arm4regs_4(bytes, p1, p2, p3, p4); in xor_neon_4() [all …]
|
/linux-6.3-rc2/arch/sparc/include/asm/ |
A D | xor_32.h | 56 const unsigned long * __restrict p3) in sparc_3() argument 95 : "r" (p1), "r" (p2), "r" (p3) in sparc_3() 101 p3 += 8; in sparc_3() 108 const unsigned long * __restrict p3, in sparc_4() argument 160 : "r" (p1), "r" (p2), "r" (p3), "r" (p4) in sparc_4() 166 p3 += 8; in sparc_4() 174 const unsigned long * __restrict p3, in sparc_5() argument 239 : "r" (p1), "r" (p2), "r" (p3), "r" (p4), "r" (p5) in sparc_5() 245 p3 += 8; in sparc_5()
|
A D | xor_64.h | 19 const unsigned long * __restrict p3); 22 const unsigned long * __restrict p3, 26 const unsigned long * __restrict p3, 44 const unsigned long * __restrict p3); 47 const unsigned long * __restrict p3, 51 const unsigned long * __restrict p3,
|
/linux-6.3-rc2/arch/hexagon/lib/ |
A D | memcpy.S | 255 p3 = cmp.gtu(back, #8); define 290 if(p3) dataF8 = memd(ptr_in+#8); 339 p3 = cmp.gtu(kernel, #0); define 349 p3 = cmp.gtu(dalign, #24); define 352 if(p3) jump .Lodd_alignment; 367 p3 = cmp.eq(kernel, rest); define 443 p3 = cmp.eq(epilogdws, #0); define 449 if(p3) jump .Lepilog60; 455 p3 = cmp.eq(kernel, #0); define 456 if(p3.new) kernel= #8; [all …]
|
/linux-6.3-rc2/tools/testing/selftests/ftrace/test.d/instances/ |
A D | instance-event.tc | 55 p3=$! 56 echo $p3 60 kill -1 $p3 65 wait $p1 $p2 $p3 107 p3=$! 108 echo $p3 123 kill -1 $p3 128 wait $p1 $p2 $p3 $p4 $p5
|
A D | instance.tc | 47 p3=$! 48 echo $p3 63 kill -1 $p3 68 wait $p1 $p2 $p3 $p4 $p5
|
/linux-6.3-rc2/arch/x86/include/asm/ |
A D | xor_avx.h | 90 const unsigned long * __restrict p3) in xor_avx_4() argument 100 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p3[i / sizeof(*p3)])); \ in xor_avx_4() 116 p3 = (unsigned long *)((uintptr_t)p3 + 512); in xor_avx_4() 125 const unsigned long * __restrict p3, in xor_avx_5() argument 138 "m" (p3[i / sizeof(*p3)])); \ in xor_avx_5() 154 p3 = (unsigned long *)((uintptr_t)p3 + 512); in xor_avx_5()
|
A D | xor.h | 149 const unsigned long * __restrict p3) in xor_sse_3() argument 199 [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3) in xor_sse_3() 209 const unsigned long * __restrict p3) in xor_sse_3_pf64() argument 237 [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3) in xor_sse_3_pf64() 247 const unsigned long * __restrict p3, in xor_sse_4() argument 305 [p2] "+r" (p2), [p3] "+r" (p3), [p4] "+r" (p4) in xor_sse_4() 315 const unsigned long * __restrict p3, in xor_sse_4_pf64() argument 346 [p2] "+r" (p2), [p3] "+r" (p3), [p4] "+r" (p4) in xor_sse_4_pf64() 356 const unsigned long * __restrict p3, in xor_sse_5() argument 422 [p3] "+r" (p3), [p4] "+r" (p4), [p5] "+r" (p5) in xor_sse_5() [all …]
|
A D | kvm_para.h | 78 unsigned long p2, unsigned long p3) in kvm_hypercall3() argument 83 return tdx_kvm_hypercall(nr, p1, p2, p3, 0); in kvm_hypercall3() 87 : "a"(nr), "b"(p1), "c"(p2), "d"(p3) in kvm_hypercall3() 93 unsigned long p2, unsigned long p3, in kvm_hypercall4() argument 99 return tdx_kvm_hypercall(nr, p1, p2, p3, p4); in kvm_hypercall4() 103 : "a"(nr), "b"(p1), "c"(p2), "d"(p3), "S"(p4) in kvm_hypercall4() 109 unsigned long p2, unsigned long p3) in kvm_sev_hypercall3() argument 115 : "a"(nr), "b"(p1), "c"(p2), "d"(p3) in kvm_sev_hypercall3()
|
A D | xor_32.h | 70 const unsigned long * __restrict p3) in xor_pII_mmx_3() argument 110 "+r" (p1), "+r" (p2), "+r" (p3) in xor_pII_mmx_3() 120 const unsigned long * __restrict p3, in xor_pII_mmx_4() argument 166 "+r" (p1), "+r" (p2), "+r" (p3), "+r" (p4) in xor_pII_mmx_4() 177 const unsigned long * __restrict p3, in xor_pII_mmx_5() argument 237 "+r" (p1), "+r" (p2), "+r" (p3) in xor_pII_mmx_5() 308 const unsigned long * __restrict p3) in xor_p5_mmx_3() argument 356 "+r" (p1), "+r" (p2), "+r" (p3) in xor_p5_mmx_3() 366 const unsigned long * __restrict p3, in xor_p5_mmx_4() argument 434 const unsigned long * __restrict p3, in xor_p5_mmx_5() argument [all …]
|
/linux-6.3-rc2/arch/s390/lib/ |
A D | xor.c | 38 const unsigned long * __restrict p3) in xor_xc_3() argument 59 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3) in xor_xc_3() 65 const unsigned long * __restrict p3, in xor_xc_4() argument 91 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3), "+a" (p4) in xor_xc_4() 97 const unsigned long * __restrict p3, in xor_xc_5() argument 128 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3), "+a" (p4), in xor_xc_5()
|
/linux-6.3-rc2/scripts/coccinelle/free/ |
A D | clk_put.cocci | 23 position p1,p2,p3; 46 return@p3 ...; } 54 p3 << clk.p3; 59 cocci.print_secs("needed clk_put",p3) 64 p3 << clk.p3; 68 coccilib.report.print_report(p3[0],msg)
|
A D | iounmap.cocci | 23 position p1,p2,p3; 46 return@p3 ...; } 54 p3 << iom.p3; 59 cocci.print_secs("needed iounmap",p3) 64 p3 << iom.p3; 68 coccilib.report.print_report(p3[0],msg)
|
/linux-6.3-rc2/arch/powerpc/lib/ |
A D | xor_vmx_glue.c | 28 const unsigned long * __restrict p3) in xor_altivec_3() argument 32 __xor_altivec_3(bytes, p1, p2, p3); in xor_altivec_3() 40 const unsigned long * __restrict p3, in xor_altivec_4() argument 45 __xor_altivec_4(bytes, p1, p2, p3, p4); in xor_altivec_4() 53 const unsigned long * __restrict p3, in xor_altivec_5() argument 59 __xor_altivec_5(bytes, p1, p2, p3, p4, p5); in xor_altivec_5()
|
A D | xor_vmx.h | 13 const unsigned long * __restrict p3); 16 const unsigned long * __restrict p3, 20 const unsigned long * __restrict p3,
|
/linux-6.3-rc2/arch/arm64/include/asm/ |
A D | xor.h | 30 const unsigned long * __restrict p3) in xor_neon_3() argument 33 xor_block_inner_neon.do_3(bytes, p1, p2, p3); in xor_neon_3() 40 const unsigned long * __restrict p3, in xor_neon_4() argument 44 xor_block_inner_neon.do_4(bytes, p1, p2, p3, p4); in xor_neon_4() 51 const unsigned long * __restrict p3, in xor_neon_5() argument 56 xor_block_inner_neon.do_5(bytes, p1, p2, p3, p4, p5); in xor_neon_5()
|
/linux-6.3-rc2/arch/alpha/include/asm/ |
A D | xor.h | 14 const unsigned long * __restrict p3); 18 const unsigned long * __restrict p3, 23 const unsigned long * __restrict p3, 33 const unsigned long * __restrict p3); 37 const unsigned long * __restrict p3, 42 const unsigned long * __restrict p3,
|
/linux-6.3-rc2/arch/hexagon/mm/ |
A D | copy_user_template.S | 35 p3=sp1loop0(.Loop8,loopcount) 40 if (p3) memd(dst++#8) = d_dbuf 65 p3=sp1loop0(.Loop4,loopcount) 70 if (p3) memw(dst++#4) = w_dbuf 91 p3=sp1loop0(.Loop2,loopcount) 96 if (p3) memh(dst++#2) = w_dbuf 107 p3=sp1loop0(.Loop1,bytes) 112 if (p3) memb(dst++#1) = w_dbuf
|
/linux-6.3-rc2/arch/arm64/lib/ |
A D | xor-neon.c | 42 const unsigned long * __restrict p3) in xor_arm64_neon_3() argument 46 uint64_t *dp3 = (uint64_t *)p3; in xor_arm64_neon_3() 78 const unsigned long * __restrict p3, in xor_arm64_neon_4() argument 83 uint64_t *dp3 = (uint64_t *)p3; in xor_arm64_neon_4() 123 const unsigned long * __restrict p3, in xor_arm64_neon_5() argument 129 uint64_t *dp3 = (uint64_t *)p3; in xor_arm64_neon_5() 197 const unsigned long * __restrict p3) in xor_arm64_eor3_3() argument 201 uint64_t *dp3 = (uint64_t *)p3; in xor_arm64_eor3_3() 232 const unsigned long * __restrict p3, in xor_arm64_eor3_4() argument 237 uint64_t *dp3 = (uint64_t *)p3; in xor_arm64_eor3_4() [all …]
|
/linux-6.3-rc2/arch/powerpc/include/asm/ |
A D | xor_altivec.h | 10 const unsigned long * __restrict p3); 13 const unsigned long * __restrict p3, 17 const unsigned long * __restrict p3,
|
/linux-6.3-rc2/arch/ia64/include/asm/ |
A D | xor.h | 11 const unsigned long * __restrict p3); 14 const unsigned long * __restrict p3, 18 const unsigned long * __restrict p3,
|
/linux-6.3-rc2/crypto/ |
A D | xor.c | 29 unsigned long *p1, *p2, *p3, *p4; in xor_blocks() local 43 p3 = (unsigned long *) srcs[2]; in xor_blocks() 45 active_template->do_4(bytes, dest, p1, p2, p3); in xor_blocks() 50 active_template->do_5(bytes, dest, p1, p2, p3, p4); in xor_blocks()
|
/linux-6.3-rc2/drivers/clk/ |
A D | clk-si5351.c | 38 unsigned long p3; member 140 params->p3 = 1; in si5351_read_parameters() 164 buf[1] = params->p3 & 0xff; in si5351_write_parameters() 426 if (hwdata->params.p3 == 0) in si5351_pll_recalc_rate() 431 rate += 512 * hwdata->params.p3; in si5351_pll_recalc_rate() 480 hwdata->params.p3 = c; in si5351_pll_round_rate() 624 rate *= 128 * hwdata->params.p3; in si5351_msynth_recalc_rate() 627 m += 512 * hwdata->params.p3; in si5351_msynth_recalc_rate() 732 hwdata->params.p3 = 1; in si5351_msynth_round_rate() 736 hwdata->params.p3 = 0; in si5351_msynth_round_rate() [all …]
|
/linux-6.3-rc2/drivers/misc/cxl/ |
A D | hcalls.c | 206 u64 p1, u64 p2, u64 p3, u64 p4, u64 *out) in cxl_h_control_function() argument 211 CXL_H9_WAIT_UNTIL_DONE(rc, retbuf, H_CONTROL_CA_FUNCTION, unit_address, op, p1, p2, p3, p4); in cxl_h_control_function() 213 unit_address, OP_STR_AFU(op), p1, p2, p3, p4, retbuf[0], rc); in cxl_h_control_function() 214 trace_cxl_hcall_control_function(unit_address, OP_STR_AFU(op), p1, p2, p3, p4, retbuf[0], rc); in cxl_h_control_function() 479 u64 p1, u64 p2, u64 p3, u64 p4, u64 *out) in cxl_h_control_facility() argument 484 CXL_H9_WAIT_UNTIL_DONE(rc, retbuf, H_CONTROL_CA_FACILITY, unit_address, op, p1, p2, p3, p4); in cxl_h_control_facility() 486 unit_address, OP_STR_CONTROL_ADAPTER(op), p1, p2, p3, p4, retbuf[0], rc); in cxl_h_control_facility() 487 …trace_cxl_hcall_control_facility(unit_address, OP_STR_CONTROL_ADAPTER(op), p1, p2, p3, p4, retbuf[… in cxl_h_control_facility()
|