| /arch/x86/include/asm/ |
| A D | xor_avx.h | 30 const unsigned long * __restrict p1) in xor_avx_2() argument 40 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p1[i / sizeof(*p1)])); \ in xor_avx_2() 50 p1 = (unsigned long *)((uintptr_t)p1 + 512); in xor_avx_2() 57 const unsigned long * __restrict p1, in xor_avx_3() argument 70 "m" (p1[i / sizeof(*p1)])); \ in xor_avx_3() 80 p1 = (unsigned long *)((uintptr_t)p1 + 512); in xor_avx_3() 104 "m" (p1[i / sizeof(*p1)])); \ in xor_avx_4() 114 p1 = (unsigned long *)((uintptr_t)p1 + 512); in xor_avx_4() 123 const unsigned long * __restrict p1, in xor_avx_5() argument 142 "m" (p1[i / sizeof(*p1)])); \ in xor_avx_5() [all …]
|
| A D | kvm_para.h | 48 static inline long kvm_hypercall1(unsigned int nr, unsigned long p1) in kvm_hypercall1() argument 53 return tdx_kvm_hypercall(nr, p1, 0, 0, 0); in kvm_hypercall1() 57 : "a"(nr), "b"(p1) in kvm_hypercall1() 62 static inline long kvm_hypercall2(unsigned int nr, unsigned long p1, in kvm_hypercall2() argument 68 return tdx_kvm_hypercall(nr, p1, p2, 0, 0); in kvm_hypercall2() 72 : "a"(nr), "b"(p1), "c"(p2) in kvm_hypercall2() 83 return tdx_kvm_hypercall(nr, p1, p2, p3, 0); in kvm_hypercall3() 87 : "a"(nr), "b"(p1), "c"(p2), "d"(p3) in kvm_hypercall3() 99 return tdx_kvm_hypercall(nr, p1, p2, p3, p4); in kvm_hypercall4() 103 : "a"(nr), "b"(p1), "c"(p2), "d"(p3), "S"(p4) in kvm_hypercall4() [all …]
|
| A D | xor.h | 60 xor_sse_2(unsigned long bytes, unsigned long * __restrict p1, in xor_sse_2() argument 104 [p1] "+r" (p1), [p2] "+r" (p2) in xor_sse_2() 139 [p1] "+r" (p1), [p2] "+r" (p2) in xor_sse_2_pf64() 147 xor_sse_3(unsigned long bytes, unsigned long * __restrict p1, in xor_sse_3() argument 199 [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3) in xor_sse_3() 237 [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3) in xor_sse_3_pf64() 245 xor_sse_4(unsigned long bytes, unsigned long * __restrict p1, in xor_sse_4() argument 304 : [cnt] "+r" (lines), [p1] "+r" (p1), in xor_sse_4() 345 : [cnt] "+r" (lines), [p1] "+r" (p1), in xor_sse_4_pf64() 421 : [cnt] "+r" (lines), [p1] "+r" (p1), [p2] "+r" (p2), in xor_sse_5() [all …]
|
| A D | xor_32.h | 24 xor_pII_mmx_2(unsigned long bytes, unsigned long * __restrict p1, in xor_pII_mmx_2() argument 60 "+r" (p1), "+r" (p2) in xor_pII_mmx_2() 68 xor_pII_mmx_3(unsigned long bytes, unsigned long * __restrict p1, in xor_pII_mmx_3() argument 110 "+r" (p1), "+r" (p2), "+r" (p3) in xor_pII_mmx_3() 118 xor_pII_mmx_4(unsigned long bytes, unsigned long * __restrict p1, in xor_pII_mmx_4() argument 166 "+r" (p1), "+r" (p2), "+r" (p3), "+r" (p4) in xor_pII_mmx_4() 237 "+r" (p1), "+r" (p2), "+r" (p3) in xor_pII_mmx_5() 298 "+r" (p1), "+r" (p2) in xor_p5_mmx_2() 356 "+r" (p1), "+r" (p2), "+r" (p3) in xor_p5_mmx_3() 424 "+r" (p1), "+r" (p2), "+r" (p3), "+r" (p4) in xor_p5_mmx_4() [all …]
|
| /arch/arm/include/asm/ |
| A D | xor.h | 61 GET_BLOCK_4(p1); in xor_arm4regs_2() 63 PUT_BLOCK_4(p1); in xor_arm4regs_2() 83 GET_BLOCK_4(p1); in xor_arm4regs_3() 86 PUT_BLOCK_4(p1); in xor_arm4regs_3() 103 GET_BLOCK_2(p1); in xor_arm4regs_4() 107 PUT_BLOCK_2(p1); in xor_arm4regs_4() 125 GET_BLOCK_2(p1); in xor_arm4regs_5() 130 PUT_BLOCK_2(p1); in xor_arm4regs_5() 160 xor_arm4regs_2(bytes, p1, p2); in xor_neon_2() 174 xor_arm4regs_3(bytes, p1, p2, p3); in xor_neon_3() [all …]
|
| /arch/sparc/include/asm/ |
| A D | xor_32.h | 16 sparc_2(unsigned long bytes, unsigned long * __restrict p1, in sparc_2() argument 44 : "r" (p1), "r" (p2) in sparc_2() 48 p1 += 8; in sparc_2() 54 sparc_3(unsigned long bytes, unsigned long * __restrict p1, in sparc_3() argument 95 : "r" (p1), "r" (p2), "r" (p3) in sparc_3() 99 p1 += 8; in sparc_3() 106 sparc_4(unsigned long bytes, unsigned long * __restrict p1, in sparc_4() argument 160 : "r" (p1), "r" (p2), "r" (p3), "r" (p4) in sparc_4() 164 p1 += 8; in sparc_4() 239 : "r" (p1), "r" (p2), "r" (p3), "r" (p4), "r" (p5) in sparc_5() [all …]
|
| A D | xor_64.h | 15 void xor_vis_2(unsigned long bytes, unsigned long * __restrict p1, 17 void xor_vis_3(unsigned long bytes, unsigned long * __restrict p1, 20 void xor_vis_4(unsigned long bytes, unsigned long * __restrict p1, 24 void xor_vis_5(unsigned long bytes, unsigned long * __restrict p1, 40 void xor_niagara_2(unsigned long bytes, unsigned long * __restrict p1, 42 void xor_niagara_3(unsigned long bytes, unsigned long * __restrict p1, 45 void xor_niagara_4(unsigned long bytes, unsigned long * __restrict p1, 49 void xor_niagara_5(unsigned long bytes, unsigned long * __restrict p1,
|
| /arch/s390/lib/ |
| A D | xor.c | 14 static void xor_xc_2(unsigned long bytes, unsigned long * __restrict p1, in xor_xc_2() argument 31 : : "d" (bytes), "a" (p1), "a" (p2) in xor_xc_2() 35 static void xor_xc_3(unsigned long bytes, unsigned long * __restrict p1, in xor_xc_3() argument 57 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3) in xor_xc_3() 61 static void xor_xc_4(unsigned long bytes, unsigned long * __restrict p1, in xor_xc_4() argument 88 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3), "+a" (p4) in xor_xc_4() 92 static void xor_xc_5(unsigned long bytes, unsigned long * __restrict p1, in xor_xc_5() argument 125 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3), "+a" (p4), in xor_xc_5()
|
| /arch/loongarch/include/asm/ |
| A D | xor_simd.h | 9 void xor_lsx_2(unsigned long bytes, unsigned long * __restrict p1, 11 void xor_lsx_3(unsigned long bytes, unsigned long * __restrict p1, 13 void xor_lsx_4(unsigned long bytes, unsigned long * __restrict p1, 16 void xor_lsx_5(unsigned long bytes, unsigned long * __restrict p1, 22 void xor_lasx_2(unsigned long bytes, unsigned long * __restrict p1, 24 void xor_lasx_3(unsigned long bytes, unsigned long * __restrict p1, 26 void xor_lasx_4(unsigned long bytes, unsigned long * __restrict p1, 29 void xor_lasx_5(unsigned long bytes, unsigned long * __restrict p1,
|
| /arch/loongarch/lib/ |
| A D | xor_simd.h | 13 void __xor_lsx_2(unsigned long bytes, unsigned long * __restrict p1, 15 void __xor_lsx_3(unsigned long bytes, unsigned long * __restrict p1, 17 void __xor_lsx_4(unsigned long bytes, unsigned long * __restrict p1, 20 void __xor_lsx_5(unsigned long bytes, unsigned long * __restrict p1, 26 void __xor_lasx_2(unsigned long bytes, unsigned long * __restrict p1, 28 void __xor_lasx_3(unsigned long bytes, unsigned long * __restrict p1, 30 void __xor_lasx_4(unsigned long bytes, unsigned long * __restrict p1, 33 void __xor_lasx_5(unsigned long bytes, unsigned long * __restrict p1,
|
| A D | xor_simd_glue.c | 15 void xor_##flavor##_2(unsigned long bytes, unsigned long * __restrict p1, \ 19 __xor_##flavor##_2(bytes, p1, p2); \ 25 void xor_##flavor##_3(unsigned long bytes, unsigned long * __restrict p1, \ 30 __xor_##flavor##_3(bytes, p1, p2, p3); \ 36 void xor_##flavor##_4(unsigned long bytes, unsigned long * __restrict p1, \ 42 __xor_##flavor##_4(bytes, p1, p2, p3, p4); \ 48 void xor_##flavor##_5(unsigned long bytes, unsigned long * __restrict p1, \ 55 __xor_##flavor##_5(bytes, p1, p2, p3, p4, p5); \
|
| /arch/powerpc/lib/ |
| A D | xor_vmx_glue.c | 15 void xor_altivec_2(unsigned long bytes, unsigned long * __restrict p1, in xor_altivec_2() argument 20 __xor_altivec_2(bytes, p1, p2); in xor_altivec_2() 26 void xor_altivec_3(unsigned long bytes, unsigned long * __restrict p1, in xor_altivec_3() argument 32 __xor_altivec_3(bytes, p1, p2, p3); in xor_altivec_3() 38 void xor_altivec_4(unsigned long bytes, unsigned long * __restrict p1, in xor_altivec_4() argument 45 __xor_altivec_4(bytes, p1, p2, p3, p4); in xor_altivec_4() 51 void xor_altivec_5(unsigned long bytes, unsigned long * __restrict p1, in xor_altivec_5() argument 59 __xor_altivec_5(bytes, p1, p2, p3, p4, p5); in xor_altivec_5()
|
| A D | xor_vmx.h | 9 void __xor_altivec_2(unsigned long bytes, unsigned long * __restrict p1, 11 void __xor_altivec_3(unsigned long bytes, unsigned long * __restrict p1, 14 void __xor_altivec_4(unsigned long bytes, unsigned long * __restrict p1, 18 void __xor_altivec_5(unsigned long bytes, unsigned long * __restrict p1,
|
| /arch/arm64/include/asm/ |
| A D | xor.h | 19 xor_neon_2(unsigned long bytes, unsigned long * __restrict p1, in xor_neon_2() argument 23 xor_block_inner_neon.do_2(bytes, p1, p2); in xor_neon_2() 28 xor_neon_3(unsigned long bytes, unsigned long * __restrict p1, in xor_neon_3() argument 33 xor_block_inner_neon.do_3(bytes, p1, p2, p3); in xor_neon_3() 38 xor_neon_4(unsigned long bytes, unsigned long * __restrict p1, in xor_neon_4() argument 44 xor_block_inner_neon.do_4(bytes, p1, p2, p3, p4); in xor_neon_4() 49 xor_neon_5(unsigned long bytes, unsigned long * __restrict p1, in xor_neon_5() argument 56 xor_block_inner_neon.do_5(bytes, p1, p2, p3, p4, p5); in xor_neon_5()
|
| /arch/riscv/include/asm/ |
| A D | xor.h | 13 static void xor_vector_2(unsigned long bytes, unsigned long *__restrict p1, in xor_vector_2() argument 17 xor_regs_2_(bytes, p1, p2); in xor_vector_2() 21 static void xor_vector_3(unsigned long bytes, unsigned long *__restrict p1, in xor_vector_3() argument 26 xor_regs_3_(bytes, p1, p2, p3); in xor_vector_3() 30 static void xor_vector_4(unsigned long bytes, unsigned long *__restrict p1, in xor_vector_4() argument 36 xor_regs_4_(bytes, p1, p2, p3, p4); in xor_vector_4() 40 static void xor_vector_5(unsigned long bytes, unsigned long *__restrict p1, in xor_vector_5() argument 47 xor_regs_5_(bytes, p1, p2, p3, p4, p5); in xor_vector_5()
|
| A D | asm-prototypes.h | 18 void xor_regs_2_(unsigned long bytes, unsigned long *__restrict p1, 20 void xor_regs_3_(unsigned long bytes, unsigned long *__restrict p1, 23 void xor_regs_4_(unsigned long bytes, unsigned long *__restrict p1, 27 void xor_regs_5_(unsigned long bytes, unsigned long *__restrict p1,
|
| /arch/alpha/include/asm/ |
| A D | xor.h | 9 xor_alpha_2(unsigned long bytes, unsigned long * __restrict p1, 12 xor_alpha_3(unsigned long bytes, unsigned long * __restrict p1, 16 xor_alpha_4(unsigned long bytes, unsigned long * __restrict p1, 21 xor_alpha_5(unsigned long bytes, unsigned long * __restrict p1, 28 xor_alpha_prefetch_2(unsigned long bytes, unsigned long * __restrict p1, 31 xor_alpha_prefetch_3(unsigned long bytes, unsigned long * __restrict p1, 35 xor_alpha_prefetch_4(unsigned long bytes, unsigned long * __restrict p1, 40 xor_alpha_prefetch_5(unsigned long bytes, unsigned long * __restrict p1,
|
| /arch/hexagon/lib/ |
| A D | memset.S | 30 p1 = cmp.gtu(r2, #7) define 59 p1 = cmp.eq(r2, #1) define 65 if p1 jumpr r31 72 p1 = cmp.eq(r2, #2) define 78 if p1 jumpr r31 85 p1 = cmp.eq(r2, #4) define 92 if p1 jumpr r31 120 if p1 jumpr r31 131 if p1 jumpr r31 141 if p1 jumpr r31 [all …]
|
| A D | divsi3.S | 11 p1 = cmp.gt(r1,#-1) define 15 p3 = xor(p0,p1) 23 p1 = cmp.gtu(r3,r4) define 27 p0 = or(p0,p1)
|
| /arch/hexagon/mm/ |
| A D | copy_user_template.S | 25 p1 = cmp.gtu(bytes,#15) define 33 if (!p1) jump .Lsmall 58 p1 = cmp.gtu(bytes,#7) define 62 if (!p1) jump .Lsmall 84 p1 = cmp.gtu(bytes,#3) define 88 if (!p1) jump .Lsmall 139 if (!p1) jump .Lsmall 151 if (!p1) jump .Lsmall 163 if (!p1) jump .Lsmall
|
| /arch/powerpc/include/asm/ |
| A D | xor_altivec.h | 6 void xor_altivec_2(unsigned long bytes, unsigned long * __restrict p1, 8 void xor_altivec_3(unsigned long bytes, unsigned long * __restrict p1, 11 void xor_altivec_4(unsigned long bytes, unsigned long * __restrict p1, 15 void xor_altivec_5(unsigned long bytes, unsigned long * __restrict p1,
|
| /arch/arm/boot/dts/nuvoton/ |
| A D | nuvoton-npcm750-runbmc-olympus.dts | 472 g1a-p1-0-hog { 478 g1a-p1-1-hog { 484 g1a-p1-2-hog { 490 g1a-p1-3-hog { 496 g1a-p1-4-hog { 502 g1a-p1-5-hog { 508 g1a-p1-6-hog { 514 g1a-p1-7-hog { 569 g1b-p1-0-hog { 575 g1b-p1-1-hog { [all …]
|
| /arch/arm/mm/ |
| A D | fault-armv.c | 223 static int __init check_writebuffer(unsigned long *p1, unsigned long *p2) in check_writebuffer() argument 229 *p1 = one; in check_writebuffer() 233 val = *p1; in check_writebuffer() 249 unsigned long *p1, *p2; in check_writebuffer_bugs() local 253 p1 = vmap(&page, 1, VM_IOREMAP, prot); in check_writebuffer_bugs() 256 if (p1 && p2) { in check_writebuffer_bugs() 257 v = check_writebuffer(p1, p2); in check_writebuffer_bugs() 263 vunmap(p1); in check_writebuffer_bugs()
|
| /arch/arm64/lib/ |
| A D | xor-neon.c | 13 static void xor_arm64_neon_2(unsigned long bytes, unsigned long * __restrict p1, in xor_arm64_neon_2() argument 16 uint64_t *dp1 = (uint64_t *)p1; in xor_arm64_neon_2() 44 uint64_t *dp1 = (uint64_t *)p1; in xor_arm64_neon_3() 81 uint64_t *dp1 = (uint64_t *)p1; in xor_arm64_neon_4() 127 uint64_t *dp1 = (uint64_t *)p1; in xor_arm64_neon_5() 195 unsigned long * __restrict p1, in xor_arm64_eor3_3() argument 199 uint64_t *dp1 = (uint64_t *)p1; in xor_arm64_eor3_3() 230 unsigned long * __restrict p1, in xor_arm64_eor3_4() argument 235 uint64_t *dp1 = (uint64_t *)p1; in xor_arm64_eor3_4() 274 unsigned long * __restrict p1, in xor_arm64_eor3_5() argument [all …]
|
| /arch/arm/boot/dts/qcom/ |
| A D | qcom-apq8084.dtsi | 259 tsens_s0_p1: s0-p1@d1 { 264 tsens_s1_p1: s1-p1@d2 { 269 tsens_s2_p1: s2-p1@d2 { 274 tsens_s3_p1: s3-p1@d3 { 279 tsens_s4_p1: s4-p1@d4 { 284 tsens_s5_p1: s5-p1@d4 { 289 tsens_s6_p1: s6-p1@d5 { 294 tsens_s7_p1: s7-p1@d6 { 299 tsens_s8_p1: s8-p1@d7 { 309 tsens_s9_p1: s9-p1@d8 { [all …]
|