| /arch/powerpc/platforms/pseries/ |
| A D | of_helpers.c | 49 const __be32 *p2; in of_read_drc_info_cell() local 67 p2 = (const __be32 *)p; in of_read_drc_info_cell() 71 p2 = of_prop_next_u32(*prop, p2, &data->drc_name_suffix_start); in of_read_drc_info_cell() 72 if (!p2) in of_read_drc_info_cell() 76 p2 = of_prop_next_u32(*prop, p2, &data->num_sequential_elems); in of_read_drc_info_cell() 77 if (!p2) in of_read_drc_info_cell() 81 p2 = of_prop_next_u32(*prop, p2, &data->sequential_inc); in of_read_drc_info_cell() 82 if (!p2) in of_read_drc_info_cell() 86 p2 = of_prop_next_u32(*prop, p2, &data->drc_power_domain); in of_read_drc_info_cell() 87 if (!p2) in of_read_drc_info_cell() [all …]
|
| /arch/sparc/include/asm/ |
| A D | xor_32.h | 17 const unsigned long * __restrict p2) in sparc_2() argument 44 : "r" (p1), "r" (p2) in sparc_2() 49 p2 += 8; in sparc_2() 55 const unsigned long * __restrict p2, in sparc_3() argument 95 : "r" (p1), "r" (p2), "r" (p3) in sparc_3() 100 p2 += 8; in sparc_3() 107 const unsigned long * __restrict p2, in sparc_4() argument 160 : "r" (p1), "r" (p2), "r" (p3), "r" (p4) in sparc_4() 165 p2 += 8; in sparc_4() 173 const unsigned long * __restrict p2, in sparc_5() argument [all …]
|
| A D | xor_64.h | 16 const unsigned long * __restrict p2); 18 const unsigned long * __restrict p2, 21 const unsigned long * __restrict p2, 25 const unsigned long * __restrict p2, 41 const unsigned long * __restrict p2); 43 const unsigned long * __restrict p2, 46 const unsigned long * __restrict p2, 50 const unsigned long * __restrict p2,
|
| /arch/arm/include/asm/ |
| A D | xor.h | 62 XOR_BLOCK_4(p2); in xor_arm4regs_2() 84 XOR_BLOCK_4(p2); in xor_arm4regs_3() 104 XOR_BLOCK_2(p2); in xor_arm4regs_4() 126 XOR_BLOCK_2(p2); in xor_arm4regs_5() 157 const unsigned long * __restrict p2) in xor_neon_2() argument 160 xor_arm4regs_2(bytes, p1, p2); in xor_neon_2() 170 const unsigned long * __restrict p2, in xor_neon_3() argument 174 xor_arm4regs_3(bytes, p1, p2, p3); in xor_neon_3() 184 const unsigned long * __restrict p2, in xor_neon_4() argument 189 xor_arm4regs_4(bytes, p1, p2, p3, p4); in xor_neon_4() [all …]
|
| /arch/x86/include/asm/ |
| A D | xor_avx.h | 58 const unsigned long * __restrict p2) in xor_avx_3() argument 68 asm volatile("vmovdqa %0, %%ymm" #reg : : "m" (p2[i / sizeof(*p2)])); \ in xor_avx_3() 81 p2 = (unsigned long *)((uintptr_t)p2 + 512); in xor_avx_3() 89 const unsigned long * __restrict p2, in xor_avx_4() argument 102 "m" (p2[i / sizeof(*p2)])); \ in xor_avx_4() 115 p2 = (unsigned long *)((uintptr_t)p2 + 512); in xor_avx_4() 124 const unsigned long * __restrict p2, in xor_avx_5() argument 140 "m" (p2[i / sizeof(*p2)])); \ in xor_avx_5() 153 p2 = (unsigned long *)((uintptr_t)p2 + 512); in xor_avx_5()
|
| A D | kvm_para.h | 63 unsigned long p2) in kvm_hypercall2() argument 68 return tdx_kvm_hypercall(nr, p1, p2, 0, 0); in kvm_hypercall2() 72 : "a"(nr), "b"(p1), "c"(p2) in kvm_hypercall2() 78 unsigned long p2, unsigned long p3) in kvm_hypercall3() argument 83 return tdx_kvm_hypercall(nr, p1, p2, p3, 0); in kvm_hypercall3() 87 : "a"(nr), "b"(p1), "c"(p2), "d"(p3) in kvm_hypercall3() 93 unsigned long p2, unsigned long p3, in kvm_hypercall4() argument 99 return tdx_kvm_hypercall(nr, p1, p2, p3, p4); in kvm_hypercall4() 103 : "a"(nr), "b"(p1), "c"(p2), "d"(p3), "S"(p4) in kvm_hypercall4() 109 unsigned long p2, unsigned long p3) in kvm_sev_hypercall3() argument [all …]
|
| A D | xor.h | 61 const unsigned long * __restrict p2) in xor_sse_2() argument 104 [p1] "+r" (p1), [p2] "+r" (p2) in xor_sse_2() 139 [p1] "+r" (p1), [p2] "+r" (p2) in xor_sse_2_pf64() 148 const unsigned long * __restrict p2, in xor_sse_3() argument 199 [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3) in xor_sse_3() 237 [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3) in xor_sse_3_pf64() 246 const unsigned long * __restrict p2, in xor_sse_4() argument 305 [p2] "+r" (p2), [p3] "+r" (p3), [p4] "+r" (p4) in xor_sse_4() 346 [p2] "+r" (p2), [p3] "+r" (p3), [p4] "+r" (p4) in xor_sse_4_pf64() 421 : [cnt] "+r" (lines), [p1] "+r" (p1), [p2] "+r" (p2), in xor_sse_5() [all …]
|
| A D | xor_32.h | 25 const unsigned long * __restrict p2) in xor_pII_mmx_2() argument 60 "+r" (p1), "+r" (p2) in xor_pII_mmx_2() 69 const unsigned long * __restrict p2, in xor_pII_mmx_3() argument 110 "+r" (p1), "+r" (p2), "+r" (p3) in xor_pII_mmx_3() 119 const unsigned long * __restrict p2, in xor_pII_mmx_4() argument 176 const unsigned long * __restrict p2, in xor_pII_mmx_5() argument 237 "+r" (p1), "+r" (p2), "+r" (p3) in xor_pII_mmx_5() 259 const unsigned long * __restrict p2) in xor_p5_mmx_2() argument 298 "+r" (p1), "+r" (p2) in xor_p5_mmx_2() 356 "+r" (p1), "+r" (p2), "+r" (p3) in xor_p5_mmx_3() [all …]
|
| /arch/s390/lib/ |
| A D | xor.c | 15 const unsigned long * __restrict p2) in xor_xc_2() argument 31 : : "d" (bytes), "a" (p1), "a" (p2) in xor_xc_2() 36 const unsigned long * __restrict p2, in xor_xc_3() argument 57 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3) in xor_xc_3() 62 const unsigned long * __restrict p2, in xor_xc_4() argument 88 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3), "+a" (p4) in xor_xc_4() 93 const unsigned long * __restrict p2, in xor_xc_5() argument 125 : "+d" (bytes), "+a" (p1), "+a" (p2), "+a" (p3), "+a" (p4), in xor_xc_5()
|
| /arch/loongarch/include/asm/ |
| A D | xor_simd.h | 10 const unsigned long * __restrict p2); 12 const unsigned long * __restrict p2, const unsigned long * __restrict p3); 14 const unsigned long * __restrict p2, const unsigned long * __restrict p3, 17 const unsigned long * __restrict p2, const unsigned long * __restrict p3, 23 const unsigned long * __restrict p2); 25 const unsigned long * __restrict p2, const unsigned long * __restrict p3); 27 const unsigned long * __restrict p2, const unsigned long * __restrict p3, 30 const unsigned long * __restrict p2, const unsigned long * __restrict p3,
|
| /arch/loongarch/lib/ |
| A D | xor_simd.h | 14 const unsigned long * __restrict p2); 16 const unsigned long * __restrict p2, const unsigned long * __restrict p3); 18 const unsigned long * __restrict p2, const unsigned long * __restrict p3, 21 const unsigned long * __restrict p2, const unsigned long * __restrict p3, 27 const unsigned long * __restrict p2); 29 const unsigned long * __restrict p2, const unsigned long * __restrict p3); 31 const unsigned long * __restrict p2, const unsigned long * __restrict p3, 34 const unsigned long * __restrict p2, const unsigned long * __restrict p3,
|
| A D | xor_simd_glue.c | 16 const unsigned long * __restrict p2) \ 19 __xor_##flavor##_2(bytes, p1, p2); \ 26 const unsigned long * __restrict p2, \ 30 __xor_##flavor##_3(bytes, p1, p2, p3); \ 37 const unsigned long * __restrict p2, \ 42 __xor_##flavor##_4(bytes, p1, p2, p3, p4); \ 49 const unsigned long * __restrict p2, \ 55 __xor_##flavor##_5(bytes, p1, p2, p3, p4, p5); \
|
| /arch/powerpc/lib/ |
| A D | xor_vmx_glue.c | 16 const unsigned long * __restrict p2) in xor_altivec_2() argument 20 __xor_altivec_2(bytes, p1, p2); in xor_altivec_2() 27 const unsigned long * __restrict p2, in xor_altivec_3() argument 32 __xor_altivec_3(bytes, p1, p2, p3); in xor_altivec_3() 39 const unsigned long * __restrict p2, in xor_altivec_4() argument 45 __xor_altivec_4(bytes, p1, p2, p3, p4); in xor_altivec_4() 52 const unsigned long * __restrict p2, in xor_altivec_5() argument 59 __xor_altivec_5(bytes, p1, p2, p3, p4, p5); in xor_altivec_5()
|
| A D | xor_vmx.h | 10 const unsigned long * __restrict p2); 12 const unsigned long * __restrict p2, 15 const unsigned long * __restrict p2, 19 const unsigned long * __restrict p2,
|
| /arch/arm64/include/asm/ |
| A D | xor.h | 20 const unsigned long * __restrict p2) in xor_neon_2() argument 23 xor_block_inner_neon.do_2(bytes, p1, p2); in xor_neon_2() 29 const unsigned long * __restrict p2, in xor_neon_3() argument 33 xor_block_inner_neon.do_3(bytes, p1, p2, p3); in xor_neon_3() 39 const unsigned long * __restrict p2, in xor_neon_4() argument 44 xor_block_inner_neon.do_4(bytes, p1, p2, p3, p4); in xor_neon_4() 50 const unsigned long * __restrict p2, in xor_neon_5() argument 56 xor_block_inner_neon.do_5(bytes, p1, p2, p3, p4, p5); in xor_neon_5()
|
| /arch/riscv/include/asm/ |
| A D | xor.h | 14 const unsigned long *__restrict p2) in xor_vector_2() argument 17 xor_regs_2_(bytes, p1, p2); in xor_vector_2() 22 const unsigned long *__restrict p2, in xor_vector_3() argument 26 xor_regs_3_(bytes, p1, p2, p3); in xor_vector_3() 31 const unsigned long *__restrict p2, in xor_vector_4() argument 36 xor_regs_4_(bytes, p1, p2, p3, p4); in xor_vector_4() 41 const unsigned long *__restrict p2, in xor_vector_5() argument 47 xor_regs_5_(bytes, p1, p2, p3, p4, p5); in xor_vector_5()
|
| A D | asm-prototypes.h | 19 const unsigned long *__restrict p2); 21 const unsigned long *__restrict p2, 24 const unsigned long *__restrict p2, 28 const unsigned long *__restrict p2,
|
| /arch/alpha/include/asm/ |
| A D | xor.h | 10 const unsigned long * __restrict p2); 13 const unsigned long * __restrict p2, 17 const unsigned long * __restrict p2, 22 const unsigned long * __restrict p2, 29 const unsigned long * __restrict p2); 32 const unsigned long * __restrict p2, 36 const unsigned long * __restrict p2, 41 const unsigned long * __restrict p2,
|
| /arch/powerpc/include/asm/ |
| A D | xor_altivec.h | 7 const unsigned long * __restrict p2); 9 const unsigned long * __restrict p2, 12 const unsigned long * __restrict p2, 16 const unsigned long * __restrict p2,
|
| A D | epapr_hcalls.h | 539 unsigned long p2) in epapr_hypercall2() argument 545 in[1] = p2; in epapr_hypercall2() 550 unsigned long p2, unsigned long p3) in epapr_hypercall3() argument 556 in[1] = p2; in epapr_hypercall3() 562 unsigned long p2, unsigned long p3, in epapr_hypercall4() argument 569 in[1] = p2; in epapr_hypercall4()
|
| /arch/arm64/lib/ |
| A D | xor-neon.c | 14 const unsigned long * __restrict p2) in xor_arm64_neon_2() argument 17 uint64_t *dp2 = (uint64_t *)p2; in xor_arm64_neon_2() 41 const unsigned long * __restrict p2, in xor_arm64_neon_3() argument 45 uint64_t *dp2 = (uint64_t *)p2; in xor_arm64_neon_3() 77 const unsigned long * __restrict p2, in xor_arm64_neon_4() argument 82 uint64_t *dp2 = (uint64_t *)p2; in xor_arm64_neon_4() 122 const unsigned long * __restrict p2, in xor_arm64_neon_5() argument 128 uint64_t *dp2 = (uint64_t *)p2; in xor_arm64_neon_5() 200 uint64_t *dp2 = (uint64_t *)p2; in xor_arm64_eor3_3() 236 uint64_t *dp2 = (uint64_t *)p2; in xor_arm64_eor3_4() [all …]
|
| /arch/arm/mm/ |
| A D | fault-armv.c | 223 static int __init check_writebuffer(unsigned long *p1, unsigned long *p2) in check_writebuffer() argument 231 *p2 = zero; in check_writebuffer() 249 unsigned long *p1, *p2; in check_writebuffer_bugs() local 254 p2 = vmap(&page, 1, VM_IOREMAP, prot); in check_writebuffer_bugs() 256 if (p1 && p2) { in check_writebuffer_bugs() 257 v = check_writebuffer(p1, p2); in check_writebuffer_bugs() 264 vunmap(p2); in check_writebuffer_bugs()
|
| /arch/hexagon/lib/ |
| A D | memcpy.S | 185 p2 = cmp.eq(len, #0); /* =0 */ define 191 p1 = or(p2, p1); 198 p2 = bitsclr(align888, #7); /* %8 < 97 */ define 202 p2 = and(p2,!p3); /* %8 < 97 */ define 203 if (p2.new) len = add(len, #-8); /* %8 < 97 */ 204 if (p2.new) jump:NT .Ldwordaligned; /* %8 < 97 */ 276 p2 = cmp.eq(kernel, #1); /* skip ovr if kernel == 0 */ define 281 if(!p2) ptr_out_p_32 = add(ptr_out_p_32, #32); 310 p2 = cmp.gtu(offset, #7); define 314 if(p2) data70 = dataF8; [all …]
|
| /arch/arm/boot/dts/qcom/ |
| A D | qcom-apq8084.dtsi | 324 tsens_s0_p2: s0-p2@da { 329 tsens_s1_p2: s1-p2@db { 334 tsens_s2_p2: s2-p2@dc { 339 tsens_s3_p2: s3-p2@dc { 344 tsens_s4_p2: s4-p2@dd { 349 tsens_s5_p2: s5-p2@de { 354 tsens_s6_p2: s6-p2@df { 359 tsens_s7_p2: s7-p2@e0 { 364 tsens_s8_p2: s8-p2@e0 { 369 tsens_s9_p2: s9-p2@e1 { [all …]
|
| /arch/powerpc/kvm/ |
| A D | trace_pr.h | 123 unsigned long long p2), 124 TP_ARGS(type, vcpu, p1, p2), 129 __field( unsigned long long, p2 ) 136 __entry->p2 = p2; 141 __entry->count, __entry->type, __entry->p1, __entry->p2)
|