| /arch/nios2/lib/ |
| A D | memmove.c | 15 unsigned long dst, src; in memmove() local 27 if (dst & 1) { in memmove() 31 if (dst & 2) { in memmove() 34 dst += 2; in memmove() 40 dst += 4; in memmove() 53 if (dst & 1) { in memmove() 55 dst--; in memmove() 59 if (dst & 2) { in memmove() 61 dst -= 2; in memmove() 67 dst -= 4; in memmove() [all …]
|
| /arch/mips/net/ |
| A D | bpf_jit_comp64.c | 206 emit(ctx, dsllv, dst, dst, src); in emit_alu_r64() 210 emit(ctx, dsrlv, dst, dst, src); in emit_alu_r64() 214 emit(ctx, dsrav, dst, dst, src); in emit_alu_r64() 218 emit(ctx, daddu, dst, dst, src); in emit_alu_r64() 280 emit(ctx, dsbh, dst, dst); in emit_bswap_r64() 281 emit(ctx, dshd, dst, dst); in emit_bswap_r64() 306 emit_sext(ctx, dst, dst); in emit_bswap_r64() 665 emit_sext(ctx, dst, dst); in build_insn() 700 emit_sext(ctx, dst, dst); in build_insn() 704 emit_sext(ctx, dst, dst); in build_insn() [all …]
|
| A D | bpf_jit_comp32.c | 228 emit(ctx, addu, lo(dst), lo(dst), src); in emit_alu_i64() 237 emit(ctx, subu, lo(dst), lo(dst), src); in emit_alu_i64() 240 emit(ctx, addiu, hi(dst), hi(dst), 1); in emit_alu_i64() 244 emit(ctx, or, lo(dst), lo(dst), src); in emit_alu_i64() 250 emit(ctx, and, lo(dst), lo(dst), src); in emit_alu_i64() 256 emit(ctx, xor, lo(dst), lo(dst), src); in emit_alu_i64() 275 emit(ctx, addu, lo(dst), lo(dst), lo(dst)); in emit_alu_r64() 358 emit(ctx, sra, hi(dst), hi(dst), 31); in emit_shift_i64() 453 emit(ctx, subu, hi(dst), hi(dst), lo(dst)); in emit_mul_i64() 565 exclude = BIT(lo(dst)) | BIT(hi(dst)); in emit_divmod_r64() [all …]
|
| A D | bpf_jit_comp.c | 319 emit(ctx, sll, dst, dst, imm); in emit_alu_i() 323 emit(ctx, srl, dst, dst, imm); in emit_alu_i() 327 emit(ctx, sra, dst, dst, imm); in emit_alu_i() 347 emit(ctx, and, dst, dst, src); in emit_alu_r() 351 emit(ctx, or, dst, dst, src); in emit_alu_r() 355 emit(ctx, xor, dst, dst, src); in emit_alu_r() 359 emit(ctx, sllv, dst, dst, src); in emit_alu_r() 363 emit(ctx, srlv, dst, dst, src); in emit_alu_r() 367 emit(ctx, srav, dst, dst, src); in emit_alu_r() 467 emit(ctx, wsbh, dst, dst); in emit_bswap_r() [all …]
|
| /arch/arm64/lib/ |
| A D | memset.S | 42 dst .req x8 label 72 neg tmp2, dst 83 add dst, dst, tmp2 109 add dst, dst, count 120 sub dst, dst, #16/* Pre-bias. */ 130 add dst, dst, #16 169 neg tmp2, dst 191 add dst, dst, #64 194 add dst, dst, tmp2 198 dc zva, dst [all …]
|
| A D | copy_template.S | 31 dst .req x6 label 42 mov dst, dstin 48 cpy1 dst, src, count 69 strb1 tmp1w, dst, #1 73 strh1 tmp1w, dst, #2 77 str1 tmp1w, dst, #4 81 str1 tmp1, dst, #8 122 str1 tmp1, dst, #8 126 str1 tmp1w, dst, #4 130 strh1 tmp1w, dst, #2 [all …]
|
| /arch/x86/crypto/ |
| A D | aria-avx.h | 16 asmlinkage void aria_aesni_avx_encrypt_16way(const void *ctx, u8 *dst, 18 asmlinkage void aria_aesni_avx_decrypt_16way(const void *ctx, u8 *dst, 20 asmlinkage void aria_aesni_avx_ctr_crypt_16way(const void *ctx, u8 *dst, 31 asmlinkage void aria_aesni_avx2_encrypt_32way(const void *ctx, u8 *dst, 33 asmlinkage void aria_aesni_avx2_decrypt_32way(const void *ctx, u8 *dst, 47 void (*aria_encrypt_16way)(const void *ctx, u8 *dst, const u8 *src); 48 void (*aria_decrypt_16way)(const void *ctx, u8 *dst, const u8 *src); 51 void (*aria_encrypt_32way)(const void *ctx, u8 *dst, const u8 *src); 52 void (*aria_decrypt_32way)(const void *ctx, u8 *dst, const u8 *src); 55 void (*aria_encrypt_64way)(const void *ctx, u8 *dst, const u8 *src); [all …]
|
| A D | serpent-sse2.h | 12 asmlinkage void __serpent_enc_blk_4way(const struct serpent_ctx *ctx, u8 *dst, 14 asmlinkage void serpent_dec_blk_4way(const struct serpent_ctx *ctx, u8 *dst, 19 __serpent_enc_blk_4way(ctx, dst, src, false); in serpent_enc_blk_xway() 23 u8 *dst, const u8 *src) in serpent_enc_blk_xway_xor() argument 25 __serpent_enc_blk_4way(ctx, dst, src, true); in serpent_enc_blk_xway_xor() 30 serpent_dec_blk_4way(ctx, dst, src); in serpent_dec_blk_xway() 39 asmlinkage void serpent_dec_blk_8way(const struct serpent_ctx *ctx, u8 *dst, 44 __serpent_enc_blk_8way(ctx, dst, src, false); in serpent_enc_blk_xway() 48 u8 *dst, const u8 *src) in serpent_enc_blk_xway_xor() argument 50 __serpent_enc_blk_8way(ctx, dst, src, true); in serpent_enc_blk_xway_xor() [all …]
|
| A D | camellia.h | 27 asmlinkage void __camellia_enc_blk(const void *ctx, u8 *dst, const u8 *src, 29 asmlinkage void camellia_dec_blk(const void *ctx, u8 *dst, const u8 *src); 32 asmlinkage void __camellia_enc_blk_2way(const void *ctx, u8 *dst, const u8 *src, 34 asmlinkage void camellia_dec_blk_2way(const void *ctx, u8 *dst, const u8 *src); 42 static inline void camellia_enc_blk(const void *ctx, u8 *dst, const u8 *src) in camellia_enc_blk() argument 44 __camellia_enc_blk(ctx, dst, src, false); in camellia_enc_blk() 49 __camellia_enc_blk(ctx, dst, src, true); in camellia_enc_blk_xor() 52 static inline void camellia_enc_blk_2way(const void *ctx, u8 *dst, in camellia_enc_blk_2way() argument 55 __camellia_enc_blk_2way(ctx, dst, src, false); in camellia_enc_blk_2way() 58 static inline void camellia_enc_blk_xor_2way(const void *ctx, u8 *dst, in camellia_enc_blk_xor_2way() argument [all …]
|
| A D | des3_ede_glue.c | 132 u64 *dst = (u64 *)walk->dst.virt.addr; in __cbc_encrypt() local 136 *dst = *src ^ *iv; in __cbc_encrypt() 137 des3_ede_enc_blk(ctx, (u8 *)dst, (u8 *)dst); in __cbc_encrypt() 138 iv = dst; in __cbc_encrypt() 141 dst += 1; in __cbc_encrypt() 173 u64 *dst = (u64 *)walk->dst.virt.addr; in __cbc_decrypt() local 188 dst -= 3 - 1; in __cbc_decrypt() 195 dst[1] ^= ivs[0]; in __cbc_decrypt() 196 dst[2] ^= ivs[1]; in __cbc_decrypt() 204 dst -= 1; in __cbc_decrypt() [all …]
|
| A D | sm4_aesni_avx_glue.c | 49 u8 *dst = walk.dst.virt.addr; in ecb_do_crypt() local 104 u8 *dst = walk.dst.virt.addr; in sm4_cbc_encrypt() local 108 sm4_crypt_block(ctx->rkey_enc, dst, dst); in sm4_cbc_encrypt() 109 iv = dst; in sm4_cbc_encrypt() 111 dst += SM4_BLOCK_SIZE; in sm4_cbc_encrypt() 137 u8 *dst = walk.dst.virt.addr; in sm4_avx_cbc_decrypt() local 143 dst += bsize; in sm4_avx_cbc_decrypt() 166 dst -= SM4_BLOCK_SIZE; in sm4_avx_cbc_decrypt() 202 u8 *dst = walk.dst.virt.addr; in sm4_avx_ctr_crypt() local 208 dst += bsize; in sm4_avx_ctr_crypt() [all …]
|
| A D | glue_helper-asm-avx.S | 18 #define store_8way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument 19 vmovdqu x0, (0*16)(dst); \ 20 vmovdqu x1, (1*16)(dst); \ 21 vmovdqu x2, (2*16)(dst); \ 22 vmovdqu x3, (3*16)(dst); \ 23 vmovdqu x4, (4*16)(dst); \ 24 vmovdqu x5, (5*16)(dst); \ 25 vmovdqu x6, (6*16)(dst); \ 26 vmovdqu x7, (7*16)(dst); 28 #define store_cbc_8way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \ argument [all …]
|
| /arch/xtensa/lib/ |
| A D | umulsidi3.S | 95 rsr dst, ACCLO 100 extui dst, src, 0, 16 102 srli dst, src, 16 109 mov dst, a12 115 mov dst, a14 165 addx2 \dst, \as, \at 168 add \dst, \tmp, \at 174 addx4 \dst, \as, \at 177 add \dst, \tmp, \at 186 add \dst, \tmp, \at [all …]
|
| /arch/parisc/include/asm/ |
| A D | elf.h | 282 dst[40] = pt->iaoq[0]; dst[41] = pt->iaoq[1]; \ 284 dst[44] = pt->sar; dst[45] = pt->iir; \ 285 dst[46] = pt->isr; dst[47] = pt->ior; \ 286 dst[48] = mfctl(22); dst[49] = mfctl(0); \ 287 dst[50] = mfctl(24); dst[51] = mfctl(25); \ 288 dst[52] = mfctl(26); dst[53] = mfctl(27); \ 289 dst[54] = mfctl(28); dst[55] = mfctl(29); \ 290 dst[56] = mfctl(30); dst[57] = mfctl(31); \ 291 dst[58] = mfctl( 8); dst[59] = mfctl( 9); \ 292 dst[60] = mfctl(12); dst[61] = mfctl(13); \ [all …]
|
| /arch/x86/lib/ |
| A D | copy_mc.c | 68 instrument_memcpy_before(dst, src, len); in copy_mc_to_kernel() 69 ret = copy_mc_fragile(dst, src, len); in copy_mc_to_kernel() 70 instrument_memcpy_after(dst, src, len, ret); in copy_mc_to_kernel() 74 instrument_memcpy_before(dst, src, len); in copy_mc_to_kernel() 75 ret = copy_mc_enhanced_fast_string(dst, src, len); in copy_mc_to_kernel() 76 instrument_memcpy_after(dst, src, len, ret); in copy_mc_to_kernel() 79 memcpy(dst, src, len); in copy_mc_to_kernel() 89 instrument_copy_to_user(dst, src, len); in copy_mc_to_user() 91 ret = copy_mc_fragile((__force void *)dst, src, len); in copy_mc_to_user() 97 instrument_copy_to_user(dst, src, len); in copy_mc_to_user() [all …]
|
| /arch/mips/cavium-octeon/ |
| A D | octeon-memcpy.S | 205 ADD dst, dst, 16*NBYTES 254 ADD dst, dst, 8*NBYTES 273 ADD dst, dst, 4*NBYTES 291 ADD dst, dst, NBYTES 301 ADD dst, dst, NBYTES 308 ADD dst, dst, NBYTES 339 ADD dst, dst, 4*NBYTES 353 ADD dst, dst, NBYTES 384 SUB dst, dst, 16*NBYTES 403 ADD dst, dst, 1 [all …]
|
| /arch/s390/include/asm/ |
| A D | pci_io.h | 96 return zpci_store(dst, val, len); in zpci_write_single() 111 *((u8 *) dst) = (u8) data; in zpci_read_single() 114 *((u16 *) dst) = (u16) data; in zpci_read_single() 117 *((u32 *) dst) = (u32) data; in zpci_read_single() 120 *((u64 *) dst) = (u64) data; in zpci_read_single() 132 int offset = dst & ZPCI_BOUNDARY_MASK; in zpci_get_max_io_size() 152 (u64) dst, n, in zpci_memcpy_fromio() 154 rc = zpci_read_single(dst, src, size); in zpci_memcpy_fromio() 158 dst += size; in zpci_memcpy_fromio() 183 dst += size; in zpci_memcpy_toio() [all …]
|
| /arch/mips/include/asm/ |
| A D | ftrace.h | 22 #define safe_load(load, src, dst, error) \ argument 44 #define safe_store(store, src, dst, error) \ argument 61 : [tmp_dst] "r" (dst), [tmp_src] "r" (src)\ 66 #define safe_load_code(dst, src, error) \ argument 67 safe_load(STR(lw), src, dst, error) 68 #define safe_store_code(src, dst, error) \ argument 69 safe_store(STR(sw), src, dst, error) 71 #define safe_load_stack(dst, src, error) \ argument 72 safe_load(STR(PTR_L), src, dst, error) 74 #define safe_store_stack(src, dst, error) \ argument [all …]
|
| /arch/sh/kernel/ |
| A D | traps_32.c | 53 dst[1] = 0xff; in sign_extend() 54 dst[2] = 0xff; in sign_extend() 55 dst[3] = 0xff; in sign_extend() 58 dst[2] = 0xff; in sign_extend() 59 dst[3] = 0xff; in sign_extend() 63 dst[2] = 0xff; in sign_extend() 64 dst[1] = 0xff; in sign_extend() 65 dst[0] = 0xff; in sign_extend() 68 dst[1] = 0xff; in sign_extend() 230 dst += 2; in handle_unaligned_ins() [all …]
|
| /arch/alpha/lib/ |
| A D | csum_partial_copy.c | 99 *dst = word; in csum_partial_cfu_aligned() 100 dst++; in csum_partial_cfu_aligned() 108 tmp = *dst; in csum_partial_cfu_aligned() 113 *dst = word | tmp; in csum_partial_cfu_aligned() 150 *dst = word; in csum_partial_cfu_dest_aligned() 151 dst++; in csum_partial_cfu_dest_aligned() 161 tmp = *dst; in csum_partial_cfu_dest_aligned() 202 dst++; in csum_partial_cfu_src_aligned() 219 dst++; in csum_partial_cfu_src_aligned() 269 dst++; in csum_partial_cfu_unaligned() [all …]
|
| /arch/s390/lib/ |
| A D | csum-partial.c | 25 memcpy(dst, src, len); in csum_copy() 26 return cksm(dst, len, sum); in csum_copy() 36 fpu_vstm(20, 23, dst); in csum_copy() 37 dst += 64; in csum_copy() 49 fpu_vstm(20, 21, dst); in csum_copy() 50 dst += 32; in csum_copy() 60 fpu_vst(20, dst); in csum_copy() 61 dst += 16; in csum_copy() 70 fpu_vstl(20, len - 1, dst); in csum_copy() 87 __wsum csum_partial_copy_nocheck(const void *src, void *dst, int len) in csum_partial_copy_nocheck() argument [all …]
|
| /arch/loongarch/net/ |
| A D | bpf_jit.c | 554 emit_insn(ctx, addd, dst, dst, src); in build_insn() 600 emit_insn(ctx, muld, dst, dst, t1); in build_insn() 685 emit_insn(ctx, and, dst, dst, src); in build_insn() 696 emit_insn(ctx, and, dst, dst, t1); in build_insn() 704 emit_insn(ctx, or, dst, dst, src); in build_insn() 715 emit_insn(ctx, or, dst, dst, t1); in build_insn() 723 emit_insn(ctx, xor, dst, dst, src); in build_insn() 734 emit_insn(ctx, xor, dst, dst, t1); in build_insn() 820 emit_insn(ctx, revb2h, dst, dst); in build_insn() 825 emit_insn(ctx, revb2w, dst, dst); in build_insn() [all …]
|
| /arch/riscv/lib/ |
| A D | riscv_v_helpers.c | 18 int __asm_vector_usercopy(void *dst, void *src, size_t n); 19 int __asm_vector_usercopy_sum_enabled(void *dst, void *src, size_t n); 20 int fallback_scalar_usercopy(void *dst, void *src, size_t n); 21 int fallback_scalar_usercopy_sum_enabled(void *dst, void *src, size_t n); 22 asmlinkage int enter_vector_usercopy(void *dst, void *src, size_t n, in enter_vector_usercopy() argument 32 remain = enable_sum ? __asm_vector_usercopy(dst, src, n) : in enter_vector_usercopy() 33 __asm_vector_usercopy_sum_enabled(dst, src, n); in enter_vector_usercopy() 38 dst += copied; in enter_vector_usercopy() 47 return enable_sum ? fallback_scalar_usercopy(dst, src, n) : in enter_vector_usercopy() 48 fallback_scalar_usercopy_sum_enabled(dst, src, n); in enter_vector_usercopy()
|
| /arch/mips/lib/ |
| A D | memcpy.S | 40 #define dst a0 macro 337 ADD dst, dst, 8*NBYTES 372 ADD dst, dst, 4*NBYTES 388 ADD dst, dst, NBYTES 406 ADD t1, dst, len # t1 is just past last byte of dst 435 ADD dst, dst, t2 473 ADD dst, dst, 4*NBYTES 490 ADD dst, dst, NBYTES 536 ADD dst, dst, 8 563 ADD dst, dst, 1 [all …]
|
| /arch/microblaze/lib/ |
| A D | memmove.c | 37 char *dst = v_dst; in memmove() local 56 dst += c; in memmove() 67 *--dst = *--src; in memmove() 71 *--dst = *--src; in memmove() 75 *--dst = *--src; in memmove() 79 i_dst = (void *)dst; in memmove() 174 dst = (void *)i_dst; in memmove() 181 *--dst = *--src; in memmove() 184 *--dst = *--src; in memmove() 187 *--dst = *--src; in memmove() [all …]
|