| /crypto/ |
| A D | cbc.c | 22 crypto_xor(iv, src, bsize); in crypto_cbc_encrypt_segment() 41 crypto_lskcipher_encrypt(tfm, src, src, bsize, NULL); in crypto_cbc_encrypt_inplace() 42 iv = src; in crypto_cbc_encrypt_inplace() 44 src += bsize; in crypto_cbc_encrypt_inplace() 61 if (src == dst) in crypto_cbc_encrypt() 82 iv = src; in crypto_cbc_decrypt_segment() 84 src += bsize; in crypto_cbc_decrypt_segment() 108 crypto_lskcipher_decrypt(tfm, src, src, bsize, NULL); in crypto_cbc_decrypt_inplace() 111 crypto_xor(src, src - bsize, bsize); in crypto_cbc_decrypt_inplace() 112 src -= bsize; in crypto_cbc_decrypt_inplace() [all …]
|
| A D | scompress.c | 31 void *src; member 78 scratch->src = NULL; in crypto_scomp_free_scratches() 104 if (scratch->src) in scomp_scratch_workfn() 180 const u8 *src; in scomp_acomp_comp_decomp() local 186 if (!req->src || !slen) in scomp_acomp_comp_decomp() 213 src = req->svirt; in scomp_acomp_comp_decomp() 215 src = NULL; in scomp_acomp_comp_decomp() 238 const u8 *src; in scomp_acomp_comp_decomp() local 241 src = scratch->src; in scomp_acomp_comp_decomp() 242 memcpy_from_sglist(scratch->src, req->src, 0, slen); in scomp_acomp_comp_decomp() [all …]
|
| A D | aegis128-core.c | 127 if (AEGIS_ALIGNED(src)) { in crypto_aegis128_ad() 142 src += AEGIS_BLOCK_SIZE; in crypto_aegis128_ad() 176 src += AEGIS_BLOCK_SIZE; in crypto_aegis128_encrypt_chunk() 192 src += AEGIS_BLOCK_SIZE; in crypto_aegis128_encrypt_chunk() 199 memcpy(msg.bytes, src, size); in crypto_aegis128_encrypt_chunk() 237 src += AEGIS_BLOCK_SIZE; in crypto_aegis128_decrypt_chunk() 253 src += AEGIS_BLOCK_SIZE; in crypto_aegis128_decrypt_chunk() 260 memcpy(msg.bytes, src, size); in crypto_aegis128_decrypt_chunk() 288 const u8 *src = walk.addr; in crypto_aegis128_process_ad() local 298 src += fill; in crypto_aegis128_process_ad() [all …]
|
| A D | tea.c | 62 y = get_unaligned_le32(&src[0]); in tea_encrypt() 63 z = get_unaligned_le32(&src[4]); in tea_encrypt() 88 y = get_unaligned_le32(&src[0]); in tea_decrypt() 89 z = get_unaligned_le32(&src[4]); in tea_decrypt() 130 y = get_unaligned_le32(&src[0]); in xtea_encrypt() 131 z = get_unaligned_le32(&src[4]); in xtea_encrypt() 148 y = get_unaligned_le32(&src[0]); in xtea_decrypt() 149 z = get_unaligned_le32(&src[4]); in xtea_decrypt() 170 y = get_unaligned_le32(&src[0]); in xeta_encrypt() 171 z = get_unaligned_le32(&src[4]); in xeta_encrypt() [all …]
|
| A D | aegis.h | 37 const u8 *src, unsigned int size); 39 const u8 *src, unsigned int size); 47 const union aegis_block *src) in crypto_aegis_block_xor() argument 49 dst->words64[0] ^= src->words64[0]; in crypto_aegis_block_xor() 50 dst->words64[1] ^= src->words64[1]; in crypto_aegis_block_xor() 54 const union aegis_block *src) in crypto_aegis_block_and() argument 56 dst->words64[0] &= src->words64[0]; in crypto_aegis_block_and() 57 dst->words64[1] &= src->words64[1]; in crypto_aegis_block_and() 61 const union aegis_block *src, in crypto_aegis_aesenc() argument 64 const u8 *s = src->bytes; in crypto_aegis_aesenc()
|
| A D | lz4.c | 31 static int __lz4_compress_crypto(const u8 *src, unsigned int slen, in __lz4_compress_crypto() argument 34 int out_len = LZ4_compress_default(src, dst, in __lz4_compress_crypto() 44 static int lz4_scompress(struct crypto_scomp *tfm, const u8 *src, in lz4_scompress() argument 48 return __lz4_compress_crypto(src, slen, dst, dlen, ctx); in lz4_scompress() 51 static int __lz4_decompress_crypto(const u8 *src, unsigned int slen, in __lz4_decompress_crypto() argument 54 int out_len = LZ4_decompress_safe(src, dst, slen, *dlen); in __lz4_decompress_crypto() 63 static int lz4_sdecompress(struct crypto_scomp *tfm, const u8 *src, in lz4_sdecompress() argument 67 return __lz4_decompress_crypto(src, slen, dst, dlen, NULL); in lz4_sdecompress()
|
| A D | lz4hc.c | 29 static int __lz4hc_compress_crypto(const u8 *src, unsigned int slen, in __lz4hc_compress_crypto() argument 32 int out_len = LZ4_compress_HC(src, dst, slen, in __lz4hc_compress_crypto() 42 static int lz4hc_scompress(struct crypto_scomp *tfm, const u8 *src, in lz4hc_scompress() argument 46 return __lz4hc_compress_crypto(src, slen, dst, dlen, ctx); in lz4hc_scompress() 49 static int __lz4hc_decompress_crypto(const u8 *src, unsigned int slen, in __lz4hc_decompress_crypto() argument 52 int out_len = LZ4_decompress_safe(src, dst, slen, *dlen); in __lz4hc_decompress_crypto() 61 static int lz4hc_sdecompress(struct crypto_scomp *tfm, const u8 *src, in lz4hc_sdecompress() argument 65 return __lz4hc_decompress_crypto(src, slen, dst, dlen, NULL); in lz4hc_sdecompress()
|
| A D | lzo-rle.c | 28 static int __lzorle_compress(const u8 *src, unsigned int slen, in __lzorle_compress() argument 34 err = lzorle1x_1_compress_safe(src, slen, dst, &tmp_len, ctx); in __lzorle_compress() 43 static int lzorle_scompress(struct crypto_scomp *tfm, const u8 *src, in lzorle_scompress() argument 47 return __lzorle_compress(src, slen, dst, dlen, ctx); in lzorle_scompress() 50 static int __lzorle_decompress(const u8 *src, unsigned int slen, in __lzorle_decompress() argument 56 err = lzo1x_decompress_safe(src, slen, dst, &tmp_len); in __lzorle_decompress() 65 static int lzorle_sdecompress(struct crypto_scomp *tfm, const u8 *src, in lzorle_sdecompress() argument 69 return __lzorle_decompress(src, slen, dst, dlen); in lzorle_sdecompress()
|
| A D | lzo.c | 28 static int __lzo_compress(const u8 *src, unsigned int slen, in __lzo_compress() argument 34 err = lzo1x_1_compress_safe(src, slen, dst, &tmp_len, ctx); in __lzo_compress() 43 static int lzo_scompress(struct crypto_scomp *tfm, const u8 *src, in lzo_scompress() argument 47 return __lzo_compress(src, slen, dst, dlen, ctx); in lzo_scompress() 50 static int __lzo_decompress(const u8 *src, unsigned int slen, in __lzo_decompress() argument 56 err = lzo1x_decompress_safe(src, slen, dst, &tmp_len); in __lzo_decompress() 65 static int lzo_sdecompress(struct crypto_scomp *tfm, const u8 *src, in lzo_sdecompress() argument 69 return __lzo_decompress(src, slen, dst, dlen); in lzo_sdecompress()
|
| A D | pcbc.c | 25 const u8 *src = walk->src.virt.addr; in crypto_pcbc_encrypt_segment() local 31 crypto_xor(iv, src, bsize); in crypto_pcbc_encrypt_segment() 33 crypto_xor_cpy(iv, dst, src, bsize); in crypto_pcbc_encrypt_segment() 35 src += bsize; in crypto_pcbc_encrypt_segment() 75 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_pcbc_encrypt() 92 const u8 *src = walk->src.virt.addr; in crypto_pcbc_decrypt_segment() local 98 crypto_cipher_decrypt_one(tfm, dst, src); in crypto_pcbc_decrypt_segment() 100 crypto_xor_cpy(iv, dst, src, bsize); in crypto_pcbc_decrypt_segment() 102 src += bsize; in crypto_pcbc_decrypt_segment() 142 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_pcbc_decrypt()
|
| A D | des_generic.c | 37 static void crypto_des_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in crypto_des_encrypt() argument 41 des_encrypt(dctx, dst, src); in crypto_des_encrypt() 44 static void crypto_des_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src) in crypto_des_decrypt() argument 48 des_decrypt(dctx, dst, src); in crypto_des_decrypt() 70 const u8 *src) in crypto_des3_ede_encrypt() argument 74 des3_ede_encrypt(dctx, dst, src); in crypto_des3_ede_encrypt() 78 const u8 *src) in crypto_des3_ede_decrypt() argument 82 des3_ede_decrypt(dctx, dst, src); in crypto_des3_ede_decrypt()
|
| A D | chacha20poly1305.c | 34 struct scatterlist src[1]; member 39 struct scatterlist src[2]; member 104 struct scatterlist *src, *dst; in chacha_decrypt() local 112 src = scatterwalk_ffwd(rctx->src, req->src, req->assoclen); in chacha_decrypt() 113 dst = src; in chacha_decrypt() 114 if (req->src != req->dst) in chacha_decrypt() 226 skcipher_request_set_crypt(&creq->req, creq->src, creq->src, in poly_genkey() 246 struct scatterlist *src, *dst; in chacha_encrypt() local 254 src = scatterwalk_ffwd(rctx->src, req->src, req->assoclen); in chacha_encrypt() 255 dst = src; in chacha_encrypt() [all …]
|
| A D | cipher.c | 59 u8 *dst, const u8 *src, bool enc) in cipher_crypt_one() argument 66 if (unlikely(((unsigned long)dst | (unsigned long)src) & alignmask)) { in cipher_crypt_one() 71 memcpy(tmp, src, bs); in cipher_crypt_one() 75 fn(crypto_cipher_tfm(tfm), dst, src); in cipher_crypt_one() 80 u8 *dst, const u8 *src) in crypto_cipher_encrypt_one() argument 82 cipher_crypt_one(tfm, dst, src, true); in crypto_cipher_encrypt_one() 87 u8 *dst, const u8 *src) in crypto_cipher_decrypt_one() argument 89 cipher_crypt_one(tfm, dst, src, false); in crypto_cipher_decrypt_one()
|
| A D | nhpoly1305.c | 94 const u8 *src, unsigned int srclen, nh_t nh_fn) in nhpoly1305_units() argument 102 nh_fn(key->nh_key, src, bytes, state->nh_hash); in nhpoly1305_units() 112 nh_fn(&key->nh_key[pos / 4], src, bytes, tmp_hash); in nhpoly1305_units() 120 src += bytes; in nhpoly1305_units() 156 const u8 *src, unsigned int srclen, in crypto_nhpoly1305_update_helper() argument 165 memcpy(&state->buffer[state->buflen], src, bytes); in crypto_nhpoly1305_update_helper() 172 src += bytes; in crypto_nhpoly1305_update_helper() 178 nhpoly1305_units(state, key, src, bytes, nh_fn); in crypto_nhpoly1305_update_helper() 179 src += bytes; in crypto_nhpoly1305_update_helper() 184 memcpy(state->buffer, src, srclen); in crypto_nhpoly1305_update_helper() [all …]
|
| A D | polyval-generic.c | 69 const u8 src[POLYVAL_BLOCK_SIZE]) in copy_and_reverse() 71 u64 a = get_unaligned((const u64 *)&src[0]); in copy_and_reverse() 72 u64 b = get_unaligned((const u64 *)&src[8]); in copy_and_reverse() 112 const u8 *src, unsigned int srclen) in polyval_update() argument 119 copy_and_reverse(tmp, src); in polyval_update() 122 src += POLYVAL_BLOCK_SIZE; in polyval_update() 129 static int polyval_finup(struct shash_desc *desc, const u8 *src, in polyval_finup() argument 137 memcpy(tmp, src, len); in polyval_finup()
|
| A D | authenc.c | 33 struct scatterlist src[2]; member 182 struct scatterlist *src, *dst; in crypto_authenc_encrypt() local 185 src = scatterwalk_ffwd(areq_ctx->src, req->src, req->assoclen); in crypto_authenc_encrypt() 186 dst = src; in crypto_authenc_encrypt() 188 if (req->src != req->dst) { in crypto_authenc_encrypt() 218 struct scatterlist *src, *dst; in crypto_authenc_decrypt_tail() local 225 src = scatterwalk_ffwd(areq_ctx->src, req->src, req->assoclen); in crypto_authenc_decrypt_tail() 226 dst = src; in crypto_authenc_decrypt_tail() 228 if (req->src != req->dst) in crypto_authenc_decrypt_tail() 234 skcipher_request_set_crypt(skreq, src, dst, in crypto_authenc_decrypt_tail() [all …]
|
| A D | ccm.c | 36 struct scatterlist src[3]; member 46 struct scatterlist src[3]; member 267 sg = scatterwalk_ffwd(pctx->src + 1, req->src, req->assoclen); in crypto_ccm_init_crypt() 268 if (sg != pctx->src + 1) in crypto_ccm_init_crypt() 302 dst = pctx->src; in crypto_ccm_encrypt() 303 if (req->src != req->dst) in crypto_ccm_encrypt() 331 dst = sg_next(req->src == req->dst ? pctx->src : pctx->dst); in crypto_ccm_decrypt_done() 364 dst = pctx->src; in crypto_ccm_decrypt() 619 sg = scatterwalk_ffwd(rctx->src + 1, req->src, req->assoclen); in crypto_rfc4309_crypt() 620 if (sg != rctx->src + 1) in crypto_rfc4309_crypt() [all …]
|
| A D | ghash-generic.c | 79 const u8 *src, unsigned int srclen) in ghash_update() argument 86 crypto_xor(dst, src, GHASH_BLOCK_SIZE); in ghash_update() 88 src += GHASH_BLOCK_SIZE; in ghash_update() 95 static void ghash_flush(struct shash_desc *desc, const u8 *src, in ghash_flush() argument 103 crypto_xor(dst, src, len); in ghash_flush() 108 static int ghash_finup(struct shash_desc *desc, const u8 *src, in ghash_finup() argument 114 ghash_flush(desc, src, len); in ghash_finup()
|
| A D | scatterwalk.c | 104 void memcpy_sglist(struct scatterlist *dst, struct scatterlist *src, in memcpy_sglist() argument 114 scatterwalk_start(&walk.in, src); in memcpy_sglist() 119 if (walk.src.virt.addr != walk.dst.virt.addr) in memcpy_sglist() 120 memcpy(walk.dst.virt.addr, walk.src.virt.addr, in memcpy_sglist() 128 struct scatterlist *src, in scatterwalk_ffwd() argument 133 return src; in scatterwalk_ffwd() 135 if (src->length > len) in scatterwalk_ffwd() 138 len -= src->length; in scatterwalk_ffwd() 139 src = sg_next(src); in scatterwalk_ffwd() 143 sg_set_page(dst, sg_page(src), src->length - len, src->offset + len); in scatterwalk_ffwd() [all …]
|
| A D | gcm.c | 37 struct scatterlist src[3]; member 57 struct scatterlist *src; member 65 struct scatterlist src[3]; member 160 sg = scatterwalk_ffwd(pctx->src + 1, req->src, req->assoclen); in crypto_gcm_init_common() 161 if (sg != pctx->src + 1) in crypto_gcm_init_common() 182 dst = req->src == req->dst ? pctx->src : pctx->dst; in crypto_gcm_init_crypt() 423 gctx->src = sg_next(req->src == req->dst ? pctx->src : pctx->dst); in gcm_encrypt_continue() 508 gctx->src = sg_next(pctx->src); in crypto_gcm_decrypt() 732 sg = scatterwalk_ffwd(rctx->src + 1, req->src, req->assoclen); in crypto_rfc4106_crypt() 733 if (sg != rctx->src + 1) in crypto_rfc4106_crypt() [all …]
|
| A D | krb5enc.c | 37 struct scatterlist src[2]; member 148 struct scatterlist *src, *dst; in krb5enc_dispatch_encrypt() local 150 src = scatterwalk_ffwd(areq_ctx->src, req->src, req->assoclen); in krb5enc_dispatch_encrypt() 151 if (req->src == req->dst) in krb5enc_dispatch_encrypt() 152 dst = src; in krb5enc_dispatch_encrypt() 159 skcipher_request_set_crypt(skreq, src, dst, req->cryptlen, req->iv); in krb5enc_dispatch_encrypt() 316 struct scatterlist *src, *dst; in krb5enc_dispatch_decrypt() local 318 src = scatterwalk_ffwd(areq_ctx->src, req->src, req->assoclen); in krb5enc_dispatch_decrypt() 319 dst = src; in krb5enc_dispatch_decrypt() 321 if (req->src != req->dst) in krb5enc_dispatch_decrypt() [all …]
|
| A D | xctr.c | 38 const u8 *src = walk->src.virt.addr; in crypto_xctr_crypt_final() local 45 crypto_xor_cpy(dst, keystream, src, nbytes); in crypto_xctr_crypt_final() 54 const u8 *src = walk->src.virt.addr; in crypto_xctr_crypt_segment() local 62 crypto_xor(dst, src, XCTR_BLOCKSIZE); in crypto_xctr_crypt_segment() 67 src += XCTR_BLOCKSIZE; in crypto_xctr_crypt_segment() 112 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_xctr_crypt()
|
| A D | ecdh_helper.c | 15 static inline u8 *ecdh_pack_data(void *dst, const void *src, size_t sz) in ecdh_pack_data() argument 17 memcpy(dst, src, sz); in ecdh_pack_data() 21 static inline const u8 *ecdh_unpack_data(void *dst, const void *src, size_t sz) in ecdh_unpack_data() argument 23 memcpy(dst, src, sz); in ecdh_unpack_data() 24 return src + sz; in ecdh_unpack_data()
|
| A D | aegis128-neon-inner.c | 23 void *memcpy(void *dest, const void *src, size_t n); 222 void crypto_aegis128_encrypt_chunk_neon(void *state, void *dst, const void *src, in crypto_aegis128_encrypt_chunk_neon() argument 234 msg = vld1q_u8(src); in crypto_aegis128_encrypt_chunk_neon() 240 src += AEGIS_BLOCK_SIZE; in crypto_aegis128_encrypt_chunk_neon() 247 const void *in = src; in crypto_aegis128_encrypt_chunk_neon() 252 in = out = memcpy(buf + AEGIS_BLOCK_SIZE - size, src, size); in crypto_aegis128_encrypt_chunk_neon() 271 void crypto_aegis128_decrypt_chunk_neon(void *state, void *dst, const void *src, in crypto_aegis128_decrypt_chunk_neon() argument 281 msg = vld1q_u8(src) ^ st.v[1] ^ (st.v[2] & st.v[3]) ^ st.v[4]; in crypto_aegis128_decrypt_chunk_neon() 286 src += AEGIS_BLOCK_SIZE; in crypto_aegis128_decrypt_chunk_neon() 293 const void *in = src; in crypto_aegis128_decrypt_chunk_neon() [all …]
|
| A D | zstd.c | 84 const void *src, void *dst, unsigned int *dlen) in zstd_compress_one() argument 92 out_len = zstd_compress_cctx(ctx->cctx, dst, req->dlen, src, req->slen, in zstd_compress_one() 143 ret = zstd_compress_one(req, ctx, walk.src.virt.addr, in zstd_compress() 152 inbuf.src = walk.src.virt.addr; in zstd_compress() 196 const void *src, void *dst, unsigned int *dlen) in zstd_decompress_one() argument 204 out_len = zstd_decompress_dctx(ctx->dctx, dst, req->dlen, src, req->slen); in zstd_decompress_one() 243 inbuf.src = walk.src.virt.addr; in zstd_decompress() 251 ret = zstd_decompress_one(req, ctx, walk.src.virt.addr, in zstd_decompress()
|