/linux-6.3-rc2/lib/crypto/ |
A D | chacha20poly1305.c | 76 poly1305_update(&poly1305_state, dst, src_len); in __chacha20poly1305_encrypt() 77 if (src_len & 0xf) in __chacha20poly1305_encrypt() 81 b.lens[1] = cpu_to_le64(src_len); in __chacha20poly1305_encrypt() 84 poly1305_final(&poly1305_state, dst + src_len); in __chacha20poly1305_encrypt() 138 if (unlikely(src_len < POLY1305_DIGEST_SIZE)) in __chacha20poly1305_decrypt() 148 dst_len = src_len - POLY1305_DIGEST_SIZE; in __chacha20poly1305_decrypt() 209 const size_t src_len, in chacha20poly1305_crypt_sg_inplace() argument 236 if (WARN_ON(src_len > INT_MAX)) in chacha20poly1305_crypt_sg_inplace() 297 if (src_len & 0xf) in chacha20poly1305_crypt_sg_inplace() 301 b.lens[1] = cpu_to_le64(src_len); in chacha20poly1305_crypt_sg_inplace() [all …]
|
/linux-6.3-rc2/arch/s390/include/asm/ |
A D | cpacf.h | 255 s.odd = (unsigned long)src_len; in cpacf_km() 266 return src_len - s.odd; in cpacf_km() 287 s.odd = (unsigned long)src_len; in cpacf_kmc() 298 return src_len - s.odd; in cpacf_kmc() 315 s.odd = (unsigned long)src_len; in cpacf_kimd() 340 s.odd = (unsigned long)src_len; in cpacf_klmd() 368 s.odd = (unsigned long)src_len; in cpacf_kmac() 379 return src_len - s.odd; in cpacf_kmac() 401 s.odd = (unsigned long)src_len; in cpacf_kmctr() 414 return src_len - s.odd; in cpacf_kmctr() [all …]
|
/linux-6.3-rc2/lib/ |
A D | decompress_unlzo.c | 105 u32 src_len, dst_len; in unlzo() local 207 src_len = get_unaligned_be32(in_buf); in unlzo() 211 if (src_len <= 0 || src_len > dst_len) { in unlzo() 217 if (fill && in_len < src_len) { in unlzo() 218 skip = fill(in_buf + in_len, src_len - in_len); in unlzo() 222 if (in_len < src_len) { in unlzo() 231 if (unlikely(dst_len == src_len)) in unlzo() 232 memcpy(out_buf, in_buf, src_len); in unlzo() 248 *posp += src_len + 12; in unlzo() 250 in_buf += src_len; in unlzo() [all …]
|
/linux-6.3-rc2/include/crypto/ |
A D | chacha20poly1305.h | 18 void chacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, 24 chacha20poly1305_decrypt(u8 *dst, const u8 *src, const size_t src_len, 28 void xchacha20poly1305_encrypt(u8 *dst, const u8 *src, const size_t src_len, 34 u8 *dst, const u8 *src, const size_t src_len, const u8 *ad, 38 bool chacha20poly1305_encrypt_sg_inplace(struct scatterlist *src, size_t src_len, 43 bool chacha20poly1305_decrypt_sg_inplace(struct scatterlist *src, size_t src_len,
|
A D | akcipher.h | 37 unsigned int src_len; member 252 unsigned int src_len, in akcipher_request_set_crypt() argument 257 req->src_len = src_len; in akcipher_request_set_crypt() 293 unsigned int src_len = req->src_len; in crypto_akcipher_encrypt() local 298 crypto_stats_akcipher_encrypt(src_len, ret, calg); in crypto_akcipher_encrypt() 317 unsigned int src_len = req->src_len; in crypto_akcipher_decrypt() local 322 crypto_stats_akcipher_decrypt(src_len, ret, calg); in crypto_akcipher_decrypt()
|
/linux-6.3-rc2/net/ipv4/ |
A D | fib_rules.c | 38 u8 src_len; member 53 if (r->dst_len || r->src_len || r->dscp) in fib4_rule_matchall() 260 if (frh->src_len) in fib4_rule_configure() 277 rule4->src_len = frh->src_len; in fib4_rule_configure() 278 rule4->srcmask = inet_make_mask(rule4->src_len); in fib4_rule_configure() 317 if (frh->src_len && (rule4->src_len != frh->src_len)) in fib4_rule_compare() 331 if (frh->src_len && (rule4->src != nla_get_in_addr(tb[FRA_SRC]))) in fib4_rule_compare() 346 frh->src_len = rule4->src_len; in fib4_rule_fill() 351 (rule4->src_len && in fib4_rule_fill()
|
/linux-6.3-rc2/net/smc/ |
A D | smc_tx.c | 405 sge[srcchunk].length = src_len; in smcr_tx_rdma_writes() 411 src_off += src_len; in smcr_tx_rdma_writes() 418 src_len = dst_len - src_len; /* remainder */ in smcr_tx_rdma_writes() 419 src_len_sum += src_len; in smcr_tx_rdma_writes() 432 src_len_sum = src_len; in smcr_tx_rdma_writes() 454 dst_off += src_len; in smcd_tx_rdma_writes() 455 src_off += src_len; in smcd_tx_rdma_writes() 462 src_len = dst_len - src_len; /* remainder */ in smcd_tx_rdma_writes() 463 src_len_sum += src_len; in smcd_tx_rdma_writes() 472 src_len_sum = src_len; in smcd_tx_rdma_writes() [all …]
|
/linux-6.3-rc2/include/linux/ |
A D | ccp.h | 183 u64 src_len; /* In bytes */ member 243 u64 src_len; /* In bytes */ member 292 u64 src_len; /* In bytes */ member 354 u64 src_len; /* In bytes */ member 382 u32 src_len; /* In bytes */ member 441 u64 src_len; /* In bytes */ member 470 u64 src_len; /* In bytes */ member
|
A D | lzo.h | 24 int lzo1x_1_compress(const unsigned char *src, size_t src_len, 28 int lzorle1x_1_compress(const unsigned char *src, size_t src_len, 32 int lzo1x_decompress_safe(const unsigned char *src, size_t src_len,
|
A D | crypto.h | 527 void crypto_stats_akcipher_encrypt(unsigned int src_len, int ret, struct crypto_alg *alg); 528 void crypto_stats_akcipher_decrypt(unsigned int src_len, int ret, struct crypto_alg *alg); 553 static inline void crypto_stats_akcipher_encrypt(unsigned int src_len, int ret, struct crypto_alg *… in crypto_stats_akcipher_encrypt() argument 555 static inline void crypto_stats_akcipher_decrypt(unsigned int src_len, int ret, struct crypto_alg *… in crypto_stats_akcipher_decrypt() argument
|
/linux-6.3-rc2/drivers/base/regmap/ |
A D | regcache-lzo.c | 21 size_t src_len; member 47 ret = lzo1x_1_compress(lzo_ctx->src, lzo_ctx->src_len, in regcache_lzo_compress() 61 ret = lzo1x_decompress_safe(lzo_ctx->src, lzo_ctx->src_len, in regcache_lzo_decompress() 181 lzo_blocks[i]->src_len = end - p; in regcache_lzo_init() 183 lzo_blocks[i]->src_len = blksize; in regcache_lzo_init() 189 lzo_blocks[i]->src_len; in regcache_lzo_init() 249 lzo_block->src_len = lzo_block->dst_len; in regcache_lzo_read() 286 lzo_block->src_len = lzo_block->dst_len; in regcache_lzo_write() 303 lzo_block->src_len = lzo_block->dst_len; in regcache_lzo_write()
|
/linux-6.3-rc2/crypto/ |
A D | rsa-pkcs1pad.c | 237 if (req->src_len > ctx->key_size - 11) in pkcs1pad_encrypt() 245 req_ctx->in_buf = kmalloc(ctx->key_size - 1 - req->src_len, in pkcs1pad_encrypt() 250 ps_end = ctx->key_size - req->src_len - 2; in pkcs1pad_encrypt() 257 ctx->key_size - 1 - req->src_len, req->src); in pkcs1pad_encrypt() 348 if (!ctx->key_size || req->src_len != ctx->key_size) in pkcs1pad_decrypt() 364 req_ctx->out_sg, req->src_len, in pkcs1pad_decrypt() 391 if (req->src_len + digest_info_size > ctx->key_size - 11) in pkcs1pad_sign() 399 req_ctx->in_buf = kmalloc(ctx->key_size - 1 - req->src_len, in pkcs1pad_sign() 414 ctx->key_size - 1 - req->src_len, req->src); in pkcs1pad_sign() 439 const unsigned int sig_size = req->src_len; in pkcs1pad_verify_complete() [all …]
|
A D | ecrdsa.c | 99 req->dst_len * 2 != req->src_len || in ecrdsa_verify() 100 WARN_ON(req->src_len > sizeof(sig)) || in ecrdsa_verify() 104 sg_copy_to_buffer(req->src, sg_nents_for_len(req->src, req->src_len), in ecrdsa_verify() 105 sig, req->src_len); in ecrdsa_verify() 108 req->src_len + req->dst_len), in ecrdsa_verify() 109 digest, req->dst_len, req->src_len); in ecrdsa_verify()
|
A D | ecdsa.c | 154 buffer = kmalloc(req->src_len + req->dst_len, GFP_KERNEL); in ecdsa_verify() 159 sg_nents_for_len(req->src, req->src_len + req->dst_len), in ecdsa_verify() 160 buffer, req->src_len + req->dst_len, 0); in ecdsa_verify() 163 buffer, req->src_len); in ecdsa_verify() 172 memcpy(&rawhash[diff], buffer + req->src_len, req->dst_len); in ecdsa_verify() 175 memcpy(&rawhash, buffer + req->src_len, keylen); in ecdsa_verify()
|
A D | sm2.c | 360 buffer = kmalloc(req->src_len + req->dst_len, GFP_KERNEL); in sm2_verify() 365 sg_nents_for_len(req->src, req->src_len + req->dst_len), in sm2_verify() 366 buffer, req->src_len + req->dst_len, 0); in sm2_verify() 371 buffer, req->src_len); in sm2_verify() 376 hash = mpi_read_raw_data(buffer + req->src_len, req->dst_len); in sm2_verify()
|
/linux-6.3-rc2/drivers/staging/rtl8723bs/os_dep/ |
A D | osdep_service.c | 123 void rtw_buf_update(u8 **buf, u32 *buf_len, u8 *src, u32 src_len) in rtw_buf_update() argument 132 if (!src || !src_len) in rtw_buf_update() 136 dup = rtw_malloc(src_len); in rtw_buf_update() 138 dup_len = src_len; in rtw_buf_update()
|
/linux-6.3-rc2/tools/bpf/bpftool/ |
A D | tracelog.c | 42 size_t src_len; in find_tracefs_mnt_single() local 47 src_len = strlen(mntpt); in find_tracefs_mnt_single() 48 if (src_len + 1 >= PATH_MAX) { in find_tracefs_mnt_single()
|
/linux-6.3-rc2/drivers/crypto/virtio/ |
A D | virtio_crypto_akcipher_algs.c | 233 unsigned int src_len = verify ? req->src_len + req->dst_len : req->src_len; in __virtio_crypto_akcipher_do_req() local 240 src_buf = kcalloc_node(src_len, 1, GFP_KERNEL, node); in __virtio_crypto_akcipher_do_req() 246 sg_copy_to_buffer(req->src, sg_nents(req->src), src_buf, src_len); in __virtio_crypto_akcipher_do_req() 247 sg_init_one(&srcdata_sg, src_buf, src_len); in __virtio_crypto_akcipher_do_req() 250 sg_copy_to_buffer(req->src, sg_nents(req->src), src_buf, src_len); in __virtio_crypto_akcipher_do_req() 251 sg_init_one(&srcdata_sg, src_buf, src_len); in __virtio_crypto_akcipher_do_req() 312 akcipher_req->para.src_data_len = cpu_to_le32(req->src_len); in virtio_crypto_rsa_do_req()
|
/linux-6.3-rc2/drivers/crypto/ccp/ |
A D | ccp-dmaengine.c | 366 unsigned int src_offset, src_len; in ccp_create_desc() local 384 src_len = sg_dma_len(src_sg); in ccp_create_desc() 391 if (!src_len) { in ccp_create_desc() 400 src_len = sg_dma_len(src_sg); in ccp_create_desc() 419 len = min(dst_len, src_len); in ccp_create_desc() 435 ccp_pt->src_len = len; in ccp_create_desc() 445 &ccp_pt->dst_dma, ccp_pt->src_len); in ccp_create_desc() 449 src_len -= len; in ccp_create_desc()
|
A D | ccp-ops.c | 487 if (aes->src_len & (AES_BLOCK_SIZE - 1)) in ccp_run_aes_cmac_cmd() 685 ilen = aes->src_len; in ccp_run_aes_gcm_cmd() 689 ilen = aes->src_len - authsize; in ccp_run_aes_gcm_cmd() 930 (aes->src_len & (AES_BLOCK_SIZE - 1))) in ccp_run_aes_cmd() 1314 (des3->src_len & (DES3_EDE_BLOCK_SIZE - 1))) in ccp_run_des3_cmd() 1535 if (!sha->final && (sha->src_len & (block_size - 1))) in ccp_run_sha_cmd() 1541 if (!sha->src_len) { in ccp_run_sha_cmd() 1638 if (sha->src_len && !sha->src) in ccp_run_sha_cmd() 1807 hmac_cmd.u.sha.src_len = block_size + digest_size; in ccp_run_sha_cmd() 2146 op.src.u.dma.length = pt->src_len; in ccp_run_passthru_nomap_cmd() [all …]
|
/linux-6.3-rc2/net/ipv6/ |
A D | fib6_rules.c | 372 if (frh->src_len) in fib6_rule_configure() 378 rule6->src.plen = frh->src_len; in fib6_rule_configure() 406 if (frh->src_len && (rule6->src.plen != frh->src_len)) in fib6_rule_compare() 415 if (frh->src_len && in fib6_rule_compare() 432 frh->src_len = rule6->src.plen; in fib6_rule_fill()
|
/linux-6.3-rc2/arch/x86/xen/ |
A D | setup.c | 674 phys_addr_t dest_off, src_off, dest_len, src_len, len; in xen_phys_memcpy() local 683 src_len = n; in xen_phys_memcpy() 684 if (src_len > (NR_FIX_BTMAPS << PAGE_SHIFT) - src_off) in xen_phys_memcpy() 685 src_len = (NR_FIX_BTMAPS << PAGE_SHIFT) - src_off; in xen_phys_memcpy() 686 len = min(dest_len, src_len); in xen_phys_memcpy() 688 from = early_memremap(src - src_off, src_len + src_off); in xen_phys_memcpy() 691 early_memunmap(from, src_len + src_off); in xen_phys_memcpy()
|
/linux-6.3-rc2/drivers/target/ |
A D | target_core_rd.c | 428 u32 src_len; in rd_execute_rw() local 460 src_len = PAGE_SIZE - rd_offset; in rd_execute_rw() 475 len = min((u32)m.length, src_len); in rd_execute_rw() 495 src_len -= len; in rd_execute_rw() 496 if (src_len) { in rd_execute_rw() 504 src_len = PAGE_SIZE; in rd_execute_rw()
|
/linux-6.3-rc2/drivers/net/wireguard/ |
A D | noise.c | 432 static void mix_hash(u8 hash[NOISE_HASH_LEN], const u8 *src, size_t src_len) in mix_hash() argument 438 blake2s_update(&blake, src, src_len); in mix_hash() 464 size_t src_len, u8 key[NOISE_SYMMETRIC_KEY_LEN], in message_encrypt() argument 467 chacha20poly1305_encrypt(dst_ciphertext, src_plaintext, src_len, hash, in message_encrypt() 470 mix_hash(hash, dst_ciphertext, noise_encrypted_len(src_len)); in message_encrypt() 474 size_t src_len, u8 key[NOISE_SYMMETRIC_KEY_LEN], in message_decrypt() argument 477 if (!chacha20poly1305_decrypt(dst_plaintext, src_ciphertext, src_len, in message_decrypt() 481 mix_hash(hash, src_ciphertext, src_len); in message_decrypt()
|
/linux-6.3-rc2/drivers/crypto/qat/qat_common/ |
A D | qat_asym_algs.c | 240 if (req->src_len > ctx->p_size) in qat_dh_compute_value() 289 if (sg_is_last(req->src) && req->src_len == ctx->p_size) { in qat_dh_compute_value() 293 int shift = ctx->p_size - req->src_len; in qat_dh_compute_value() 300 req->src, 0, req->src_len, 0); in qat_dh_compute_value() 668 if (req->src_len > ctx->key_sz) in qat_rsa_enc() 697 if (sg_is_last(req->src) && req->src_len == ctx->key_sz) { in qat_rsa_enc() 701 int shift = ctx->key_sz - req->src_len; in qat_rsa_enc() 708 0, req->src_len, 0); in qat_rsa_enc() 802 if (req->src_len > ctx->key_sz) in qat_rsa_dec() 845 int shift = ctx->key_sz - req->src_len; in qat_rsa_dec() [all …]
|