| /linux/crypto/ |
| A D | rsa-pkcs1pad.c | 229 req->dst_len = ctx->key_size; in pkcs1pad_encrypt_sign_complete() 264 req->dst_len = ctx->key_size; in pkcs1pad_encrypt() 302 unsigned int dst_len; in pkcs1pad_decrypt_complete() local 310 dst_len = req_ctx->child_req.dst_len; in pkcs1pad_decrypt_complete() 320 dst_len--; in pkcs1pad_decrypt_complete() 336 if (req->dst_len < dst_len - pos) in pkcs1pad_decrypt_complete() 338 req->dst_len = dst_len - pos; in pkcs1pad_decrypt_complete() 464 unsigned int dst_len; in pkcs1pad_verify_complete() local 472 dst_len = req_ctx->child_req.dst_len; in pkcs1pad_verify_complete() 482 dst_len--; in pkcs1pad_verify_complete() [all …]
|
| A D | ecrdsa.c | 77 unsigned int ndigits = req->dst_len / sizeof(u64); in ecrdsa_verify() 96 req->dst_len != ctx->digest_len || in ecrdsa_verify() 97 req->dst_len != ctx->curve->g.ndigits * sizeof(u64) || in ecrdsa_verify() 99 req->dst_len * 2 != req->src_len || in ecrdsa_verify() 101 WARN_ON(req->dst_len > sizeof(digest))) in ecrdsa_verify() 108 req->src_len + req->dst_len), in ecrdsa_verify() 109 digest, req->dst_len, req->src_len); in ecrdsa_verify()
|
| A D | ecdsa.c | 144 buffer = kmalloc(req->src_len + req->dst_len, GFP_KERNEL); in ecdsa_verify() 149 sg_nents_for_len(req->src, req->src_len + req->dst_len), in ecdsa_verify() 150 buffer, req->src_len + req->dst_len, 0); in ecdsa_verify() 157 if (bufsize > req->dst_len) in ecdsa_verify() 158 bufsize = req->dst_len; in ecdsa_verify()
|
| A D | jitterentropy-kcapi.c | 172 int jent_read_random_block(void *hash_state, char *dst, unsigned int dst_len) in jent_read_random_block() argument 182 if (!ret && dst_len) in jent_read_random_block() 183 memcpy(dst, jent_block, dst_len); in jent_read_random_block()
|
| /linux/fs/bcachefs/ |
| A D | compress.c | 167 src_len, dst_len, dst_len); in __bio_uncompress() 168 if (ret != dst_len) in __bio_uncompress() 176 .avail_out = dst_len, in __bio_uncompress() 202 dst_data, dst_len, in __bio_uncompress() 207 if (ret != dst_len) in __bio_uncompress() 310 &len, dst_len, in attempt_compress() 319 src_len, dst_len, in attempt_compress() 427 *dst_len = ret; in __bio_compress() 459 pad = round_up(*dst_len, block_bytes(c)) - *dst_len; in __bio_compress() 462 *dst_len += pad; in __bio_compress() [all …]
|
| /linux/lib/ |
| A D | decompress_unlzo.c | 105 u32 src_len, dst_len; in unlzo() local 181 dst_len = get_unaligned_be32(in_buf); in unlzo() 186 if (dst_len == 0) { in unlzo() 192 if (dst_len > LZO_BLOCK_SIZE) { in unlzo() 211 if (src_len <= 0 || src_len > dst_len) { in unlzo() 226 tmp = dst_len; in unlzo() 231 if (unlikely(dst_len == src_len)) in unlzo() 237 if (r != LZO_E_OK || dst_len != tmp) { in unlzo() 243 if (flush && flush(out_buf, dst_len) != dst_len) in unlzo() 246 out_buf += dst_len; in unlzo()
|
| /linux/net/smc/ |
| A D | smc_tx.c | 398 if (src_len_sum == dst_len) in smcr_tx_rdma_writes() 411 dst_len = len - dst_len; /* remainder */ in smcr_tx_rdma_writes() 412 dst_len_sum += dst_len; in smcr_tx_rdma_writes() 442 if (src_len_sum == dst_len) in smcd_tx_rdma_writes() 452 dst_len = len - dst_len; /* remainder */ in smcd_tx_rdma_writes() 453 dst_len_sum += dst_len; in smcd_tx_rdma_writes() 506 dst_len = min_t(size_t, in smc_tx_rdma_writes() 513 dst_len = len; in smc_tx_rdma_writes() 518 src_len = dst_len; in smc_tx_rdma_writes() 526 dst_off, dst_len); in smc_tx_rdma_writes() [all …]
|
| /linux/net/ipv4/ |
| A D | fib_rules.c | 37 u8 dst_len; member 54 if (r->dst_len || r->src_len || r->dscp) in fib4_rule_matchall() 290 if (frh->dst_len) in fib4_rule_configure() 306 rule4->dst_len = frh->dst_len; in fib4_rule_configure() 307 rule4->dstmask = inet_make_mask(rule4->dst_len); in fib4_rule_configure() 347 if (frh->dst_len && (rule4->dst_len != frh->dst_len)) in fib4_rule_compare() 371 if (frh->dst_len && (rule4->dst != nla_get_in_addr(tb[FRA_DST]))) in fib4_rule_compare() 382 frh->dst_len = rule4->dst_len; in fib4_rule_fill() 394 if ((rule4->dst_len && in fib4_rule_fill()
|
| /linux/drivers/block/zram/ |
| A D | backend_lz4.c | 78 req->dst_len, params->level, in lz4_compress() 87 req->dst_len, params->level); in lz4_compress() 91 req->dst_len = ret; in lz4_compress() 103 req->dst_len); in lz4_decompress() 112 req->dst_len); in lz4_decompress()
|
| A D | backend_lz4hc.c | 78 req->dst_len, params->level, in lz4hc_compress() 88 req->src_len, req->dst_len); in lz4hc_compress() 92 req->dst_len = ret; in lz4hc_compress() 104 req->dst_len); in lz4hc_decompress() 113 req->dst_len); in lz4hc_decompress()
|
| A D | backend_842.c | 35 unsigned int dlen = req->dst_len; in compress_842() 41 req->dst_len = dlen; in compress_842() 48 unsigned int dlen = req->dst_len; in decompress_842()
|
| A D | backend_zstd.c | 186 ret = zstd_compress_cctx(zctx->cctx, req->dst, req->dst_len, in zstd_compress() 190 req->dst_len, req->src, in zstd_compress() 195 req->dst_len = ret; in zstd_compress() 207 ret = zstd_decompress_dctx(zctx->dctx, req->dst, req->dst_len, in zstd_decompress() 211 req->dst_len, req->src, in zstd_decompress()
|
| A D | zcomp.c | 122 const void *src, unsigned int *dst_len) in zcomp_compress() argument 128 .dst_len = 2 * PAGE_SIZE, in zcomp_compress() 134 *dst_len = req.dst_len; in zcomp_compress() 145 .dst_len = PAGE_SIZE, in zcomp_decompress()
|
| A D | backend_deflate.c | 102 deflate->avail_out = req->dst_len; in deflate_compress() 108 req->dst_len = deflate->total_out; in deflate_compress() 129 inflate->avail_out = req->dst_len; in deflate_decompress()
|
| A D | backend_lzo.c | 37 &req->dst_len, ctx->context); in lzo_compress() 47 req->dst, &req->dst_len); in lzo_decompress()
|
| A D | backend_lzorle.c | 37 &req->dst_len, ctx->context); in lzorle_compress() 47 req->dst, &req->dst_len); in lzorle_decompress()
|
| A D | zcomp.h | 45 size_t dst_len; member 83 const void *src, unsigned int *dst_len);
|
| /linux/include/linux/ |
| A D | lzo.h | 25 unsigned char *dst, size_t *dst_len, void *wrkmem); 29 unsigned char *dst, size_t *dst_len, void *wrkmem); 33 unsigned char *dst, size_t *dst_len);
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/lag/ |
| A D | mp.c | 103 static void mlx5_lag_fib_set(struct lag_mp *mp, struct fib_info *fi, u32 dst, int dst_len) in mlx5_lag_fib_set() argument 108 mp->fib.dst_len = dst_len; in mlx5_lag_fib_set() 167 (mp->fib.dst != fen_info->dst || mp->fib.dst_len != fen_info->dst_len) && in mlx5_lag_fib_route_event() 176 if (mp->fib.dst == fen_info->dst && mp->fib.dst_len == fen_info->dst_len) in mlx5_lag_fib_route_event() 193 mlx5_lag_fib_set(mp, fi, fen_info->dst, fen_info->dst_len); in mlx5_lag_fib_route_event() 208 mlx5_lag_fib_set(mp, fi, fen_info->dst, fen_info->dst_len); in mlx5_lag_fib_route_event()
|
| /linux/lib/crypto/ |
| A D | chacha20poly1305.c | 130 size_t dst_len; in __chacha20poly1305_decrypt() local 148 dst_len = src_len - POLY1305_DIGEST_SIZE; in __chacha20poly1305_decrypt() 149 poly1305_update(&poly1305_state, src, dst_len); in __chacha20poly1305_decrypt() 150 if (dst_len & 0xf) in __chacha20poly1305_decrypt() 151 poly1305_update(&poly1305_state, pad0, 0x10 - (dst_len & 0xf)); in __chacha20poly1305_decrypt() 154 b.lens[1] = cpu_to_le64(dst_len); in __chacha20poly1305_decrypt() 159 ret = crypto_memneq(b.mac, src + dst_len, POLY1305_DIGEST_SIZE); in __chacha20poly1305_decrypt() 161 chacha20_crypt(chacha_state, dst, src, dst_len); in __chacha20poly1305_decrypt()
|
| /linux/include/crypto/ |
| A D | akcipher.h | 40 unsigned int dst_len; member 255 unsigned int dst_len) in akcipher_request_set_crypt() argument 260 req->dst_len = dst_len; in akcipher_request_set_crypt()
|
| /linux/drivers/crypto/virtio/ |
| A D | virtio_crypto_skcipher_algs.c | 337 u64 dst_len; in __virtio_crypto_skcipher_do_req() local 387 dst_len = virtio_crypto_alg_sg_nents_length(req->dst); in __virtio_crypto_skcipher_do_req() 388 if (unlikely(dst_len > U32_MAX)) { in __virtio_crypto_skcipher_do_req() 394 dst_len = min_t(unsigned int, req->cryptlen, dst_len); in __virtio_crypto_skcipher_do_req() 396 req->cryptlen, dst_len); in __virtio_crypto_skcipher_do_req() 398 if (unlikely(req->cryptlen + dst_len + ivsize + in __virtio_crypto_skcipher_do_req() 406 cpu_to_le32((uint32_t)dst_len); in __virtio_crypto_skcipher_do_req()
|
| /linux/security/keys/ |
| A D | dh.c | 271 if (copy_from_user(outbuf + req->dst_len, kdfcopy->otherinfo, in __keyctl_dh_compute() 278 req->dst_len + kdfcopy->otherinfolen); in __keyctl_dh_compute() 279 } else if (copy_to_user(buffer, outbuf, req->dst_len) == 0) { in __keyctl_dh_compute() 280 ret = req->dst_len; in __keyctl_dh_compute()
|
| /linux/drivers/misc/mei/ |
| A D | vsc-tp.c | 161 int ret, offset = 0, cpy_len, src_len, dst_len = sizeof(struct vsc_tp_packet); in vsc_tp_xfer_helper() local 189 cpy_len = min(src_len, dst_len); in vsc_tp_xfer_helper() 195 dst_len -= cpy_len; in vsc_tp_xfer_helper() 206 dst_len = min(ilen, le16_to_cpu(ack.len)); in vsc_tp_xfer_helper() 209 dst_len = sizeof(recv_crc); in vsc_tp_xfer_helper() 225 dst_len = sizeof(recv_crc); in vsc_tp_xfer_helper()
|
| /linux/net/sched/ |
| A D | act_tunnel_key.c | 88 tunnel_key_copy_geneve_opt(const struct nlattr *nla, void *dst, int dst_len, in tunnel_key_copy_geneve_opt() argument 123 WARN_ON(dst_len < opt_len); in tunnel_key_copy_geneve_opt() 140 tunnel_key_copy_vxlan_opt(const struct nlattr *nla, void *dst, int dst_len, in tunnel_key_copy_vxlan_opt() argument 167 tunnel_key_copy_erspan_opt(const struct nlattr *nla, void *dst, int dst_len, in tunnel_key_copy_erspan_opt() argument 220 int dst_len, struct netlink_ext_ack *extack) in tunnel_key_copy_opts() argument 238 dst_len, extack); in tunnel_key_copy_opts() 247 dst_len -= opt_len; in tunnel_key_copy_opts() 258 dst_len, extack); in tunnel_key_copy_opts() 270 dst_len, extack); in tunnel_key_copy_opts()
|