Home
last modified time | relevance | path

Searched refs:areq (Results 1 – 25 of 71) sorted by relevance

123

/linux-6.3-rc2/drivers/crypto/allwinner/sun4i-ss/
A Dsun4i-ss-cipher.c44 if (!areq->src || !areq->dst) { in sun4i_ss_opti_poll()
65 if (areq->iv) { in sun4i_ss_opti_poll()
80 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_opti_poll()
108 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_opti_poll()
137 scatterwalk_map_and_copy(areq->iv, areq->dst, areq->cryptlen - ivsize, in sun4i_ss_opti_poll()
164 areq->base.complete, areq->base.data); in sun4i_ss_cipher_poll_fallback()
166 areq->cryptlen, areq->iv); in sun4i_ss_cipher_poll_fallback()
212 if (!areq->src || !areq->dst) { in sun4i_ss_cipher_poll()
272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_ss_cipher_poll()
326 sg_miter_start(&mo, areq->dst, sg_nents(areq->dst), in sun4i_ss_cipher_poll()
[all …]
A Dsun4i-ss-hash.c91 sun4i_hash_init(areq); in sun4i_hash_import_md5()
134 sun4i_hash_init(areq); in sun4i_hash_import_sha1()
220 copied = sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in sun4i_hash()
247 if (end > areq->nbytes || areq->nbytes - end > 63) { in sun4i_hash()
249 end, areq->nbytes); in sun4i_hash()
255 if (areq->nbytes < 4) in sun4i_hash()
272 sg_miter_start(&mi, areq->src, sg_nents(areq->src), in sun4i_hash()
513 return sun4i_hash(areq); in sun4i_hash_final()
521 return sun4i_hash(areq); in sun4i_hash_update()
530 return sun4i_hash(areq); in sun4i_hash_finup()
[all …]
A Dsun4i-ss.h201 int sun4i_hash_init(struct ahash_request *areq);
202 int sun4i_hash_update(struct ahash_request *areq);
203 int sun4i_hash_final(struct ahash_request *areq);
204 int sun4i_hash_finup(struct ahash_request *areq);
205 int sun4i_hash_digest(struct ahash_request *areq);
211 int sun4i_ss_cbc_aes_encrypt(struct skcipher_request *areq);
212 int sun4i_ss_cbc_aes_decrypt(struct skcipher_request *areq);
213 int sun4i_ss_ecb_aes_encrypt(struct skcipher_request *areq);
214 int sun4i_ss_ecb_aes_decrypt(struct skcipher_request *areq);
216 int sun4i_ss_cbc_des_encrypt(struct skcipher_request *areq);
[all …]
/linux-6.3-rc2/drivers/crypto/allwinner/sun8i-ce/
A Dsun8i-ce-cipher.c34 if (sg_nents_for_len(areq->src, areq->cryptlen) > MAX_SG || in sun8i_ce_cipher_need_fallback()
35 sg_nents_for_len(areq->dst, areq->cryptlen) > MAX_SG) { in sun8i_ce_cipher_need_fallback()
56 sg = areq->src; in sun8i_ce_cipher_need_fallback()
105 areq->base.complete, areq->base.data); in sun8i_ce_cipher_fallback()
107 areq->cryptlen, areq->iv); in sun8i_ce_cipher_fallback()
133 int ns = sg_nents_for_len(areq->src, areq->cryptlen); in sun8i_ce_cipher_prepare()
134 int nd = sg_nents_for_len(areq->dst, areq->cryptlen); in sun8i_ce_cipher_prepare()
209 if (areq->src == areq->dst) { in sun8i_ce_cipher_prepare()
268 if (areq->src == areq->dst) { in sun8i_ce_cipher_prepare()
285 scatterwalk_map_and_copy(areq->iv, areq->dst, offset, in sun8i_ce_cipher_prepare()
[all …]
A Dsun8i-ce-hash.c145 rctx->fallback_req.src = areq->src; in sun8i_ce_hash_update()
165 rctx->fallback_req.src = areq->src; in sun8i_ce_hash_finup()
209 if (areq->nbytes == 0) { in sun8i_ce_hash_need_fallback()
214 if (sg_nents_for_len(areq->src, areq->nbytes) > MAX_SG - 1) { in sun8i_ce_hash_need_fallback()
218 sg = areq->src; in sun8i_ce_hash_need_fallback()
247 nr_sgs = sg_nents_for_len(areq->src, areq->nbytes); in sun8i_ce_hash_digest()
346 int ns = sg_nents_for_len(areq->src, areq->nbytes); in sun8i_ce_hash_run()
378 dev_dbg(ce->dev, "%s %s len=%d\n", __func__, crypto_tfm_alg_name(areq->base.tfm), areq->nbytes); in sun8i_ce_hash_run()
398 len = areq->nbytes; in sun8i_ce_hash_run()
419 byte_count = areq->nbytes; in sun8i_ce_hash_run()
[all …]
A Dsun8i-ce.h350 int sun8i_ce_enqueue(struct crypto_async_request *areq, u32 type);
358 int sun8i_ce_skdecrypt(struct skcipher_request *areq);
359 int sun8i_ce_skencrypt(struct skcipher_request *areq);
367 int sun8i_ce_hash_init(struct ahash_request *areq);
368 int sun8i_ce_hash_export(struct ahash_request *areq, void *out);
369 int sun8i_ce_hash_import(struct ahash_request *areq, const void *in);
370 int sun8i_ce_hash(struct ahash_request *areq);
371 int sun8i_ce_hash_final(struct ahash_request *areq);
372 int sun8i_ce_hash_update(struct ahash_request *areq);
373 int sun8i_ce_hash_finup(struct ahash_request *areq);
[all …]
/linux-6.3-rc2/drivers/crypto/amlogic/
A Damlogic-gxl-cipher.c73 areq->base.complete, areq->base.data); in meson_cipher_do_fallback()
74 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in meson_cipher_do_fallback()
75 areq->cryptlen, areq->iv); in meson_cipher_do_fallback()
108 areq->cryptlen, in meson_cipher()
177 if (areq->src == areq->dst) { in meson_cipher()
178 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher()
187 nr_sgs = dma_map_sg(mc->dev, areq->src, sg_nents(areq->src), in meson_cipher()
194 nr_sgd = dma_map_sg(mc->dev, areq->dst, sg_nents(areq->dst), in meson_cipher()
203 src_sg = areq->src; in meson_cipher()
238 if (areq->src == areq->dst) { in meson_cipher()
[all …]
/linux-6.3-rc2/drivers/crypto/gemini/
A Dsl3516-ce-cipher.c30 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sl3516_ce_need_fallback()
49 sg = areq->src; in sl3516_ce_need_fallback()
65 sg = areq->dst; in sl3516_ce_need_fallback()
113 areq->base.complete, areq->base.data); in sl3516_ce_cipher_fallback()
115 areq->cryptlen, areq->iv); in sl3516_ce_cipher_fallback()
149 if (areq->src == areq->dst) { in sl3516_ce_cipher()
150 nr_sgs = dma_map_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher()
246 if (areq->src == areq->dst) { in sl3516_ce_cipher()
247 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher()
250 dma_unmap_sg(ce->dev, areq->src, sg_nents(areq->src), in sl3516_ce_cipher()
[all …]
/linux-6.3-rc2/drivers/crypto/allwinner/sun8i-ss/
A Dsun8i-ss-cipher.c33 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sun8i_ss_need_fallback()
38 if (sg_nents_for_len(areq->src, areq->cryptlen) > 8 || in sun8i_ss_need_fallback()
39 sg_nents_for_len(areq->dst, areq->cryptlen) > 8) { in sun8i_ss_need_fallback()
45 sg = areq->src; in sun8i_ss_need_fallback()
60 sg = areq->dst; in sun8i_ss_need_fallback()
105 areq->base.complete, areq->base.data); in sun8i_ss_cipher_fallback()
107 areq->cryptlen, areq->iv); in sun8i_ss_cipher_fallback()
192 int nsgs = sg_nents_for_len(areq->src, areq->cryptlen); in sun8i_ss_cipher()
193 int nsgd = sg_nents_for_len(areq->dst, areq->cryptlen); in sun8i_ss_cipher()
225 if (areq->src == areq->dst) { in sun8i_ss_cipher()
[all …]
A Dsun8i-ss-hash.c235 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_update()
255 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_finup()
280 rctx->fallback_req.src = areq->src; in sun8i_ss_hash_digest_fb()
358 if (areq->nbytes == 0) { in sun8i_ss_hash_need_fallback()
374 sg = areq->src; in sun8i_ss_hash_need_fallback()
529 nr_sgs = dma_map_sg(ss->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in sun8i_ss_hash_run()
544 len = areq->nbytes; in sun8i_ss_hash_run()
545 sg = areq->src; in sun8i_ss_hash_run()
578 byte_count = areq->nbytes; in sun8i_ss_hash_run()
599 byte_count = areq->nbytes + bs; in sun8i_ss_hash_run()
[all …]
A Dsun8i-ss.h296 int sun8i_ss_enqueue(struct crypto_async_request *areq, u32 type);
304 int sun8i_ss_skdecrypt(struct skcipher_request *areq);
305 int sun8i_ss_skencrypt(struct skcipher_request *areq);
318 int sun8i_ss_hash_init(struct ahash_request *areq);
319 int sun8i_ss_hash_export(struct ahash_request *areq, void *out);
320 int sun8i_ss_hash_import(struct ahash_request *areq, const void *in);
321 int sun8i_ss_hash_final(struct ahash_request *areq);
322 int sun8i_ss_hash_update(struct ahash_request *areq);
323 int sun8i_ss_hash_finup(struct ahash_request *areq);
324 int sun8i_ss_hash_digest(struct ahash_request *areq);
/linux-6.3-rc2/drivers/crypto/cavium/nitrox/
A Dnitrox_aead.c229 rctx->srclen = areq->assoclen + areq->cryptlen; in nitrox_aes_gcm_enc()
235 rctx->src = areq->src; in nitrox_aes_gcm_enc()
244 areq); in nitrox_aes_gcm_enc()
263 rctx->srclen = areq->cryptlen + areq->assoclen; in nitrox_aes_gcm_dec()
278 areq); in nitrox_aes_gcm_dec()
400 if (areq->assoclen != 16 && areq->assoclen != 20) in nitrox_rfc4106_set_aead_rctx_sglist()
406 sg = scatterwalk_ffwd(rctx->src + 1, areq->src, areq->assoclen); in nitrox_rfc4106_set_aead_rctx_sglist()
410 if (areq->src != areq->dst) { in nitrox_rfc4106_set_aead_rctx_sglist()
413 sg = scatterwalk_ffwd(rctx->dst + 1, areq->dst, areq->assoclen); in nitrox_rfc4106_set_aead_rctx_sglist()
419 aead_rctx->dst = (areq->src == areq->dst) ? rctx->src : rctx->dst; in nitrox_rfc4106_set_aead_rctx_sglist()
[all …]
/linux-6.3-rc2/drivers/crypto/inside-secure/
A Dsafexcel_hash.c285 areq->nbytes = 0; in safexcel_handle_req_result()
343 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in safexcel_ahash_send_req()
425 areq->nbytes), in safexcel_ahash_send_req()
679 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in safexcel_ahash_cache()
681 areq->nbytes, 0); in safexcel_ahash_cache()
748 if (!areq->nbytes) in safexcel_ahash_update()
799 memcpy(areq->result, in safexcel_ahash_final()
827 aes_encrypt(ctx->aes, areq->result, areq->result); in safexcel_ahash_final()
1131 if (!areq) { in __safexcel_hmac_setkey()
1930 return safexcel_crc32_init(areq) ?: safexcel_ahash_finup(areq); in safexcel_crc32_digest()
[all …]
/linux-6.3-rc2/crypto/
A Dalgif_skcipher.c73 if (IS_ERR(areq)) in _skcipher_recvmsg()
74 return PTR_ERR(areq); in _skcipher_recvmsg()
93 if (!areq->tsgl_entries) in _skcipher_recvmsg()
94 areq->tsgl_entries = 1; in _skcipher_recvmsg()
95 areq->tsgl = sock_kmalloc(sk, array_size(sizeof(*areq->tsgl), in _skcipher_recvmsg()
96 areq->tsgl_entries), in _skcipher_recvmsg()
98 if (!areq->tsgl) { in _skcipher_recvmsg()
102 sg_init_table(areq->tsgl, areq->tsgl_entries); in _skcipher_recvmsg()
107 skcipher_request_set_crypt(&areq->cra_u.skcipher_req, areq->tsgl, in _skcipher_recvmsg()
113 areq->iocb = msg->msg_iocb; in _skcipher_recvmsg()
[all …]
A Dalgif_aead.c155 if (IS_ERR(areq)) in _aead_recvmsg()
156 return PTR_ERR(areq); in _aead_recvmsg()
252 if (!areq->tsgl_entries) in _aead_recvmsg()
253 areq->tsgl_entries = 1; in _aead_recvmsg()
254 areq->tsgl = sock_kmalloc(sk, array_size(sizeof(*areq->tsgl), in _aead_recvmsg()
255 areq->tsgl_entries), in _aead_recvmsg()
257 if (!areq->tsgl) { in _aead_recvmsg()
261 sg_init_table(areq->tsgl, areq->tsgl_entries); in _aead_recvmsg()
273 areq->tsgl); in _aead_recvmsg()
276 rsgl_src = areq->tsgl; in _aead_recvmsg()
[all …]
A Daf_alg.c810 tsgl = areq->tsgl; in af_alg_free_areq_sgls()
1183 sock_kfree_s(sk, areq, areq->areqlen); in af_alg_free_resources()
1255 if (unlikely(!areq)) in af_alg_alloc_areq()
1258 areq->areqlen = areqlen; in af_alg_alloc_areq()
1259 areq->sk = sk; in af_alg_alloc_areq()
1260 areq->last_rsgl = NULL; in af_alg_alloc_areq()
1262 areq->tsgl = NULL; in af_alg_alloc_areq()
1263 areq->tsgl_entries = 0; in af_alg_alloc_areq()
1265 return areq; in af_alg_alloc_areq()
1320 if (areq->last_rsgl) in af_alg_get_rsgl()
[all …]
/linux-6.3-rc2/drivers/crypto/qat/qat_common/
A Dqat_comp_algs.c87 areq->dlen = dlen; in qat_comp_resubmit()
153 scatterwalk_map_and_copy(&qat_produced_adler, areq->dst, areq->dlen, in qat_comp_rfc1950_callback()
202 areq->slen, areq->dlen, consumed, produced, cmp_err, xlt_err); in qat_comp_generic_callback()
204 areq->dlen = 0; in qat_comp_generic_callback()
243 areq->dlen = produced; in qat_comp_generic_callback()
329 if (!areq->src || !slen) in qat_comp_alg_compress_decompress()
332 if (areq->dst && !dlen) in qat_comp_alg_compress_decompress()
342 if (!areq->dst) { in qat_comp_alg_compress_decompress()
347 if (!areq->dst) in qat_comp_alg_compress_decompress()
351 areq->dlen = dlen; in qat_comp_alg_compress_decompress()
[all …]
/linux-6.3-rc2/drivers/crypto/
A Dtalitos.c1004 talitos_sg_unmap(dev, edesc, areq->src, areq->dst, in ipsec_esp_unmap()
1226 (areq->src == areq->dst) ? in ipsec_esp()
1264 if (areq->src != areq->dst) { in ipsec_esp()
1426 return talitos_edesc_alloc(ctx->dev, areq->src, areq->dst, in aead_edesc_alloc()
1538 talitos_sg_unmap(dev, edesc, areq->src, areq->dst, areq->cryptlen, 0); in common_nonsnoop_unmap()
1599 (areq->src == areq->dst) ? in common_nonsnoop()
1610 if (areq->src != areq->dst) { in common_nonsnoop()
1647 areq->iv, 0, areq->cryptlen, 0, ivsize, 0, in skcipher_edesc_alloc()
2099 return ahash_process_req(areq, areq->nbytes); in ahash_update()
2117 return ahash_process_req(areq, areq->nbytes); in ahash_finup()
[all …]
A Datmel-i2c.h163 void (*cbk)(struct atmel_i2c_work_data *work_data, void *areq,
165 void *areq; member
174 void *areq, int status),
175 void *areq);
/linux-6.3-rc2/drivers/crypto/rockchip/
A Drk3288_crypto_skcipher.c75 areq->base.complete, areq->base.data); in rk_cipher_fallback()
76 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in rk_cipher_fallback()
77 areq->cryptlen, areq->iv); in rk_cipher_fallback()
304 u8 *ivtouse = areq->iv; in rk_cipher_run()
305 unsigned int len = areq->cryptlen; in rk_cipher_run()
327 sgs = areq->src; in rk_cipher_run()
328 sgd = areq->dst; in rk_cipher_run()
360 rk_cipher_hw_init(rkc, areq); in rk_cipher_run()
398 if (areq->iv && ivsize > 0) { in rk_cipher_run()
399 offset = areq->cryptlen - ivsize; in rk_cipher_run()
[all …]
A Drk3288_crypto_ahash.c38 static int rk_ahash_digest_fb(struct ahash_request *areq) in rk_ahash_digest_fb() argument
40 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); in rk_ahash_digest_fb()
41 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in rk_ahash_digest_fb()
49 rctx->fallback_req.base.flags = areq->base.flags & in rk_ahash_digest_fb()
52 rctx->fallback_req.nbytes = areq->nbytes; in rk_ahash_digest_fb()
53 rctx->fallback_req.src = areq->src; in rk_ahash_digest_fb()
54 rctx->fallback_req.result = areq->result; in rk_ahash_digest_fb()
234 ret = dma_map_sg(rkc->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in rk_hash_prepare()
260 struct scatterlist *sg = areq->src; in rk_hash_run()
290 rk_ahash_reg_init(areq, rkc); in rk_hash_run()
[all …]
/linux-6.3-rc2/drivers/net/ethernet/fungible/funcore/
A Dfun_dev.c226 const struct fun_dev_params *areq) in fun_enable_admin_queue() argument
229 .cqe_size_log2 = areq->cqe_size_log2, in fun_enable_admin_queue()
230 .sqe_size_log2 = areq->sqe_size_log2, in fun_enable_admin_queue()
231 .cq_depth = areq->cq_depth, in fun_enable_admin_queue()
232 .sq_depth = areq->sq_depth, in fun_enable_admin_queue()
233 .rq_depth = areq->rq_depth, in fun_enable_admin_queue()
235 unsigned int ntags = areq->sq_depth - 1; in fun_enable_admin_queue()
268 fdev->adminq_cb = areq->event_cb; in fun_enable_admin_queue()
277 rc = fun_enable_ctrl(fdev, areq->cqe_size_log2, areq->sqe_size_log2); in fun_enable_admin_queue()
281 if (areq->rq_depth) { in fun_enable_admin_queue()
[all …]
/linux-6.3-rc2/drivers/crypto/stm32/
A Dstm32-cryp.c183 struct aead_request *areq; member
537 memcpy(iv, cryp->areq->iv, 12); in stm32_cryp_gcm_init()
552 if (cryp->areq->assoclen) { in stm32_cryp_gcm_init()
653 if (cryp->areq->assoclen) in stm32_cryp_ccm_init()
684 if (cryp->areq->assoclen) { in stm32_cryp_ccm_init()
829 void *areq);
845 void *areq);
1175 if (!req && !areq) in stm32_cryp_prepare_req()
1197 cryp->areq = NULL; in stm32_cryp_prepare_req()
1219 cryp->areq = areq; in stm32_cryp_prepare_req()
[all …]
/linux-6.3-rc2/drivers/crypto/xilinx/
A Dzynqmp-aes-gcm.c207 struct aead_request *areq = in zynqmp_handle_aes_req() local
211 struct zynqmp_aead_req_ctx *rq_ctx = aead_request_ctx(areq); in zynqmp_handle_aes_req()
216 need_fallback = zynqmp_fallback_check(tfm_ctx, areq); in zynqmp_handle_aes_req()
221 aead_request_set_callback(subreq, areq->base.flags, in zynqmp_handle_aes_req()
223 aead_request_set_crypt(subreq, areq->src, areq->dst, in zynqmp_handle_aes_req()
224 areq->cryptlen, areq->iv); in zynqmp_handle_aes_req()
225 aead_request_set_ad(subreq, areq->assoclen); in zynqmp_handle_aes_req()
231 err = zynqmp_aes_aead_cipher(areq); in zynqmp_handle_aes_req()
234 crypto_finalize_aead_request(engine, areq, err); in zynqmp_handle_aes_req()
/linux-6.3-rc2/drivers/crypto/axis/
A Dartpec6_crypto.c1439 sg_pcopy_to_buffer(areq->src, sg_nents(areq->src), in artpec6_crypto_prepare_hash()
1676 req_ctx = skcipher_request_ctx(areq); in artpec6_crypto_prepare_crypto()
1799 areq->cryptlen; in artpec6_crypto_prepare_crypto()
1904 input_length = areq->cryptlen; in artpec6_crypto_prepare_aead()
1910 __cpu_to_be64(8*areq->assoclen); in artpec6_crypto_prepare_aead()
1930 count = areq->assoclen; in artpec6_crypto_prepare_aead()
1968 size_t output_len = areq->cryptlen; in artpec6_crypto_prepare_aead()
2193 sg_pcopy_to_buffer(areq->src, in artpec6_crypto_complete_aead()
2194 sg_nents(areq->src), in artpec6_crypto_complete_aead()
2197 areq->assoclen + areq->cryptlen - in artpec6_crypto_complete_aead()
[all …]

Completed in 89 milliseconds

123