Lines Matching refs:areq
62 static int rk_cipher_fallback(struct skcipher_request *areq) in rk_cipher_fallback() argument
64 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in rk_cipher_fallback()
66 struct rk_cipher_rctx *rctx = skcipher_request_ctx(areq); in rk_cipher_fallback()
74 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in rk_cipher_fallback()
75 areq->base.complete, areq->base.data); in rk_cipher_fallback()
76 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in rk_cipher_fallback()
77 areq->cryptlen, areq->iv); in rk_cipher_fallback()
295 struct skcipher_request *areq = container_of(async_req, struct skcipher_request, base); in rk_cipher_run() local
296 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in rk_cipher_run()
297 struct rk_cipher_rctx *rctx = skcipher_request_ctx(areq); in rk_cipher_run()
304 u8 *ivtouse = areq->iv; in rk_cipher_run()
305 unsigned int len = areq->cryptlen; in rk_cipher_run()
319 if (areq->iv && crypto_skcipher_ivsize(tfm) > 0) { in rk_cipher_run()
321 offset = areq->cryptlen - ivsize; in rk_cipher_run()
322 scatterwalk_map_and_copy(rctx->backup_iv, areq->src, in rk_cipher_run()
327 sgs = areq->src; in rk_cipher_run()
328 sgd = areq->dst; in rk_cipher_run()
360 rk_cipher_hw_init(rkc, areq); in rk_cipher_run()
398 if (areq->iv && ivsize > 0) { in rk_cipher_run()
399 offset = areq->cryptlen - ivsize; in rk_cipher_run()
401 memcpy(areq->iv, rctx->backup_iv, ivsize); in rk_cipher_run()
404 scatterwalk_map_and_copy(areq->iv, areq->dst, offset, in rk_cipher_run()
413 crypto_finalize_skcipher_request(engine, areq, err); in rk_cipher_run()