| /drivers/crypto/marvell/cesa/ |
| A D | cipher.c | 96 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 98 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 104 sreq->size = len; in mv_cesa_skcipher_std_step() 110 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 112 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 115 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op.desc)); in mv_cesa_skcipher_std_step() 117 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op.desc)); in mv_cesa_skcipher_std_step() 136 sreq->offset); in mv_cesa_skcipher_std_process() 138 sreq->offset += len; in mv_cesa_skcipher_std_process() 184 sreq->size = 0; in mv_cesa_skcipher_std_prepare() [all …]
|
| A D | hash.c | 164 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_step() local 181 if (!sreq->offset) { in mv_cesa_ahash_std_step() 197 len = min_t(size_t, req->nbytes + creq->cache_ptr - sreq->offset, in mv_cesa_ahash_std_step() 206 sreq->offset += mv_cesa_sg_copy_to_sram( in mv_cesa_ahash_std_step() 209 len - creq->cache_ptr, sreq->offset); in mv_cesa_ahash_std_step() 215 if (creq->last_req && sreq->offset == req->nbytes && in mv_cesa_ahash_std_step() 290 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_process() local 292 if (sreq->offset < (req->nbytes - creq->cache_ptr)) in mv_cesa_ahash_std_process() 309 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_prepare() local 311 sreq->offset = 0; in mv_cesa_ahash_std_prepare()
|
| /drivers/crypto/inside-secure/ |
| A D | safexcel_cipher.c | 644 if (sreq->nr_src > 0) in safexcel_handle_req_result() 648 if (sreq->nr_src > 0) in safexcel_handle_req_result() 651 if (sreq->nr_dst > 0) in safexcel_handle_req_result() 737 sreq->nr_src = max(sreq->nr_src, sreq->nr_dst); in safexcel_send_req() 738 sreq->nr_dst = sreq->nr_src; in safexcel_send_req() 902 if (sreq->nr_src > 0) in safexcel_send_req() 906 if (sreq->nr_src > 0) in safexcel_send_req() 909 if (sreq->nr_dst > 0) in safexcel_send_req() 985 if (sreq->needs_inv) { in safexcel_skcipher_handle_result() 1008 if (sreq->needs_inv) { in safexcel_aead_handle_result() [all …]
|
| A D | safexcel_hash.c | 251 if (sreq->nents) { in safexcel_handle_req_result() 255 sreq->nents = 0; in safexcel_handle_req_result() 259 dma_unmap_single(priv->dev, sreq->result_dma, sreq->digest_sz, in safexcel_handle_req_result() 265 dma_unmap_single(priv->dev, sreq->cache_dma, sreq->cache_sz, in safexcel_handle_req_result() 272 if (sreq->hmac && in safexcel_handle_req_result() 275 memcpy(sreq->cache, sreq->state, in safexcel_handle_req_result() 278 memcpy(sreq->state, &ctx->base.opad, sreq->digest_sz); in safexcel_handle_req_result() 280 sreq->len = sreq->block_sz + in safexcel_handle_req_result() 282 sreq->processed = sreq->block_sz; in safexcel_handle_req_result() 283 sreq->hmac = 0; in safexcel_handle_req_result() [all …]
|
| /drivers/net/wireless/mediatek/mt76/ |
| A D | mt76_connac_mcu.c | 1740 if (!sreq->ssids[i].ssid_len) in mt76_connac_mcu_hw_scan() 1754 if (!sreq->n_ssids) in mt76_connac_mcu_hw_scan() 1760 if (sreq->n_channels == 0 || sreq->n_channels > 64) { in mt76_connac_mcu_hw_scan() 1790 if (sreq->ie_len > 0) { in mt76_connac_mcu_hw_scan() 1791 memcpy(req->ies, sreq->ie, sreq->ie_len); in mt76_connac_mcu_hw_scan() 1801 sreq->mac_addr_mask); in mt76_connac_mcu_hw_scan() 1869 sreq->mac_addr_mask); in mt76_connac_mcu_sched_scan_req() 1878 ssid = &sreq->ssids[i]; in mt76_connac_mcu_sched_scan_req() 1885 match = &sreq->match_sets[i]; in mt76_connac_mcu_sched_scan_req() 1915 if (sreq->ie_len > 0) { in mt76_connac_mcu_sched_scan_req() [all …]
|
| A D | mt76_connac_mcu.h | 1985 struct cfg80211_scan_request *sreq); 1992 struct cfg80211_sched_scan_request *sreq);
|
| /drivers/crypto/hisilicon/sec2/ |
| A D | sec_crypto.c | 124 struct skcipher_request *sreq, bool encrypt); 2220 u32 cryptlen = sreq->c_req.sk_req->cryptlen; in sec_skcipher_cryptlen_check() 2263 sreq->c_req.c_len = sk_req->cryptlen; in sec_skcipher_param_check() 2266 sreq->use_pbuf = true; in sec_skcipher_param_check() 2268 sreq->use_pbuf = false; in sec_skcipher_param_check() 2277 return sec_skcipher_cryptlen_check(ctx, sreq); in sec_skcipher_param_check() 2303 skcipher_request_set_crypt(subreq, sreq->src, sreq->dst, in sec_skcipher_soft_crypto() 2304 sreq->cryptlen, sreq->iv); in sec_skcipher_soft_crypto() 2499 if (unlikely(sec_aead_spec_check(ctx, sreq))) { in sec_aead_param_check() 2506 sreq->use_pbuf = true; in sec_aead_param_check() [all …]
|
| /drivers/crypto/marvell/octeontx/ |
| A D | otx_cptvf_algs.c | 148 struct skcipher_request *sreq; in output_iv_copyback() local 154 sreq = container_of(areq, struct skcipher_request, base); in output_iv_copyback() 155 stfm = crypto_skcipher_reqtfm(sreq); in output_iv_copyback() 159 rctx = skcipher_request_ctx_dma(sreq); in output_iv_copyback() 162 start = sreq->cryptlen - ivsize; in output_iv_copyback() 165 scatterwalk_map_and_copy(sreq->iv, sreq->dst, start, in output_iv_copyback() 168 if (sreq->src != sreq->dst) { in output_iv_copyback() 169 scatterwalk_map_and_copy(sreq->iv, sreq->src, in output_iv_copyback() 172 memcpy(sreq->iv, req_info->iv_out, ivsize); in output_iv_copyback()
|
| /drivers/net/wireless/mediatek/mt76/mt7925/ |
| A D | mcu.c | 2911 req->scan_type = sreq->n_ssids ? 1 : 0; in mt7925_mcu_hw_scan() 2916 for (i = 0; i < sreq->n_ssids; i++) { in mt7925_mcu_hw_scan() 2917 if (!sreq->ssids[i].ssid_len) in mt7925_mcu_hw_scan() 2924 sreq->ssids[i].ssid_len); in mt7925_mcu_hw_scan() 2930 if (sreq->n_6ghz_params) { in mt7925_mcu_hw_scan() 2984 sreq->mac_addr_mask); in mt7925_mcu_hw_scan() 3043 req->intervals_num = sreq->n_scan_plans; in mt7925_mcu_sched_scan_req() 3050 ssid->ssids_num = sreq->n_ssids; in mt7925_mcu_sched_scan_req() 3053 cfg_ssid = &sreq->ssids[i]; in mt7925_mcu_sched_scan_req() 3060 match->match_num = sreq->n_match_sets; in mt7925_mcu_sched_scan_req() [all …]
|
| A D | mcu.h | 676 struct cfg80211_sched_scan_request *sreq,
|
| /drivers/crypto/marvell/octeontx2/ |
| A D | otx2_cptvf_algs.c | 132 struct skcipher_request *sreq; in output_iv_copyback() local 137 sreq = container_of(areq, struct skcipher_request, base); in output_iv_copyback() 138 stfm = crypto_skcipher_reqtfm(sreq); in output_iv_copyback() 142 rctx = skcipher_request_ctx_dma(sreq); in output_iv_copyback() 145 start = sreq->cryptlen - ivsize; in output_iv_copyback() 148 scatterwalk_map_and_copy(sreq->iv, sreq->dst, start, in output_iv_copyback() 151 if (sreq->src != sreq->dst) { in output_iv_copyback() 152 scatterwalk_map_and_copy(sreq->iv, sreq->src, in output_iv_copyback() 155 memcpy(sreq->iv, req_info->iv_out, ivsize); in output_iv_copyback()
|
| /drivers/crypto/intel/qat/qat_common/ |
| A D | qat_algs.c | 683 struct skcipher_request *sreq = qat_req->skcipher_req; in qat_alg_update_iv_ctr_mode() local 688 memcpy(qat_req->iv, sreq->iv, AES_BLOCK_SIZE); in qat_alg_update_iv_ctr_mode() 694 iv_lo += DIV_ROUND_UP(sreq->cryptlen, AES_BLOCK_SIZE); in qat_alg_update_iv_ctr_mode() 704 struct skcipher_request *sreq = qat_req->skcipher_req; in qat_alg_update_iv_cbc_mode() local 705 int offset = sreq->cryptlen - AES_BLOCK_SIZE; in qat_alg_update_iv_cbc_mode() 709 sgl = sreq->dst; in qat_alg_update_iv_cbc_mode() 711 sgl = sreq->src; in qat_alg_update_iv_cbc_mode() 741 struct skcipher_request *sreq = qat_req->skcipher_req; in qat_skcipher_alg_callback() local 752 memcpy(sreq->iv, qat_req->iv, AES_BLOCK_SIZE); in qat_skcipher_alg_callback() 754 skcipher_request_complete(sreq, res); in qat_skcipher_alg_callback()
|
| /drivers/crypto/starfive/ |
| A D | jh7110-aes.c | 283 starfive_aes_write_iv(ctx, (void *)cryp->req.sreq->iv); in starfive_aes_hw_init() 333 starfive_aes_get_iv(cryp, (void *)cryp->req.sreq->iv); in starfive_aes_finish_req() 338 crypto_finalize_skcipher_request(cryp->engine, cryp->req.sreq, in starfive_aes_finish_req() 578 cryp->req.sreq = req; in starfive_aes_do_one_req()
|
| A D | jh7110-cryp.h | 206 struct skcipher_request *sreq; member
|