| /linux/fs/nfs/ |
| A D | fscache.c | 287 netfs->sreq = sreq; in nfs_netfs_alloc() 296 struct inode *inode = sreq->rreq->inode; in nfs_netfs_issue_read() 297 struct nfs_open_context *ctx = sreq->rreq->netfs_priv; in nfs_netfs_issue_read() 303 start = (sreq->start + sreq->transferred) >> PAGE_SHIFT; in nfs_netfs_issue_read() 304 last = ((sreq->start + sreq->len - sreq->transferred - 1) >> PAGE_SHIFT); in nfs_netfs_issue_read() 309 netfs = nfs_netfs_alloc(sreq); in nfs_netfs_issue_read() 311 return netfs_read_subreq_terminated(sreq, -ENOMEM, false); in nfs_netfs_issue_read() 354 struct netfs_io_subrequest *sreq; in nfs_netfs_read_completion() local 359 sreq = netfs->sreq; in nfs_netfs_read_completion() 361 sreq->rreq->origin != NETFS_DIO_READ) in nfs_netfs_read_completion() [all …]
|
| A D | fscache.h | 46 struct netfs_io_subrequest *sreq; member 75 netfs->sreq->transferred = min_t(s64, netfs->sreq->len, in nfs_netfs_put() 77 netfs_read_subreq_terminated(netfs->sreq, netfs->error, false); in nfs_netfs_put()
|
| /linux/drivers/crypto/marvell/cesa/ |
| A D | cipher.c | 93 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 95 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 101 sreq->size = len; in mv_cesa_skcipher_std_step() 107 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 109 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op)); in mv_cesa_skcipher_std_step() 112 memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op.desc)); in mv_cesa_skcipher_std_step() 114 memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op.desc)); in mv_cesa_skcipher_std_step() 133 sreq->offset); in mv_cesa_skcipher_std_process() 135 sreq->offset += len; in mv_cesa_skcipher_std_process() 181 sreq->size = 0; in mv_cesa_skcipher_std_prepare() [all …]
|
| A D | hash.c | 161 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_step() local 178 if (!sreq->offset) { in mv_cesa_ahash_std_step() 194 len = min_t(size_t, req->nbytes + creq->cache_ptr - sreq->offset, in mv_cesa_ahash_std_step() 203 sreq->offset += mv_cesa_sg_copy_to_sram( in mv_cesa_ahash_std_step() 206 len - creq->cache_ptr, sreq->offset); in mv_cesa_ahash_std_step() 212 if (creq->last_req && sreq->offset == req->nbytes && in mv_cesa_ahash_std_step() 287 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_process() local 289 if (sreq->offset < (req->nbytes - creq->cache_ptr)) in mv_cesa_ahash_std_process() 306 struct mv_cesa_ahash_std_req *sreq = &creq->req.std; in mv_cesa_ahash_std_prepare() local 308 sreq->offset = 0; in mv_cesa_ahash_std_prepare()
|
| /linux/include/trace/events/ |
| A D | netfs.h | 311 TP_ARGS(sreq, what), 328 __entry->error = sreq->error; 329 __entry->flags = sreq->flags; 332 __entry->len = sreq->len; 334 __entry->start = sreq->start; 367 __entry->index = sreq ? sreq->debug_index : -1; 369 __entry->flags = sreq ? sreq->flags : 0; 370 __entry->source = sreq ? sreq->source : NETFS_INVALID_READ; 372 __entry->len = sreq ? sreq->len : rreq->len; 373 __entry->transferred = sreq ? sreq->transferred : 0; [all …]
|
| /linux/drivers/crypto/inside-secure/ |
| A D | safexcel_cipher.c | 644 if (sreq->nr_src > 0) in safexcel_handle_req_result() 648 if (sreq->nr_src > 0) in safexcel_handle_req_result() 651 if (sreq->nr_dst > 0) in safexcel_handle_req_result() 737 sreq->nr_src = max(sreq->nr_src, sreq->nr_dst); in safexcel_send_req() 738 sreq->nr_dst = sreq->nr_src; in safexcel_send_req() 902 if (sreq->nr_src > 0) in safexcel_send_req() 906 if (sreq->nr_src > 0) in safexcel_send_req() 909 if (sreq->nr_dst > 0) in safexcel_send_req() 985 if (sreq->needs_inv) { in safexcel_skcipher_handle_result() 1008 if (sreq->needs_inv) { in safexcel_aead_handle_result() [all …]
|
| A D | safexcel_hash.c | 251 if (sreq->nents) { in safexcel_handle_req_result() 253 sreq->nents = 0; in safexcel_handle_req_result() 257 dma_unmap_single(priv->dev, sreq->result_dma, sreq->digest_sz, in safexcel_handle_req_result() 263 dma_unmap_single(priv->dev, sreq->cache_dma, sreq->cache_sz, in safexcel_handle_req_result() 270 if (sreq->hmac && in safexcel_handle_req_result() 273 memcpy(sreq->cache, sreq->state, in safexcel_handle_req_result() 276 memcpy(sreq->state, &ctx->base.opad, sreq->digest_sz); in safexcel_handle_req_result() 278 sreq->len = sreq->block_sz + in safexcel_handle_req_result() 280 sreq->processed = sreq->block_sz; in safexcel_handle_req_result() 281 sreq->hmac = 0; in safexcel_handle_req_result() [all …]
|
| /linux/drivers/net/wireless/mediatek/mt76/ |
| A D | mt76_connac_mcu.c | 1696 if (!sreq->ssids[i].ssid_len) in mt76_connac_mcu_hw_scan() 1710 if (!sreq->n_ssids) in mt76_connac_mcu_hw_scan() 1716 if (sreq->n_channels == 0 || sreq->n_channels > 64) { in mt76_connac_mcu_hw_scan() 1746 if (sreq->ie_len > 0) { in mt76_connac_mcu_hw_scan() 1747 memcpy(req->ies, sreq->ie, sreq->ie_len); in mt76_connac_mcu_hw_scan() 1757 sreq->mac_addr_mask); in mt76_connac_mcu_hw_scan() 1825 sreq->mac_addr_mask); in mt76_connac_mcu_sched_scan_req() 1834 ssid = &sreq->ssids[i]; in mt76_connac_mcu_sched_scan_req() 1841 match = &sreq->match_sets[i]; in mt76_connac_mcu_sched_scan_req() 1871 if (sreq->ie_len > 0) { in mt76_connac_mcu_sched_scan_req() [all …]
|
| A D | mt76_connac_mcu.h | 1973 struct cfg80211_sched_scan_request *sreq);
|
| /linux/drivers/crypto/hisilicon/sec2/ |
| A D | sec_crypto.c | 2011 struct sec_req *sreq) in sec_skcipher_cryptlen_check() argument 2013 u32 cryptlen = sreq->c_req.sk_req->cryptlen; in sec_skcipher_cryptlen_check() 2056 sreq->c_req.c_len = sk_req->cryptlen; in sec_skcipher_param_check() 2059 sreq->use_pbuf = true; in sec_skcipher_param_check() 2061 sreq->use_pbuf = false; in sec_skcipher_param_check() 2096 skcipher_request_set_crypt(subreq, sreq->src, sreq->dst, in sec_skcipher_soft_crypto() 2097 sreq->cryptlen, sreq->iv); in sec_skcipher_soft_crypto() 2266 if (sreq->c_req.encrypt) in sec_aead_spec_check() 2267 sreq->c_req.c_len = req->cryptlen; in sec_aead_spec_check() 2312 sreq->use_pbuf = true; in sec_aead_param_check() [all …]
|
| /linux/drivers/net/wireless/mediatek/mt76/mt7925/ |
| A D | mcu.c | 2691 if (!sreq->n_ssids) in mt7925_mcu_hw_scan() 2700 if (!sreq->ssids[i].ssid_len) in mt7925_mcu_hw_scan() 2705 sreq->ssids[i].ssid_len); in mt7925_mcu_hw_scan() 2740 if (sreq->ie_len > 0) { in mt7925_mcu_hw_scan() 2741 memcpy(ie->ies, sreq->ie, sreq->ie_len); in mt7925_mcu_hw_scan() 2751 sreq->mac_addr_mask); in mt7925_mcu_hw_scan() 2813 ssid->ssids_num = sreq->n_ssids; in mt7925_mcu_sched_scan_req() 2816 cfg_ssid = &sreq->ssids[i]; in mt7925_mcu_sched_scan_req() 2825 cfg_match = &sreq->match_sets[i]; in mt7925_mcu_sched_scan_req() 2856 if (sreq->ie_len > 0) { in mt7925_mcu_sched_scan_req() [all …]
|
| A D | mcu.h | 626 struct cfg80211_sched_scan_request *sreq);
|
| /linux/drivers/crypto/marvell/octeontx/ |
| A D | otx_cptvf_algs.c | 148 struct skcipher_request *sreq; in output_iv_copyback() local 154 sreq = container_of(areq, struct skcipher_request, base); in output_iv_copyback() 155 stfm = crypto_skcipher_reqtfm(sreq); in output_iv_copyback() 159 rctx = skcipher_request_ctx_dma(sreq); in output_iv_copyback() 162 start = sreq->cryptlen - ivsize; in output_iv_copyback() 165 scatterwalk_map_and_copy(sreq->iv, sreq->dst, start, in output_iv_copyback() 168 if (sreq->src != sreq->dst) { in output_iv_copyback() 169 scatterwalk_map_and_copy(sreq->iv, sreq->src, in output_iv_copyback() 172 memcpy(sreq->iv, req_info->iv_out, ivsize); in output_iv_copyback()
|
| /linux/drivers/crypto/marvell/octeontx2/ |
| A D | otx2_cptvf_algs.c | 132 struct skcipher_request *sreq; in output_iv_copyback() local 137 sreq = container_of(areq, struct skcipher_request, base); in output_iv_copyback() 138 stfm = crypto_skcipher_reqtfm(sreq); in output_iv_copyback() 142 rctx = skcipher_request_ctx_dma(sreq); in output_iv_copyback() 145 start = sreq->cryptlen - ivsize; in output_iv_copyback() 148 scatterwalk_map_and_copy(sreq->iv, sreq->dst, start, in output_iv_copyback() 151 if (sreq->src != sreq->dst) { in output_iv_copyback() 152 scatterwalk_map_and_copy(sreq->iv, sreq->src, in output_iv_copyback() 155 memcpy(sreq->iv, req_info->iv_out, ivsize); in output_iv_copyback()
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| A D | qat_algs.c | 683 struct skcipher_request *sreq = qat_req->skcipher_req; in qat_alg_update_iv_ctr_mode() local 688 memcpy(qat_req->iv, sreq->iv, AES_BLOCK_SIZE); in qat_alg_update_iv_ctr_mode() 694 iv_lo += DIV_ROUND_UP(sreq->cryptlen, AES_BLOCK_SIZE); in qat_alg_update_iv_ctr_mode() 704 struct skcipher_request *sreq = qat_req->skcipher_req; in qat_alg_update_iv_cbc_mode() local 705 int offset = sreq->cryptlen - AES_BLOCK_SIZE; in qat_alg_update_iv_cbc_mode() 709 sgl = sreq->dst; in qat_alg_update_iv_cbc_mode() 711 sgl = sreq->src; in qat_alg_update_iv_cbc_mode() 741 struct skcipher_request *sreq = qat_req->skcipher_req; in qat_skcipher_alg_callback() local 752 memcpy(sreq->iv, qat_req->iv, AES_BLOCK_SIZE); in qat_skcipher_alg_callback() 754 skcipher_request_complete(sreq, res); in qat_skcipher_alg_callback()
|
| /linux/drivers/crypto/starfive/ |
| A D | jh7110-cryp.h | 206 struct skcipher_request *sreq; member
|
| A D | jh7110-aes.c | 283 starfive_aes_write_iv(ctx, (void *)cryp->req.sreq->iv); in starfive_aes_hw_init() 333 starfive_aes_get_iv(cryp, (void *)cryp->req.sreq->iv); in starfive_aes_finish_req() 338 crypto_finalize_skcipher_request(cryp->engine, cryp->req.sreq, in starfive_aes_finish_req() 578 cryp->req.sreq = req; in starfive_aes_do_one_req()
|
| /linux/include/linux/netfilter/ |
| A D | nf_conntrack_pptp.h | 288 struct PptpStartSessionRequest sreq; member
|
| /linux/arch/s390/kernel/ |
| A D | perf_cpum_sf.c | 164 struct hws_lsctl_request_block sreq; in sf_disable() local 166 memset(&sreq, 0, sizeof(sreq)); in sf_disable() 167 lsctl(&sreq); in sf_disable()
|