| /drivers/crypto/marvell/cesa/ |
| A D | hash.c | 33 if (!creq->last_req) in mv_cesa_ahash_req_iter_init() 191 creq->cache, creq->cache_ptr); in mv_cesa_ahash_std_step() 194 creq->cache, creq->cache_ptr); in mv_cesa_ahash_std_step() 246 i = mv_cesa_ahash_pad_req(creq, creq->cache); in mv_cesa_ahash_std_step() 443 memset(creq, 0, sizeof(*creq)); in mv_cesa_ahash_init() 452 creq->len = 0; in mv_cesa_ahash_init() 480 creq->cache + creq->cache_ptr, in mv_cesa_ahash_cache_req() 532 memcpy(ahashdreq->cache, creq->cache, creq->cache_ptr); in mv_cesa_ahash_dma_add_cache() 728 creq->cache_ptr = req->nbytes + creq->cache_ptr - in mv_cesa_ahash_dma_req_init() 851 memcpy(cache, creq->cache, creq->cache_ptr); in mv_cesa_ahash_export() [all …]
|
| A D | cipher.c | 72 mv_cesa_dma_cleanup(&creq->base); in mv_cesa_skcipher_dma_cleanup() 150 struct mv_cesa_req *basereq = &creq->base; in mv_cesa_skcipher_process() 164 mv_cesa_dma_step(&creq->base); in mv_cesa_skcipher_step() 173 struct mv_cesa_req *basereq = &creq->base; in mv_cesa_skcipher_dma_prepare() 194 creq->base.engine = engine; in mv_cesa_skcipher_prepare() 223 basereq = &creq->base; in mv_cesa_skcipher_complete() 320 struct mv_cesa_req *basereq = &creq->base; in mv_cesa_skcipher_dma_req_init() 436 if (creq->src_nents < 0) { in mv_cesa_skcipher_req_init() 438 return creq->src_nents; in mv_cesa_skcipher_req_init() 441 if (creq->dst_nents < 0) { in mv_cesa_skcipher_req_init() [all …]
|
| A D | cesa.c | 170 struct mv_cesa_req *creq) in mv_cesa_queue_req() argument 173 struct mv_cesa_engine *engine = creq->engine; in mv_cesa_queue_req() 177 if ((mv_cesa_req_get_type(creq) == CESA_DMA_REQ) && in mv_cesa_queue_req() 179 mv_cesa_tdma_chain(engine, creq); in mv_cesa_queue_req()
|
| A D | cesa.h | 734 struct mv_cesa_req *creq);
|
| /drivers/infiniband/hw/bnxt_re/ |
| A D | qplib_rcfw.c | 734 struct bnxt_qplib_creq_ctx *creq = &rcfw->creq; in bnxt_qplib_service_creq() local 797 creq = &rcfw->creq; in bnxt_qplib_creq_irq() 798 hwq = &creq->hwq; in bnxt_qplib_creq_irq() 935 creq = &rcfw->creq; in bnxt_qplib_alloc_rcfw_channel() 985 creq = &rcfw->creq; in bnxt_qplib_rcfw_stop_irq() 1009 creq = &rcfw->creq; in bnxt_qplib_disable_rcfw_channel() 1020 creq->msix_vec = 0; in bnxt_qplib_disable_rcfw_channel() 1030 creq = &rcfw->creq; in bnxt_qplib_rcfw_start_irq() 1033 if (creq->requested) in bnxt_qplib_rcfw_start_irq() 1138 creq = &rcfw->creq; in bnxt_qplib_start_rcfw() [all …]
|
| A D | main.c | 2048 bnxt_re_net_ring_free(rdev, rdev->rcfw.creq.ring_id, type); in bnxt_re_dev_uninit() 2078 struct bnxt_qplib_creq_ctx *creq; in bnxt_re_dev_init() local 2142 creq = &rdev->rcfw.creq; in bnxt_re_dev_init() 2143 rattr.dma_arr = creq->hwq.pbl[PBL_LVL_0].pg_map_arr; in bnxt_re_dev_init() 2144 rattr.pages = creq->hwq.pbl[creq->hwq.level].pg_count; in bnxt_re_dev_init() 2149 rc = bnxt_re_net_ring_alloc(rdev, &rattr, &creq->ring_id); in bnxt_re_dev_init() 2259 bnxt_re_net_ring_free(rdev, rdev->rcfw.creq.ring_id, type); in bnxt_re_dev_init()
|
| A D | qplib_rcfw.h | 225 struct bnxt_qplib_creq_ctx creq; member
|
| /drivers/crypto/cavium/nitrox/ |
| A D | nitrox_req.h | 213 struct se_crypto_request creq; member 623 struct se_crypto_request *creq = &nkreq->creq; in alloc_src_req_buf() local 648 struct se_crypto_request *creq = &nkreq->creq; in nitrox_creq_set_src_sg() local 651 sg = creq->src; in nitrox_creq_set_src_sg() 670 struct se_crypto_request *creq = &nkreq->creq; in alloc_dst_req_buf() local 681 struct se_crypto_request *creq = &nkreq->creq; in nitrox_creq_set_orh() local 684 set_orh_value(creq->orh); in nitrox_creq_set_orh() 689 struct se_crypto_request *creq = &nkreq->creq; in nitrox_creq_set_comp() local 692 set_comp_value(creq->comp); in nitrox_creq_set_comp() 704 struct se_crypto_request *creq = &nkreq->creq; in nitrox_creq_set_dst_sg() local [all …]
|
| A D | nitrox_aead.c | 153 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_set_creq() local 157 creq->flags = rctx->flags; in nitrox_set_creq() 161 creq->ctrl.value = 0; in nitrox_set_creq() 162 creq->opcode = FLEXI_CRYPTO_ENCRYPT_HMAC; in nitrox_set_creq() 163 creq->ctrl.s.arg = rctx->ctrl_arg; in nitrox_set_creq() 170 creq->gph.param3 = cpu_to_be16(param3.param); in nitrox_set_creq() 172 creq->ctx_handle = rctx->ctx_handle; in nitrox_set_creq() 218 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_aes_gcm_enc() local 252 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_aes_gcm_dec() local 446 struct se_crypto_request *creq = &aead_rctx->nkreq.creq; in nitrox_rfc4106_enc() local [all …]
|
| A D | nitrox_skcipher.c | 92 if (nkreq->creq.ctrl.s.arg == ENCRYPT) { in nitrox_cbc_cipher_callback() 253 struct se_crypto_request *creq; in nitrox_skcipher_crypt() local 256 creq = &nkreq->creq; in nitrox_skcipher_crypt() 257 creq->flags = skreq->base.flags; in nitrox_skcipher_crypt() 262 creq->ctrl.value = 0; in nitrox_skcipher_crypt() 263 creq->opcode = FLEXI_CRYPTO_ENCRYPT_HMAC; in nitrox_skcipher_crypt() 264 creq->ctrl.s.arg = (enc ? ENCRYPT : DECRYPT); in nitrox_skcipher_crypt() 267 creq->gph.param1 = 0; in nitrox_skcipher_crypt() 269 creq->gph.param2 = cpu_to_be16(ivsize); in nitrox_skcipher_crypt() 270 creq->gph.param3 = 0; in nitrox_skcipher_crypt() [all …]
|
| A D | nitrox_reqmgr.c | 214 struct se_crypto_request *creq) in softreq_map_iobuf() argument 218 ret = dma_map_inbufs(sr, creq); in softreq_map_iobuf() 222 ret = dma_map_outbufs(sr, creq); in softreq_map_iobuf()
|
| /drivers/crypto/ccree/ |
| A D | cc_request_mgr.c | 47 struct cc_crypto_req creq; member 348 struct cc_crypto_req *creq; in cc_proc_backlog() local 362 creq = &bli->creq; in cc_proc_backlog() 363 req = creq->user_arg; in cc_proc_backlog() 370 creq->user_cb(dev, req, -EINPROGRESS); in cc_proc_backlog() 387 cc_do_send_request(drvdata, &bli->creq, bli->desc, bli->len, in cc_proc_backlog() 435 memcpy(&bli->creq, cc_req, sizeof(*cc_req)); in cc_send_request()
|
| /drivers/net/wireless/marvell/libertas/ |
| A D | cfg.c | 1279 creq = kzalloc(sizeof(*creq) + sizeof(struct cfg80211_ssid) + in _new_connect_scan_req() 1282 if (!creq) in _new_connect_scan_req() 1286 creq->ssids = (void *)&creq->channels[n_channels]; in _new_connect_scan_req() 1287 creq->n_channels = n_channels; in _new_connect_scan_req() 1288 creq->n_ssids = 1; in _new_connect_scan_req() 1310 creq->n_channels = i; in _new_connect_scan_req() 1317 kfree(creq); in _new_connect_scan_req() 1318 creq = NULL; in _new_connect_scan_req() 1321 return creq; in _new_connect_scan_req() 1336 struct cfg80211_scan_request *creq; in lbs_cfg_connect() local [all …]
|
| /drivers/gpu/drm/vmwgfx/ |
| A D | vmwgfx_surface.c | 1451 rep->creq = rep_ext.creq.base; in vmw_gb_surface_reference_ioctl() 1726 rep->creq.base.format = metadata->format; in vmw_gb_surface_reference_internal() 1727 rep->creq.base.mip_levels = metadata->mip_levels[0]; in vmw_gb_surface_reference_internal() 1728 rep->creq.base.drm_surface_flags = 0; in vmw_gb_surface_reference_internal() 1731 rep->creq.base.array_size = metadata->array_size; in vmw_gb_surface_reference_internal() 1732 rep->creq.base.buffer_handle = backup_handle; in vmw_gb_surface_reference_internal() 1733 rep->creq.base.base_size = metadata->base_size; in vmw_gb_surface_reference_internal() 1741 rep->creq.version = drm_vmw_gb_surface_v1; in vmw_gb_surface_reference_internal() 1742 rep->creq.svga3d_flags_upper_32_bits = in vmw_gb_surface_reference_internal() 1745 rep->creq.quality_level = metadata->quality_level; in vmw_gb_surface_reference_internal() [all …]
|
| /drivers/crypto/inside-secure/ |
| A D | safexcel_cipher.c | 1630 struct safexcel_cipher_req *creq = aead_request_ctx(req); in safexcel_aead_encrypt() local 1632 return safexcel_queue_req(&req->base, creq, SAFEXCEL_ENCRYPT); in safexcel_aead_encrypt() 1637 struct safexcel_cipher_req *creq = aead_request_ctx(req); in safexcel_aead_decrypt() local 2685 struct safexcel_cipher_req *creq = aead_request_ctx(req); in safexcel_ccm_encrypt() local 2695 struct safexcel_cipher_req *creq = aead_request_ctx(req); in safexcel_ccm_decrypt() local 2824 struct safexcel_cipher_req *creq = aead_request_ctx(req); in safexcel_aead_chachapoly_crypt() local 2840 return safexcel_queue_req(&req->base, creq, dir); in safexcel_aead_chachapoly_crypt() 3269 struct safexcel_cipher_req *creq = aead_request_ctx(req); in safexcel_aead_sm4cbc_sm3_encrypt() local 3283 struct safexcel_cipher_req *creq = aead_request_ctx(req); in safexcel_aead_sm4cbc_sm3_decrypt() local 3555 struct safexcel_cipher_req *creq = aead_request_ctx(req); in safexcel_rfc4309_ccm_encrypt() local [all …]
|
| /drivers/usb/gadget/function/ |
| A D | f_fs.c | 3780 const struct usb_ctrlrequest *creq) in ffs_func_setup() argument 3787 pr_vdebug("creq->bRequestType = %02x\n", creq->bRequestType); in ffs_func_setup() 3788 pr_vdebug("creq->bRequest = %02x\n", creq->bRequest); in ffs_func_setup() 3806 switch (creq->bRequestType & USB_RECIP_MASK) { in ffs_func_setup() 3814 ret = ffs_func_revmap_ep(func, le16_to_cpu(creq->wIndex)); in ffs_func_setup() 3823 ret = le16_to_cpu(creq->wIndex); in ffs_func_setup() 3829 ffs->ev.setup = *creq; in ffs_func_setup() 3838 const struct usb_ctrlrequest *creq, in ffs_func_req_match() argument 3846 switch (creq->bRequestType & USB_RECIP_MASK) { in ffs_func_req_match() 3849 le16_to_cpu(creq->wIndex)) >= 0); in ffs_func_req_match() [all …]
|
| /drivers/net/ethernet/marvell/octeontx2/nic/ |
| A D | otx2_tc.c | 369 struct nix_mcast_grp_create_req *creq; in otx2_tc_update_mcast() local 375 creq = otx2_mbox_alloc_msg_nix_mcast_grp_create(&nic->mbox); in otx2_tc_update_mcast() 376 if (!creq) { in otx2_tc_update_mcast() 381 creq->dir = NIX_MCAST_INGRESS; in otx2_tc_update_mcast() 391 &creq->hdr); in otx2_tc_update_mcast()
|
| /drivers/usb/gadget/udc/ |
| A D | aspeed_udc.c | 216 void __iomem *creq; member 1001 memcpy_fromio(&crq, udc->creq, sizeof(crq)); in ast_udc_getstatus() 1041 memcpy_fromio(&crq, udc->creq, sizeof(crq)); in ast_udc_ep0_handle_setup() 1529 udc->creq = udc->reg + AST_UDC_SETUP0; in ast_udc_probe()
|