| /linux/arch/powerpc/crypto/ |
| A D | aes-gcm-p10-glue.c | 93 gctx->aadLen = alen; in set_aad() 104 memset(gctx->aad_hash, 0, 16); in set_aad() 107 memcpy(gctx->aad_hash, nXi, 16); in set_aad() 124 gctx->Plen = 0; in gcmp10_init() 133 memcpy(gctx->iv, iv, 16); in gcmp10_init() 135 gctx->aadLen = assoclen; in gcmp10_init() 136 memset(gctx->aad_hash, 0, 16); in gcmp10_init() 138 set_aad(gctx, hash, assoc, assoclen); in gcmp10_init() 149 if (len == 0 && gctx->aadLen == 0) { in finish_tag() 150 memcpy(hash->Htable, gctx->ivtag, 16); in finish_tag() [all …]
|
| /linux/drivers/gpu/drm/radeon/ |
| A D | atom.c | 196 val = gctx->card->reg_read(gctx->card, idx); in atom_get_src_int() 209 if (!gctx->iio[gctx->io_mode & 0x7F]) { in atom_get_src_int() 216 gctx->iio[gctx->io_mode & 0x7F], in atom_get_src_int() 327 val = gctx->card->pll_read(gctx->card, idx); in atom_get_src_int() 334 val = gctx->card->mc_read(gctx->card, idx); in atom_get_src_int() 475 gctx->card->reg_write(gctx->card, idx, in atom_put_dst() 478 gctx->card->reg_write(gctx->card, idx, val); in atom_put_dst() 491 if (!gctx->iio[gctx->io_mode & 0xFF]) { in atom_put_dst() 496 atom_iio_execute(gctx, gctx->iio[gctx->io_mode & 0xFF], in atom_put_dst() 561 gctx->card->pll_write(gctx->card, idx, val); in atom_put_dst() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| A D | atom.c | 198 val = gctx->card->reg_read(gctx->card, idx); in atom_get_src_int() 211 if (!gctx->iio[gctx->io_mode & 0x7F]) { in atom_get_src_int() 218 gctx->iio[gctx->io_mode & 0x7F], in atom_get_src_int() 329 val = gctx->card->pll_read(gctx->card, idx); in atom_get_src_int() 336 val = gctx->card->mc_read(gctx->card, idx); in atom_get_src_int() 476 gctx->card->reg_write(gctx->card, idx, in atom_put_dst() 479 gctx->card->reg_write(gctx->card, idx, val); in atom_put_dst() 492 if (!gctx->iio[gctx->io_mode & 0xFF]) { in atom_put_dst() 497 atom_iio_execute(gctx, gctx->iio[gctx->io_mode & 0xFF], in atom_put_dst() 562 gctx->card->pll_write(gctx->card, idx, val); in atom_put_dst() [all …]
|
| /linux/net/sunrpc/auth_gss/ |
| A D | gss_krb5_mech.c | 523 static u32 gss_krb5_get_mic(struct gss_ctx *gctx, struct xdr_buf *text, in gss_krb5_get_mic() argument 526 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_krb5_get_mic() 544 static u32 gss_krb5_verify_mic(struct gss_ctx *gctx, in gss_krb5_verify_mic() argument 548 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_krb5_verify_mic() 565 static u32 gss_krb5_wrap(struct gss_ctx *gctx, int offset, in gss_krb5_wrap() argument 568 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_krb5_wrap() 587 static u32 gss_krb5_unwrap(struct gss_ctx *gctx, int offset, in gss_krb5_unwrap() argument 590 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_krb5_unwrap() 593 &gctx->slack, &gctx->align); in gss_krb5_unwrap()
|
| /linux/crypto/ |
| A D | gcm.c | 231 lengths.b = cpu_to_be64(gctx->cryptlen * 8); in gcm_hash_len() 246 return gctx->complete(req, flags); in gcm_hash_len_continue() 291 remain = gcm_remain(gctx->cryptlen); in gcm_hash_crypt_continue() 320 if (gctx->cryptlen) in gcm_hash_assoc_remain_continue() 322 gctx->src, gctx->cryptlen, flags) ?: in gcm_hash_assoc_remain_continue() 428 gctx->cryptlen = req->cryptlen; in gcm_encrypt_continue() 429 gctx->complete = gcm_enc_copy_hash; in gcm_encrypt_continue() 494 crypto_gcm_init_crypt(req, gctx->cryptlen); in gcm_dec_hash_continue() 512 gctx->src = sg_next(pctx->src); in crypto_gcm_decrypt() 513 gctx->cryptlen = cryptlen; in crypto_gcm_decrypt() [all …]
|
| /linux/drivers/accel/ivpu/ |
| A D | ivpu_fw.c | 292 fw->mem = ivpu_bo_create(vdev, &vdev->gctx, &fw_range, fw->runtime_size, in ivpu_fw_mem_init() 299 ret = ivpu_mmu_context_set_pages_ro(vdev, &vdev->gctx, fw->read_only_addr, in ivpu_fw_mem_init() 328 fw->mem_shave_nn = ivpu_bo_create(vdev, &vdev->gctx, &vdev->hw->ranges.shave, in ivpu_fw_mem_init()
|
| A D | ivpu_drv.h | 130 struct ivpu_mmu_context gctx; member
|
| A D | ivpu_mmu_context.c | 576 return ivpu_mmu_context_init(vdev, &vdev->gctx, IVPU_GLOBAL_CONTEXT_MMU_SSID); in ivpu_mmu_global_context_init() 581 return ivpu_mmu_context_fini(vdev, &vdev->gctx); in ivpu_mmu_global_context_fini()
|
| A D | ivpu_gem.c | 338 return ivpu_bo_create(vdev, &vdev->gctx, &vdev->hw->ranges.global, size, flags); in ivpu_bo_create_global()
|
| A D | ivpu_mmu.c | 765 ret = ivpu_mmu_cd_add(vdev, 0, vdev->gctx.pgtable.pgd_dma); in ivpu_mmu_cd_add_gbl()
|
| /linux/drivers/crypto/chelsio/ |
| A D | chcr_algo.c | 114 static inline struct chcr_gcm_ctx *GCM_CTX(struct chcr_aead_ctx *gctx) in GCM_CTX() argument 116 return gctx->ctx->gcm; in GCM_CTX() 119 static inline struct chcr_authenc_ctx *AUTHENC_CTX(struct chcr_aead_ctx *gctx) in AUTHENC_CTX() argument 121 return gctx->ctx->authenc; in AUTHENC_CTX() 3499 struct chcr_gcm_ctx *gctx = GCM_CTX(aeadctx); in chcr_gcm_setkey() local 3545 memset(gctx->ghash_h, 0, AEAD_H_SIZE); in chcr_gcm_setkey() 3546 aes_encrypt(&aes, gctx->ghash_h, gctx->ghash_h); in chcr_gcm_setkey()
|