| /linux/drivers/media/rc/ |
| A D | imon.c | 529 if (!ictx || ictx->disconnected || !refcount_inc_not_zero(&ictx->users)) { in display_open() 1369 ictx->last_keycode = ictx->kc; in imon_mouse_event() 1536 ictx->kc = ictx->last_keycode; in imon_parse_press_type() 1541 ictx->kc = ictx->last_keycode; in imon_parse_press_type() 1546 ictx->kc = ictx->last_keycode; in imon_parse_press_type() 1615 ictx->pad_mouse = !ictx->pad_mouse; in imon_incoming_packet() 1674 ictx->last_keycode = ictx->kc; in imon_incoming_packet() 2032 ictx->vendor, ictx->product); in imon_init_idev() 2081 ictx->vendor, ictx->product); in imon_init_touch() 2271 ictx->idev = imon_init_idev(ictx); in imon_init_intf0() [all …]
|
| /linux/drivers/iommu/iommufd/ |
| A D | vfio_compat.c | 19 xa_lock(&ictx->objects); in get_compat_ioas() 20 if (!ictx->vfio_ioas || !iommufd_lock_obj(&ictx->vfio_ioas->obj)) in get_compat_ioas() 22 ioas = ictx->vfio_ioas; in get_compat_ioas() 24 xa_unlock(&ictx->objects); in get_compat_ioas() 59 xa_lock(&ictx->objects); in iommufd_vfio_compat_set_no_iommu() 60 if (!ictx->vfio_ioas) { in iommufd_vfio_compat_set_no_iommu() 61 ictx->no_iommu_mode = 1; in iommufd_vfio_compat_set_no_iommu() 89 xa_lock(&ictx->objects); in iommufd_vfio_compat_ioas_create() 99 if (ictx->vfio_ioas && iommufd_lock_obj(&ictx->vfio_ioas->obj)) { in iommufd_vfio_compat_ioas_create() 101 iommufd_put_object(ictx, &ictx->vfio_ioas->obj); in iommufd_vfio_compat_ioas_create() [all …]
|
| A D | main.c | 115 xa_lock(&ictx->objects); in iommufd_get_object() 237 ictx = kzalloc(sizeof(*ictx), GFP_KERNEL_ACCOUNT); in iommufd_fops_open() 238 if (!ictx) in iommufd_fops_open() 252 xa_init(&ictx->groups); in iommufd_fops_open() 253 ictx->file = filp; in iommufd_fops_open() 290 kfree(ictx); in iommufd_fops_release() 407 ucmd.ictx = ictx; in iommufd_fops_ioctl() 443 get_file(ictx->file); in iommufd_ctx_get() 462 iommufd_ctx_get(ictx); in iommufd_ctx_from_file() 463 return ictx; in iommufd_ctx_from_file() [all …]
|
| A D | iommufd_private.h | 106 struct iommufd_ctx *ictx; member 175 wake_up_interruptible_all(&ictx->destroy_wait); in iommufd_put_object() 225 iommufd_object_remove(ictx, obj, obj->id, 0); in iommufd_object_put_and_try_destroy() 241 #define iommufd_object_alloc(ictx, ptr, type) \ argument 242 __iommufd_object_alloc(ictx, ptr, type, obj) 268 return container_of(iommufd_get_object(ictx, id, in iommufd_get_ioas() 283 struct iommufd_ctx *ictx); 401 struct iommufd_ctx *ictx; member 415 struct iommufd_ctx *ictx; member 439 struct iommufd_ctx *ictx; member [all …]
|
| A D | ioas.c | 46 ioas = iommufd_ioas_alloc(ucmd->ictx); in iommufd_ioas_alloc_ioctl() 54 iommufd_object_finalize(ucmd->ictx, &ioas->obj); in iommufd_ioas_alloc_ioctl() 108 iommufd_put_object(ucmd->ictx, &ioas->obj); in iommufd_ioas_iova_ranges() 178 iommufd_put_object(ucmd->ictx, &ioas->obj); in iommufd_ioas_allow_iovas() 235 iommufd_put_object(ucmd->ictx, &ioas->obj); in iommufd_ioas_map() 326 iommufd_put_object(ucmd->ictx, &ioas->obj); in iommufd_ioas_unmap() 331 struct iommufd_ctx *ictx) in iommufd_option_rlimit_mode() argument 346 xa_lock(&ictx->objects); in iommufd_option_rlimit_mode() 347 if (!xa_empty(&ictx->objects)) { in iommufd_option_rlimit_mode() 357 xa_unlock(&ictx->objects); in iommufd_option_rlimit_mode() [all …]
|
| A D | device.c | 75 xa_lock(&ictx->groups); in iommufd_get_group() 78 xa_unlock(&ictx->groups); in iommufd_get_group() 82 xa_unlock(&ictx->groups); in iommufd_get_group() 101 new_igroup->ictx = ictx; in iommufd_get_group() 108 xa_lock(&ictx->groups); in iommufd_get_group() 208 idev->ictx = ictx; in iommufd_device_bind() 210 iommufd_ctx_get(ictx); in iommufd_device_bind() 252 if (!ictx || !group) in iommufd_ctx_has_group() 286 return idev->ictx; in iommufd_device_to_ictx() 849 access->ictx = ictx; in iommufd_access_create() [all …]
|
| A D | hw_pagetable.c | 127 ictx, hwpt_paging, IOMMUFD_OBJ_HWPT_PAGING, common.obj); in iommufd_hwpt_paging_alloc() 198 iommufd_object_abort_and_destroy(ictx, &hwpt->obj); in iommufd_hwpt_paging_alloc() 215 iommufd_hwpt_nested_alloc(struct iommufd_ctx *ictx, in iommufd_hwpt_nested_alloc() argument 259 iommufd_object_abort_and_destroy(ictx, &hwpt->obj); in iommufd_hwpt_nested_alloc() 299 ucmd->ictx, ioas, idev, cmd->flags, false, in iommufd_hwpt_alloc() 310 ucmd->ictx, in iommufd_hwpt_alloc() 336 iommufd_put_object(ucmd->ictx, &fault->obj); in iommufd_hwpt_alloc() 343 iommufd_object_finalize(ucmd->ictx, &hwpt->obj); in iommufd_hwpt_alloc() 352 iommufd_put_object(ucmd->ictx, pt_obj); in iommufd_hwpt_alloc() 354 iommufd_put_object(ucmd->ictx, &idev->obj); in iommufd_hwpt_alloc() [all …]
|
| A D | fault.c | 352 iommufd_ctx_put(fault->ictx); in iommufd_fault_fops_release() 376 fault = iommufd_object_alloc(ucmd->ictx, fault, IOMMUFD_OBJ_FAULT); in iommufd_fault_alloc() 380 fault->ictx = ucmd->ictx; in iommufd_fault_alloc() 394 iommufd_ctx_get(fault->ictx); in iommufd_fault_alloc() 409 iommufd_object_finalize(ucmd->ictx, &fault->obj); in iommufd_fault_alloc() 419 iommufd_ctx_put(fault->ictx); in iommufd_fault_alloc() 421 iommufd_object_abort_and_destroy(ucmd->ictx, &fault->obj); in iommufd_fault_alloc()
|
| /linux/fs/netfs/ |
| A D | locking.c | 34 if (!test_bit(NETFS_ICTX_ODIRECT, &ictx->flags)) in netfs_block_o_direct() 36 clear_bit(NETFS_ICTX_ODIRECT, &ictx->flags); in netfs_block_o_direct() 59 struct netfs_inode *ictx = netfs_inode(inode); in netfs_start_io_read() local 71 if (netfs_block_o_direct(ictx) < 0) { in netfs_start_io_read() 104 struct netfs_inode *ictx = netfs_inode(inode); in netfs_start_io_write() local 108 if (netfs_block_o_direct(ictx) < 0) { in netfs_start_io_write() 134 struct netfs_inode *ictx = netfs_inode(inode); in netfs_block_buffered() local 137 if (!test_bit(NETFS_ICTX_ODIRECT, &ictx->flags)) { in netfs_block_buffered() 138 set_bit(NETFS_ICTX_ODIRECT, &ictx->flags); in netfs_block_buffered() 143 clear_bit(NETFS_ICTX_ODIRECT, &ictx->flags); in netfs_block_buffered() [all …]
|
| A D | write_issue.c | 96 struct netfs_inode *ictx; in netfs_create_write_req() local 107 ictx = netfs_inode(wreq->inode); in netfs_create_write_req() 108 if (is_buffered && netfs_is_cache_enabled(ictx)) in netfs_create_write_req() 546 if (!mutex_trylock(&ictx->wb_lock)) { in netfs_writepages() 552 mutex_lock(&ictx->wb_lock); in netfs_writepages() 588 mutex_unlock(&ictx->wb_lock); in netfs_writepages() 597 mutex_unlock(&ictx->wb_lock); in netfs_writepages() 611 mutex_lock(&ictx->wb_lock); in netfs_begin_writethrough() 616 mutex_unlock(&ictx->wb_lock); in netfs_begin_writethrough() 666 struct netfs_inode *ictx = netfs_inode(wreq->inode); in netfs_end_writethrough() local [all …]
|
| A D | direct_write.c | 141 struct netfs_inode *ictx = netfs_inode(inode); in netfs_unbuffered_write_iter() local 188 if (end > ictx->zero_point) in netfs_unbuffered_write_iter() 189 ictx->zero_point = end; in netfs_unbuffered_write_iter() 191 fscache_invalidate(netfs_i_cookie(ictx), NULL, i_size_read(inode), in netfs_unbuffered_write_iter()
|
| /linux/arch/s390/crypto/ |
| A D | sha3_512_s390.c | 50 const struct sha3_state *ictx = in; in sha3_512_import() local 52 if (unlikely(ictx->rsizw)) in sha3_512_import() 54 sctx->count = ictx->rsiz; in sha3_512_import() 56 memcpy(sctx->state, ictx->st, sizeof(ictx->st)); in sha3_512_import() 57 memcpy(sctx->buf, ictx->buf, sizeof(ictx->buf)); in sha3_512_import() 58 sctx->first_message_part = ictx->partial; in sha3_512_import() 67 const struct sha3_state *ictx = in; in sha3_384_import() local 69 if (unlikely(ictx->rsizw)) in sha3_384_import() 71 sctx->count = ictx->rsiz; in sha3_384_import() 73 memcpy(sctx->state, ictx->st, sizeof(ictx->st)); in sha3_384_import() [all …]
|
| A D | sha3_256_s390.c | 49 const struct sha3_state *ictx = in; in sha3_256_import() local 51 sctx->count = ictx->rsiz; in sha3_256_import() 52 memcpy(sctx->state, ictx->st, sizeof(ictx->st)); in sha3_256_import() 53 memcpy(sctx->buf, ictx->buf, sizeof(ictx->buf)); in sha3_256_import() 54 sctx->first_message_part = ictx->partial; in sha3_256_import() 63 const struct sha3_state *ictx = in; in sha3_224_import() local 65 sctx->count = ictx->rsiz; in sha3_224_import() 66 memcpy(sctx->state, ictx->st, sizeof(ictx->st)); in sha3_224_import() 67 memcpy(sctx->buf, ictx->buf, sizeof(ictx->buf)); in sha3_224_import() 68 sctx->first_message_part = ictx->partial; in sha3_224_import()
|
| A D | sha1_s390.c | 59 const struct sha1_state *ictx = in; in s390_sha1_import() local 61 sctx->count = ictx->count; in s390_sha1_import() 62 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in s390_sha1_import() 63 memcpy(sctx->buf, ictx->buffer, sizeof(ictx->buffer)); in s390_sha1_import()
|
| A D | sha512_s390.c | 54 const struct sha512_state *ictx = in; in sha512_import() local 56 if (unlikely(ictx->count[1])) in sha512_import() 58 sctx->count = ictx->count[0]; in sha512_import() 60 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in sha512_import() 61 memcpy(sctx->buf, ictx->buf, sizeof(ictx->buf)); in sha512_import()
|
| A D | sha256_s390.c | 52 const struct sha256_state *ictx = in; in sha256_import() local 54 sctx->count = ictx->count; in sha256_import() 55 memcpy(sctx->state, ictx->state, sizeof(ictx->state)); in sha256_import() 56 memcpy(sctx->buf, ictx->buf, sizeof(ictx->buf)); in sha256_import()
|
| /linux/include/linux/ |
| A D | iommufd.h | 21 struct iommufd_device *iommufd_device_bind(struct iommufd_ctx *ictx, 48 iommufd_access_create(struct iommufd_ctx *ictx, 55 void iommufd_ctx_get(struct iommufd_ctx *ictx); 60 void iommufd_ctx_put(struct iommufd_ctx *ictx); 61 bool iommufd_ctx_has_group(struct iommufd_ctx *ictx, struct iommu_group *group); 70 int iommufd_vfio_compat_ioas_get_id(struct iommufd_ctx *ictx, u32 *out_ioas_id); 71 int iommufd_vfio_compat_ioas_create(struct iommufd_ctx *ictx); 72 int iommufd_vfio_compat_set_no_iommu(struct iommufd_ctx *ictx); 79 static inline void iommufd_ctx_put(struct iommufd_ctx *ictx) in iommufd_ctx_put() argument 104 static inline int iommufd_vfio_compat_ioas_create(struct iommufd_ctx *ictx) in iommufd_vfio_compat_ioas_create() argument [all …]
|
| A D | vfio.h | 114 struct iommufd_ctx *ictx, u32 *out_device_id); 136 int vfio_iommufd_get_dev_id(struct vfio_device *vdev, struct iommufd_ctx *ictx); 138 struct iommufd_ctx *ictx, u32 *out_device_id); 143 struct iommufd_ctx *ictx, u32 *out_device_id); 155 vfio_iommufd_get_dev_id(struct vfio_device *vdev, struct iommufd_ctx *ictx) in vfio_iommufd_get_dev_id() argument 161 ((int (*)(struct vfio_device *vdev, struct iommufd_ctx *ictx, \ 170 ((int (*)(struct vfio_device *vdev, struct iommufd_ctx *ictx, \
|
| /linux/drivers/vfio/ |
| A D | iommufd.c | 14 struct iommufd_ctx *ictx) in vfio_iommufd_device_has_compat_ioas() argument 18 return !iommufd_vfio_compat_ioas_get_id(ictx, &ioas_id); in vfio_iommufd_device_has_compat_ioas() 24 struct iommufd_ctx *ictx = df->iommufd; in vfio_df_iommufd_bind() local 28 return vdev->ops->bind_iommufd(vdev, ictx, &df->devid); in vfio_df_iommufd_bind() 32 struct iommufd_ctx *ictx) in vfio_iommufd_compat_attach_ioas() argument 43 ret = iommufd_vfio_compat_ioas_get_id(ictx, &ioas_id); in vfio_iommufd_compat_attach_ioas() 90 if (vfio_iommufd_device_ictx(vdev) == ictx) in vfio_iommufd_get_dev_id() 97 if (iommufd_ctx_has_group(ictx, group)) in vfio_iommufd_get_dev_id() 114 struct iommufd_ctx *ictx, u32 *out_device_id) in vfio_iommufd_physical_bind() argument 118 idev = iommufd_device_bind(ictx, vdev->dev, out_device_id); in vfio_iommufd_physical_bind() [all …]
|
| /linux/crypto/ |
| A D | essiv.c | 312 err = essiv_init_tfm(ictx, tctx); in essiv_skcipher_init_tfm() 334 aead = crypto_spawn_aead(&ictx->u.aead_spawn); in essiv_aead_init_tfm() 345 err = essiv_init_tfm(ictx, tctx); in essiv_aead_init_tfm() 385 crypto_drop_aead(&ictx->u.aead_spawn); in essiv_aead_free_instance() 453 struct essiv_instance_ctx *ictx; in essiv_create() local 480 sizeof(*ictx), GFP_KERNEL); in essiv_create() 485 ictx = crypto_instance_ctx(inst); in essiv_create() 493 &ictx->u.skcipher_spawn); in essiv_create() 500 sizeof(*ictx), GFP_KERNEL); in essiv_create() 505 ictx = crypto_instance_ctx(inst); in essiv_create() [all …]
|
| A D | hctr2.c | 345 xctr = crypto_spawn_skcipher(&ictx->xctr_spawn); in hctr2_init_tfm() 355 polyval = crypto_spawn_shash(&ictx->polyval_spawn); in hctr2_init_tfm() 398 crypto_drop_cipher(&ictx->blockcipher_spawn); in hctr2_free_instance() 399 crypto_drop_skcipher(&ictx->xctr_spawn); in hctr2_free_instance() 400 crypto_drop_shash(&ictx->polyval_spawn); in hctr2_free_instance() 412 struct hctr2_instance_ctx *ictx; in hctr2_create_common() local 423 inst = kzalloc(sizeof(*inst) + sizeof(*ictx), GFP_KERNEL); in hctr2_create_common() 426 ictx = skcipher_instance_ctx(inst); in hctr2_create_common() 429 err = crypto_grab_skcipher(&ictx->xctr_spawn, in hctr2_create_common() 448 err = crypto_grab_cipher(&ictx->blockcipher_spawn, in hctr2_create_common() [all …]
|
| A D | authenc.c | 117 struct authenc_instance_ctx *ictx = aead_instance_ctx(inst); in authenc_geniv_ahash_done() local 137 struct authenc_instance_ctx *ictx = aead_instance_ctx(inst); in crypto_authenc_genicv() local 193 struct authenc_instance_ctx *ictx = aead_instance_ctx(inst); in crypto_authenc_encrypt() local 198 ictx->reqoff); in crypto_authenc_encrypt() 231 struct authenc_instance_ctx *ictx = aead_instance_ctx(inst); in crypto_authenc_decrypt_tail() local 235 ictx->reqoff); in crypto_authenc_decrypt_tail() 279 struct authenc_instance_ctx *ictx = aead_instance_ctx(inst); in crypto_authenc_decrypt() local 302 struct authenc_instance_ctx *ictx = aead_instance_ctx(inst); in crypto_authenc_init_tfm() local 309 auth = crypto_spawn_ahash(&ictx->auth); in crypto_authenc_init_tfm() 313 enc = crypto_spawn_skcipher(&ictx->enc); in crypto_authenc_init_tfm() [all …]
|
| A D | adiantum.c | 448 hash = crypto_spawn_shash(&ictx->hash_spawn); in adiantum_init_tfm() 492 crypto_drop_skcipher(&ictx->streamcipher_spawn); in adiantum_free_instance() 493 crypto_drop_cipher(&ictx->blockcipher_spawn); in adiantum_free_instance() 494 crypto_drop_shash(&ictx->hash_spawn); in adiantum_free_instance() 527 struct adiantum_instance_ctx *ictx; in adiantum_create() local 537 inst = kzalloc(sizeof(*inst) + sizeof(*ictx), GFP_KERNEL); in adiantum_create() 540 ictx = skcipher_instance_ctx(inst); in adiantum_create() 543 err = crypto_grab_skcipher(&ictx->streamcipher_spawn, in adiantum_create() 551 err = crypto_grab_cipher(&ictx->blockcipher_spawn, in adiantum_create() 562 err = crypto_grab_shash(&ictx->hash_spawn, in adiantum_create() [all …]
|
| A D | pcrypt.c | 101 struct pcrypt_instance_ctx *ictx; in pcrypt_aead_encrypt() local 103 ictx = pcrypt_tfm_ictx(aead); in pcrypt_aead_encrypt() 117 err = padata_do_parallel(ictx->psenc, padata, &ctx->cb_cpu); in pcrypt_aead_encrypt() 150 struct pcrypt_instance_ctx *ictx; in pcrypt_aead_decrypt() local 152 ictx = pcrypt_tfm_ictx(aead); in pcrypt_aead_decrypt() 166 err = padata_do_parallel(ictx->psdec, padata, &ctx->cb_cpu); in pcrypt_aead_decrypt() 179 struct pcrypt_instance_ctx *ictx = aead_instance_ctx(inst); in pcrypt_aead_init_tfm() local 183 cpu_index = (unsigned int)atomic_inc_return(&ictx->tfm_count) % in pcrypt_aead_init_tfm() 190 cipher = crypto_spawn_aead(&ictx->spawn); in pcrypt_aead_init_tfm()
|
| /linux/drivers/crypto/allwinner/sun4i-ss/ |
| A D | sun4i-ss-hash.c | 88 const struct md5_state *ictx = in; in sun4i_hash_import_md5() local 93 op->byte_count = ictx->byte_count & ~0x3F; in sun4i_hash_import_md5() 94 op->len = ictx->byte_count & 0x3F; in sun4i_hash_import_md5() 96 memcpy(op->buf, ictx->block, op->len); in sun4i_hash_import_md5() 99 op->hash[i] = ictx->hash[i]; in sun4i_hash_import_md5() 131 const struct sha1_state *ictx = in; in sun4i_hash_import_sha1() local 136 op->byte_count = ictx->count & ~0x3F; in sun4i_hash_import_sha1() 137 op->len = ictx->count & 0x3F; in sun4i_hash_import_sha1() 139 memcpy(op->buf, ictx->buffer, op->len); in sun4i_hash_import_sha1() 142 op->hash[i] = ictx->state[i]; in sun4i_hash_import_sha1()
|