| /drivers/infiniband/core/ |
| A D | roce_gid_mgmt.c | 647 for (i = 0; i < ARRAY_SIZE(work->cmds) && work->cmds[i].cb; i++) { in netdevice_event_work_handler() 650 work->cmds[i].cb, in netdevice_event_work_handler() 651 work->cmds[i].ndev); in netdevice_event_work_handler() 669 memcpy(ndev_work->cmds, cmds, sizeof(ndev_work->cmds)); in netdevice_queue_work() 670 for (i = 0; i < ARRAY_SIZE(ndev_work->cmds) && ndev_work->cmds[i].cb; i++) { in netdevice_queue_work() 707 cmds[1] = add_cmd; in ndev_event_unlink() 730 cmds[0].ndev = event_ndev; in ndev_event_link() 787 cmds[2] = add_cmd; in netdevice_event() 792 cmds[0] = del_cmd; in netdevice_event() 801 cmds[2] = add_cmd; in netdevice_event() [all …]
|
| /drivers/media/pci/saa7164/ |
| A D | saa7164-cmd.c | 18 if (dev->cmds[i].inuse == 0) { in saa7164_cmd_alloc_seqno() 19 dev->cmds[i].inuse = 1; in saa7164_cmd_alloc_seqno() 20 dev->cmds[i].signalled = 0; in saa7164_cmd_alloc_seqno() 21 dev->cmds[i].timeout = 0; in saa7164_cmd_alloc_seqno() 22 ret = dev->cmds[i].seqno; in saa7164_cmd_alloc_seqno() 36 dev->cmds[seqno].inuse = 0; in saa7164_cmd_free_seqno() 37 dev->cmds[seqno].signalled = 0; in saa7164_cmd_free_seqno() 38 dev->cmds[seqno].timeout = 0; in saa7164_cmd_free_seqno() 48 dev->cmds[seqno].timeout = 1; in saa7164_cmd_timeout_seqno() 178 u32 cmds; in saa7164_cmd_set() local [all …]
|
| /drivers/gpu/drm/i915/gt/ |
| A D | gen7_renderclear.c | 383 struct batch_chunk cmds, state; in emit_batch() local 397 gen7_emit_pipeline_flush(&cmds); in emit_batch() 398 gen7_emit_pipeline_invalidate(&cmds); in emit_batch() 401 batch_add(&cmds, 0xffff0000 | in emit_batch() 408 gen7_emit_pipeline_invalidate(&cmds); in emit_batch() 409 gen7_emit_pipeline_flush(&cmds); in emit_batch() 412 gen7_emit_pipeline_invalidate(&cmds); in emit_batch() 414 batch_add(&cmds, MI_NOOP); in emit_batch() 415 gen7_emit_pipeline_invalidate(&cmds); in emit_batch() 417 gen7_emit_pipeline_flush(&cmds); in emit_batch() [all …]
|
| /drivers/gpu/drm/vmwgfx/ |
| A D | vmwgfx_overlay.c | 105 } *cmds; in vmw_overlay_send_put() local 121 if (!cmds) in vmw_overlay_send_put() 124 items = (typeof(items))&cmds[1]; in vmw_overlay_send_put() 131 cmds->header.streamId = arg->stream_id; in vmw_overlay_send_put() 185 } *cmds; in vmw_overlay_send_stop() local 189 cmds = VMW_CMD_RESERVE(dev_priv, sizeof(*cmds)); in vmw_overlay_send_stop() 190 if (cmds) in vmw_overlay_send_stop() 201 fill_escape(&cmds->escape, sizeof(cmds->body)); in vmw_overlay_send_stop() 203 cmds->body.header.streamId = stream_id; in vmw_overlay_send_stop() 205 cmds->body.items[0].value = false; in vmw_overlay_send_stop() [all …]
|
| /drivers/net/wireless/intel/iwlwifi/fw/ |
| A D | notif-wait.c | 47 if (w->cmds[i] == rec_id || in iwl_notification_wait() 48 (!iwl_cmd_groupid(w->cmds[i]) && in iwl_notification_wait() 49 DEF_ID(w->cmds[i]) == rec_id)) { in iwl_notification_wait() 85 const u16 *cmds, int n_cmds, in iwl_init_notification_wait() argument 96 memcpy(wait_entry->cmds, cmds, n_cmds * sizeof(u16)); in iwl_init_notification_wait()
|
| A D | notif-wait.h | 53 u16 cmds[MAX_NOTIF_CMDS]; member 83 const u16 *cmds, int n_cmds,
|
| /drivers/iommu/arm/arm-smmu-v3/ |
| A D | arm-smmu-v3-iommufd.c | 345 struct arm_vsmmu_invalidation_cmd *cmds; in arm_vsmmu_cache_invalidate() local 350 cmds = kcalloc(array->entry_num, sizeof(*cmds), GFP_KERNEL); in arm_vsmmu_cache_invalidate() 351 if (!cmds) in arm_vsmmu_cache_invalidate() 353 cur = cmds; in arm_vsmmu_cache_invalidate() 354 end = cmds + array->entry_num; in arm_vsmmu_cache_invalidate() 356 static_assert(sizeof(*cmds) == 2 * sizeof(u64)); in arm_vsmmu_cache_invalidate() 358 cmds, sizeof(*cmds), array, in arm_vsmmu_cache_invalidate() 363 last = cmds; in arm_vsmmu_cache_invalidate() 384 array->entry_num = cur - cmds; in arm_vsmmu_cache_invalidate() 385 kfree(cmds); in arm_vsmmu_cache_invalidate()
|
| A D | arm-smmu-v3.c | 942 cmds->num = 0; in arm_smmu_cmdq_batch_init() 956 arm_smmu_cmdq_issue_cmdlist(smmu, cmds->cmdq, cmds->cmds, in arm_smmu_cmdq_batch_add() 957 cmds->num, true); in arm_smmu_cmdq_batch_add() 962 arm_smmu_cmdq_issue_cmdlist(smmu, cmds->cmdq, cmds->cmds, in arm_smmu_cmdq_batch_add() 963 cmds->num, false); in arm_smmu_cmdq_batch_add() 968 if (unlikely(arm_smmu_cmdq_build_cmd(&cmds->cmds[index], cmd))) { in arm_smmu_cmdq_batch_add() 974 cmds->num++; in arm_smmu_cmdq_batch_add() 980 return arm_smmu_cmdq_issue_cmdlist(smmu, cmds->cmdq, cmds->cmds, in arm_smmu_cmdq_batch_submit() 981 cmds->num, true); in arm_smmu_cmdq_batch_submit() 1224 struct arm_smmu_cmdq_batch cmds; in arm_smmu_sync_cd() local [all …]
|
| /drivers/gpu/drm/i915/gt/uc/ |
| A D | intel_guc_ct.c | 180 ctb->cmds = cmds; in guc_ct_buffer_init() 265 u32 *cmds; in intel_guc_ct_init() local 333 u32 base, desc, cmds, size; in intel_guc_ct_enable() local 357 cmds = base + ptrdiff(ct->ctbs.recv.cmds, blob); in intel_guc_ct_enable() 364 cmds = base + ptrdiff(ct->ctbs.send.cmds, blob); in intel_guc_ct_enable() 442 u32 *cmds = ctb->cmds; in ct_write() local 483 cmds[tail] = header; in ct_write() 486 cmds[tail] = hxg; in ct_write() 490 cmds[tail] = action[i]; in ct_write() 893 u32 *cmds = ctb->cmds; in ct_read() local [all …]
|
| /drivers/gpu/host1x/ |
| A D | job.c | 66 job->cmds = num_cmdbufs ? mem : NULL; in host1x_job_alloc() 220 if (job->cmds[i].is_wait) in pin_job() 223 g = &job->cmds[i].gather; in pin_job() 541 if (job->cmds[i].is_wait) in copy_gathers() 544 g = &job->cmds[i].gather; in copy_gathers() 570 if (job->cmds[i].is_wait) in copy_gathers() 572 g = &job->cmds[i].gather; in copy_gathers() 619 if (job->cmds[i].is_wait) in host1x_job_pin() 621 g = &job->cmds[i].gather; in host1x_job_pin() 632 if (!job->cmds[j].is_wait && in host1x_job_pin() [all …]
|
| /drivers/dma/ |
| A D | bcm-sba-raid.c | 115 struct brcm_sba_command cmds[]; member 608 msg->sba.cmds = cmds; in sba_fillup_interrupt_msg() 696 msg->sba.cmds = cmds; in sba_fillup_memcpy_msg() 761 struct brcm_sba_command *cmds, in sba_fillup_xor_msg() argument 835 msg->sba.cmds = cmds; in sba_fillup_xor_msg() 905 struct brcm_sba_command *cmds, in sba_fillup_pq_msg() argument 1046 msg->sba.cmds = cmds; in sba_fillup_pq_msg() 1068 req->cmds, &req->msg, in sba_prep_dma_pq_req() 1300 msg->sba.cmds = cmds; in sba_fillup_pq_single_msg() 1504 req->cmds[j].cmd = 0; in sba_prealloc_channel_resources() [all …]
|
| /drivers/vdpa/mlx5/core/ |
| A D | mr.c | 213 struct mlx5_vdpa_async_cmd *cmds; in create_direct_keys() local 218 cmds = kvcalloc(mr->num_directs, sizeof(*cmds), GFP_KERNEL); in create_direct_keys() 219 if (!cmds) in create_direct_keys() 234 cmds[i].out = cmd_mem->out; in create_direct_keys() 235 cmds[i].outlen = sizeof(cmd_mem->out); in create_direct_keys() 236 cmds[i].in = cmd_mem->in; in create_direct_keys() 275 kvfree(cmds); in create_direct_keys() 290 cmds = kvcalloc(mr->num_directs, sizeof(*cmds), GFP_KERNEL); in DEFINE_FREE() 292 if (!cmds || !cmd_mem) in DEFINE_FREE() 296 cmds[i].out = cmd_mem[i].out; in DEFINE_FREE() [all …]
|
| A D | resources.c | 332 struct mlx5_vdpa_async_cmd *cmds, in issue_async_cmd() argument 337 struct mlx5_vdpa_async_cmd *cmd = &cmds[issued]; in issue_async_cmd() 349 wait_for_completion(&cmds[*completed].cmd_done); in issue_async_cmd() 367 struct mlx5_vdpa_async_cmd *cmds, in mlx5_vdpa_exec_async_cmds() argument 375 init_completion(&cmds[i].cmd_done); in mlx5_vdpa_exec_async_cmds() 379 err = issue_async_cmd(mvdev, cmds, issued, &completed); in mlx5_vdpa_exec_async_cmds() 390 wait_for_completion(&cmds[completed++].cmd_done); in mlx5_vdpa_exec_async_cmds()
|
| /drivers/crypto/ccp/ |
| A D | ccp-crypto-main.c | 57 struct list_head cmds; member 106 list_for_each_entry_continue(tmp, &req_queue.cmds, entry) { in ccp_crypto_cmd_complete() 117 if (req_queue.backlog != &req_queue.cmds) { in ccp_crypto_cmd_complete() 228 list_for_each_entry(tmp, &req_queue.cmds, entry) { in ccp_crypto_enqueue_cmd() 244 if (req_queue.backlog == &req_queue.cmds) in ccp_crypto_enqueue_cmd() 250 list_add_tail(&crypto_cmd->entry, &req_queue.cmds); in ccp_crypto_enqueue_cmd() 409 INIT_LIST_HEAD(&req_queue.cmds); in ccp_crypto_init() 410 req_queue.backlog = &req_queue.cmds; in ccp_crypto_init()
|
| /drivers/interconnect/qcom/ |
| A D | bcm-voter.c | 281 struct tcs_cmd cmds[MAX_BCMS]; in qcom_icc_bcm_voter_commit() local 309 tcs_list_gen(voter, QCOM_ICC_BUCKET_AMC, cmds, commit_idx); in qcom_icc_bcm_voter_commit() 316 cmds, commit_idx); in qcom_icc_bcm_voter_commit() 345 tcs_list_gen(voter, QCOM_ICC_BUCKET_WAKE, cmds, commit_idx); in qcom_icc_bcm_voter_commit() 347 ret = rpmh_write_batch(voter->dev, RPMH_WAKE_ONLY_STATE, cmds, commit_idx); in qcom_icc_bcm_voter_commit() 353 tcs_list_gen(voter, QCOM_ICC_BUCKET_SLEEP, cmds, commit_idx); in qcom_icc_bcm_voter_commit() 355 ret = rpmh_write_batch(voter->dev, RPMH_SLEEP_STATE, cmds, commit_idx); in qcom_icc_bcm_voter_commit()
|
| /drivers/iio/adc/ |
| A D | ti-adc108s102.c | 116 unsigned int bit, cmds; in adc108s102_update_scan_mode() local 122 cmds = 0; in adc108s102_update_scan_mode() 124 st->tx_buf[cmds++] = cpu_to_be16(ADC108S102_CMD(bit)); in adc108s102_update_scan_mode() 127 st->tx_buf[cmds++] = 0x00; in adc108s102_update_scan_mode() 132 st->ring_xfer.len = cmds * sizeof(st->tx_buf[0]); in adc108s102_update_scan_mode()
|
| /drivers/media/platform/renesas/vsp1/ |
| A D | vsp1_dl.c | 161 struct vsp1_dl_ext_cmd *cmds; member 446 pool->cmds = kcalloc(num_cmds, sizeof(*pool->cmds), GFP_KERNEL); in vsp1_dl_cmd_pool_create() 447 if (!pool->cmds) { in vsp1_dl_cmd_pool_create() 460 kfree(pool->cmds); in vsp1_dl_cmd_pool_create() 466 struct vsp1_dl_ext_cmd *cmd = &pool->cmds[i]; in vsp1_dl_cmd_pool_create() 480 cmd->cmds = pool->mem + cmd_offset; in vsp1_dl_cmd_pool_create() 535 kfree(pool->cmds); in vsp1_dl_ext_cmd_pool_destroy() 764 cmd->cmds[0].opcode = cmd->opcode; in vsp1_dl_ext_cmd_fill_header() 765 cmd->cmds[0].flags = cmd->flags; in vsp1_dl_ext_cmd_fill_header() 766 cmd->cmds[0].address_set = cmd->data_dma; in vsp1_dl_ext_cmd_fill_header() [all …]
|
| /drivers/nvme/target/ |
| A D | rdma.c | 104 struct nvmet_rdma_cmd *cmds; member 129 struct nvmet_rdma_cmd *cmds; member 367 struct nvmet_rdma_cmd *cmds; in nvmet_rdma_alloc_cmds() local 371 if (!cmds) in nvmet_rdma_alloc_cmds() 380 return cmds; in nvmet_rdma_alloc_cmds() 385 kfree(cmds); in nvmet_rdma_alloc_cmds() 397 kfree(cmds); in nvmet_rdma_free_cmds() 1114 if (IS_ERR(nsrq->cmds)) { in nvmet_rdma_init_srq() 1115 ret = PTR_ERR(nsrq->cmds); in nvmet_rdma_init_srq() 1123 nsrq->cmds[i].nsrq = nsrq; in nvmet_rdma_init_srq() [all …]
|
| /drivers/mailbox/ |
| A D | bcm-flexrm-mailbox.c | 709 if (!msg->sba.cmds || !msg->sba.cmds_count) in flexrm_sba_sanity_check() 717 if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_B) && in flexrm_sba_sanity_check() 718 (msg->sba.cmds[i].data_len > SRCT_LENGTH_MASK)) in flexrm_sba_sanity_check() 720 if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_C) && in flexrm_sba_sanity_check() 721 (msg->sba.cmds[i].data_len > SRCT_LENGTH_MASK)) in flexrm_sba_sanity_check() 724 (msg->sba.cmds[i].resp_len > DSTT_LENGTH_MASK)) in flexrm_sba_sanity_check() 727 (msg->sba.cmds[i].data_len > DSTT_LENGTH_MASK)) in flexrm_sba_sanity_check() 742 if ((msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_B) || in flexrm_sba_estimate_nonheader_desc_count() 743 (msg->sba.cmds[i].flags & BRCM_SBA_CMD_TYPE_C)) in flexrm_sba_estimate_nonheader_desc_count() 746 if (msg->sba.cmds[i].flags & BRCM_SBA_CMD_HAS_RESP) in flexrm_sba_estimate_nonheader_desc_count() [all …]
|
| /drivers/gpu/drm/xe/ |
| A D | xe_guc_ct.c | 1563 &g2h->cmds, sizeof(u32) * head, in g2h_read() 1566 &g2h->cmds, 0, in g2h_read() 1570 &g2h->cmds, sizeof(u32) * head, in g2h_read() 1757 xe_map_wr_ring_u32(xe, cmds, idx, size, lo); in xe_fixup_u64_in_cmds() 1773 struct iosys_map *cmds, u32 head, in ct_fixup_ggtt_in_message() argument 1792 xe_fixup_u64_in_cmds(xe, cmds, size, head + in ct_fixup_ggtt_in_message() 1795 xe_fixup_u64_in_cmds(xe, cmds, size, head + in ct_fixup_ggtt_in_message() 1798 xe_fixup_u64_in_cmds(xe, cmds, size, head + in ct_fixup_ggtt_in_message() 1804 n = xe_map_rd_ring_u32(xe, cmds, head + in ct_fixup_ggtt_in_message() 1808 xe_fixup_u64_in_cmds(xe, cmds, size, head + in ct_fixup_ggtt_in_message() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | fs_core.c | 1411 root->cmds->destroy_flow_table(root, ft); in __mlx5_create_flow_table() 2591 err = root->cmds->update_root_ft(root, in update_root_ft_destroy() 3029 root_ns->cmds = cmds; in create_root_ns() 4226 const struct mlx5_flow_cmds *cmds; in mlx5_flow_namespace_set_mode() local 4241 cmds = mlx5_fs_cmd_get_dr_cmds(); in mlx5_flow_namespace_set_mode() 4243 cmds = mlx5_fs_cmd_get_hws_cmds(); in mlx5_flow_namespace_set_mode() 4245 cmds = mlx5_fs_cmd_get_fw_cmds(); in mlx5_flow_namespace_set_mode() 4246 if (!cmds) in mlx5_flow_namespace_set_mode() 4249 err = cmds->create_ns(root); in mlx5_flow_namespace_set_mode() 4256 root->cmds->destroy_ns(root); in mlx5_flow_namespace_set_mode() [all …]
|
| /drivers/gpu/drm/tegra/ |
| A D | submit.c | 397 struct drm_tegra_submit_cmd *cmds; in submit_create_job() local 405 cmds = alloc_copy_user_array(u64_to_user_ptr(args->cmds_ptr), args->num_cmds, in submit_create_job() 406 sizeof(*cmds)); in submit_create_job() 407 if (IS_ERR(cmds)) { in submit_create_job() 409 return ERR_CAST(cmds); in submit_create_job() 428 struct drm_tegra_submit_cmd *cmd = &cmds[i]; in submit_create_job() 485 kvfree(cmds); in submit_create_job()
|
| /drivers/gpu/drm/amd/display/dc/ |
| A D | dc_dmub_srv.c | 1847 union dmub_rb_cmd cmds[MAX_PLANES]; in dc_dmub_srv_fams2_passthrough_flip() local 1856 memset(cmds, 0, sizeof(union dmub_rb_cmd) * MAX_PLANES); in dc_dmub_srv_fams2_passthrough_flip() 1870 cmds[num_cmds].fams2_flip.header.payload_bytes = in dc_dmub_srv_fams2_passthrough_flip() 1874 cmds[num_cmds].fams2_flip.header.multi_cmd_pending = 1; in dc_dmub_srv_fams2_passthrough_flip() 1891 cmds[num_cmds].fams2_flip.flip_info.addr_info.meta_addr_lo = in dc_dmub_srv_fams2_passthrough_flip() 1893 cmds[num_cmds].fams2_flip.flip_info.addr_info.meta_addr_hi = in dc_dmub_srv_fams2_passthrough_flip() 1895 cmds[num_cmds].fams2_flip.flip_info.addr_info.surf_addr_lo = in dc_dmub_srv_fams2_passthrough_flip() 1897 cmds[num_cmds].fams2_flip.flip_info.addr_info.surf_addr_hi = in dc_dmub_srv_fams2_passthrough_flip() 1907 cmds[num_cmds].fams2_flip.flip_info.addr_info.meta_addr_lo = in dc_dmub_srv_fams2_passthrough_flip() 1909 cmds[num_cmds].fams2_flip.flip_info.addr_info.meta_addr_hi = in dc_dmub_srv_fams2_passthrough_flip() [all …]
|
| /drivers/gpu/drm/panel/ |
| A D | panel-novatek-nt35510.c | 200 u32 cmds; member 725 if (nt->conf->cmds & NT35510_CMD_SETVCMOFF) { in nt35510_setup_power() 912 if (nt->conf->cmds & NT35510_CMD_CORRECT_GAMMA) { in nt35510_power_on() 1022 if (nt->conf->cmds & NT35510_CMD_CONTROL_DISPLAY) { in nt35510_prepare() 1168 if (nt->conf->cmds & NT35510_CMD_CONTROL_DISPLAY) in nt35510_probe() 1250 .cmds = NT35510_CMD_CORRECT_GAMMA, 1325 .cmds = NT35510_CMD_CONTROL_DISPLAY | NT35510_CMD_SETVCMOFF,
|
| /drivers/soc/qcom/ |
| A D | trace-rpmh.h | 31 __entry->addr = r->cmds[0].addr; 32 __entry->data = r->cmds[0].data;
|