Home
last modified time | relevance | path

Searched refs:num_chunks (Results 1 – 25 of 28) sorted by relevance

12

/drivers/staging/media/atomisp/pci/
A Dsh_css_defs.h282 #define __ISP_MIN_INTERNAL_WIDTH(num_chunks, pipelining, mode) \ argument
283 ((num_chunks) * (pipelining) * (1 << _ISP_LOG_VECTOR_STEP(mode)) * \
292 #define __ISP_CHUNK_STRIDE_DDR(c_subsampling, num_chunks) \ argument
293 ((c_subsampling) * (num_chunks) * HIVE_ISP_DDR_WORD_BYTES)
299 num_chunks, \ argument
304 __ISP_MIN_INTERNAL_WIDTH(num_chunks, \
310 __ISP_CHUNK_STRIDE_DDR(c_subsampling, num_chunks) \
A Dia_css_acc_types.h111 u32 num_chunks; member
/drivers/net/ethernet/intel/idpf/
A Didpf_virtchnl.c1197 u16 num_chunks = le16_to_cpu(chunks->num_chunks); in idpf_vport_get_q_reg() local
1672 k += num_chunks; in idpf_send_config_tx_queues_msg()
1674 num_chunks = min(num_chunks, totqs); in idpf_send_config_tx_queues_msg()
1826 num_chunks = min(num_chunks, totqs); in idpf_send_config_rx_queues_msg()
1960 eq->chunks.num_chunks = cpu_to_le16(num_chunks); in idpf_send_ena_dis_queues_msg()
1972 num_chunks = min(num_chunks, num_q); in idpf_send_ena_dis_queues_msg()
2099 num_chunks = min(num_chunks, num_q); in idpf_send_map_unmap_queue_vector_msg()
2185 u16 num_chunks; in idpf_send_delete_queues_msg() local
2196 num_chunks = le16_to_cpu(chunks->num_chunks); in idpf_send_delete_queues_msg()
2204 eq->chunks.num_chunks = cpu_to_le16(num_chunks); in idpf_send_delete_queues_msg()
[all …]
A Dvirtchnl2.h597 __le16 num_chunks; member
599 struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks);
1202 __le16 num_chunks; member
1204 struct virtchnl2_queue_chunk chunks[] __counted_by_le(num_chunks);
/drivers/gpu/drm/xe/
A Dxe_guc_log.c83 snapshot->num_chunks = DIV_ROUND_UP(snapshot->size, GUC_LOG_CHUNK_SIZE); in xe_guc_log_snapshot_alloc()
85 snapshot->copy = kcalloc(snapshot->num_chunks, sizeof(*snapshot->copy), in xe_guc_log_snapshot_alloc()
91 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_alloc()
103 for (i = 0; i < snapshot->num_chunks; i++) in xe_guc_log_snapshot_alloc()
126 for (i = 0; i < snapshot->num_chunks; i++) in xe_guc_log_snapshot_free()
160 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_capture()
209 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_print()
212 char suffix = i == snapshot->num_chunks - 1 ? '\n' : 0; in xe_guc_log_snapshot_print()
A Dxe_hmm.c26 unsigned long num_chunks = 0; in xe_alloc_sg() local
51 num_chunks++; in xe_alloc_sg()
55 return sg_alloc_table(st, num_chunks, GFP_KERNEL); in xe_alloc_sg()
A Dxe_guc_log_types.h26 int num_chunks; member
/drivers/platform/x86/intel/ifs/
A Dload.c123 int i, num_chunks, chunk_size; in copy_hashes_authenticate_chunks() local
135 num_chunks = hashes_status.num_chunks; in copy_hashes_authenticate_chunks()
149 for (i = 0; i < num_chunks; i++) { in copy_hashes_authenticate_chunks()
171 return gen >= IFS_GEN_STRIDE_AWARE ? status.chunks_in_stride : status.num_chunks; in get_num_chunks()
187 int i, num_chunks, chunk_size; in copy_hashes_authenticate_chunks_gen2() local
206 num_chunks = get_num_chunks(ifsd->generation, hashes_status); in copy_hashes_authenticate_chunks_gen2()
215 num_chunks = ifsd->valid_chunks; in copy_hashes_authenticate_chunks_gen2()
234 for (i = 0; i < num_chunks; i++) { in copy_hashes_authenticate_chunks_gen2()
A Difs.h189 u32 num_chunks :8; member
202 u16 num_chunks; member
/drivers/crypto/intel/qat/qat_common/
A Dicp_qat_uclo.h260 unsigned short num_chunks; member
275 short num_chunks; member
520 unsigned short num_chunks; member
546 unsigned short num_chunks; member
562 unsigned short num_chunks; member
A Dqat_uclo.c117 if (suof_hdr->num_chunks <= 0x1) { in qat_uclo_check_suof_format()
458 for (i = 0; i < obj_hdr->num_chunks; i++) { in qat_uclo_find_chunk()
505 for (i = 0; i < file_hdr->num_chunks; i++) { in qat_uclo_map_chunk()
1208 suof_handle->img_table.num_simgs = suof_ptr->num_chunks - 1; in qat_uclo_map_suof()
1899 uobj_chunk_num = uobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof()
1901 sobj_chunk_num = sobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof()
1981 if (mof_hdr->num_chunks <= 0x1) { in qat_uclo_check_mof_format()
2024 chunks_num = mof_ptr->num_chunks; in qat_uclo_map_mof_obj()
/drivers/accel/habanalabs/common/
A Dcommand_submission.c1351 u32 cs_type_flags, num_chunks; in hl_cs_sanity_checks() local
1382 num_chunks = args->in.num_chunks_execute; in hl_cs_sanity_checks()
1393 if (!num_chunks) { in hl_cs_sanity_checks()
1409 void __user *chunks, u32 num_chunks, in hl_cs_copy_chunk_array() argument
1414 if (num_chunks > HL_MAX_JOBS_PER_CS) { in hl_cs_copy_chunk_array()
1543 for (i = 0 ; i < num_chunks ; i++) { in cs_ioctl_default()
1695 u32 num_chunks, tmp; in hl_cs_ctx_switch() local
1732 if (!num_chunks) { in hl_cs_ctx_switch()
1751 if (num_chunks) { in hl_cs_ctx_switch()
2199 void __user *chunks, u32 num_chunks, in cs_ioctl_signal_wait() argument
[all …]
/drivers/infiniband/sw/siw/
A Dsiw_mem.c341 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local
348 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get()
355 kcalloc(num_chunks, sizeof(struct siw_page_chunk), GFP_KERNEL); in siw_umem_get()
/drivers/gpu/drm/lima/
A Dlima_sched.c336 dt->num_chunks++; in lima_sched_build_error_task_list()
343 dt->num_chunks++; in lima_sched_build_error_task_list()
349 dt->num_chunks++; in lima_sched_build_error_task_list()
388 dt->num_chunks++; in lima_sched_build_error_task_list()
A Dlima_dump.h47 __u32 num_chunks; member
/drivers/gpu/drm/
A Ddrm_gpusvm.c441 const unsigned long *chunk_sizes, int num_chunks) in drm_gpusvm_init() argument
443 if (!ops->invalidate || !num_chunks) in drm_gpusvm_init()
455 gpusvm->num_chunks = num_chunks; in drm_gpusvm_init()
806 for (; i < gpusvm->num_chunks; ++i) { in drm_gpusvm_range_chunk_size()
817 if (i == gpusvm->num_chunks) in drm_gpusvm_range_chunk_size()
/drivers/gpu/drm/radeon/
A Dradeon_cs.c277 if (!cs->num_chunks) { in radeon_cs_parser_init()
289 p->chunks_array = kvmalloc_array(cs->num_chunks, sizeof(uint64_t), GFP_KERNEL); in radeon_cs_parser_init()
295 sizeof(uint64_t)*cs->num_chunks)) { in radeon_cs_parser_init()
299 p->nchunks = cs->num_chunks; in radeon_cs_parser_init()
/drivers/crypto/marvell/octeontx/
A Dotx_cptvf.h37 u32 num_chunks; /* Number of command chunks */ member
A Dotx_cptvf_main.c179 queue->num_chunks = 0; in free_command_queues()
226 i, queue->num_chunks); in alloc_command_queues()
231 if (queue->num_chunks == 0) { in alloc_command_queues()
238 queue->num_chunks++; in alloc_command_queues()
/drivers/net/dsa/sja1105/
A Dsja1105_spi.c44 int num_chunks; in sja1105_xfer() local
47 num_chunks = DIV_ROUND_UP(len, priv->max_xfer_len); in sja1105_xfer()
56 for (i = 0; i < num_chunks; i++) { in sja1105_xfer()
/drivers/vfio/pci/mlx5/
A Dmain.c385 int num_chunks; in mlx5vf_prep_stop_copy() local
408 num_chunks = mvdev->chunk_mode ? MAX_NUM_CHUNKS : 1; in mlx5vf_prep_stop_copy()
409 for (i = 0; i < num_chunks; i++) { in mlx5vf_prep_stop_copy()
444 for (i = 0; i < num_chunks; i++) { in mlx5vf_prep_stop_copy()
/drivers/mtd/nand/raw/
A Dmxc_nand.c254 u16 num_chunks = mtd->writesize / 512; in copy_spare() local
261 oob_chunk_size = (host->used_oobsize / num_chunks) & ~1; in copy_spare()
264 for (i = 0; i < num_chunks - 1; i++) in copy_spare()
274 for (i = 0; i < num_chunks - 1; i++) in copy_spare()
/drivers/gpu/drm/amd/amdgpu/
A Damdgpu_cs.c51 if (cs->in.num_chunks == 0) in amdgpu_cs_parser_init()
188 chunk_array = kvmalloc_array(cs->in.num_chunks, sizeof(uint64_t), in amdgpu_cs_pass1()
196 sizeof(uint64_t)*cs->in.num_chunks)) { in amdgpu_cs_pass1()
201 p->nchunks = cs->in.num_chunks; in amdgpu_cs_pass1()
/drivers/media/i2c/
A Dthp7312.c1669 unsigned int num_chunks; in thp7312_fw_load_to_ram() local
1677 num_chunks = DIV_ROUND_UP(bank_size, THP7312_FW_DOWNLOAD_UNIT); in thp7312_fw_load_to_ram()
1680 __func__, bank_size, i, num_chunks); in thp7312_fw_load_to_ram()
1682 for (j = 0 ; j < num_chunks; j++) { in thp7312_fw_load_to_ram()
/drivers/net/wireless/intel/ipw2x00/
A Dipw2200.c3765 if (le32_to_cpu(bd->u.data.num_chunks) > NUM_TFD_CHUNKS) { in ipw_queue_tx_free_tfd()
3767 le32_to_cpu(bd->u.data.num_chunks)); in ipw_queue_tx_free_tfd()
3773 for (i = 0; i < le32_to_cpu(bd->u.data.num_chunks); i++) { in ipw_queue_tx_free_tfd()
10168 tfd->u.data.num_chunks = cpu_to_le32(min((u8) (NUM_TFD_CHUNKS - 2), in ipw_tx_skb()
10171 txb->nr_frags, le32_to_cpu(tfd->u.data.num_chunks)); in ipw_tx_skb()
10172 for (i = 0; i < le32_to_cpu(tfd->u.data.num_chunks); i++) { in ipw_tx_skb()
10174 i, le32_to_cpu(tfd->u.data.num_chunks), in ipw_tx_skb()
10177 i, tfd->u.data.num_chunks, in ipw_tx_skb()
10221 le32_add_cpu(&tfd->u.data.num_chunks, 1); in ipw_tx_skb()

Completed in 92 milliseconds

12