/linux-6.3-rc2/drivers/gpu/drm/amd/amdkfd/ |
A D | kfd_kernel_queue.c | 42 enum kfd_queue_type type, unsigned int queue_size) in kq_initialize() argument 52 queue_size); in kq_initialize() 85 retval = kfd_gtt_sa_allocate(dev, queue_size, &kq->pq); in kq_initialize() 87 pr_err("Failed to init pq queues size %d\n", queue_size); in kq_initialize() 124 memset(kq->pq_kernel_addr, 0, queue_size); in kq_initialize() 128 prop.queue_size = queue_size; in kq_initialize() 246 queue_size_dwords = kq->queue->properties.queue_size / 4; in kq_acquire_packet_buffer() 317 (kq->queue->properties.queue_size / 4); in kq_rollback_packet()
|
A D | kfd_queue.c | 35 pr_debug("Queue Size: %llu\n", q->queue_size); in print_queue_properties() 52 pr_debug("Queue Size: %llu\n", q->properties.queue_size); in print_queue()
|
A D | kfd_mqd_manager_cik.c | 166 uint32_t wptr_mask = (uint32_t)((p->queue_size / 4) - 1); in load_mqd() 192 m->cp_hqd_pq_control |= order_base_2(q->queue_size / 4) - 1; in __update_mqd() 238 m->sdma_rlc_rb_cntl = order_base_2(q->queue_size / 4) in update_mqd_sdma() 359 m->cp_hqd_pq_control |= order_base_2(q->queue_size / 4) - 1; in update_mqd_hiq()
|
/linux-6.3-rc2/drivers/net/ethernet/microsoft/mana/ |
A D | hw_channel.c | 239 enum gdma_queue_type type, u64 queue_size, in mana_hwc_create_gdma_wq() argument 249 spec.queue_size = queue_size; in mana_hwc_create_gdma_wq() 255 u64 queue_size, in mana_hwc_create_gdma_cq() argument 264 spec.queue_size = queue_size; in mana_hwc_create_gdma_cq() 273 u64 queue_size, in mana_hwc_create_gdma_eq() argument 281 spec.queue_size = queue_size; in mana_hwc_create_gdma_eq() 470 u32 queue_size; in mana_hwc_create_wq() local 476 queue_size = roundup_pow_of_two(GDMA_MAX_RQE_SIZE * q_depth); in mana_hwc_create_wq() 478 queue_size = roundup_pow_of_two(GDMA_MAX_SQE_SIZE * q_depth); in mana_hwc_create_wq() 480 if (queue_size < MINIMUM_SUPPORTED_PAGE_SIZE) in mana_hwc_create_wq() [all …]
|
A D | gdma_main.c | 202 req.queue_size = queue->queue_size; in mana_gd_create_hw_eq() 311 u32 num_cqe = cq->queue_size / GDMA_CQE_SIZE; in mana_gd_ring_cq() 321 u32 head = eq->head % (eq->queue_size / GDMA_EQE_SIZE); in mana_gd_process_eqe() 384 num_eqe = eq->queue_size / GDMA_EQE_SIZE; in mana_gd_process_eq_events() 645 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi); in mana_gd_create_hwc_queue() 652 queue->queue_size = spec->queue_size; in mana_gd_create_hwc_queue() 783 queue->queue_size = spec->queue_size; in mana_gd_create_mana_eq() 830 queue->queue_size = spec->queue_size; in mana_gd_create_mana_wq_cq() 977 u32 wq_size = wq->queue_size; in mana_gd_wq_avail_space() 1045 end_ptr = base_ptr + wq->queue_size; in mana_gd_write_sgl() [all …]
|
A D | mana_en.c | 812 req.wq_size = wq_spec->queue_size; in mana_create_wq_obj() 813 req.cq_size = cq_spec->queue_size; in mana_create_wq_obj() 914 spec.queue_size = EQ_SIZE; in mana_create_eq() 1507 spec.queue_size = txq_size; in mana_create_txq() 1521 spec.queue_size = cq_size; in mana_create_txq() 1533 wq_spec.queue_size = txq->gdma_sq->queue_size; in mana_create_txq() 1536 cq_spec.queue_size = cq->gdma_cq->queue_size; in mana_create_txq() 1741 spec.queue_size = rq_size; in mana_create_rxq() 1754 spec.queue_size = cq_size; in mana_create_rxq() 1765 wq_spec.queue_size = rxq->gdma_rq->queue_size; in mana_create_rxq() [all …]
|
/linux-6.3-rc2/drivers/firmware/tegra/ |
A D | ivc.c | 562 unsigned tegra_ivc_total_queue_size(unsigned queue_size) in tegra_ivc_total_queue_size() argument 564 if (!IS_ALIGNED(queue_size, TEGRA_IVC_ALIGN)) { in tegra_ivc_total_queue_size() 566 __func__, queue_size, TEGRA_IVC_ALIGN); in tegra_ivc_total_queue_size() 570 return queue_size + sizeof(struct tegra_ivc_header); in tegra_ivc_total_queue_size() 652 size_t queue_size; in tegra_ivc_init() local 670 queue_size = tegra_ivc_total_queue_size(num_frames * frame_size); in tegra_ivc_init() 673 ivc->rx.phys = dma_map_single(peer, iosys_map_get_vaddr(rx), queue_size, in tegra_ivc_init() 678 ivc->tx.phys = dma_map_single(peer, iosys_map_get_vaddr(tx), queue_size, in tegra_ivc_init() 681 dma_unmap_single(peer, ivc->rx.phys, queue_size, in tegra_ivc_init()
|
A D | bpmp-tegra186.c | 107 size_t message_size, queue_size; in tegra186_bpmp_channel_init() local 118 queue_size = tegra_ivc_total_queue_size(message_size); in tegra186_bpmp_channel_init() 119 offset = queue_size * index; in tegra186_bpmp_channel_init()
|
/linux-6.3-rc2/sound/firewire/ |
A D | amdtp-stream.c | 643 if (++s->packet_index >= s->queue_size) in queue_packet() 899 unsigned int queue_size) in compute_ohci_it_cycle() argument 902 return increment_ohci_cycle_count(cycle, queue_size); in compute_ohci_it_cycle() 912 unsigned int queue_size = s->queue_size; in generate_tx_packet_descs() local 981 packet_index = (packet_index + 1) % queue_size; in generate_tx_packet_descs() 1250 const unsigned int queue_size = s->queue_size; in process_rx_packets_intermediately() local 1696 s->queue_size = queue_size; in amdtp_stream_start() 1720 queue_size * 3 / 2); in amdtp_stream_start() 1748 s->ctx_data.rx.seq.size = queue_size; in amdtp_stream_start() 1776 for (i = 0; i < s->queue_size; ++i) { in amdtp_stream_start() [all …]
|
/linux-6.3-rc2/kernel/bpf/ |
A D | queue_stack_maps.c | 72 u64 size, queue_size; in queue_stack_map_alloc() local 75 queue_size = sizeof(*qs) + size * attr->value_size; in queue_stack_map_alloc() 77 qs = bpf_map_area_alloc(queue_size, numa_node); in queue_stack_map_alloc()
|
/linux-6.3-rc2/drivers/net/wireless/intel/iwlwifi/pcie/ |
A D | rx.c | 122 WARN_ON(rxq->queue_size & (rxq->queue_size - 1)); in iwl_rxq_space() 685 free_size * rxq->queue_size, in iwl_pcie_free_rxq_dma() 696 rxq->queue_size, in iwl_pcie_free_rxq_dma() 716 rxq->queue_size = trans->cfg->num_rbds; in iwl_pcie_alloc_rxq_dma() 718 rxq->queue_size = RX_QUEUE_SIZE; in iwl_pcie_alloc_rxq_dma() 734 rxq->queue_size, in iwl_pcie_alloc_rxq_dma() 1120 queue_size = trans->trans_cfg->mq_rx_supported ? in _iwl_pcie_rx_init() 1124 num_alloc = queue_size + allocator_pool_size; in _iwl_pcie_rx_init() 1497 r &= (rxq->queue_size - 1); in iwl_pcie_rx_handle() 1547 i = (i + 1) & (rxq->queue_size - 1); in iwl_pcie_rx_handle() [all …]
|
/linux-6.3-rc2/include/linux/ |
A D | vmw_vmci_defs.h | 866 u64 queue_size) in vmci_q_header_add_producer_tail() argument 868 vmci_qp_add_pointer(&q_header->producer_tail, add, queue_size); in vmci_q_header_add_producer_tail() 878 u64 queue_size) in vmci_q_header_add_consumer_head() argument 880 vmci_qp_add_pointer(&q_header->consumer_head, add, queue_size); in vmci_q_header_add_consumer_head()
|
/linux-6.3-rc2/drivers/nvme/host/ |
A D | rdma.c | 88 int queue_size; member 269 init_attr.cap.max_recv_wr = queue->queue_size + 1; in nvme_rdma_create_qp() 534 queue->queue_size, in nvme_rdma_create_queue_ib() 540 queue->queue_size, nvme_rdma_queue_idx(queue)); in nvme_rdma_create_queue_ib() 546 queue->queue_size, IB_MR_TYPE_INTEGRITY, in nvme_rdma_create_queue_ib() 551 queue->queue_size, nvme_rdma_queue_idx(queue)); in nvme_rdma_create_queue_ib() 575 int idx, size_t queue_size) in nvme_rdma_alloc_queue() argument 595 queue->queue_size = queue_size; in nvme_rdma_alloc_queue() 1801 for (i = 0; i < queue->queue_size; i++) { in nvme_rdma_conn_established() 1893 priv.hrqsize = cpu_to_le16(queue->queue_size); in nvme_rdma_route_resolved() [all …]
|
/linux-6.3-rc2/sound/core/ |
A D | timer.c | 77 int queue_size; member 1337 if (tu->qused >= tu->queue_size) { in snd_timer_user_interrupt() 1341 tu->qtail %= tu->queue_size; in snd_timer_user_interrupt() 1355 if (tu->qused >= tu->queue_size) { in snd_timer_user_append_to_tqueue() 1359 tu->qtail %= tu->queue_size; in snd_timer_user_append_to_tqueue() 1478 tu->queue_size = size; in realloc_user_queue() 1859 if (params.queue_size > 0 && in snd_timer_user_params() 1860 (params.queue_size < 32 || params.queue_size > 1024)) { in snd_timer_user_params() 1893 if (params.queue_size > 0 && in snd_timer_user_params() 1894 (unsigned int)tu->queue_size != params.queue_size) { in snd_timer_user_params() [all …]
|
/linux-6.3-rc2/drivers/usb/host/ |
A D | u132-hcd.c | 162 u16 queue_size; member 1912 endp->queue_size = 1; in create_endpoint_and_queue_int() 1935 endp->queue_size -= 1; in queue_int_on_old_endpoint() 2011 endp->queue_size = 1; in create_endpoint_and_queue_bulk() 2032 endp->queue_size -= 1; in queue_bulk_on_old_endpoint() 2100 endp->queue_size = 1; in create_endpoint_and_queue_control() 2121 endp->queue_size = 1; in create_endpoint_and_queue_control() 2220 endp->queue_size -= 1; in queue_control_on_old_endpoint() 2391 endp->queue_size -= 1; in dequeue_from_overflow_chain() 2441 u16 queue_size = endp->queue_size; in u132_endp_urb_dequeue() local [all …]
|
/linux-6.3-rc2/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_amdkfd_gfx_v11.c | 222 uint32_t queue_size = in hqd_load_v11() local 225 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in hqd_load_v11() 227 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in hqd_load_v11() 228 guessed_wptr += queue_size; in hqd_load_v11() 229 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in hqd_load_v11()
|
A D | amdgpu_amdkfd_gfx_v10.c | 250 uint32_t queue_size = in kgd_hqd_load() local 253 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in kgd_hqd_load() 255 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in kgd_hqd_load() 256 guessed_wptr += queue_size; in kgd_hqd_load() 257 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in kgd_hqd_load()
|
A D | amdgpu_amdkfd_gfx_v9.c | 264 uint32_t queue_size = in kgd_gfx_v9_hqd_load() local 267 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in kgd_gfx_v9_hqd_load() 269 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in kgd_gfx_v9_hqd_load() 270 guessed_wptr += queue_size; in kgd_gfx_v9_hqd_load() 271 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in kgd_gfx_v9_hqd_load()
|
A D | amdgpu_amdkfd_gfx_v10_3.c | 237 uint32_t queue_size = in hqd_load_v10_3() local 240 uint64_t guessed_wptr = m->cp_hqd_pq_rptr & (queue_size - 1); in hqd_load_v10_3() 242 if ((m->cp_hqd_pq_wptr_lo & (queue_size - 1)) < guessed_wptr) in hqd_load_v10_3() 243 guessed_wptr += queue_size; in hqd_load_v10_3() 244 guessed_wptr += m->cp_hqd_pq_wptr_lo & ~(queue_size - 1); in hqd_load_v10_3()
|
/linux-6.3-rc2/drivers/misc/genwqe/ |
A D | card_ddcb.c | 1026 unsigned int queue_size; in setup_ddcb_queue() local 1032 queue_size = roundup(GENWQE_DDCB_MAX * sizeof(struct ddcb), PAGE_SIZE); in setup_ddcb_queue() 1042 queue->ddcb_vaddr = __genwqe_alloc_consistent(cd, queue_size, in setup_ddcb_queue() 1090 __genwqe_free_consistent(cd, queue_size, queue->ddcb_vaddr, in setup_ddcb_queue() 1105 unsigned int queue_size; in free_ddcb_queue() local 1107 queue_size = roundup(queue->ddcb_max * sizeof(struct ddcb), PAGE_SIZE); in free_ddcb_queue() 1113 __genwqe_free_consistent(cd, queue_size, queue->ddcb_vaddr, in free_ddcb_queue()
|
/linux-6.3-rc2/drivers/infiniband/hw/mana/ |
A D | qp.c | 191 wq_spec.queue_size = wq->wq_buf_size; in mana_ib_create_qp_rss() 194 cq_spec.queue_size = cq->cqe * COMP_ENTRY_SIZE; in mana_ib_create_qp_rss() 348 wq_spec.queue_size = ucmd.sq_buf_size; in mana_ib_create_qp_raw() 351 cq_spec.queue_size = send_cq->cqe * COMP_ENTRY_SIZE; in mana_ib_create_qp_raw()
|
/linux-6.3-rc2/drivers/platform/chrome/wilco_ec/ |
A D | event.c | 103 static int queue_size = 64; variable 104 module_param(queue_size, int, 0644); 470 dev_data->events = event_queue_new(queue_size); in event_device_add()
|
/linux-6.3-rc2/include/soc/tegra/ |
A D | ivc.h | 95 unsigned tegra_ivc_total_queue_size(unsigned queue_size);
|
/linux-6.3-rc2/drivers/i2c/busses/ |
A D | i2c-qcom-cci.c | 122 u16 queue_size[NUM_QUEUES]; member 323 if (val == cci->data->queue_size[queue]) in cci_validate_queue() 700 .queue_size = { 64, 16 }, 734 .queue_size = { 64, 16 }, 768 .queue_size = { 64, 16 },
|
/linux-6.3-rc2/drivers/nvme/target/ |
A D | loop.c | 571 if (opts->queue_size > ctrl->ctrl.maxcmd) { in nvme_loop_create_ctrl() 575 opts->queue_size, ctrl->ctrl.maxcmd); in nvme_loop_create_ctrl() 576 opts->queue_size = ctrl->ctrl.maxcmd; in nvme_loop_create_ctrl() 578 ctrl->ctrl.sqsize = opts->queue_size - 1; in nvme_loop_create_ctrl()
|