Home
last modified time | relevance | path

Searched refs:q_id (Results 1 – 25 of 68) sorted by relevance

123

/linux-6.3-rc2/drivers/net/ethernet/huawei/hinic/
A Dhinic_hw_eqs.c557 ceq_ctrl.q_id = eq->q_id; in set_ceq_ctrl_reg()
731 eq->q_id = q_id; in init_eq()
874 for (q_id = 0; q_id < num_aeqs; q_id++) { in hinic_aeqs_init()
899 int q_id; in hinic_aeqs_free() local
901 for (q_id = 0; q_id < aeqs->num_aeqs ; q_id++) in hinic_aeqs_free()
928 for (q_id = 0; q_id < num_ceqs; q_id++) { in hinic_ceqs_init()
953 int q_id; in hinic_ceqs_free() local
955 for (q_id = 0; q_id < ceqs->num_ceqs; q_id++) in hinic_ceqs_free()
963 int q_id; in hinic_dump_ceq_info() local
965 for (q_id = 0; q_id < hwdev->func_to_io.ceqs.num_ceqs; q_id++) { in hinic_dump_ceq_info()
[all …]
A Dhinic_hw_csr.h87 #define HINIC_CSR_AEQ_HI_PHYS_ADDR_REG(q_id, pg_num) \ argument
88 (HINIC_CSR_AEQ_MTT_OFF(q_id) + \
91 #define HINIC_CSR_CEQ_HI_PHYS_ADDR_REG(q_id, pg_num) \ argument
92 (HINIC_CSR_CEQ_MTT_OFF(q_id) + \
95 #define HINIC_CSR_AEQ_LO_PHYS_ADDR_REG(q_id, pg_num) \ argument
96 (HINIC_CSR_AEQ_MTT_OFF(q_id) + \
99 #define HINIC_CSR_CEQ_LO_PHYS_ADDR_REG(q_id, pg_num) \ argument
100 (HINIC_CSR_CEQ_MTT_OFF(q_id) + \
A Dhinic_hw_io.c30 #define CI_ADDR(base_addr, q_id) ((base_addr) + \ argument
31 (q_id) * CI_Q_ADDR_SIZE)
132 base_qpn + qp->q_id); in write_sq_ctxts()
176 base_qpn + qp->q_id); in write_rq_ctxts()
272 struct hinic_qp *qp, int q_id, in init_qp() argument
281 qp->q_id = q_id; in init_qp()
306 func_to_io->sq_db[q_id] = db_base; in init_qp()
308 qp->sq.qid = q_id; in init_qp()
311 CI_ADDR(func_to_io->ci_addr_base, q_id), in init_qp()
318 qp->rq.qid = q_id; in init_qp()
[all …]
A Dhinic_tx.c494 u16 prod_idx, q_id = skb->queue_mapping; in hinic_lb_xmit_frame() local
502 txq = &nic_dev->txqs[q_id]; in hinic_lb_xmit_frame()
514 netif_stop_subqueue(netdev, qp->q_id); in hinic_lb_xmit_frame()
518 netif_wake_subqueue(nic_dev->netdev, qp->q_id); in hinic_lb_xmit_frame()
537 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_lb_xmit_frame()
555 u16 prod_idx, q_id = skb->queue_mapping; in hinic_xmit_frame() local
563 txq = &nic_dev->txqs[q_id]; in hinic_xmit_frame()
595 netif_stop_subqueue(netdev, qp->q_id); in hinic_xmit_frame()
602 netif_wake_subqueue(nic_dev->netdev, qp->q_id); in hinic_xmit_frame()
626 netdev_txq = netdev_get_tx_queue(netdev, q_id); in hinic_xmit_frame()
[all …]
A Dhinic_rx.c421 skb_record_rx_queue(skb, qp->q_id); in rxq_recv()
532 intr_coal = &nic_dev->rx_intr_coalesce[qp->q_id]; in rx_request_irq()
549 cpumask_set_cpu(qp->q_id % num_online_cpus(), &rq->affinity_mask); in rx_request_irq()
594 "%s_rxq%d", netdev->name, qp->q_id); in hinic_init_rxq()
A Dhinic_main.c818 u16 num_sqs, q_id; in hinic_tx_timeout() local
824 for (q_id = 0; q_id < num_sqs; q_id++) { in hinic_tx_timeout()
825 if (!netif_xmit_stopped(netdev_get_tx_queue(netdev, q_id))) in hinic_tx_timeout()
828 sq = hinic_hwdev_get_sq(nic_dev->hwdev, q_id); in hinic_tx_timeout()
833 q_id, sw_pi, hw_ci, sw_ci, in hinic_tx_timeout()
834 nic_dev->txqs[q_id].napi.state); in hinic_tx_timeout()
A Dhinic_hw_qp.c42 #define SQ_CTXT_OFFSET(max_sqs, max_rqs, q_id) \ argument
43 (((max_rqs) + (max_sqs)) * CTXT_RSVD + (q_id) * Q_CTXT_SIZE)
45 #define RQ_CTXT_OFFSET(max_sqs, max_rqs, q_id) \ argument
47 (max_sqs + (q_id)) * Q_CTXT_SIZE)
625 HINIC_SQ_DB_INFO_SET(qp->q_id, QID)); in sq_prepare_db()
A Dhinic_hw_cmdq.c508 enum hinic_set_arm_qtype q_type, u32 q_id) in hinic_set_arm_bit() argument
517 arm_bit.q_id = q_id; in hinic_set_arm_bit()
521 dev_err(&pdev->dev, "Failed to set arm for qid %d\n", q_id); in hinic_set_arm_bit()
A Dhinic_ethtool.c677 static int set_queue_coalesce(struct hinic_dev *nic_dev, u16 q_id, in set_queue_coalesce() argument
687 intr_coal = set_rx_coal ? &nic_dev->rx_intr_coalesce[q_id] : in set_queue_coalesce()
688 &nic_dev->tx_intr_coalesce[q_id]; in set_queue_coalesce()
697 q_id >= nic_dev->num_qps) in set_queue_coalesce()
700 msix_idx = set_rx_coal ? nic_dev->rxqs[q_id].rq->msix_entry : in set_queue_coalesce()
701 nic_dev->txqs[q_id].sq->msix_entry; in set_queue_coalesce()
711 set_rx_coal ? "rx" : "tx", q_id); in set_queue_coalesce()
/linux-6.3-rc2/tools/cgroup/
A Diocost_monitor.py64 def __init__(self, root_blkcg, q_id, include_dying=False): argument
67 self.walk(root_blkcg, q_id, '')
72 def walk(self, blkcg, q_id, parent_path): argument
80 address=radix_tree_lookup(blkcg.blkg_tree.address_of_(), q_id))
88 self.walk(c, q_id, path)
217 q_id = None variable
225 q_id = blkg.q.id.value_() variable
251 for path, blkg in BlkgIterator(blkcg_root, q_id):
/linux-6.3-rc2/drivers/block/
A Dublk_drv.c119 int q_id; member
338 return ublk_get_queue(ub, q_id)->io_cmd_buf; in ublk_queue_cmd_buf()
783 __func__, io->cmd->cmd_op, ubq->q_id, in __ublk_rq_task_work()
986 int q_id, ret = 0; in ublk_ch_mmap() local
1010 if (sz != ublk_queue_cmd_buf_size(ub, q_id)) in ublk_ch_mmap()
1020 u32 qid = ub_cmd->q_id, tag = ub_cmd->tag; in ublk_commit_completion()
1271 __func__, cmd->cmd_op, ub_cmd->q_id, tag, in ublk_ch_uring_cmd()
1277 ubq = ublk_get_queue(ub, ub_cmd->q_id); in ublk_ch_uring_cmd()
1278 if (!ubq || ub_cmd->q_id != ubq->q_id) in ublk_ch_uring_cmd()
1388 ubq->q_id = q_id; in ublk_init_queue()
[all …]
/linux-6.3-rc2/fs/xfs/scrub/
A Dquota.c94 offset = dq->q_id / qi->qi_dqperchunk; in xchk_quota_item()
95 if (dq->q_id && dq->q_id <= sqi->last_id) in xchk_quota_item()
98 sqi->last_id = dq->q_id; in xchk_quota_item()
148 if (dq->q_id == 0) in xchk_quota_item()
/linux-6.3-rc2/drivers/net/ethernet/intel/ice/
A Dice_base.c713 u16 q_id, q_base; in ice_vsi_map_rings_to_vectors() local
723 for (q_id = q_base; q_id < (q_base + tx_rings_per_v); q_id++) { in ice_vsi_map_rings_to_vectors()
724 struct ice_tx_ring *tx_ring = vsi->tx_rings[q_id]; in ice_vsi_map_rings_to_vectors()
740 for (q_id = q_base; q_id < (q_base + rx_rings_per_v); q_id++) { in ice_vsi_map_rings_to_vectors()
741 struct ice_rx_ring *rx_ring = vsi->rx_rings[q_id]; in ice_vsi_map_rings_to_vectors()
969 &txq_meta->q_id, &txq_meta->q_teid, rst_src, in ice_vsi_stop_tx_ring()
1011 txq_meta->q_id = ring->reg_idx; in ice_fill_txq_meta()
A Dice_eswitch.c43 rule_info.sw_act.fwd_id.q_id = hw->func_caps.common_cap.rxq_first_id + in ice_eswitch_add_vf_mac_rule()
176 int q_id; in ice_eswitch_remap_rings_to_vectors() local
178 ice_for_each_txq(vsi, q_id) { in ice_eswitch_remap_rings_to_vectors()
185 vf = ice_get_vf_by_id(pf, q_id); in ice_eswitch_remap_rings_to_vectors()
191 tx_ring = vsi->tx_rings[q_id]; in ice_eswitch_remap_rings_to_vectors()
192 rx_ring = vsi->rx_rings[q_id]; in ice_eswitch_remap_rings_to_vectors()
A Dice_idc.c119 u16 q_id; in ice_del_rdma_qset() local
130 q_id = qset->qs_handle; in ice_del_rdma_qset()
135 return ice_dis_vsi_rdma_qset(vsi->port_info, 1, &teid, &q_id); in ice_del_rdma_qset()
A Dice_switch.h108 u16 q_id:11; member
160 u16 q_id:11; member
/linux-6.3-rc2/fs/xfs/
A Dxfs_dquot.c77 ASSERT(dq->q_id); in xfs_qm_adjust_dqlimits()
164 ASSERT(dq->q_id); in xfs_qm_adjust_dqtimers()
478 dqp->q_id = id; in xfs_dquot_alloc()
535 if (be32_to_cpu(ddqp->d_id) != dqp->q_id) in xfs_dquot_check_type()
574 __this_address, dqp->q_id); in xfs_dquot_from_disk()
618 ddqp->d_id = cpu_to_be32(dqp->q_id); in xfs_dquot_to_disk()
1178 if (dqp->q_id == 0) in xfs_qm_dqflush_check()
1197 if (dqp->q_id == 0) in xfs_qm_dqflush_check()
1247 dqp->q_id, fa); in xfs_qm_dqflush()
1322 if (d1->q_id > d2->q_id) { in xfs_dqlock2()
[all …]
A Dxfs_trans_dquot.c60 if (dqp->q_id != 0 && in xfs_trans_log_dquot()
424 if (dqp->q_id) { in xfs_trans_apply_dquot_deltas()
556 quota_send_warning(make_kqid(&init_user_ns, qtype, dqp->q_id), in xfs_quota_warn()
643 if ((flags & XFS_QMOPT_FORCE_RES) == 0 && dqp->q_id && in xfs_trans_dqresv()
A Dxfs_qm.c85 next_index = dqp->q_id + 1; in xfs_qm_dquot_walk()
166 radix_tree_delete(xfs_dquot_tree(qi, xfs_dquot_type(dqp)), dqp->q_id); in xfs_qm_dqpurge()
1117 if (dqp->q_id) { in xfs_qm_quotacheck_dqadjust()
1624 radix_tree_delete(xfs_dquot_tree(qi, xfs_dquot_type(dqp)), dqp->q_id); in xfs_qm_dqfree_one()
1887 ASSERT(i_uid_read(VFS_I(ip)) == udqp->q_id); in xfs_qm_vop_create_dqattach()
1894 ASSERT(i_gid_read(VFS_I(ip)) == gdqp->q_id); in xfs_qm_vop_create_dqattach()
1901 ASSERT(ip->i_projid == pdqp->q_id); in xfs_qm_vop_create_dqattach()
/linux-6.3-rc2/drivers/net/wireless/intel/iwlwifi/pcie/
A Dtx.c87 txq_id != trans->txqs.cmd.q_id && in iwl_pcie_txq_inc_wr_ptr()
217 if (txq_id != trans->txqs.cmd.q_id) { in iwl_pcie_txq_unmap()
229 txq_id == trans->txqs.cmd.q_id) in iwl_pcie_txq_unmap()
265 if (txq_id == trans->txqs.cmd.q_id) in iwl_pcie_txq_free()
328 iwl_trans_ac_txq_enable(trans, trans->txqs.cmd.q_id, in iwl_pcie_tx_start()
536 bool cmd_queue = (txq_id == trans->txqs.cmd.q_id); in iwl_pcie_tx_alloc()
752 if (txq_id == trans->txqs.cmd.q_id && in iwl_trans_pcie_txq_enable()
760 if (txq_id != trans->txqs.cmd.q_id) in iwl_trans_pcie_txq_enable()
830 if (txq_id == trans->txqs.cmd.q_id && in iwl_trans_pcie_txq_enable()
1030 cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) | in iwl_pcie_enqueue_hcmd()
[all …]
A Dtx-gen2.c31 struct iwl_txq *txq = trans->txqs.txq[trans->txqs.cmd.q_id]; in iwl_pcie_gen2_enqueue_hcmd()
143 cpu_to_le16(QUEUE_TO_SEQ(trans->txqs.cmd.q_id) | in iwl_pcie_gen2_enqueue_hcmd()
191 cmd_size, txq->write_ptr, idx, trans->txqs.cmd.q_id); in iwl_pcie_gen2_enqueue_hcmd()
/linux-6.3-rc2/drivers/crypto/hisilicon/sec/
A Dsec_drv.c688 int q_id; in sec_isr_handle() local
693 q_id = outorder_msg->data & SEC_OUT_BD_INFO_Q_ID_M; in sec_isr_handle()
694 msg = msg_ring->vaddr + q_id; in sec_isr_handle()
701 set_bit(q_id, queue->unprocessed); in sec_isr_handle()
702 if (q_id == queue->expected) in sec_isr_handle()
719 q_id = outorder_msg->data & SEC_OUT_BD_INFO_Q_ID_M; in sec_isr_handle()
720 msg = msg_ring->vaddr + q_id; in sec_isr_handle()
/linux-6.3-rc2/drivers/net/ethernet/hisilicon/hns3/hns3pf/
A Dhclge_tm.h257 int hclge_tm_get_q_to_qs_map(struct hclge_dev *hdev, u16 q_id, u16 *qset_id);
258 int hclge_tm_get_q_to_tc(struct hclge_dev *hdev, u16 q_id, u8 *tc_id);
/linux-6.3-rc2/drivers/gpu/drm/amd/amdkfd/
A Dkfd_process_queue_manager.c267 retval = assign_queue_slot_by_qid(pqm, q_data->q_id); in pqm_create_queue()
268 *qid = q_data->q_id; in pqm_create_queue()
660 q_data->q_id = q->properties.queue_id; in criu_checkpoint_queue()
690 pr_debug("Dumping Queue: gpu_id:%x queue_id:%u\n", q_data->gpu_id, q_data->q_id); in criu_checkpoint_queue()
884 ret = pqm_set_gws(&p->pqm, q_data->q_id, pdd->dev->gws); in kfd_criu_restore_queue()
/linux-6.3-rc2/include/uapi/linux/
A Dublk_cmd.h205 __u16 q_id; member

Completed in 53 milliseconds

123