Searched refs:cq (Results 1 – 11 of 11) sorted by relevance
| /include/linux/mlx5/ |
| A D | cq.h | 142 static inline void mlx5_cq_set_ci(struct mlx5_core_cq *cq) in mlx5_cq_set_ci() argument 144 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx5_cq_set_ci() 160 sn = cq->arm_sn & 3; in mlx5_cq_arm() 163 *cq->arm_db = cpu_to_be32(sn << 28 | cmd | ci); in mlx5_cq_arm() 171 doorbell[1] = cpu_to_be32(cq->cqn); in mlx5_cq_arm() 176 static inline void mlx5_cq_hold(struct mlx5_core_cq *cq) in mlx5_cq_hold() argument 178 refcount_inc(&cq->refcount); in mlx5_cq_hold() 181 static inline void mlx5_cq_put(struct mlx5_core_cq *cq) in mlx5_cq_put() argument 183 if (refcount_dec_and_test(&cq->refcount)) in mlx5_cq_put() 184 complete(&cq->free); in mlx5_cq_put() [all …]
|
| /include/trace/events/ |
| A D | rdma_core.h | 51 struct ib_cq *cq 54 TP_ARGS(cq), 62 cq->interrupt = true; 72 struct ib_cq *cq 75 TP_ARGS(cq), 83 cq->interrupt = false; 93 const struct ib_cq *cq 96 TP_ARGS(cq), 146 const struct ib_cq *cq 149 TP_ARGS(cq), [all …]
|
| /include/linux/mlx4/ |
| A D | cq.h | 147 static inline void mlx4_cq_arm(struct mlx4_cq *cq, u32 cmd, in mlx4_cq_arm() argument 155 sn = cq->arm_sn & 3; in mlx4_cq_arm() 156 ci = cq->cons_index & 0xffffff; in mlx4_cq_arm() 158 *cq->arm_db = cpu_to_be32(sn << 28 | cmd | ci); in mlx4_cq_arm() 166 doorbell[0] = cpu_to_be32(sn << 28 | cmd | cq->cqn); in mlx4_cq_arm() 172 static inline void mlx4_cq_set_ci(struct mlx4_cq *cq) in mlx4_cq_set_ci() argument 174 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx4_cq_set_ci() 182 int mlx4_cq_modify(struct mlx4_dev *dev, struct mlx4_cq *cq, 184 int mlx4_cq_resize(struct mlx4_dev *dev, struct mlx4_cq *cq,
|
| A D | device.h | 832 int cq; member 1147 struct mlx4_uar *uar, u64 db_rec, struct mlx4_cq *cq, 1150 void mlx4_cq_free(struct mlx4_dev *dev, struct mlx4_cq *cq);
|
| /include/rdma/ |
| A D | rdmavt_cq.h | 65 bool rvt_cq_enter(struct rvt_cq *cq, struct ib_wc *entry, bool solicited);
|
| A D | rdmavt_qp.h | 786 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.recv_cq); in rvt_recv_cq() local 788 if (unlikely(!rvt_cq_enter(cq, wc, solicited))) in rvt_recv_cq() 806 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq); in rvt_send_cq() local 808 if (unlikely(!rvt_cq_enter(cq, wc, solicited))) in rvt_send_cq() 923 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_tail() local 926 RDMA_READ_UAPI_ATOMIC(cq->queue->tail) : in ib_cq_tail() 939 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_head() local 942 RDMA_READ_UAPI_ATOMIC(cq->queue->head) : in ib_cq_head()
|
| A D | ib_verbs.h | 774 struct ib_cq *cq; member 1094 struct ib_cq *cq; member 1652 struct ib_cq *cq; member 1703 struct ib_cq *cq; member 1723 struct ib_cq *cq; member 2512 int (*create_cq_umem)(struct ib_cq *cq, 2523 int (*pre_destroy_cq)(struct ib_cq *cq); 4008 void ib_free_cq(struct ib_cq *cq); 4066 int ret = ib_destroy_cq_user(cq, NULL); in ib_destroy_cq() 4086 return cq->device->ops.poll_cq(cq, num_entries, wc); in ib_poll_cq() [all …]
|
| /include/net/mana/ |
| A D | gdma.h | 107 } cq; member 346 } cq; member 371 } cq; member 464 int mana_gd_poll_cq(struct gdma_queue *cq, struct gdma_comp *comp, int num_cqe); 466 void mana_gd_ring_cq(struct gdma_queue *cq, u8 arm_bit);
|
| A D | hw_channel.h | 191 struct hwc_cq *cq; member
|
| /include/net/ |
| A D | xsk_buff_pool.h | 65 struct xsk_queue *cq; member
|
| /include/linux/ |
| A D | io_uring_types.h | 142 struct io_uring sq, cq; member
|
Completed in 50 milliseconds