| /linux/drivers/infiniband/sw/rxe/ |
| A D | rxe_req.c | 44 unsigned int wqe_index; in req_retry() local 55 qp->req.wqe_index = cons; in req_retry() 59 for (wqe_index = cons; wqe_index != prod; in req_retry() 60 wqe_index = queue_next_index(q, wqe_index)) { in req_retry() 61 wqe = queue_addr_from_index(qp->sq.queue, wqe_index); in req_retry() 115 unsigned int index = qp->req.wqe_index; in req_next_wqe() 554 qp->req.wqe_index = queue_next_index(qp->sq.queue, in update_state() 555 qp->req.wqe_index); in update_state() 605 qp->req.wqe_index = queue_next_index(qp->sq.queue, qp->req.wqe_index); in rxe_do_local_ops() 636 qp->req.wqe_index = queue_get_consumer(q, in rxe_requester() [all …]
|
| A D | rxe_verbs.h | 118 int wqe_index; member
|
| A D | rxe_qp.c | 249 qp->req.wqe_index = queue_get_producer(qp->sq.queue, in rxe_qp_init_req()
|
| /linux/drivers/infiniband/hw/mthca/ |
| A D | mthca_cq.c | 486 int wqe_index; in mthca_poll_one() local 537 entry->wr_id = (*cur_qp)->wrid[wqe_index + in mthca_poll_one() 543 wqe_index = wqe >> srq->wqe_shift; in mthca_poll_one() 544 entry->wr_id = srq->wrid[wqe_index]; in mthca_poll_one() 550 wqe_index = wqe >> wq->wqe_shift; in mthca_poll_one() 556 if (unlikely(wqe_index < 0)) in mthca_poll_one() 557 wqe_index = wq->max - 1; in mthca_poll_one() 558 entry->wr_id = (*cur_qp)->wrid[wqe_index]; in mthca_poll_one() 562 if (wq->last_comp < wqe_index) in mthca_poll_one() 563 wq->tail += wqe_index - wq->last_comp; in mthca_poll_one() [all …]
|
| /linux/include/linux/mlx4/ |
| A D | cq.h | 57 __be16 wqe_index; member 66 __be16 wqe_index; member 82 __be16 wqe_index; member
|
| /linux/drivers/net/ethernet/ibm/ehea/ |
| A D | ehea_qmr.h | 308 int *wqe_index) in ehea_get_swqe() argument 313 *wqe_index = (queue->current_q_offset) >> (7 + EHEA_SG_SQ); in ehea_get_swqe() 325 static inline struct ehea_cqe *ehea_poll_rq1(struct ehea_qp *qp, int *wqe_index) in ehea_poll_rq1() argument 329 *wqe_index = (queue->current_q_offset) >> (7 + EHEA_SG_RQ1); in ehea_poll_rq1()
|
| A D | ehea_main.c | 590 int arr_len, int wqe_index) in get_skb_by_index_ll() argument 596 x = wqe_index + 1; in get_skb_by_index_ll() 609 skb = skb_array[wqe_index]; in get_skb_by_index_ll() 610 skb_array[wqe_index] = NULL; in get_skb_by_index_ll() 671 cqe = ehea_poll_rq1(qp, &wqe_index); in ehea_proc_rwqes() 679 last_wqe_index = wqe_index; in ehea_proc_rwqes() 686 wqe_index); in ehea_proc_rwqes() 738 cqe = ehea_poll_rq1(qp, &wqe_index); in ehea_proc_rwqes() 881 int wqe_index; in ehea_poll() local 2477 int wqe_index; in ehea_purge_sq() local [all …]
|
| /linux/drivers/infiniband/hw/mlx5/ |
| A D | odp.c | 64 u16 wqe_index; member 1056 u16 wqe_index = pfault->wqe.wqe_index; in mlx5_ib_mr_initiator_pfault_handler() local 1070 wqe_index, qpn); in mlx5_ib_mr_initiator_pfault_handler() 1194 u16 wqe_index = pfault->wqe.wqe_index; in mlx5_ib_mr_wqe_pfault_handler() local 1225 ret = mlx5_ib_read_wqe_sq(qp, wqe_index, wqe, PAGE_SIZE, in mlx5_ib_mr_wqe_pfault_handler() 1232 ret = mlx5_ib_read_wqe_rq(qp, wqe_index, wqe, PAGE_SIZE, in mlx5_ib_mr_wqe_pfault_handler() 1241 ret = mlx5_ib_read_wqe_srq(srq, wqe_index, wqe, PAGE_SIZE, in mlx5_ib_mr_wqe_pfault_handler() 1269 ret, wqe_index, pfault->token); in mlx5_ib_mr_wqe_pfault_handler() 1441 pfault->wqe.wqe_index = in mlx5_ib_eq_pf_process() 1442 be16_to_cpu(pf_eqe->wqe.wqe_index); in mlx5_ib_eq_pf_process() [all …]
|
| A D | srq.c | 389 void mlx5_ib_free_srq_wqe(struct mlx5_ib_srq *srq, int wqe_index) in mlx5_ib_free_srq_wqe() argument 397 next->next_wqe_index = cpu_to_be16(wqe_index); in mlx5_ib_free_srq_wqe() 398 srq->tail = wqe_index; in mlx5_ib_free_srq_wqe()
|
| A D | qp.c | 107 size_t buflen, int wqe_index, in mlx5_ib_read_user_wqe_common() argument 112 size_t offset = wq_offset + ((wqe_index % wq_wqe_cnt) << wq_wqe_shift); in mlx5_ib_read_user_wqe_common() 142 wqe_index = wqe_index & qp->sq.fbc.sz_m1; in mlx5_ib_read_kernel_wqe_sq() 145 p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, wqe_index); in mlx5_ib_read_kernel_wqe_sq() 161 wqe_index = (wqe_index + 1) & qp->sq.fbc.sz_m1; in mlx5_ib_read_kernel_wqe_sq() 162 p = mlx5_frag_buf_get_wqe(&qp->sq.fbc, wqe_index); in mlx5_ib_read_kernel_wqe_sq() 182 ret = mlx5_ib_read_user_wqe_common(umem, buffer, buflen, wqe_index, in mlx5_ib_read_user_wqe_sq() 229 return mlx5_ib_read_kernel_wqe_sq(qp, wqe_index, buffer, in mlx5_ib_read_wqe_sq() 232 return mlx5_ib_read_user_wqe_sq(qp, wqe_index, buffer, buflen, bc); in mlx5_ib_read_wqe_sq() 244 ret = mlx5_ib_read_user_wqe_common(umem, buffer, buflen, wqe_index, in mlx5_ib_read_user_wqe_rq() [all …]
|
| A D | mlx5_ib.h | 1237 void mlx5_ib_free_srq_wqe(struct mlx5_ib_srq *srq, int wqe_index); 1264 int mlx5_ib_read_wqe_sq(struct mlx5_ib_qp *qp, int wqe_index, void *buffer, 1266 int mlx5_ib_read_wqe_rq(struct mlx5_ib_qp *qp, int wqe_index, void *buffer, 1268 int mlx5_ib_read_wqe_srq(struct mlx5_ib_srq *srq, int wqe_index, void *buffer,
|
| /linux/drivers/infiniband/hw/mlx4/ |
| A D | srq.c | 293 void mlx4_ib_free_srq_wqe(struct mlx4_ib_srq *srq, int wqe_index) in mlx4_ib_free_srq_wqe() argument 301 next->next_wqe_index = cpu_to_be16(wqe_index); in mlx4_ib_free_srq_wqe() 302 srq->tail = wqe_index; in mlx4_ib_free_srq_wqe()
|
| A D | cq.c | 517 be32_to_cpu(cqe->my_qpn), be16_to_cpu(cqe->wqe_index), in mlx4_ib_handle_error_cqe() 734 wqe_ctr = be16_to_cpu(cqe->wqe_index); in mlx4_ib_poll_one() 741 wqe_ctr = be16_to_cpu(cqe->wqe_index); in mlx4_ib_poll_one() 746 wqe_ctr = be16_to_cpu(cqe->wqe_index); in mlx4_ib_poll_one() 944 mlx4_ib_free_srq_wqe(srq, be16_to_cpu(cqe->wqe_index)); in __mlx4_ib_cq_clean()
|
| A D | mlx4_ib.h | 791 void mlx4_ib_free_srq_wqe(struct mlx4_ib_srq *srq, int wqe_index);
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| A D | en_tx.c | 402 u16 wqe_index; in mlx4_en_handle_err_cqe() local 410 wqe_index = be16_to_cpu(err_cqe->wqe_index) & ring->size_mask; in mlx4_en_handle_err_cqe() 411 tx_info = &ring->tx_info[wqe_index]; in mlx4_en_handle_err_cqe() 414 wqe_index, desc_size); in mlx4_en_handle_err_cqe() 415 tx_desc = ring->buf + (wqe_index << LOG_TXBB_SIZE); in mlx4_en_handle_err_cqe() 477 new_index = be16_to_cpu(cqe->wqe_index) & size_mask; in mlx4_en_process_tx_cq()
|
| A D | en_netdev.c | 1738 cq->buf->wqe_index = cpu_to_be16(0xffff); in mlx4_en_start_port()
|
| /linux/include/linux/mlx5/ |
| A D | device.h | 661 __be16 wqe_index; member
|
| A D | mlx5_ifc.h | 2933 u8 wqe_index[0x10]; member
|
| /linux/drivers/infiniband/hw/hns/ |
| A D | hns_roce_hw_v2.c | 906 static void hns_roce_free_srq_wqe(struct hns_roce_srq *srq, u32 wqe_index) in hns_roce_free_srq_wqe() argument 911 bitmap_clear(srq->idx_que.bitmap, wqe_index, 1); in hns_roce_free_srq_wqe() 3283 int wqe_index; in __hns_roce_v2_cq_clean() local 3300 wqe_index = hr_reg_read(cqe, CQE_WQE_IDX); in __hns_roce_v2_cq_clean() 3301 hns_roce_free_srq_wqe(srq, wqe_index); in __hns_roce_v2_cq_clean()
|
| /linux/drivers/scsi/elx/libefc_sli/ |
| A D | sli4.h | 983 __le16 wqe_index; member
|