| /linux/drivers/infiniband/hw/mthca/ |
| A D | mthca_cq.c | 289 for (prod_index = cq->cons_index; in mthca_cq_clean() 318 cq->cons_index += nfreed; in mthca_cq_clean() 337 cq->cons_index &= cq->ibcq.cqe; in mthca_cq_resize_copy_cqes() 339 cq->cons_index -= cq->ibcq.cqe + 1; in mthca_cq_resize_copy_cqes() 384 cq->cqn, cq->cons_index); in handle_error_cqe() 648 ++cq->cons_index; in mthca_poll_one() 698 cq->cons_index &= cq->ibcq.cqe; in mthca_poll_cq() 743 db_rec[0] = cpu_to_be32(cq->cons_index); in mthca_arbel_arm_cq() 761 mthca_write64(dbhi, cq->cons_index, in mthca_arbel_arm_cq() 859 cq->cons_index = 0; in mthca_init_cq() [all …]
|
| A D | mthca_eq.c | 237 eqe = get_eqe(eq, eq->cons_index); in next_eqe_sw() 363 ++eq->cons_index; in mthca_eq_int() 379 set_eq_ci(dev, eq, eq->cons_index); in mthca_eq_int() 411 dev->eq_table.eq[i].cons_index); in mthca_tavor_interrupt() 424 tavor_set_eq_ci(dev, eq, eq->cons_index); in mthca_tavor_msi_x_interrupt() 444 dev->eq_table.eq[i].cons_index); in mthca_arbel_interrupt() 458 arbel_set_eq_ci(dev, eq, eq->cons_index); in mthca_arbel_msi_x_interrupt() 555 eq->cons_index = 0; in mthca_create_eq()
|
| A D | mthca_provider.h | 91 u32 cons_index; member 188 u32 cons_index; member
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| A D | eq.h | 30 u32 cons_index; member 63 struct mlx5_eqe *eqe = get_eqe(eq, eq->cons_index & eq->fbc.sz_m1); in next_eqe_sw() 65 return (eqe->owner ^ (eq->cons_index >> eq->fbc.log_sz)) & 1 ? NULL : eqe; in next_eqe_sw() 71 u32 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in eq_update_ci()
|
| /linux/include/linux/mlx5/ |
| A D | cq.h | 51 u32 cons_index; member 144 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx5_cq_set_ci() 154 u32 cons_index) in mlx5_cq_arm() argument 161 ci = cons_index & 0xffffff; in mlx5_cq_arm()
|
| /linux/include/linux/mlx4/ |
| A D | cq.h | 156 ci = cq->cons_index & 0xffffff; in mlx4_cq_arm() 174 *cq->set_ci_db = cpu_to_be32(cq->cons_index & 0xffffff); in mlx4_cq_set_ci()
|
| A D | device.h | 732 u32 cons_index; member
|
| /linux/drivers/infiniband/hw/mlx4/ |
| A D | cq.c | 90 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw() 348 i = cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes() 352 return i - cq->mcq.cons_index; in mlx4_ib_get_outstanding_cqes() 362 i = cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes() 377 ++cq->mcq.cons_index; in mlx4_ib_cq_resize_copy_cqes() 681 ++cq->mcq.cons_index; in mlx4_ib_poll_one() 931 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); ++prod_index) in __mlx4_ib_cq_clean() 932 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe) in __mlx4_ib_cq_clean() 939 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) { in __mlx4_ib_cq_clean() 959 cq->mcq.cons_index += nfreed; in __mlx4_ib_cq_clean()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| A D | eq.c | 543 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 573 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 698 eq->eqn, eq->cons_index, ret); in mlx4_eq_int() 783 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 786 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 808 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 810 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 820 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 823 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 827 ++eq->cons_index; in mlx4_eq_int() [all …]
|
| A D | en_tx.c | 439 u32 cons_index = mcq->cons_index; in mlx4_en_process_tx_cq() local 456 index = cons_index & size_mask; in mlx4_en_process_tx_cq() 465 cons_index & size) && (done < budget)) { in mlx4_en_process_tx_cq() 508 ++cons_index; in mlx4_en_process_tx_cq() 509 index = cons_index & size_mask; in mlx4_en_process_tx_cq() 517 mcq->cons_index = cons_index; in mlx4_en_process_tx_cq()
|
| A D | en_rx.c | 743 index = cq->mcq.cons_index & ring->size_mask; in mlx4_en_process_rx_cq() 748 cq->mcq.cons_index & cq->size)) { in mlx4_en_process_rx_cq() 961 ++cq->mcq.cons_index; in mlx4_en_process_rx_cq() 962 index = (cq->mcq.cons_index) & ring->size_mask; in mlx4_en_process_rx_cq() 979 ring->cons = cq->mcq.cons_index; in mlx4_en_process_rx_cq()
|
| A D | cq.c | 410 cq->cons_index = 0; in mlx4_cq_alloc()
|
| A D | mlx4.h | 397 u32 cons_index; member
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | eq.c | 143 ++eq->cons_index; in mlx5_eq_comp_int() 166 count_eqe = eq->core.cons_index; in mlx5_eq_poll_irq_disabled() 168 count_eqe = eq->core.cons_index - count_eqe; in mlx5_eq_poll_irq_disabled() 232 ++eq->cons_index; in mlx5_eq_async_int() 286 eq->cons_index = 0; in create_map_eq() 795 u32 ci = eq->cons_index + cc; in mlx5_eq_get_eqe() 816 eq->cons_index += cc; in mlx5_eq_update_ci() 817 val = (eq->cons_index & 0xffffff) | (eq->eqn << 24); in mlx5_eq_update_ci()
|
| A D | cq.c | 109 cq->cons_index = 0; in mlx5_create_cq()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| A D | health.c | 58 devlink_fmsg_u32_pair_put(fmsg, "ci", eq->core.cons_index); in mlx5e_health_eq_diag_fmsg() 137 eq->core.eqn, eq->core.cons_index, eq->core.irqn); in mlx5e_health_channel_eq_recover()
|
| /linux/drivers/infiniband/hw/mlx5/ |
| A D | cq.c | 99 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw() 472 ++cq->mcq.cons_index; in mlx5_poll_one() 666 uar_page, to_mcq(ibcq)->mcq.cons_index); in mlx5_ib_arm_cq() 1097 for (prod_index = cq->mcq.cons_index; get_sw_cqe(cq, prod_index); prod_index++) in __mlx5_ib_cq_clean() 1098 if (prod_index == cq->mcq.cons_index + cq->ibcq.cqe) in __mlx5_ib_cq_clean() 1104 while ((int) --prod_index - (int) cq->mcq.cons_index >= 0) { in __mlx5_ib_cq_clean() 1122 cq->mcq.cons_index += nfreed; in __mlx5_ib_cq_clean() 1236 i = cq->mcq.cons_index; in copy_resize_cqes() 1267 ++cq->mcq.cons_index; in copy_resize_cqes()
|
| /linux/drivers/infiniband/hw/hns/ |
| A D | hns_roce_restrack.c | 24 if (rdma_nl_put_driver_u32(msg, "cons_index", hr_cq->cons_index)) in hns_roce_fill_res_cq_entry()
|
| A D | hns_roce_hw_v2.c | 3571 hr_cq->cons_index += nfreed; in __hns_roce_v2_cq_clean() 3930 cqe = get_sw_cqe_v2(hr_cq, hr_cq->cons_index); in hns_roce_v2_poll_one() 3934 ++hr_cq->cons_index; in hns_roce_v2_poll_one() 6070 hr_reg_write(&eq_db, EQ_DB_CI, eq->cons_index); in update_eq_db() 6080 (eq->cons_index & (eq->entries - 1)) * in next_aeqe_sw_v2() 6084 !!(eq->cons_index & eq->entries)) ? aeqe : NULL; in next_aeqe_sw_v2() 6139 event_type, eq->eqn, eq->cons_index); in hns_roce_v2_aeq_int() 6145 ++eq->cons_index; in hns_roce_v2_aeq_int() 6165 (eq->cons_index & (eq->entries - 1)) * in next_ceqe_sw_v2() 6430 eq->cons_index = 0; in init_eq_config() [all …]
|
| A D | hns_roce_device.h | 432 u32 cons_index; member 713 u32 cons_index; member
|
| A D | hns_roce_cq.c | 420 hr_cq->cons_index = 0; in hns_roce_create_cq()
|
| /linux/drivers/vfio/pci/mlx5/ |
| A D | cmd.c | 1148 cq->mcq.cons_index); in mlx5vf_create_cq() 1691 cqe = get_sw_cqe(cq, cq->mcq.cons_index); in mlx5vf_cq_poll_one() 1695 ++cq->mcq.cons_index; in mlx5vf_cq_poll_one() 1751 cq->mcq.cons_index); in mlx5vf_tracker_read_and_clear()
|
| /linux/drivers/vdpa/mlx5/net/ |
| A D | mlx5_vnet.c | 496 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw() 507 vcq->mcq.cons_index++; in mlx5_vdpa_poll_one() 552 mlx5_cq_arm(&mvq->cq.mcq, MLX5_CQ_DB_REQ_NOT, uar_page, mvq->cq.mcq.cons_index); in mlx5_vdpa_cq_comp() 619 mlx5_cq_arm(&mvq->cq.mcq, MLX5_CQ_DB_REQ_NOT, uar_page, mvq->cq.mcq.cons_index); in cq_create()
|