Home
last modified time | relevance | path

Searched refs:eqc (Results 1 – 8 of 8) sorted by relevance

/drivers/infiniband/hw/erdma/
A Derdma_eq.c190 struct erdma_eq_cb *eqc = &dev->ceqs[ceqn]; in erdma_set_ceq_irq() local
193 snprintf(eqc->irq.name, ERDMA_IRQNAME_SIZE, "erdma-ceq%u@pci:%s", ceqn, in erdma_set_ceq_irq()
195 eqc->irq.msix_vector = pci_irq_vector(dev->pdev, ceqn + 1); in erdma_set_ceq_irq()
201 &eqc->irq.affinity_hint_mask); in erdma_set_ceq_irq()
203 err = request_irq(eqc->irq.msix_vector, erdma_intr_ceq_handler, 0, in erdma_set_ceq_irq()
204 eqc->irq.name, eqc); in erdma_set_ceq_irq()
210 irq_set_affinity_hint(eqc->irq.msix_vector, in erdma_set_ceq_irq()
211 &eqc->irq.affinity_hint_mask); in erdma_set_ceq_irq()
218 struct erdma_eq_cb *eqc = &dev->ceqs[ceqn]; in erdma_free_ceq_irq() local
220 irq_set_affinity_hint(eqc->irq.msix_vector, NULL); in erdma_free_ceq_irq()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/
A Deq.c266 void *eqc; in create_map_eq() local
310 eqc = MLX5_ADDR_OF(create_eq_in, in, eq_context_entry); in create_map_eq()
311 MLX5_SET(eqc, eqc, log_eq_size, eq->fbc.log_sz); in create_map_eq()
312 MLX5_SET(eqc, eqc, uar_page, priv->uar->index); in create_map_eq()
313 MLX5_SET(eqc, eqc, intr, vecidx); in create_map_eq()
314 MLX5_SET(eqc, eqc, log_page_size, in create_map_eq()
A Ddebugfs.c408 param = 1 << MLX5_GET(eqc, ctx, log_eq_size); in eq_read_field()
411 param = MLX5_GET(eqc, ctx, intr); in eq_read_field()
414 param = MLX5_GET(eqc, ctx, log_page_size) + 12; in eq_read_field()
/drivers/infiniband/hw/hns/
A Dhns_roce_hw_v2.c6567 struct hns_roce_eq_context *eqc; in config_eqc() local
6571 eqc = mb_buf; in config_eqc()
6587 hr_reg_write(eqc, EQC_EQE_HOP_NUM, eq->hop_num); in config_eqc()
6589 hr_reg_write(eqc, EQC_COALESCE, eq->coalesce); in config_eqc()
6590 hr_reg_write(eqc, EQC_ARM_ST, eq->arm_st); in config_eqc()
6591 hr_reg_write(eqc, EQC_EQN, eq->eqn); in config_eqc()
6593 hr_reg_write(eqc, EQC_EQE_BA_PG_SZ, in config_eqc()
6595 hr_reg_write(eqc, EQC_EQE_BUF_PG_SZ, in config_eqc()
6611 hr_reg_write(eqc, EQC_EQE_BA_L, bt_ba >> 3); in config_eqc()
6612 hr_reg_write(eqc, EQC_EQE_BA_H, bt_ba >> 35); in config_eqc()
[all …]
/drivers/crypto/hisilicon/
A Ddebugfs.c246 struct qm_eqc eqc; in qm_eqc_aeqc_dump() local
260 xeqc = &eqc; in qm_eqc_aeqc_dump()
273 eqc.base_h = cpu_to_le32(QM_XQC_ADDR_MASK); in qm_eqc_aeqc_dump()
274 eqc.base_l = cpu_to_le32(QM_XQC_ADDR_MASK); in qm_eqc_aeqc_dump()
A Dqm.c693 tmp_xqc = qm->xqc_buf.eqc; in qm_set_and_get_xqc()
3153 struct qm_eqc eqc = {0}; in qm_eq_ctx_cfg() local
3155 eqc.base_l = cpu_to_le32(lower_32_bits(qm->eqe_dma)); in qm_eq_ctx_cfg()
3156 eqc.base_h = cpu_to_le32(upper_32_bits(qm->eqe_dma)); in qm_eq_ctx_cfg()
3158 eqc.dw3 = cpu_to_le32(QM_EQE_AEQE_SIZE); in qm_eq_ctx_cfg()
3159 eqc.dw6 = cpu_to_le32(((u32)qm->eq_depth - 1) | (1 << QM_EQC_PHASE_SHIFT)); in qm_eq_ctx_cfg()
3161 return qm_set_and_get_xqc(qm, QM_MB_CMD_EQC, &eqc, 0, 0); in qm_eq_ctx_cfg()
5515 QM_XQC_BUF_INIT(xqc_buf, eqc); in hisi_qm_alloc_rsv_buf()
/drivers/s390/block/
A Dscm_blk.c381 switch (scmrq->aob->response.eqc) { in scm_blk_handle_error()
/drivers/net/ethernet/mellanox/mlx4/
A Dresource_tracker.c3061 static int eq_get_mtt_addr(struct mlx4_eq_context *eqc) in eq_get_mtt_addr() argument
3063 return be32_to_cpu(eqc->mtt_base_addr_l) & 0xfffffff8; in eq_get_mtt_addr()
3066 static int eq_get_mtt_size(struct mlx4_eq_context *eqc) in eq_get_mtt_size() argument
3068 int log_eq_size = eqc->log_eq_size & 0x1f; in eq_get_mtt_size()
3069 int page_shift = (eqc->log_page_size & 0x3f) + 12; in eq_get_mtt_size()
3102 struct mlx4_eq_context *eqc = inbox->buf; in mlx4_SW2HW_EQ_wrapper() local
3103 int mtt_base = eq_get_mtt_addr(eqc) / dev->caps.mtt_entry_sz; in mlx4_SW2HW_EQ_wrapper()
3104 int mtt_size = eq_get_mtt_size(eqc); in mlx4_SW2HW_EQ_wrapper()

Completed in 47 milliseconds