Home
last modified time | relevance | path

Searched refs:q_no (Results 1 – 25 of 34) sorted by relevance

12

/linux/drivers/net/ethernet/cavium/liquidio/
A Dcn23xx_vf_device.c57 for (q_no = 0; q_no < num_queues; q_no++) { in cn23xx_vf_reset_io_queues()
67 for (q_no = 0; q_no < num_queues; q_no++) { in cn23xx_vf_reset_io_queues()
111 for (q_no = 0; q_no < (oct->sriov_info.rings_per_vf); q_no++) { in cn23xx_vf_setup_global_input_regs()
156 for (q_no = 0; q_no < (oct->sriov_info.rings_per_vf); q_no++) { in cn23xx_vf_setup_global_output_regs()
320 for (q_no = 0; q_no < oct->num_iqs; q_no++) { in cn23xx_enable_vf_io_queues()
341 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in cn23xx_enable_vf_io_queues()
449 for (q_no = 0 ; q_no < oct->num_iqs ; q_no++) in cn23xx_octeon_pfvf_handshake()
550 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in cn23xx_enable_vf_interrupt()
565 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in cn23xx_enable_vf_interrupt()
590 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in cn23xx_disable_vf_interrupt()
[all …]
A Dcn23xx_pf_device.c359 for (q_no = srn; q_no < ern; q_no++) { in cn23xx_reset_io_queues()
367 for (q_no = srn; q_no < ern; q_no++) { in cn23xx_reset_io_queues()
379 q_no); in cn23xx_reset_io_queues()
422 for (q_no = 0; q_no < ern; q_no++) { in cn23xx_pf_setup_global_input_regs()
443 for (q_no = srn; q_no < ern; q_no++) { in cn23xx_pf_setup_global_input_regs()
493 for (q_no = srn; q_no < ern; q_no++) { in cn23xx_pf_setup_global_output_regs()
730 mbox->q_no = q_no; in cn23xx_setup_pf_mbox()
797 for (q_no = srn; q_no < ern; q_no++) { in cn23xx_enable_io_queues()
851 for (q_no = srn; q_no < ern; q_no++) { in cn23xx_enable_io_queues()
876 for (q_no = srn; q_no < ern; q_no++) { in cn23xx_disable_io_queues()
[all …]
A Docteon_droq.c209 vfree(oct->droq[q_no]); in octeon_delete_droq()
210 oct->droq[q_no] = NULL; in octeon_delete_droq()
218 u32 q_no, in octeon_init_droq() argument
230 droq = oct->droq[q_no]; in octeon_init_droq()
234 droq->q_no = q_no; in octeon_init_droq()
836 value |= (1 << q_no); in octeon_enable_irq()
839 value |= (1 << q_no); in octeon_enable_irq()
849 lio_enable_irq(oct->droq[q_no], oct->instr_queue[q_no]); in octeon_enable_irq()
853 lio_enable_irq(oct->droq[q_no], oct->instr_queue[q_no]); in octeon_enable_irq()
931 if (oct->droq[q_no]) { in octeon_create_droq()
[all …]
A Docteon_droq.h248 u32 q_no; member
338 u32 q_no,
350 int octeon_delete_droq(struct octeon_device *oct_dev, u32 q_no);
365 u32 q_no,
376 int octeon_unregister_droq_ops(struct octeon_device *oct, u32 q_no);
400 int octeon_create_droq(struct octeon_device *oct, u32 q_no,
410 int octeon_enable_irq(struct octeon_device *oct, u32 q_no);
A Dlio_core.c454 int q_no = wk->ctxul; in octnet_poll_check_rxq_oom_status() local
469 int q, q_no; in setup_rx_oom_poll_fn() local
472 q_no = lio->linfo.rxpciq[q].s.q_no; in setup_rx_oom_poll_fn()
484 wq->wk.ctxul = q_no; in setup_rx_oom_poll_fn()
496 int q_no; in cleanup_rx_oom_poll_fn() local
498 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in cleanup_rx_oom_poll_fn()
564 writel(oct->droq[q_no]->max_count, oct->droq[q_no]->pkts_credit_reg); in octeon_setup_droq()
761 iq_no = droq->q_no; in liquidio_napi_poll()
823 int q, q_no; in liquidio_setup_io_queues() local
842 q_no = lio->linfo.rxpciq[q].s.q_no; in liquidio_setup_io_queues()
[all …]
A Docteon_mailbox.c65 mbox->mbox_req.q_no = mbox->q_no; in octeon_mbox_read()
77 mbox->mbox_resp.q_no = mbox->q_no; in octeon_mbox_read()
134 struct octeon_mbox *mbox = oct->mbox[mbox_cmd->q_no]; in octeon_mbox_write()
262 mbox->q_no); in octeon_mbox_process_cmd()
263 pcie_flr(oct->sriov_info.dpiring_to_vfpcidev_lut[mbox->q_no]); in octeon_mbox_process_cmd()
355 int octeon_mbox_cancel(struct octeon_device *oct, int q_no) in octeon_mbox_cancel() argument
357 struct octeon_mbox *mbox = oct->mbox[q_no]; in octeon_mbox_cancel()
A Docteon_device.c901 txpciq.s.q_no = iq_no; in octeon_setup_instr_queues()
962 u32 q_no; in octeon_set_io_queues_off() local
968 for (q_no = 0; q_no < oct->sriov_info.rings_per_vf; q_no++) { in octeon_set_io_queues_off()
982 q_no); in octeon_set_io_queues_off()
995 "unable to reset qno %u\n", q_no); in octeon_set_io_queues_off()
1005 u32 q_no, in octeon_set_droq_pkt_op() argument
1015 reg_val = reg_val | (1 << q_no); in octeon_set_droq_pkt_op()
1017 reg_val = reg_val & (~(1 << q_no)); in octeon_set_droq_pkt_op()
1286 (oct->io_qmask.iq & BIT_ULL(q_no))) in octeon_get_tx_qsize()
1296 (oct->io_qmask.oq & BIT_ULL(q_no))) in octeon_get_rx_qsize()
[all …]
A Docteon_mailbox.h68 u32 q_no; member
91 u32 q_no; member
123 int octeon_mbox_cancel(struct octeon_device *oct, int q_no);
A Dlio_ethtool.c2252 int q_no; in oct_cfg_rx_intrcnt() local
2259 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in oct_cfg_rx_intrcnt()
2274 int q_no; in oct_cfg_rx_intrcnt() local
2281 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in oct_cfg_rx_intrcnt()
2329 int q_no; in oct_cfg_rx_intrtime() local
2337 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in oct_cfg_rx_intrtime()
2351 int q_no; in oct_cfg_rx_intrtime() local
2360 for (q_no = 0; q_no < oct->num_oqs; q_no++) { in oct_cfg_rx_intrtime()
2395 int q_no; in oct_cfg_tx_intrcnt() local
2403 for (q_no = 0; q_no < oct->num_iqs; q_no++) { in oct_cfg_tx_intrcnt()
[all …]
A Docteon_nic.h85 u32 q_no; member
112 static inline int octnet_iq_is_full(struct octeon_device *oct, u32 q_no) in octnet_iq_is_full() argument
114 return ((u32)atomic_read(&oct->instr_queue[q_no]->instr_pending) in octnet_iq_is_full()
115 >= (oct->instr_queue[q_no]->max_count - 2)); in octnet_iq_is_full()
A Dcn66xx_regs.h473 #define CN6XXX_DPI_DMA_ENG_ENB(q_no) \ argument
474 (CN6XXX_DPI_DMA_ENG0_ENB + ((q_no) * 8))
478 #define CN6XXX_DPI_DMA_ENG_BUF(q_no) \ argument
479 (CN6XXX_DPI_DMA_ENG0_BUF + ((q_no) * 8))
A Dlio_main.c155 int q_no; in octeon_droq_bh() local
161 for (q_no = 0; q_no < MAX_OCTEON_OUTPUT_QUEUES(oct); q_no++) { in octeon_droq_bh()
162 if (!(oct->io_qmask.oq & BIT_ULL(q_no))) in octeon_droq_bh()
166 lio_enable_irq(oct->droq[q_no], NULL); in octeon_droq_bh()
465 lio->oct_dev->num_iqs].s.q_no; in check_txq_status()
1289 lio->linfo.rxpciq[j].s.q_no); in liquidio_stop_nic_module()
2053 iq_no = lio->linfo.txpciq[i].s.q_no; in liquidio_get_stats64()
2069 oq_no = lio->linfo.rxpciq[i].s.q_no; in liquidio_get_stats64()
2263 sc->iq_no = ndata->q_no; in send_nic_timestamp_pkt()
2346 ndata.q_no = iq_no; in liquidio_xmit()
[all …]
A Dlio_vf_main.c622 sc->iq_no = lio->linfo.txpciq[0].s.q_no; in send_rx_ctrl_cmd()
731 lio->linfo.rxpciq[j].s.q_no); in liquidio_stop_nic_module()
1188 iq_no = lio->linfo.txpciq[i].s.q_no; in liquidio_get_stats64()
1204 oq_no = lio->linfo.rxpciq[i].s.q_no; in liquidio_get_stats64()
1379 sc->iq_no = ndata->q_no; in send_nic_timestamp_pkt()
1428 iq_no = lio->linfo.txpciq[q_idx].s.q_no; in liquidio_xmit()
1455 ndata.q_no = iq_no; in liquidio_xmit()
1457 if (octnet_iq_is_full(oct, ndata.q_no)) { in liquidio_xmit()
1460 ndata.q_no); in liquidio_xmit()
2162 lio->txq = lio->linfo.txpciq[0].s.q_no; in setup_nic_devices()
[all …]
A Dliquidio_common.h728 u64 q_no:8; member
744 u64 q_no:8;
756 u64 q_no:8; member
760 u64 q_no:8;
A Docteon_device.h851 int octeon_get_tx_qsize(struct octeon_device *oct, u32 q_no);
853 int octeon_get_rx_qsize(struct octeon_device *oct, u32 q_no);
865 void octeon_set_droq_pkt_op(struct octeon_device *oct, u32 q_no, u32 enable);
/linux/drivers/net/ethernet/marvell/octeon_ep/
A Doctep_tx.c94 netif_wake_subqueue(iq->netdev, iq->q_no); in octep_iq_process_completions()
184 oct->iq[q_no] = iq; in octep_setup_iq()
189 iq->q_no = q_no; in octep_setup_iq()
193 iq->netdev_q = netdev_get_tx_queue(iq->netdev, q_no); in octep_setup_iq()
201 "Failed to allocate DMA memory for IQ-%d\n", q_no); in octep_setup_iq()
213 q_no); in octep_setup_iq()
222 "Failed to allocate buff info for IQ-%d\n", q_no); in octep_setup_iq()
238 oct->hw_ops.setup_iq_regs(oct, q_no); in octep_setup_iq()
250 oct->iq[q_no] = NULL; in octep_setup_iq()
266 int q_no = iq->q_no; in octep_free_iq() local
[all …]
A Doctep_rx.c50 oq->q_no); in octep_oq_fill_ring_buffers()
99 oq->q_no); in octep_oq_refill()
123 static int octep_setup_oq(struct octep_device *oct, int q_no) in octep_setup_oq() argument
131 oct->oq[q_no] = oq; in octep_setup_oq()
136 oq->q_no = q_no; in octep_setup_oq()
157 "Failed to allocate DMA memory for OQ-%d !!\n", q_no); in octep_setup_oq()
164 "Failed to allocate buffer info for OQ-%d\n", q_no); in octep_setup_oq()
172 oct->hw_ops.setup_oq_regs(oct, q_no); in octep_setup_oq()
186 oct->oq[q_no] = NULL; in octep_setup_oq()
228 int q_no = oq->q_no; in octep_free_oq() local
[all …]
A Doctep_cnxk_pf.c132 dev_dbg(&oct->pdev->dev, "Reset PF IQ-%d\n", q_no); in cnxk_reset_iq()
135 q_no += conf->pf_ring_cfg.srn; in cnxk_reset_iq()
138 octep_write_csr64(oct, CNXK_SDP_R_IN_ENABLE(q_no), val); in cnxk_reset_iq()
141 octep_write_csr64(oct, CNXK_SDP_R_IN_CNTS(q_no), val); in cnxk_reset_iq()
143 octep_write_csr64(oct, CNXK_SDP_R_IN_PKT_CNT(q_no), val); in cnxk_reset_iq()
144 octep_write_csr64(oct, CNXK_SDP_R_IN_BYTE_CNT(q_no), val); in cnxk_reset_iq()
159 q_no += CFG_GET_PORTS_PF_SRN(oct->conf); in cnxk_reset_oq()
162 octep_write_csr64(oct, CNXK_SDP_R_OUT_ENABLE(q_no), val); in cnxk_reset_oq()
168 val = octep_read_csr(oct, CNXK_SDP_R_OUT_CNTS(q_no)); in cnxk_reset_oq()
169 octep_write_csr(oct, CNXK_SDP_R_OUT_CNTS(q_no), val); in cnxk_reset_oq()
[all …]
A Doctep_cn9k_pf.c112 dev_dbg(&oct->pdev->dev, "Reset PF IQ-%d\n", q_no); in cn93_reset_iq()
115 q_no += conf->pf_ring_cfg.srn; in cn93_reset_iq()
118 octep_write_csr64(oct, CN93_SDP_R_IN_ENABLE(q_no), val); in cn93_reset_iq()
121 octep_write_csr64(oct, CN93_SDP_R_IN_CNTS(q_no), val); in cn93_reset_iq()
123 octep_write_csr64(oct, CN93_SDP_R_IN_PKT_CNT(q_no), val); in cn93_reset_iq()
124 octep_write_csr64(oct, CN93_SDP_R_IN_BYTE_CNT(q_no), val); in cn93_reset_iq()
139 q_no += CFG_GET_PORTS_PF_SRN(oct->conf); in cn93_reset_oq()
142 octep_write_csr64(oct, CN93_SDP_R_OUT_ENABLE(q_no), val); in cn93_reset_oq()
145 val = octep_read_csr(oct, CN93_SDP_R_OUT_CNTS(q_no)); in cn93_reset_oq()
146 octep_write_csr(oct, CN93_SDP_R_OUT_CNTS(q_no), val); in cn93_reset_oq()
[all …]
A Doctep_main.c567 netdev_dbg(iq->netdev, "enabling intr for Q-%u\n", iq->q_no); in octep_enable_ioq_irq()
810 netif_stop_subqueue(iq->netdev, iq->q_no); in octep_iq_full_check()
824 netif_start_subqueue(iq->netdev, iq->q_no); in octep_iq_full_check()
855 u16 q_no, wi; in octep_start_xmit() local
860 q_no = skb_get_queue_mapping(skb); in octep_start_xmit()
861 if (q_no >= oct->num_iqs) { in octep_start_xmit()
862 netdev_err(netdev, "Invalid Tx skb->queue_mapping=%d\n", q_no); in octep_start_xmit()
863 q_no = q_no % oct->num_iqs; in octep_start_xmit()
866 iq = oct->iq[q_no]; in octep_start_xmit()
/linux/drivers/net/ethernet/marvell/octeon_ep_vf/
A Doctep_vf_tx.c90 netif_subqueue_completed_wake(iq->netdev, iq->q_no, compl_pkts, in octep_vf_iq_process_completions()
183 oct->iq[q_no] = iq; in octep_vf_setup_iq()
188 iq->q_no = q_no; in octep_vf_setup_iq()
192 iq->netdev_q = netdev_get_tx_queue(iq->netdev, q_no); in octep_vf_setup_iq()
200 "Failed to allocate DMA memory for IQ-%d\n", q_no); in octep_vf_setup_iq()
212 q_no); in octep_vf_setup_iq()
221 "Failed to allocate buff info for IQ-%d\n", q_no); in octep_vf_setup_iq()
237 oct->hw_ops.setup_iq_regs(oct, q_no); in octep_vf_setup_iq()
249 oct->iq[q_no] = NULL; in octep_vf_setup_iq()
265 int q_no = iq->q_no; in octep_vf_free_iq() local
[all …]
A Doctep_vf_cn9k.c81 static void cn93_vf_reset_iq(struct octep_vf_device *oct, int q_no) in cn93_vf_reset_iq() argument
85 dev_dbg(&oct->pdev->dev, "Reset VF IQ-%d\n", q_no); in cn93_vf_reset_iq()
88 octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_ENABLE(q_no), val); in cn93_vf_reset_iq()
92 octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_PKT_CNT(q_no), val); in cn93_vf_reset_iq()
93 octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_BYTE_CNT(q_no), val); in cn93_vf_reset_iq()
100 val = octep_vf_read_csr64(oct, CN93_VF_SDP_R_IN_CNTS(q_no)); in cn93_vf_reset_iq()
101 octep_vf_write_csr64(oct, CN93_VF_SDP_R_IN_CNTS(q_no), in cn93_vf_reset_iq()
111 octep_vf_write_csr64(oct, CN93_VF_SDP_R_OUT_ENABLE(q_no), val); in cn93_vf_reset_oq()
114 val = octep_vf_read_csr(oct, CN93_VF_SDP_R_OUT_CNTS(q_no)); in cn93_vf_reset_oq()
115 octep_vf_write_csr(oct, CN93_VF_SDP_R_OUT_CNTS(q_no), val); in cn93_vf_reset_oq()
[all …]
A Doctep_vf_cnxk.c84 static void cnxk_vf_reset_iq(struct octep_vf_device *oct, int q_no) in cnxk_vf_reset_iq() argument
88 dev_dbg(&oct->pdev->dev, "Reset VF IQ-%d\n", q_no); in cnxk_vf_reset_iq()
91 octep_vf_write_csr64(oct, CNXK_VF_SDP_R_IN_ENABLE(q_no), val); in cnxk_vf_reset_iq()
94 octep_vf_write_csr64(oct, CNXK_VF_SDP_R_IN_INT_LEVELS(q_no), val); in cnxk_vf_reset_iq()
95 octep_vf_write_csr64(oct, CNXK_VF_SDP_R_IN_PKT_CNT(q_no), val); in cnxk_vf_reset_iq()
96 octep_vf_write_csr64(oct, CNXK_VF_SDP_R_IN_BYTE_CNT(q_no), val); in cnxk_vf_reset_iq()
103 val = octep_vf_read_csr64(oct, CNXK_VF_SDP_R_IN_CNTS(q_no)); in cnxk_vf_reset_iq()
113 octep_vf_write_csr64(oct, CNXK_VF_SDP_R_OUT_ENABLE(q_no), val); in cnxk_vf_reset_oq()
116 val = octep_vf_read_csr(oct, CNXK_VF_SDP_R_OUT_CNTS(q_no)); in cnxk_vf_reset_oq()
117 octep_vf_write_csr(oct, CNXK_VF_SDP_R_OUT_CNTS(q_no), val); in cnxk_vf_reset_oq()
[all …]
A Doctep_vf_rx.c50 oq->q_no); in octep_vf_oq_fill_ring_buffers()
99 oq->q_no); in octep_vf_oq_refill()
123 static int octep_vf_setup_oq(struct octep_vf_device *oct, int q_no) in octep_vf_setup_oq() argument
131 oct->oq[q_no] = oq; in octep_vf_setup_oq()
136 oq->q_no = q_no; in octep_vf_setup_oq()
157 "Failed to allocate DMA memory for OQ-%d !!\n", q_no); in octep_vf_setup_oq()
165 "Failed to allocate buffer info for OQ-%d\n", q_no); in octep_vf_setup_oq()
173 oct->hw_ops.setup_oq_regs(oct, q_no); in octep_vf_setup_oq()
187 oct->oq[q_no] = NULL; in octep_vf_setup_oq()
229 int q_no = oq->q_no; in octep_vf_free_oq() local
[all …]
A Doctep_vf_main.c300 netdev_dbg(iq->netdev, "enabling intr for Q-%u\n", iq->q_no); in octep_vf_enable_ioq_irq()
565 ret = netif_subqueue_maybe_stop(iq->netdev, iq->q_no, IQ_INSTR_SPACE(iq), in octep_vf_iq_full_check()
609 u16 q_no, wi; in octep_vf_start_xmit() local
614 q_no = skb_get_queue_mapping(skb); in octep_vf_start_xmit()
615 if (q_no >= oct->num_iqs) { in octep_vf_start_xmit()
616 netdev_err(netdev, "Invalid Tx skb->queue_mapping=%d\n", q_no); in octep_vf_start_xmit()
617 q_no = q_no % oct->num_iqs; in octep_vf_start_xmit()
620 iq = oct->iq[q_no]; in octep_vf_start_xmit()

Completed in 79 milliseconds

12