| /drivers/net/ethernet/chelsio/cxgb3/ |
| A D | sge.c | 1281 qs = &pi->qs[qidx]; in t3_eth_xmit() 1836 struct sge_qset *qs = &adap->sge.qs[queue_set(skb)]; in t3_offload_tx() local 2100 skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); in rx_eth() 2164 qs->nomem = 0; in lro_add_page() 2198 skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]); in lro_add_page() 2373 fl = (len & F_RSPD_FLQ) ? &qs->fl[1] : &qs->fl[0]; in process_responses() 2454 restart_tx(qs); in process_responses() 2956 &qs->rspq.lock : &adap->sge.qs[0].rspq.lock; in sge_timer_rx() 2979 if (qs->fl[0].credits < qs->fl[0].size) in sge_timer_rx() 2981 if (qs->fl[1].credits < qs->fl[1].size) in sge_timer_rx() [all …]
|
| A D | cxgb3_main.c | 595 struct sge_qset *qs = &adap->sge.qs[i]; in ring_dbs() local 597 if (qs->adap) in ring_dbs() 608 struct sge_qset *qs = &adap->sge.qs[i]; in init_napi() local 610 if (qs->adap) in init_napi() 611 netif_napi_add(qs->netdev, &qs->napi, qs->napi.poll); in init_napi() 664 pi->qs = &adap->sge.qs[pi->first_qset]; in setup_sge_qsets() 2017 qs = &adapter->sge.qs[i]; in set_coalesce() 2620 struct sge_qset *qs = &adapter->sge.qs[qidx]; in cxgb_netpoll() local 2624 source = qs; in cxgb_netpoll() 2768 struct sge_qset *qs = &adapter->sge.qs[0]; in t3_adap_check_task() local [all …]
|
| A D | adapter.h | 68 struct sge_qset *qs; member 216 struct sge_qset qs[SGE_QSETS]; member 325 void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p);
|
| /drivers/net/ethernet/cavium/thunder/ |
| A D | nicvf_queues.c | 382 struct queue_set *qs = nic->qs; in nicvf_refill_rbdr() local 957 struct queue_set *qs = nic->qs; in nicvf_qset_config() local 960 if (!qs) { in nicvf_qset_config() 971 mbx.qs.num = qs->vnic_id; in nicvf_qset_config() 992 struct queue_set *qs = nic->qs; in nicvf_free_resources() local 1010 struct queue_set *qs = nic->qs; in nicvf_alloc_resources() local 1042 if (!qs) in nicvf_set_qset_resources() 1044 nic->qs = qs; in nicvf_set_qset_resources() 1050 qs->cq_cnt = max_t(u8, qs->rq_cnt, qs->sq_cnt); in nicvf_set_qset_resources() 1067 struct queue_set *qs = nic->qs; in nicvf_config_data_transfer() local [all …]
|
| A D | nicvf_ethtool.c | 217 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings() 282 (nic->qs->rq_cnt + nic->qs->sq_cnt); in nicvf_get_sset_count() 290 (snic->qs->rq_cnt + snic->qs->sq_cnt); in nicvf_get_sset_count() 475 struct queue_set *qs = nic->qs; in nicvf_get_ringparam() local 478 ring->rx_pending = qs->cq_len; in nicvf_get_ringparam() 480 ring->tx_pending = qs->sq_len; in nicvf_get_ringparam() 489 struct queue_set *qs = nic->qs; in nicvf_set_ringparam() local 504 if ((tx_count == qs->sq_len) && (rx_count == qs->cq_len)) in nicvf_set_ringparam() 508 qs->sq_len = rounddown_pow_of_two(tx_count); in nicvf_set_ringparam() 509 qs->cq_len = rounddown_pow_of_two(rx_count); in nicvf_set_ringparam() [all …]
|
| A D | nicvf_main.c | 319 mbx.cpi_cfg.rq_cnt = nic->qs->rq_cnt; in nicvf_config_cpi() 456 nic->snicvf[sqs]->qs->cq_cnt = in nicvf_request_sqs() 457 max(nic->snicvf[sqs]->qs->rq_cnt, nic->snicvf[sqs]->qs->sq_cnt); in nicvf_request_sqs() 667 sq = &nic->qs->sq[cqe_tx->sq_idx]; in nicvf_snd_pkt_handler() 852 struct queue_set *qs = nic->qs; in nicvf_cq_intr_handler() local 988 struct queue_set *qs = nic->qs; in nicvf_handle_qs_err() local 1288 sq = &snic->qs->sq[qid]; in nicvf_xmit() 1328 struct queue_set *qs = nic->qs; in nicvf_stop() local 1453 struct queue_set *qs = nic->qs; in nicvf_open() local 1656 struct queue_set *qs = nic->qs; in nicvf_update_stats() local [all …]
|
| A D | nic.h | 283 struct queue_set *qs; member 587 struct qs_cfg_msg qs; member
|
| A D | nicvf_queues.h | 335 void nicvf_cmp_queue_config(struct nicvf *nic, struct queue_set *qs,
|
| A D | nic_main.c | 981 (mbx.qs.num << NIC_QS_ID_SHIFT); in nic_handle_mbx_intr() 982 cfg = mbx.qs.cfg; in nic_handle_mbx_intr() 1026 nic_tx_channel_cfg(nic, mbx.qs.num, &mbx.sq); in nic_handle_mbx_intr()
|
| /drivers/soc/qcom/ |
| A D | socinfo.c | 788 struct qcom_socinfo *qs; in qcom_socinfo_probe() local 799 qs = devm_kzalloc(&pdev->dev, sizeof(*qs), GFP_KERNEL); in qcom_socinfo_probe() 800 if (!qs) in qcom_socinfo_probe() 803 qs->attr.family = "Snapdragon"; in qcom_socinfo_probe() 811 if (!qs->attr.soc_id || !qs->attr.revision) in qcom_socinfo_probe() 818 if (!qs->attr.serial_number) in qcom_socinfo_probe() 822 qs->soc_dev = soc_device_register(&qs->attr); in qcom_socinfo_probe() 823 if (IS_ERR(qs->soc_dev)) in qcom_socinfo_probe() 824 return PTR_ERR(qs->soc_dev); in qcom_socinfo_probe() 831 platform_set_drvdata(pdev, qs); in qcom_socinfo_probe() [all …]
|
| /drivers/net/ethernet/chelsio/cxgb4vf/ |
| A D | cxgb4vf_main.c | 369 for (qs = 0, msi = MSIX_IQFLINT; qs < pi->nqsets; qs++, msi++) { in name_msix_vecs() 634 int qs; in setup_sge_queues() local 636 for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { in setup_sge_queues() 667 for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { in setup_sge_queues() 709 for (qs = 0; qs < pi->nqsets; qs++) in setup_rss() 710 rss[qs] = rxq[qs].rspq.abs_id; in setup_rss() 1640 for (qs = pi->first_qset; qs < pi->first_qset + pi->nqsets; qs++) { in cxgb4vf_set_ringparam() 1814 for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) { in collect_sge_port_stats() 2053 for (qs = 0; qs < n; ++qs) \ in sge_qinfo_show() 2189 for (qs = 0; qs < n; ++qs) \ in sge_qstats_show() [all …]
|
| A D | sge.c | 2568 int qs; in t4vf_free_sge_resources() local 2570 for (qs = 0; qs < adapter->sge.ethqsets; qs++, rxq++, txq++) { in t4vf_free_sge_resources()
|
| /drivers/net/ethernet/hisilicon/hns/ |
| A D | hns_ae_adapt.c | 112 ae_handle->qs[i] = &ring_pair_cb->q; in hns_ae_get_handle() 113 ae_handle->qs[i]->rx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle() 114 ae_handle->qs[i]->tx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle() 158 ret = hns_rcb_wait_tx_ring_clean(handle->qs[i]); in hns_ae_wait_flow_down() 190 hns_rcb_ring_enable_hw(handle->qs[i], val); in hns_ae_ring_enable_all() 316 q = handle->qs[i]; in hns_ae_set_mtu() 344 hns_rcb_int_clr_hw(handle->qs[k], in hns_ae_start() 347 hns_rcbv2_int_clr_hw(handle->qs[k], in hns_ae_start() 640 queue = handle->qs[idx]; in hns_ae_update_stats() 719 hns_rcb_get_stats(handle->qs[idx], p); in hns_ae_get_stats() [all …]
|
| A D | hnae.c | 299 hnae_fini_queue(handle->qs[i]); in hnae_reinit_handle() 305 ret = hnae_init_queue(handle, handle->qs[i], handle->dev); in hnae_reinit_handle() 312 hnae_fini_queue(handle->qs[j]); in hnae_reinit_handle() 351 ret = hnae_init_queue(handle, handle->qs[i], dev); in hnae_get_handle() 364 hnae_fini_queue(handle->qs[j]); in hnae_get_handle() 378 hnae_fini_queue(h->qs[i]); in hnae_put_handle()
|
| A D | hns_dsaf_rcb.c | 36 void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag) in hns_rcb_wait_fbd_clean() argument 45 fbd_num += dsaf_read_dev(qs[i], in hns_rcb_wait_fbd_clean() 48 fbd_num += dsaf_read_dev(qs[i], in hns_rcb_wait_fbd_clean() 57 dev_err(qs[i]->handle->owner_dev, in hns_rcb_wait_fbd_clean() 61 int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs) in hns_rcb_wait_tx_ring_clean() argument 66 tail = dsaf_read_dev(&qs->tx_ring, RCB_REG_TAIL); in hns_rcb_wait_tx_ring_clean() 69 head = dsaf_read_dev(&qs->tx_ring, RCB_REG_HEAD); in hns_rcb_wait_tx_ring_clean() 77 dev_err(qs->dev->dev, "rcb wait timeout, head not equal to tail.\n"); in hns_rcb_wait_tx_ring_clean()
|
| A D | hns_dsaf_rcb.h | 133 void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag); 134 int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs);
|
| A D | hns_enet.c | 1664 ring = &h->qs[i]->rx_ring; in hns_nic_clear_all_rx_fetch() 1907 tx_bytes += h->qs[idx]->tx_ring.stats.tx_bytes; in hns_nic_get_stats64() 1908 tx_pkts += h->qs[idx]->tx_ring.stats.tx_pkts; in hns_nic_get_stats64() 1909 rx_bytes += h->qs[idx]->rx_ring.stats.rx_bytes; in hns_nic_get_stats64() 1910 rx_pkts += h->qs[idx]->rx_ring.stats.rx_pkts; in hns_nic_get_stats64() 2008 i, h->qs[i]->tx_ring.next_to_clean); in hns_nic_dump() 2010 i, h->qs[i]->tx_ring.next_to_use); in hns_nic_dump() 2012 i, h->qs[i]->rx_ring.next_to_clean); in hns_nic_dump() 2014 i, h->qs[i]->rx_ring.next_to_use); in hns_nic_dump() 2123 rd->ring = &h->qs[i]->tx_ring; in hns_nic_init_ring_data() [all …]
|
| A D | hnae.h | 561 struct hnae_queue *qs[]; /* flexible array of all queues */ member 665 ring = &h->qs[i]->rx_ring; in hnae_reinit_all_ring_desc() 680 ring = &h->qs[i]->rx_ring; in hnae_reinit_all_ring_page_off()
|
| /drivers/scsi/elx/efct/ |
| A D | efct_hw_queues.c | 199 struct sli4_queue *qs[SLI4_MAX_CQ_SET_COUNT]; in efct_hw_new_cq_set() local 217 qs[i] = cq->queue; in efct_hw_new_cq_set() 222 if (sli_cq_alloc_set(sli4, qs, num_cqs, entry_count, assefct)) { in efct_hw_new_cq_set() 315 struct sli4_queue *qs[SLI4_MAX_RQ_SET_COUNT * 2] = { NULL }; in efct_hw_new_rq_set() local 342 qs[q_count] = rq->hdr; in efct_hw_new_rq_set() 349 qs[q_count + 1] = rq->data; in efct_hw_new_rq_set() 354 if (sli_fc_rq_set_alloc(&hw->sli, num_rq_pairs, qs, in efct_hw_new_rq_set()
|
| /drivers/net/ethernet/marvell/octeontx2/nic/ |
| A D | otx2_ethtool.c | 373 struct otx2_qset *qs = &pfvf->qset; in otx2_get_ringparam() local 376 ring->rx_pending = qs->rqe_cnt ? qs->rqe_cnt : Q_COUNT(Q_SIZE_256); in otx2_get_ringparam() 378 ring->tx_pending = qs->sqe_cnt ? qs->sqe_cnt : Q_COUNT(Q_SIZE_4K); in otx2_get_ringparam() 393 struct otx2_qset *qs = &pfvf->qset; in otx2_set_ringparam() local 431 if (tx_count == qs->sqe_cnt && rx_count == qs->rqe_cnt && in otx2_set_ringparam() 439 qs->sqe_cnt = tx_count; in otx2_set_ringparam() 440 qs->rqe_cnt = rx_count; in otx2_set_ringparam()
|
| /drivers/md/ |
| A D | dm-cache-policy-smq.c | 258 struct ilist qs[MAX_LEVELS]; member 279 l_init(q->qs + i); in q_init() 303 l_add_tail(q->es, q->qs + e->level, e); in q_push() 313 l_add_head(q->es, q->qs + e->level, e); in q_push_front() 323 l_add_before(q->es, q->qs + e->level, old, e); in q_push_before() 328 l_del(q->es, q->qs + e->level, e); in q_del() 344 for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) { in q_peek() 378 for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) in __redist_pop_from() 380 l_del(q->es, q->qs + e->level, e); in __redist_pop_from() 438 l = q->qs + level; in q_redistribute() [all …]
|
| /drivers/s390/cio/ |
| A D | qdio.h | 97 : [ccq] "+&d" (_ccq), [qs] "+&d" (_queuestart) in do_sqbs() 116 : [ccq] "+&d" (_ccq), [qs] "+&d" (_queuestart), in do_eqbs()
|
| /drivers/scsi/elx/libefc_sli/ |
| A D | sli4.c | 624 qs[i]->id = i + le16_to_cpu(rsp->q_id); in sli_fc_rq_set_alloc() 625 if ((qs[i]->id & 1) == 0) in sli_fc_rq_set_alloc() 626 qs[i]->u.flag |= SLI4_QUEUE_FLAG_HDR; in sli_fc_rq_set_alloc() 628 qs[i]->u.flag &= ~SLI4_QUEUE_FLAG_HDR; in sli_fc_rq_set_alloc() 630 qs[i]->db_regaddr = db_regaddr; in sli_fc_rq_set_alloc() 639 __sli_queue_destroy(sli4, qs[i]); in sli_fc_rq_set_alloc() 817 struct sli4_queue *qs[], u32 num_cqs, in sli_cmd_cq_set_create() argument 829 n_cqe = qs[0]->dma.size / SLI4_CQE_BYTES; in sli_cmd_cq_set_create() 957 qs[i]->id = le16_to_cpu(res->q_id) + i; in sli_cq_alloc_set() 958 qs[i]->db_regaddr = db_regaddr; in sli_cq_alloc_set() [all …]
|
| /drivers/net/ethernet/brocade/bna/ |
| A D | bfi_enet.h | 494 struct bfi_enet_rxq qs; /* small/header buffers */ member
|
| /drivers/net/phy/ |
| A D | sfp.c | 562 static bool sfp_match(const char *qs, const char *str, size_t len) in sfp_match() argument 564 if (!qs) in sfp_match() 566 if (strlen(qs) != len) in sfp_match() 568 return !strncmp(qs, str, len); in sfp_match()
|