/linux-6.3-rc2/drivers/block/null_blk/ |
A D | main.c | 731 wake_up(&nq->wait); in put_tag() 739 tag = find_first_zero_bit(nq->tag_map, nq->queue_depth); in get_tag() 759 tag = get_tag(nq); in __alloc_cmd() 764 cmd->nq = nq; in __alloc_cmd() 1677 cmd->nq = nq; in null_queue_rq() 1711 kfree(nq->cmds); in cleanup_queue() 1839 nq->cmds = kcalloc(nq->queue_depth, sizeof(*cmd), GFP_KERNEL); in setup_commands() 1840 if (!nq->cmds) in setup_commands() 1843 nq->tag_map = bitmap_zalloc(nq->queue_depth, GFP_KERNEL); in setup_commands() 1844 if (!nq->tag_map) { in setup_commands() [all …]
|
A D | null_blk.h | 26 struct nullb_queue *nq; member
|
A D | zoned.c | 372 struct nullb_device *dev = cmd->nq->dev; in null_zone_write() 608 struct nullb_device *dev = cmd->nq->dev; in null_zone_mgmt() 686 dev = cmd->nq->dev; in null_process_zoned_cmd()
|
/linux-6.3-rc2/drivers/infiniband/hw/bnxt_re/ |
A D | qplib_fp.c | 159 struct bnxt_qplib_nq *nq = nq_work->nq; in bnxt_qpn_cqn_sched_task() local 167 nq->cqn_handler(nq, cq); in bnxt_qpn_cqn_sched_task() 301 struct bnxt_qplib_nq *nq = from_tasklet(nq, t, nq_tasklet); in bnxt_qplib_service_nq() local 343 if (!nq->cqn_handler(nq, (cq))) in bnxt_qplib_service_nq() 364 if (!nq->srqn_handler(nq, in bnxt_qplib_service_nq() 417 free_irq(nq->msix_vec, nq); in bnxt_qplib_nq_stop_irq() 457 rc = request_irq(nq->msix_vec, bnxt_qplib_nq_irq, 0, nq->name, nq); in bnxt_qplib_nq_start_irq() 463 rc = irq_set_affinity_hint(nq->msix_vec, &nq->mask); in bnxt_qplib_nq_start_irq() 548 bnxt_qplib_free_hwq(nq->res, &nq->hwq); in bnxt_qplib_free_nq() 1934 nq_work->nq = qp->scq->nq; in bnxt_qplib_post_send() [all …]
|
A D | main.c | 251 struct bnxt_qplib_nq *nq; in bnxt_re_stop_irq() local 255 nq = &rdev->nq[indx - 1]; in bnxt_re_stop_irq() 256 bnxt_qplib_nq_stop_irq(nq, false); in bnxt_re_stop_irq() 267 struct bnxt_qplib_nq *nq; in bnxt_re_start_irq() local 289 nq = &rdev->nq[indx - 1]; in bnxt_re_start_irq() 812 bnxt_qplib_free_nq(&rdev->nq[i]); in bnxt_re_free_nq_res() 813 rdev->nq[i].res = NULL; in bnxt_re_free_nq_res() 858 struct bnxt_qplib_nq *nq; in bnxt_re_alloc_res() local 860 nq = &rdev->nq[i]; in bnxt_re_alloc_res() 870 rattr.pages = nq->hwq.pbl[rdev->nq[i].hwq.level].pg_count; in bnxt_re_alloc_res() [all …]
|
A D | qplib_fp.h | 404 struct bnxt_qplib_nq *nq; member 466 typedef int (*cqn_handler_t)(struct bnxt_qplib_nq *nq, 468 typedef int (*srqn_handler_t)(struct bnxt_qplib_nq *nq, 491 struct bnxt_qplib_nq *nq; member 495 void bnxt_qplib_nq_stop_irq(struct bnxt_qplib_nq *nq, bool kill); 496 void bnxt_qplib_disable_nq(struct bnxt_qplib_nq *nq); 497 int bnxt_qplib_nq_start_irq(struct bnxt_qplib_nq *nq, int nq_indx, 499 int bnxt_qplib_enable_nq(struct pci_dev *pdev, struct bnxt_qplib_nq *nq, 540 void bnxt_qplib_free_nq(struct bnxt_qplib_nq *nq); 541 int bnxt_qplib_alloc_nq(struct bnxt_qplib_res *res, struct bnxt_qplib_nq *nq);
|
A D | ib_verbs.c | 1563 nq = qplib_srq->cq->nq; in bnxt_re_destroy_srq() 1567 if (nq) in bnxt_re_destroy_srq() 1568 nq->budget--; in bnxt_re_destroy_srq() 1650 nq = &rdev->nq[0]; in bnxt_re_create_srq() 1676 if (nq) in bnxt_re_create_srq() 1677 nq->budget++; in bnxt_re_create_srq() 2799 nq = cq->qplib_cq.nq; in bnxt_re_destroy_cq() 2805 nq->budget--; in bnxt_re_destroy_cq() 2873 nq = &rdev->nq[nq_alloc_cnt % (rdev->num_msix - 1)]; in bnxt_re_create_cq() 2876 cq->qplib_cq.nq = nq; in bnxt_re_create_cq() [all …]
|
A D | bnxt_re.h | 148 struct bnxt_qplib_nq nq[BNXT_RE_MAX_MSIX]; member
|
/linux-6.3-rc2/lib/crypto/ |
A D | curve25519-hacl64.c | 551 point_swap_conditional(nq, nqpq, bit0); in ladder_smallloop_cmult_small_loop_step() 552 addanddouble_fmonty(nq2, nqpq2, nq, nqpq, q); in ladder_smallloop_cmult_small_loop_step() 572 ladder_smallloop_cmult_small_loop_double_step(nq, nqpq, nq2, in ladder_smallloop_cmult_small_loop() 585 ladder_smallloop_cmult_small_loop(nq, nqpq, nq2, nqpq2, q, in ladder_bigloop_cmult_big_loop() 593 u64 *nq = point_buf; in ladder_cmult() local 598 nq[0] = 1; in ladder_cmult() 600 point_copy(result, nq); in ladder_cmult() 778 u64 *nq = buf; in curve25519_generic() local 779 u64 *x = nq; in curve25519_generic() 781 ladder_cmult(nq, scalar, q); in curve25519_generic() [all …]
|
/linux-6.3-rc2/fs/xfs/ |
A D | xfs_trans_dquot.c | 78 struct xfs_dqtrx *oq, *nq; in xfs_trans_dup_dqinfo() local 97 nq = &nqa[i]; in xfs_trans_dup_dqinfo() 102 nq->qt_dquot = oq->qt_dquot; in xfs_trans_dup_dqinfo() 103 nq->qt_bcount_delta = nq->qt_icount_delta = 0; in xfs_trans_dup_dqinfo() 104 nq->qt_rtbcount_delta = 0; in xfs_trans_dup_dqinfo() 109 nq->qt_blk_res = oq->qt_blk_res - blk_res_used; in xfs_trans_dup_dqinfo() 112 nq->qt_rtblk_res = oq->qt_rtblk_res - in xfs_trans_dup_dqinfo() 116 nq->qt_ino_res = oq->qt_ino_res - oq->qt_ino_res_used; in xfs_trans_dup_dqinfo()
|
/linux-6.3-rc2/arch/x86/crypto/ |
A D | curve25519-x86_64.c | 977 u64 *nq = p01_tmp1; in point_add_and_double() local 981 u64 *x2 = nq; in point_add_and_double() 982 u64 *z2 = nq + (u32)4U; in point_add_and_double() 1024 fmul2(nq, dc1, ab1, tmp2); in point_add_and_double() 1028 static void point_double(u64 *nq, u64 *tmp1, u64 *tmp2) in point_double() argument 1030 u64 *x2 = nq; in point_double() 1031 u64 *z2 = nq + (u32)4U; in point_double() 1048 fmul2(nq, dc, ab, tmp2); in point_double()
|
/linux-6.3-rc2/drivers/net/ethernet/engleder/ |
A D | tsnep_main.c | 634 struct netdev_queue *nq; in tsnep_tx_poll() local 640 __netif_tx_lock(nq, smp_processor_id()); in tsnep_tx_poll() 704 netif_tx_queue_stopped(nq)) { in tsnep_tx_poll() 705 netif_tx_wake_queue(nq); in tsnep_tx_poll() 708 __netif_tx_unlock(nq); in tsnep_tx_poll() 716 struct netdev_queue *nq; in tsnep_tx_pending() local 730 __netif_tx_unlock(nq); in tsnep_tx_pending() 1639 struct netdev_queue *nq; in tsnep_netdev_xdp_xmit() local 1650 __netif_tx_lock(nq, cpu); in tsnep_netdev_xdp_xmit() 1661 txq_trans_cond_update(nq); in tsnep_netdev_xdp_xmit() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/chelsio/cxgb4/ |
A D | cxgb4_uld.c | 111 unsigned int nq = rxq_info->nrxq + rxq_info->nciq; in alloc_uld_rxqs() local 125 for (i = 0; i < nq; i++, q++) { in alloc_uld_rxqs() 403 int nq = txq_info->ntxq; in free_sge_txq_uld() local 406 for (i = 0; i < nq; i++) { in free_sge_txq_uld() 426 int nq = txq_info->ntxq; in alloc_sge_txq_uld() local 429 j = nq / adap->params.nports; in alloc_sge_txq_uld() 430 for (i = 0; i < nq; i++) { in alloc_sge_txq_uld()
|
/linux-6.3-rc2/drivers/net/ethernet/marvell/ |
A D | mvneta.c | 1922 netif_tx_wake_queue(nq); in mvneta_txq_done() 2171 struct netdev_queue *nq; in mvneta_xdp_xmit_back() local 2184 __netif_tx_lock(nq, cpu); in mvneta_xdp_xmit_back() 2199 __netif_tx_unlock(nq); in mvneta_xdp_xmit_back() 2213 struct netdev_queue *nq; in mvneta_xdp_xmit() local 2225 __netif_tx_lock(nq, cpu); in mvneta_xdp_xmit() 2237 __netif_tx_unlock(nq); in mvneta_xdp_xmit() 2916 netif_tx_stop_queue(nq); in mvneta_tx() 2959 struct netdev_queue *nq; in mvneta_tx_done_gbe() local 2966 __netif_tx_lock(nq, cpu); in mvneta_tx_done_gbe() [all …]
|
A D | mv643xx_eth.c | 498 if (netif_tx_queue_stopped(nq)) { in txq_maybe_wake() 499 __netif_tx_lock(nq, smp_processor_id()); in txq_maybe_wake() 501 netif_tx_wake_queue(nq); in txq_maybe_wake() 502 __netif_tx_unlock(nq); in txq_maybe_wake() 998 struct netdev_queue *nq; in mv643xx_eth_xmit() local 1002 nq = netdev_get_tx_queue(dev, queue); in mv643xx_eth_xmit() 1021 netif_tx_stop_queue(nq); in mv643xx_eth_xmit() 1039 __netif_tx_lock(nq, smp_processor_id()); in txq_kick() 1052 __netif_tx_unlock(nq); in txq_kick() 1063 __netif_tx_lock_bh(nq); in txq_reclaim() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/chelsio/cxgb4vf/ |
A D | t4vf_hw.c | 1272 int nq = min(n, 32); in t4vf_config_rss_range() local 1279 cmd.niqid = cpu_to_be16(nq); in t4vf_config_rss_range() 1285 start += nq; in t4vf_config_rss_range() 1286 n -= nq; in t4vf_config_rss_range() 1293 while (nq > 0) { in t4vf_config_rss_range() 1302 int nqbuf = min(3, nq); in t4vf_config_rss_range() 1304 nq -= nqbuf; in t4vf_config_rss_range()
|
/linux-6.3-rc2/arch/s390/mm/ |
A D | pgtable.c | 812 unsigned char key, bool nq) in set_guest_storage_key() argument 867 page_set_storage_key(paddr, skey, !nq); in set_guest_storage_key() 891 bool nq, bool mr, bool mc) in cond_set_guest_storage_key() argument 910 rc = set_guest_storage_key(current->mm, addr, key, nq); in cond_set_guest_storage_key()
|
/linux-6.3-rc2/drivers/net/ |
A D | tap.c | 190 struct tap_queue *nq; in tap_disable_queue() local 201 nq = rtnl_dereference(tap->taps[tap->numvtaps - 1]); in tap_disable_queue() 202 nq->queue_index = index; in tap_disable_queue() 204 rcu_assign_pointer(tap->taps[index], nq); in tap_disable_queue()
|
/linux-6.3-rc2/drivers/net/ethernet/freescale/ |
A D | fec_main.c | 886 struct netdev_queue *nq; in fec_enet_start_xmit() local 891 nq = netdev_get_tx_queue(ndev, queue); in fec_enet_start_xmit() 902 netif_tx_stop_queue(nq); in fec_enet_start_xmit() 1382 struct netdev_queue *nq; in fec_enet_tx_queue() local 1390 nq = netdev_get_tx_queue(ndev, queue_id); in fec_enet_tx_queue() 1470 if (netif_tx_queue_stopped(nq)) { in fec_enet_tx_queue() 1473 netif_tx_wake_queue(nq); in fec_enet_tx_queue() 3851 struct netdev_queue *nq; in fec_enet_xdp_xmit() local 3857 nq = netdev_get_tx_queue(fep->netdev, queue); in fec_enet_xdp_xmit() 3859 __netif_tx_lock(nq, cpu); in fec_enet_xdp_xmit() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/intel/igc/ |
A D | igc_main.c | 2366 struct netdev_queue *nq; in igc_xdp_xmit_back() local 2374 nq = txring_txq(ring); in igc_xdp_xmit_back() 2376 __netif_tx_lock(nq, cpu); in igc_xdp_xmit_back() 2378 __netif_tx_unlock(nq); in igc_xdp_xmit_back() 2445 struct netdev_queue *nq; in igc_finalize_xdp() local 2450 nq = txring_txq(ring); in igc_finalize_xdp() 2454 __netif_tx_unlock(nq); in igc_finalize_xdp() 2820 __netif_tx_unlock(nq); in igc_xdp_xmit_zc() 6291 struct netdev_queue *nq; in igc_xdp_xmit() local 6302 nq = txring_txq(ring); in igc_xdp_xmit() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/aquantia/atlantic/ |
A D | aq_nic.c | 837 struct netdev_queue *nq; in aq_nic_xmit_xdpf() local 848 nq = netdev_get_tx_queue(ndev, tx_ring->idx); in aq_nic_xmit_xdpf() 849 __netif_tx_lock(nq, cpu); in aq_nic_xmit_xdpf() 862 __netif_tx_unlock(nq); in aq_nic_xmit_xdpf()
|
/linux-6.3-rc2/net/sched/ |
A D | sch_api.c | 320 struct netdev_queue *nq; in qdisc_lookup_rcu() local 329 nq = dev_ingress_queue_rcu(dev); in qdisc_lookup_rcu() 330 if (nq) in qdisc_lookup_rcu() 331 q = qdisc_match_from_root(nq->qdisc_sleeping, handle); in qdisc_lookup_rcu()
|
/linux-6.3-rc2/arch/s390/kvm/ |
A D | priv.c | 1041 bool mr = false, mc = false, nq; in handle_pfmf() local 1071 nq = vcpu->run->s.regs.gprs[reg1] & PFMF_NQ; in handle_pfmf() 1123 key, NULL, nq, mr, mc); in handle_pfmf()
|
/linux-6.3-rc2/drivers/net/ethernet/stmicro/stmmac/ |
A D | stmmac_main.c | 2419 txq_trans_cond_update(nq); in stmmac_xdp_xmit_zc() 4787 struct netdev_queue *nq; in stmmac_xdp_xmit_back() local 4795 nq = netdev_get_tx_queue(priv->dev, queue); in stmmac_xdp_xmit_back() 4797 __netif_tx_lock(nq, cpu); in stmmac_xdp_xmit_back() 4799 txq_trans_cond_update(nq); in stmmac_xdp_xmit_back() 4805 __netif_tx_unlock(nq); in stmmac_xdp_xmit_back() 6412 struct netdev_queue *nq; in stmmac_xdp_xmit() local 6423 nq = netdev_get_tx_queue(priv->dev, queue); in stmmac_xdp_xmit() 6425 __netif_tx_lock(nq, cpu); in stmmac_xdp_xmit() 6427 txq_trans_cond_update(nq); in stmmac_xdp_xmit() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/marvell/mvpp2/ |
A D | mvpp2_main.c | 2837 struct netdev_queue *nq = netdev_get_tx_queue(port->dev, txq->log_id); in mvpp2_txq_done() local 2850 if (netif_tx_queue_stopped(nq)) in mvpp2_txq_done() 2852 netif_tx_wake_queue(nq); in mvpp2_txq_done() 3350 static void mvpp2_isr_handle_ptp_queue(struct mvpp2_port *port, int nq) in mvpp2_isr_handle_ptp_queue() argument 3360 if (nq) in mvpp2_isr_handle_ptp_queue() 3363 queue = &port->tx_hwtstamp_queue[nq]; in mvpp2_isr_handle_ptp_queue() 3620 struct netdev_queue *nq; in mvpp2_xdp_finish_tx() local 3624 nq = netdev_get_tx_queue(port->dev, txq_id); in mvpp2_xdp_finish_tx() 3636 netif_tx_stop_queue(nq); in mvpp2_xdp_finish_tx() 4404 struct netdev_queue *nq = netdev_get_tx_queue(dev, txq_id); in mvpp2_tx() local [all …]
|