Home
last modified time | relevance | path

Searched refs:sde (Results 1 – 25 of 31) sorted by relevance

12

/linux-6.3-rc2/drivers/infiniband/hw/hfi1/
A Dsdma.c530 return sde->tx_ring[sde->tx_head & sde->sdma_mask]; in get_txhead()
547 head = sde->descq_head & sde->sdma_mask; in sdma_flush_descq()
555 sde->tx_ring[sde->tx_head++ & sde->sdma_mask] = NULL; in sdma_flush_descq()
849 sde = map->sde[selector & map->mask]; in sdma_select_user_engine()
865 map->sde[map->ctr + i] = map->sde[i]; in sdma_populate_sde_map()
875 if (map->sde[i] == sde) { in sdma_cleanup_sde_map()
954 rht_node->map[vl]->sde[0] = sde; in sdma_set_cpu_to_sde_map()
1395 sde->imask = sde->int_mask | sde->progress_mask | in sdma_init()
1812 sde->tx_ring[sde->tx_head++ & sde->sdma_mask] = NULL; in sdma_make_progress()
2304 WARN_ON_ONCE(sde->tx_ring[sde->tx_tail & sde->sdma_mask]); in submit_tx()
[all …]
A Dvnic_sdma.c55 sde->dd, in build_vnic_ulp_payload()
66 ret = sdma_txadd_page(sde->dd, in build_vnic_ulp_payload()
106 sde->dd, in build_vnic_tx_desc()
114 ret = build_vnic_ulp_payload(sde, tx); in build_vnic_tx_desc()
130 struct sdma_engine *sde = vnic_sdma->sde; in hfi1_vnic_send_dma() local
137 if (unlikely(!sde || !sdma_running(sde))) in hfi1_vnic_send_dma()
150 ret = build_vnic_tx_desc(sde, tx, pbc); in hfi1_vnic_send_dma()
194 write_seqlock(&sde->waitlock); in hfi1_vnic_sdma_sleep()
195 if (sdma_progress(sde, seq, txreq)) { in hfi1_vnic_sdma_sleep()
196 write_sequnlock(&sde->waitlock); in hfi1_vnic_sdma_sleep()
[all …]
A Dsdma.h393 return sde->descq_tail == sde->descq_head; in sdma_empty()
398 return sde->descq_cnt - in sdma_descq_freecnt()
399 (sde->descq_tail - in sdma_descq_freecnt()
400 READ_ONCE(sde->descq_head)) - 1; in sdma_descq_freecnt()
405 return sde->descq_cnt - sdma_descq_freecnt(sde); in sdma_descq_inprocess()
860 sde->desc_avail = sdma_descq_freecnt(sde); in sdma_progress()
861 if (tx->num_desc > sde->desc_avail) in sdma_progress()
945 struct sdma_engine *sde[]; member
987 struct sdma_engine *sde) in sdma_engine_progress_schedule() argument
989 if (!sde || sdma_descq_inprocess(sde) < (sde->descq_cnt / 8)) in sdma_engine_progress_schedule()
[all …]
A Dtrace_tx.h333 TP_ARGS(sde, aidx),
433 TP_ARGS(sde, sn),
454 TP_ARGS(sde, sn)
459 TP_ARGS(sde, sn)
630 __entry->sde = sde;
702 __field(u8, sde)
713 __entry->sde = sde;
724 __entry->sde,
915 __entry->sde = txq->sde;
934 __entry->sde ? __entry->sde->this_idx : 0,
[all …]
A Dmsix.c185 int msix_request_sdma_irq(struct sdma_engine *sde) in msix_request_sdma_irq() argument
191 sde->dd->unit, sde->this_idx); in msix_request_sdma_irq()
192 nr = msix_request_irq(sde->dd, sde, sdma_interrupt, NULL, in msix_request_sdma_irq()
196 sde->msix_intr = nr; in msix_request_sdma_irq()
197 remap_sdma_interrupts(sde->dd, sde->this_idx, nr); in msix_request_sdma_irq()
259 struct sdma_engine *sde = &dd->per_sdma[i]; in msix_request_irqs() local
261 ret = msix_request_sdma_irq(sde); in msix_request_irqs()
264 enable_sdma_srcs(sde->dd, i); in msix_request_irqs()
A Dsysfs.c571 ssize_t (*show)(struct sdma_engine *sde, char *buf);
572 ssize_t (*store)(struct sdma_engine *sde, const char *buf, size_t cnt);
579 struct sdma_engine *sde = in sde_show() local
585 return sde_attr->show(sde, buf); in sde_show()
593 struct sdma_engine *sde = in sde_store() local
602 return sde_attr->store(sde, buf, count); in sde_store()
620 return sdma_get_cpu_to_sde_map(sde, buf); in sde_show_cpu_to_sde_map()
623 static ssize_t sde_store_cpu_to_sde_map(struct sdma_engine *sde, in sde_store_cpu_to_sde_map() argument
626 return sdma_set_cpu_to_sde_map(sde, buf, count); in sde_store_cpu_to_sde_map()
629 static ssize_t sde_show_vl(struct sdma_engine *sde, char *buf) in sde_show_vl() argument
[all …]
A Dqp.c26 struct sdma_engine *sde,
437 struct sdma_engine *sde, in iowait_sleep() argument
461 write_seqlock(&sde->waitlock); in iowait_sleep()
472 &sde->dmawait); in iowait_sleep()
477 write_sequnlock(&sde->waitlock); in iowait_sleep()
487 write_sequnlock(&sde->waitlock); in iowait_sleep()
542 struct sdma_engine *sde; in qp_to_sdma_engine() local
553 return sde; in qp_to_sdma_engine()
603 struct sdma_engine *sde; in qp_iter_print() local
649 sde, in qp_iter_print()
[all …]
A Dipoib_tx.c126 tx->txq->sde->this_idx); in hfi1_ipoib_free_tx()
388 txq->sde = in hfi1_ipoib_send_dma_common()
409 ret = sdma_send_txlist(txq->sde, in hfi1_ipoib_submit_tx_list()
443 ret = sdma_send_txreq(txq->sde, in hfi1_ipoib_submit_tx()
621 write_seqlock(&sde->waitlock); in hfi1_ipoib_sdma_sleep()
625 write_sequnlock(&sde->waitlock); in hfi1_ipoib_sdma_sleep()
643 write_sequnlock(&sde->waitlock); in hfi1_ipoib_sdma_sleep()
647 write_sequnlock(&sde->waitlock); in hfi1_ipoib_sdma_sleep()
717 txq->sde = NULL; in hfi1_ipoib_txreq_init()
849 txq->sde, in hfi1_ipoib_tx_timeout()
[all …]
A Dverbs_txreq.h23 struct sdma_engine *sde; member
50 tx->sde = priv->s_sde; in get_txreq()
A Duser_sdma.c62 struct sdma_engine *sde,
85 struct sdma_engine *sde, in defer_packet_queue() argument
94 write_seqlock(&sde->waitlock); in defer_packet_queue()
95 trace_hfi1_usdma_defer(pq, sde, &pq->busy); in defer_packet_queue()
96 if (sdma_progress(sde, seq, txreq)) in defer_packet_queue()
105 pq->busy.lock = &sde->waitlock; in defer_packet_queue()
109 write_sequnlock(&sde->waitlock); in defer_packet_queue()
112 write_sequnlock(&sde->waitlock); in defer_packet_queue()
540 if (!req->sde || !sdma_running(req->sde)) { in hfi1_user_sdma_process_request()
897 ret = sdma_send_txlist(req->sde, in user_sdma_send_pkts()
[all …]
A Daffinity.c743 struct sdma_engine *sde = msix->arg; in hfi1_update_sdma_affinity() local
744 struct hfi1_devdata *dd = sde->dd; in hfi1_update_sdma_affinity()
749 if (cpu > num_online_cpus() || cpu == sde->cpu) in hfi1_update_sdma_affinity()
757 old_cpu = sde->cpu; in hfi1_update_sdma_affinity()
758 sde->cpu = cpu; in hfi1_update_sdma_affinity()
763 sde->this_idx, cpu); in hfi1_update_sdma_affinity()
841 struct sdma_engine *sde = NULL; in get_irq_affinity() local
853 sde = (struct sdma_engine *)msix->arg; in get_irq_affinity()
854 scnprintf(extra, 64, "engine %u", sde->this_idx); in get_irq_affinity()
904 sde->cpu = cpu; in get_irq_affinity()
A Dmsix.h17 int msix_request_sdma_irq(struct sdma_engine *sde);
A Diowait.h101 struct sdma_engine *sde,
133 int (*sleep)(struct sdma_engine *sde,
A Dvnic.h44 struct sdma_engine *sde; member
A Dipoib.h111 struct sdma_engine *sde; member
A Diowait.c44 int (*sleep)(struct sdma_engine *sde, in iowait_init() argument
A Dvnic_main.c284 struct sdma_engine *sde; in hfi1_vnic_select_queue() local
287 sde = sdma_select_engine_vl(vinfo->dd, mdata->entropy, mdata->vl); in hfi1_vnic_select_queue()
288 return sde->this_idx; in hfi1_vnic_select_queue()
A Dverbs.c655 struct sdma_engine *sde, in build_verbs_ulp_payload() argument
670 sde->dd, in build_verbs_ulp_payload()
719 struct sdma_engine *sde, in build_verbs_tx_desc() argument
753 sde->dd, in build_verbs_tx_desc()
774 ret = build_verbs_ulp_payload(sde, length, tx); in build_verbs_tx_desc()
781 ret = sdma_txadd_daddr(sde->dd, &tx->txreq, in build_verbs_tx_desc()
782 sde->dd->sdma_pad_phys, extra_bytes); in build_verbs_tx_desc()
848 ret = build_verbs_tx_desc(tx->sde, len, tx, ahg_info, pbc); in hfi1_verbs_send_dma()
852 ret = sdma_send_txreq(tx->sde, ps->wait, &tx->txreq, ps->pkts_sent); in hfi1_verbs_send_dma()
A Duser_sdma.h145 struct sdma_engine *sde; member
/linux-6.3-rc2/arch/arm64/boot/dts/qcom/
A Dsm8550-mtp.dts492 sde_dsi_active: sde-dsi-active-state {
499 sde_dsi_suspend: sde-dsi-suspend-state {
506 sde_te_active: sde-te-active-state {
513 sde_te_suspend: sde-te-suspend-state {
A Dsdm845-shift-axolotl.dts578 sde_dsi_active: sde-dsi-active-state {
585 sde_dsi_suspend: sde-dsi-suspend-state {
592 sde_te_active: sde-te-active-state {
599 sde_te_suspend: sde-te-suspend-state {
A Dsdm845-xiaomi-polaris.dts607 sde_dsi_active: sde-dsi-active-state {
614 sde_dsi_suspend: sde-dsi-suspend-state {
/linux-6.3-rc2/drivers/net/fddi/skfp/
A Dsmt.c74 static void smt_fill_sde(struct s_smc *smc, struct smt_p_sde *sde);
938 smt_fill_sde(smc,&nif->sde) ; /* set station descriptor */ in smt_send_nif()
1030 smt_fill_sde(smc,&sif->sde) ; /* set station descriptor */ in smt_send_sif_config()
1157 static void smt_fill_sde(struct s_smc *smc, struct smt_p_sde *sde) in smt_fill_sde() argument
1159 SMTSETPARA(sde,SMT_P_SDE) ; in smt_fill_sde()
1160 sde->sde_non_master = smc->mib.fddiSMTNonMaster_Ct ; in smt_fill_sde()
1161 sde->sde_master = smc->mib.fddiSMTMaster_Ct ; in smt_fill_sde()
1162 sde->sde_mac_count = NUMMACS ; /* only 1 MAC */ in smt_fill_sde()
1164 sde->sde_type = SMT_SDE_CONCENTRATOR ; in smt_fill_sde()
1166 sde->sde_type = SMT_SDE_STATION ; in smt_fill_sde()
/linux-6.3-rc2/drivers/net/fddi/skfp/h/
A Dsmt.h727 struct smt_p_sde sde ; /* station descriptor */ member
740 struct smt_p_sde sde ; /* station descriptor */ member
/linux-6.3-rc2/arch/mips/include/asm/octeon/
A Dcvmx-pciercx-defs.h264 __BITFIELD_FIELD(uint32_t sde:1,

Completed in 50 milliseconds

12