| /drivers/infiniband/hw/hfi1/ |
| A D | init.c | 179 rcd->dd->rcd[rcd->ctxt] = NULL; in hfi1_rcd_free() 182 hfi1_free_ctxtdata(rcd->dd, rcd); in hfi1_rcd_free() 238 dd->rcd[ctxt] = rcd; in allocate_rcd_index() 291 rcd = dd->rcd[ctxt]; in hfi1_rcd_get_by_index() 337 rcd->do_interrupt = rcd->slow_handler; in hfi1_create_ctxtdata() 404 rcd->ctxt, rcd->egrbufs.count); in hfi1_create_ctxtdata() 436 rcd->ctxt, rcd->egrbufs.size); in hfi1_create_ctxtdata() 1093 rcd->rcvhdrq, rcd->rcvhdrq_dma); in hfi1_free_ctxtdata() 1786 set_hdrq_regs(rcd->dd, rcd->ctxt, rcd->rcvhdrqentsize, in hfi1_create_rcvhdrq() 1922 rcd->ctxt, rcd->egrbufs.alloced, in hfi1_setup_eagerbufs() [all …]
|
| A D | aspm.c | 172 struct hfi1_ctxtdata *rcd = timer_container_of(rcd, t, aspm_timer); in aspm_ctx_timer_function() local 176 aspm_enable_dec(rcd->dd); in aspm_ctx_timer_function() 193 if (rcd) { in aspm_disable_all() 198 hfi1_rcd_put(rcd); in aspm_disable_all() 220 if (rcd) { in aspm_enable_all() 225 hfi1_rcd_put(rcd); in aspm_enable_all() 234 rcd->aspm_intr_supported = rcd->dd->aspm_supported && in aspm_ctx_init() 236 rcd->ctxt < rcd->dd->first_dyn_alloc_ctxt; in aspm_ctx_init() 249 if (rcd) in aspm_init() 250 aspm_ctx_init(rcd); in aspm_init() [all …]
|
| A D | driver.c | 373 packet->rcd = rcd; in init_packet() 504 struct hfi1_ctxtdata *rcd = packet->rcd; in init_ps_mdata() local 506 mdata->rcd = rcd; in init_ps_mdata() 572 struct hfi1_ctxtdata *rcd = packet->rcd; in __prescan_rxq() local 638 struct hfi1_ctxtdata *rcd = packet->rcd; in process_rcv_qp_work() local 942 rcd->do_interrupt = rcd->slow_handler; in set_all_slowpath() 1535 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_setup_bypass_packet() local 1608 struct hfi1_ctxtdata *rcd = packet->rcd; in show_eflags_errs() local 1626 struct hfi1_ctxtdata *rcd = packet->rcd; in handle_eflags() local 1628 rcv_hdrerr(rcd, rcd->ppd, packet); in handle_eflags() [all …]
|
| A D | exp_rcv.c | 26 hfi1_exp_tid_set_init(&rcd->tid_used_list); in hfi1_exp_tid_group_init() 36 struct hfi1_devdata *dd = rcd->dd; in hfi1_alloc_ctxt_rcv_groups() 43 rcd->groups = in hfi1_alloc_ctxt_rcv_groups() 45 GFP_KERNEL, rcd->numa_id); in hfi1_alloc_ctxt_rcv_groups() 46 if (!rcd->groups) in hfi1_alloc_ctxt_rcv_groups() 48 tidbase = rcd->expected_base; in hfi1_alloc_ctxt_rcv_groups() 50 grp = &rcd->groups[i]; in hfi1_alloc_ctxt_rcv_groups() 73 kfree(rcd->groups); in hfi1_free_ctxt_rcv_groups() 74 rcd->groups = NULL; in hfi1_free_ctxt_rcv_groups() 75 hfi1_exp_tid_group_init(rcd); in hfi1_free_ctxt_rcv_groups() [all …]
|
| A D | msix.c | 131 int nr = msix_request_irq(rcd->dd, rcd, handler, thread, in msix_request_rcd_irq_common() 132 rcd->is_vnic ? IRQ_NETDEVCTXT : IRQ_RCVCTXT, in msix_request_rcd_irq_common() 141 rcd->ireg = (IS_RCVAVAIL_START + rcd->ctxt) / 64; in msix_request_rcd_irq_common() 142 rcd->imask = ((u64)1) << ((IS_RCVAVAIL_START + rcd->ctxt) % 64); in msix_request_rcd_irq_common() 143 rcd->msix_intr = nr; in msix_request_rcd_irq_common() 144 remap_intr(rcd->dd, IS_RCVAVAIL_START + rcd->ctxt, nr); in msix_request_rcd_irq_common() 159 rcd->dd->unit, rcd->ctxt); in msix_request_rcd_irq() 175 rcd->dd->unit, rcd->ctxt); in msix_netdev_request_rcd_irq() 270 if (rcd) in msix_request_irqs() 271 ret = msix_request_rcd_irq(rcd); in msix_request_irqs() [all …]
|
| A D | hfi.h | 1465 return rcd->head; in hfi1_rcd_head() 1475 rcd->head = head; in hfi1_set_rcd_head() 1481 return (__le32 *)rcd->rcvhdrq + rcd->head + rcd->rhf_offset; in get_rhf_addr() 1511 return rcd->seq_cnt; in hfi1_seq_cnt() 1522 rcd->seq_cnt = cnt; in hfi1_set_seq_cnt() 1546 rcd->seq_cnt = hfi1_seq_incr_wrap(rcd->seq_cnt); in hfi1_seq_incr() 1574 return rcd->do_interrupt == rcd->slow_handler; in hfi1_is_slowpath() 1586 return rcd->do_interrupt == rcd->fast_handler; in hfi1_is_fastpath() 1595 if (unlikely(!rcd)) in hfi1_set_fast() 1598 rcd->do_interrupt = rcd->fast_handler; in hfi1_set_fast() [all …]
|
| A D | tid_rdma.c | 306 hfi1_set_ctxt_jkey(rcd->dd, rcd, rcd->jkey); in hfi1_kern_exp_rcv_init() 750 write_uctxt_csr(rcd->dd, rcd->ctxt, in kern_set_hw_flow() 809 fqp = first_qp(rcd, &rcd->flow_queue); in hfi1_kern_setup_hw_flow() 837 fqp = first_qp(rcd, &rcd->flow_queue); in hfi1_kern_clear_hw_flow() 1466 struct hfi1_ctxtdata *rcd = req->rcd; in hfi1_kern_exp_rcv_setup() local 1530 fqp = first_qp(rcd, &rcd->rarr_queue); in hfi1_kern_exp_rcv_setup() 1557 struct hfi1_ctxtdata *rcd = req->rcd; in hfi1_kern_exp_rcv_clear() local 1574 fqp = first_qp(rcd, &rcd->rarr_queue); in hfi1_kern_exp_rcv_clear() 1670 req->rcd = qpriv->rcd; in hfi1_init_trdma_req() 3468 struct hfi1_ctxtdata *rcd = qpriv->rcd; in hfi1_tid_write_alloc_resources() local [all …]
|
| A D | netdev_rx.c | 209 hfi1_rcd_get(rxq->rcd); in hfi1_netdev_rxq_init() 213 i, rxq->rcd->ctxt); in hfi1_netdev_rxq_init() 232 if (rxq->rcd) { in hfi1_netdev_rxq_init() 234 hfi1_rcd_put(rxq->rcd); in hfi1_netdev_rxq_init() 235 rxq->rcd = NULL; in hfi1_netdev_rxq_init() 254 hfi1_rcd_put(rxq->rcd); in hfi1_netdev_rxq_deinit() 255 rxq->rcd = NULL; in hfi1_netdev_rxq_deinit() 271 rxq->rcd->ctxt); in enable_queues() 275 rxq->rcd); in enable_queues() 289 rxq->rcd->ctxt); in disable_queues() [all …]
|
| A D | intr.c | 171 void handle_user_interrupt(struct hfi1_ctxtdata *rcd) in handle_user_interrupt() argument 173 struct hfi1_devdata *dd = rcd->dd; in handle_user_interrupt() 177 if (bitmap_empty(rcd->in_use_ctxts, HFI1_MAX_SHARED_CTXTS)) in handle_user_interrupt() 180 if (test_and_clear_bit(HFI1_CTXT_WAITING_RCV, &rcd->event_flags)) { in handle_user_interrupt() 181 wake_up_interruptible(&rcd->wait); in handle_user_interrupt() 182 hfi1_rcvctrl(dd, HFI1_RCVCTRL_INTRAVAIL_DIS, rcd); in handle_user_interrupt() 184 &rcd->event_flags)) { in handle_user_interrupt() 185 rcd->urgent++; in handle_user_interrupt() 186 wake_up_interruptible(&rcd->wait); in handle_user_interrupt()
|
| A D | trace_rx.h | 27 TP_STRUCT__entry(DD_DEV_ENTRY(packet->rcd->dd) 36 TP_fast_assign(DD_DEV_ASSIGN(packet->rcd->dd); 38 __entry->ctxt = packet->rcd->ctxt; 59 TP_PROTO(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd), 60 TP_ARGS(dd, rcd), 67 __entry->ctxt = rcd->ctxt; 68 __entry->slow_path = hfi1_is_slowpath(rcd); 69 __entry->dma_rtail = get_dma_rtail_setting(rcd);
|
| A D | exp_rcv.h | 151 hfi1_tid_group_to_idx(struct hfi1_ctxtdata *rcd, struct tid_group *grp) in hfi1_tid_group_to_idx() argument 153 return grp - &rcd->groups[0]; in hfi1_tid_group_to_idx() 162 hfi1_idx_to_tid_group(struct hfi1_ctxtdata *rcd, u16 idx) in hfi1_idx_to_tid_group() argument 164 return &rcd->groups[idx]; in hfi1_idx_to_tid_group() 167 int hfi1_alloc_ctxt_rcv_groups(struct hfi1_ctxtdata *rcd); 168 void hfi1_free_ctxt_rcv_groups(struct hfi1_ctxtdata *rcd); 169 void hfi1_exp_tid_group_init(struct hfi1_ctxtdata *rcd);
|
| A D | aspm.h | 22 void __aspm_ctx_disable(struct hfi1_ctxtdata *rcd); 26 static inline void aspm_ctx_disable(struct hfi1_ctxtdata *rcd) in aspm_ctx_disable() argument 29 if (likely(!rcd->aspm_intr_supported)) in aspm_ctx_disable() 32 __aspm_ctx_disable(rcd); in aspm_ctx_disable()
|
| A D | debugfs.c | 69 struct hfi1_ctxtdata *rcd; in _opcode_stats_seq_show() local 72 rcd = hfi1_rcd_get_by_index(dd, j); in _opcode_stats_seq_show() 73 if (rcd) { in _opcode_stats_seq_show() 77 hfi1_rcd_put(rcd); in _opcode_stats_seq_show() 160 struct hfi1_ctxtdata *rcd; in _ctx_stats_seq_show() local 171 if (!rcd) in _ctx_stats_seq_show() 177 hfi1_rcd_put(rcd); in _ctx_stats_seq_show() 325 struct hfi1_ctxtdata *rcd; in _rcds_seq_show() local 330 if (rcd) in _rcds_seq_show() 331 seqfile_dump_rcd(s, rcd); in _rcds_seq_show() [all …]
|
| A D | fault.c | 50 struct hfi1_ctxtdata *rcd; in _fault_stats_seq_show() local 53 rcd = hfi1_rcd_get_by_index(dd, j); in _fault_stats_seq_show() 54 if (rcd) { in _fault_stats_seq_show() 55 n_packets += rcd->opstats->stats[i].n_packets; in _fault_stats_seq_show() 56 n_bytes += rcd->opstats->stats[i].n_bytes; in _fault_stats_seq_show() 58 hfi1_rcd_put(rcd); in _fault_stats_seq_show() 311 struct hfi1_ibdev *ibd = &packet->rcd->dd->verbs_dev; in hfi1_dbg_should_fault_rx()
|
| A D | tid_rdma.h | 99 struct hfi1_ctxtdata *rcd; member 209 int hfi1_kern_exp_rcv_init(struct hfi1_ctxtdata *rcd, int reinit); 236 int hfi1_kern_setup_hw_flow(struct hfi1_ctxtdata *rcd, struct rvt_qp *qp); 237 void hfi1_kern_clear_hw_flow(struct hfi1_ctxtdata *rcd, struct rvt_qp *qp); 238 void hfi1_kern_init_ctxt_generations(struct hfi1_ctxtdata *rcd); 255 bool hfi1_handle_kdeth_eflags(struct hfi1_ctxtdata *rcd,
|
| A D | rc.h | 25 static inline void rc_defered_ack(struct hfi1_ctxtdata *rcd, in rc_defered_ack() argument 31 list_add_tail(&qp->rspwait, &rcd->qp_wait_list); in rc_defered_ack() 55 struct hfi1_ctxtdata *rcd);
|
| A D | chip.c | 8372 write_csr(rcd->dd, CCE_INT_FORCE + (8 * rcd->ireg), rcd->imask); in force_recv_intr() 8457 struct hfi1_ctxtdata *rcd = rxq->rcd; in hfi1_netdev_rx_napi() local 8460 work_done = rcd->do_interrupt(rcd, budget); in hfi1_netdev_rx_napi() 8507 disposition = rcd->do_interrupt(rcd, 0); in receive_context_interrupt() 8530 (void)rcd->do_interrupt(rcd, 1); in receive_context_thread() 9977 for (i = rcd->eager_base; i < rcd->eager_base + in hfi1_clear_tids() 12004 if (!rcd) in hfi1_rcvctrl() 12032 memset(rcd->rcvhdrq, 0, rcvhdrq_size(rcd)); in hfi1_rcvctrl() 14754 if (!rcd || !rcd->sc) in hfi1_set_ctxt_jkey() 14788 if (!rcd || !rcd->sc) in hfi1_clear_ctxt_jkey() [all …]
|
| A D | verbs.c | 329 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_kdeth_eager_rcv() local 332 struct hfi1_pportdata *ppd = rcd->ppd; in hfi1_kdeth_eager_rcv() 383 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_kdeth_expected_rcv() local 386 struct hfi1_pportdata *ppd = rcd->ppd; in hfi1_kdeth_expected_rcv() 437 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_do_pkey_check() local 438 struct hfi1_pportdata *ppd = rcd->ppd; in hfi1_do_pkey_check() 457 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_handle_packet() local 458 struct hfi1_pportdata *ppd = rcd->ppd; in hfi1_handle_packet() 537 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_ib_rcv() local 545 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_16B_rcv() local [all …]
|
| A D | rc.c | 1346 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_send_rc_ack() local 1987 ibp = rcd_to_iport(rcd); in do_rc_ack() 2024 set_restart_qp(qp, rcd); in do_rc_ack() 2304 struct hfi1_ctxtdata *rcd = packet->rcd; in rc_rcv_resp() local 2467 ibp = rcd_to_iport(rcd); in rc_rcv_resp() 2538 rc_defered_ack(rcd, qp); in rc_rcv_error() 2765 struct hfi1_ctxtdata *rcd = packet->rcd; in hfi1_rc_rcv() local 3170 rc_defered_ack(rcd, qp); in hfi1_rc_rcv() 3178 rc_defered_ack(rcd, qp); in hfi1_rc_rcv() 3186 rc_defered_ack(rcd, qp); in hfi1_rc_rcv() [all …]
|
| A D | msix.h | 16 int msix_request_rcd_irq(struct hfi1_ctxtdata *rcd); 22 int msix_netdev_request_rcd_irq(struct hfi1_ctxtdata *rcd);
|
| A D | trace_misc.h | 73 TP_STRUCT__entry(DD_DEV_ENTRY(packet->rcd->ppd->dd) 81 TP_fast_assign(DD_DEV_ASSIGN(packet->rcd->ppd->dd); 83 __entry->ctxt = packet->rcd->ctxt;
|
| /drivers/acpi/apei/ |
| A D | erst.c | 1074 struct cper_pstore_record *rcd; in erst_reader() local 1081 if (!rcd) { in erst_reader() 1096 len = erst_read_record(record_id, &rcd->hdr, rcd_len, sizeof(*rcd), in erst_reader() 1109 memcpy(record->buf, rcd->data, len - sizeof(*rcd)); in erst_reader() 1130 kfree(rcd); in erst_reader() 1140 memset(rcd, 0, sizeof(*rcd)); in erst_writer() 1144 rcd->hdr.section_count = 1; in erst_writer() 1149 rcd->hdr.record_length = sizeof(*rcd) + record->size; in erst_writer() 1155 rcd->sec_hdr.section_offset = sizeof(*rcd); in erst_writer() 1176 ret = erst_write(&rcd->hdr); in erst_writer() [all …]
|
| /drivers/net/ethernet/meta/fbnic/ |
| A D | fbnic_txrx.c | 1015 if (!FIELD_GET(FBNIC_RCD_OPT_META_TS, rcd)) in fbnic_rx_tstamp() 1027 u64 rcd, struct sk_buff *skb, in fbnic_populate_skb_fields() argument 1039 fbnic_skb_hash_type(rcd)); in fbnic_populate_skb_fields() 1044 static bool fbnic_rcd_metadata_err(u64 rcd) in fbnic_rcd_metadata_err() argument 1067 u64 rcd; in fbnic_clean_rcq() local 1074 rcd = le64_to_cpu(*raw_rcd); in fbnic_clean_rcq() 1079 fbnic_pkt_prepare(nv, rcd, pkt, qt); in fbnic_clean_rcq() 1084 fbnic_add_rx_frag(nv, rcd, pkt, qt); in fbnic_clean_rcq() 1092 fbnic_rx_tstamp(nv, rcd, pkt); in fbnic_clean_rcq() 1097 if (likely(!fbnic_rcd_metadata_err(rcd))) in fbnic_clean_rcq() [all …]
|
| /drivers/net/vmxnet3/ |
| A D | vmxnet3_drv.c | 370 # define vmxnet3_getRxComp(dstrcd, rcd, tmp) (dstrcd) = (rcd) argument 1460 WARN_ON_ONCE(!(gdesc->rcd.tcp || gdesc->rcd.udp) && in vmxnet3_rx_csum() 1473 WARN_ON_ONCE(!(gdesc->rcd.tcp || gdesc->rcd.udp) && in vmxnet3_rx_csum() 1498 if (!rcd->fcs) in vmxnet3_rx_error() 1642 BUG_ON(rcd->rqID != rq->qid && rcd->rqID != rq->qid2 && in vmxnet3_rq_rx_complete() 1654 if (unlikely(rcd->eop && rcd->err)) { in vmxnet3_rq_rx_complete() 1659 if (rcd->sop && rcd->eop && vmxnet3_xdp_enabled(adapter)) { in vmxnet3_rq_rx_complete() 1698 BUG_ON(!(rcd->sop && rcd->eop)); in vmxnet3_rq_rx_complete() 1729 rcd->len, in vmxnet3_rq_rx_complete() 1828 if (rcd->len) { in vmxnet3_rq_rx_complete() [all …]
|
| /drivers/cpufreq/ |
| A D | sa1110-cpufreq.c | 130 static inline void set_mdcas(u_int *mdcas, int delayed, u_int rcd) in set_mdcas() argument 134 rcd = 2 * rcd - 1; in set_mdcas() 135 shift = delayed + 1 + rcd; in set_mdcas() 137 mdcas[0] = (1 << rcd) - 1; in set_mdcas()
|