| /drivers/net/team/ |
| A D | team_mode_loadbalance.c | 72 struct lb_priv_ex *ex; /* priv extension */ member 244 if (!lb_priv->ex->orig_fprog) { in lb_bpf_func_get() 301 if (lb_priv->ex->orig_fprog) { in lb_bpf_func_set() 308 lb_priv->ex->orig_fprog = fprog; in lb_bpf_func_set() 322 if (!lb_priv->ex->orig_fprog) in lb_bpf_func_free() 484 s_info = &lb_priv->ex->stats.info[j]; in lb_stats_refresh() 597 lb_priv->ex = kzalloc(sizeof(*lb_priv->ex), GFP_KERNEL); in lb_init() 598 if (!lb_priv->ex) in lb_init() 600 lb_priv->ex->team = team; in lb_init() 625 kfree(lb_priv->ex); in lb_init() [all …]
|
| /drivers/scsi/libsas/ |
| A D | sas_expander.c | 41 if (!ex->parent_port) { in sas_ex_add_parent_port() 44 BUG_ON(!ex->parent_port); in sas_ex_add_parent_port() 436 ex->ex_phy = kcalloc(ex->num_phys, sizeof(*ex->ex_phy), GFP_KERNEL); in sas_expander_discover() 437 if (!ex->ex_phy) in sas_expander_discover() 446 kfree(ex->ex_phy); in sas_expander_discover() 447 ex->ex_phy = NULL; in sas_expander_discover() 1127 int i = 0, end = ex->num_phys; in sas_ex_discover_devices() 1592 if (level == ex->level) in sas_ex_level_discovery() 1893 if (level > ex->level) in sas_discover_bfs_by_root_level() 1907 int level = ex->level+1; in sas_discover_bfs_by_root() [all …]
|
| /drivers/block/ |
| A D | rbd.c | 1302 obj_req->ex.oe_off, obj_req->ex.oe_len); in rbd_osd_submit() 1333 return obj_req->ex.oe_off + obj_req->ex.oe_len == in rbd_obj_is_tail() 2124 if (ex->fe_off + ex->fe_len > overlap) in prune_extents() 2125 ex->fe_len = overlap - ex->fe_off; in prune_extents() 2280 obj_req->ex.oe_off, obj_req->ex.oe_len, in __rbd_osd_setup_discard_ops() 2351 obj_req->ex.oe_off, obj_req->ex.oe_len, in __rbd_osd_setup_zeroout_ops() 2616 container_of(ex, struct rbd_obj_request, ex); in set_bio_pos() 2627 container_of(ex, struct rbd_obj_request, ex); in count_bio_bvecs() 2640 container_of(ex, struct rbd_obj_request, ex); in copy_bio_bvecs() 2679 container_of(ex, struct rbd_obj_request, ex); in set_bvec_pos() [all …]
|
| /drivers/net/ethernet/nvidia/ |
| A D | forcedeth.c | 1876 if (unlikely(np->put_rx.ex++ == np->last_rx.ex)) in nv_alloc_rx_optimized() 1877 np->put_rx.ex = np->rx_ring.ex; in nv_alloc_rx_optimized() 1911 np->last_rx.ex = &np->rx_ring.ex[np->rx_ring_size-1]; in nv_init_rx() 1942 np->last_tx.ex = &np->tx_ring.ex[np->tx_ring_size-1]; in nv_init_tx() 2670 while ((np->get_tx.ex != np->put_tx.ex) && in nv_tx_done_optimized() 2704 if (unlikely(np->get_tx.ex++ == np->last_tx.ex)) in nv_tx_done_optimized() 2705 np->get_tx.ex = np->tx_ring.ex; in nv_tx_done_optimized() 2987 while ((np->get_rx.ex != np->put_rx.ex) && in nv_rx_process_optimized() 3056 if (unlikely(np->get_rx.ex++ == np->last_rx.ex)) in nv_rx_process_optimized() 3057 np->get_rx.ex = np->rx_ring.ex; in nv_rx_process_optimized() [all …]
|
| /drivers/infiniband/hw/hfi1/ |
| A D | uc.c | 99 qp, wqe->wr.ex.invalidate_rkey); in hfi1_make_uc_req() 132 ohdr->u.imm_data = wqe->wr.ex.imm_data; in hfi1_make_uc_req() 161 ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; in hfi1_make_uc_req() 191 ohdr->u.imm_data = wqe->wr.ex.imm_data; in hfi1_make_uc_req() 217 ohdr->u.imm_data = wqe->wr.ex.imm_data; in hfi1_make_uc_req() 389 wc.ex.imm_data = ohdr->u.imm_data; in hfi1_uc_rcv() 394 wc.ex.imm_data = 0; in hfi1_uc_rcv() 469 wc.ex.imm_data = ohdr->u.rc.imm_data; in hfi1_uc_rcv() 484 wc.ex.imm_data = ohdr->u.imm_data; in hfi1_uc_rcv()
|
| A D | ud.c | 112 wc.ex.imm_data = swqe->wr.ex.imm_data; in ud_loopback() 233 ohdr->u.ud.imm_data = wqe->wr.ex.imm_data; in hfi1_make_bth_deth() 922 wc.ex.imm_data = packet->ohdr->u.ud.imm_data; in hfi1_ud_rcv() 925 wc.ex.imm_data = 0; in hfi1_ud_rcv()
|
| A D | rc.c | 521 wqe->wr.ex.invalidate_rkey); in hfi1_make_rc_req() 569 ohdr->u.imm_data = wqe->wr.ex.imm_data; in hfi1_make_rc_req() 575 wqe->wr.ex.invalidate_rkey); in hfi1_make_rc_req() 612 ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; in hfi1_make_rc_req() 925 ohdr->u.imm_data = wqe->wr.ex.imm_data; in hfi1_make_rc_req() 930 ohdr->u.ieth = cpu_to_be32(wqe->wr.ex.invalidate_rkey); in hfi1_make_rc_req() 970 ohdr->u.imm_data = wqe->wr.ex.imm_data; in hfi1_make_rc_req() 2905 wc.ex.imm_data = ohdr->u.imm_data; in hfi1_rc_rcv() 2913 wc.ex.invalidate_rkey = rkey; in hfi1_rc_rcv() 2922 wc.ex.imm_data = 0; in hfi1_rc_rcv() [all …]
|
| /drivers/net/ethernet/freescale/fman/ |
| A D | mac.c | 45 enum fman_mac_exceptions ex) in mac_exception() argument 47 if (ex == FM_MAC_EX_10G_RX_FIFO_OVFL) { in mac_exception() 51 dev_err(mac_dev->dev, "10G MAC got RX FIFO Error = %x\n", ex); in mac_exception() 55 __func__, ex); in mac_exception()
|
| /drivers/infiniband/sw/rxe/ |
| A D | rxe_verbs.c | 766 wr->ex.imm_data = ibwr->ex.imm_data; in init_send_wr() 778 wr->ex.imm_data = ibwr->ex.imm_data; in init_send_wr() 786 wr->ex.imm_data = ibwr->ex.imm_data; in init_send_wr() 789 wr->ex.invalidate_rkey = ibwr->ex.invalidate_rkey; in init_send_wr() 792 wr->ex.invalidate_rkey = ibwr->ex.invalidate_rkey; in init_send_wr() 806 wr->ex.invalidate_rkey = ibwr->ex.invalidate_rkey; in init_send_wr()
|
| A D | rxe_req.c | 471 immdt_set_imm(pkt, ibwr->ex.imm_data); in init_req_packet() 474 ieth_set_rkey(pkt, ibwr->ex.invalidate_rkey); in init_req_packet() 601 rkey = wqe->wr.ex.invalidate_rkey; in rxe_do_local_ops()
|
| /drivers/thermal/tegra/ |
| A D | soctherm.c | 851 ex |= cp | gp | pl | me; in soctherm_thermal_isr_thread() 852 if (ex) { in soctherm_thermal_isr_thread() 853 writel(ex, ts->regs + THERMCTL_INTR_STATUS); in soctherm_thermal_isr_thread() 854 st &= ~ex; in soctherm_thermal_isr_thread() 882 ex |= TH_INTR_IGNORE_MASK; in soctherm_thermal_isr_thread() 883 st &= ~ex; in soctherm_thermal_isr_thread() 995 u32 st, ex, oc1, oc2, oc3, oc4; in soctherm_edp_isr_thread() local 1004 ex = oc1 | oc2 | oc3 | oc4; in soctherm_edp_isr_thread() 1006 pr_err("soctherm: OC ALARM 0x%08x\n", ex); in soctherm_edp_isr_thread() 1007 if (ex) { in soctherm_edp_isr_thread() [all …]
|
| /drivers/infiniband/hw/erdma/ |
| A D | erdma_qp.c | 410 sqe->imm_data = wr->ex.imm_data; in init_send_sqe_rc() 413 sqe->invalid_stag = cpu_to_le32(wr->ex.invalidate_rkey); in init_send_sqe_rc() 428 sqe->imm_data = wr->ex.imm_data; in init_send_sqe_ud() 495 write_sqe->imm_data = send_wr->ex.imm_data; in erdma_push_one_sqe() 515 cpu_to_le32(send_wr->ex.invalidate_rkey); in erdma_push_one_sqe() 588 regmr_sge->stag = cpu_to_le32(send_wr->ex.invalidate_rkey); in erdma_push_one_sqe()
|
| A D | erdma_cq.c | 180 wc->ex.imm_data = cpu_to_be32(le32_to_cpu(cqe->imm_data)); in erdma_poll_one_cqe() 183 wc->ex.invalidate_rkey = be32_to_cpu(cqe->inv_rkey); in erdma_poll_one_cqe()
|
| /drivers/infiniband/hw/mlx5/ |
| A D | ib_virt.c | 138 goto ex; in mlx5_ib_get_vf_stats() 146 ex: in mlx5_ib_get_vf_stats()
|
| A D | cq.c | 209 wc->ex.imm_data = cqe->immediate; in handle_responder() 221 wc->ex.imm_data = cqe->immediate; in handle_responder() 226 wc->ex.invalidate_rkey = be32_to_cpu(cqe->inval_rkey); in handle_responder() 1216 goto ex; in resize_kernel() 1222 ex: in resize_kernel() 1323 goto ex; in mlx5_ib_resize_cq() 1341 goto ex; in mlx5_ib_resize_cq() 1426 ex: in mlx5_ib_resize_cq()
|
| /drivers/gpu/drm/ci/ |
| A D | build-igt.sh | 4 set -ex
|
| A D | igt_runner.sh | 4 set -ex
|
| A D | lava-submit.sh | 42 set -ex
|
| A D | build.sh | 4 set -ex
|
| /drivers/infiniband/sw/siw/ |
| A D | siw_cq.c | 68 wc->ex.invalidate_rkey = cqe->inval_stag; in siw_reap_cqe()
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | vport.c | 701 goto ex; in mlx5_query_hca_vport_context() 710 goto ex; in mlx5_query_hca_vport_context() 748 ex: in mlx5_query_hca_vport_context() 1076 goto ex; in mlx5_core_modify_hca_vport_context() 1096 ex: in mlx5_core_modify_hca_vport_context()
|
| /drivers/hid/intel-thc-hid/ |
| A D | Kconfig | 12 interfaces with Touch Devices (ex: touchscreen, touchpad etc.). It
|
| /drivers/infiniband/sw/rdmavt/ |
| A D | trace_cq.h | 92 __entry->imm = be32_to_cpu(wc->ex.imm_data);
|
| A D | cq.c | 86 uqueue[head].ex.imm_data = entry->ex.imm_data; in rvt_cq_enter()
|
| /drivers/infiniband/hw/vmw_pvrdma/ |
| A D | pvrdma_qp.c | 758 wqe_hdr->ex.imm_data = wr->ex.imm_data; in pvrdma_post_send() 799 wqe_hdr->ex.invalidate_rkey = in pvrdma_post_send() 800 wr->ex.invalidate_rkey; in pvrdma_post_send()
|