Home
last modified time | relevance | path

Searched refs:flow (Results 1 – 25 of 214) sorted by relevance

123456789

/drivers/net/wireless/broadcom/brcm80211/brcmfmac/
A Dflowring.c85 hash = flow->hash; in brcmf_flowring_lookup()
130 hash = flow->hash; in brcmf_flowring_create()
363 flow = kzalloc(sizeof(*flow), GFP_KERNEL); in brcmf_flowring_attach()
364 if (flow) { in brcmf_flowring_attach()
365 flow->dev = dev; in brcmf_flowring_attach()
372 flow->rings = kcalloc(nrofrings, sizeof(*flow->rings), in brcmf_flowring_attach()
375 kfree(flow); in brcmf_flowring_attach()
376 flow = NULL; in brcmf_flowring_attach()
380 return flow; in brcmf_flowring_attach()
404 kfree(flow); in brcmf_flowring_detach()
[all …]
A Dflowring.h50 u32 brcmf_flowring_lookup(struct brcmf_flowring *flow, u8 da[ETH_ALEN],
52 u32 brcmf_flowring_create(struct brcmf_flowring *flow, u8 da[ETH_ALEN],
54 void brcmf_flowring_delete(struct brcmf_flowring *flow, u16 flowid);
55 void brcmf_flowring_open(struct brcmf_flowring *flow, u16 flowid);
56 u8 brcmf_flowring_tid(struct brcmf_flowring *flow, u16 flowid);
57 u32 brcmf_flowring_enqueue(struct brcmf_flowring *flow, u16 flowid,
60 void brcmf_flowring_reinsert(struct brcmf_flowring *flow, u16 flowid,
62 u32 brcmf_flowring_qlen(struct brcmf_flowring *flow, u16 flowid);
63 u8 brcmf_flowring_ifidx_get(struct brcmf_flowring *flow, u16 flowid);
65 void brcmf_flowring_detach(struct brcmf_flowring *flow);
[all …]
/drivers/gpu/ipu-v3/
A Dipu-dp.c46 u32 flow; member
151 flow->base + DP_CSC_0); in ipu_dp_csc_init()
153 flow->base + DP_CSC_1); in ipu_dp_csc_init()
161 flow->base + DP_CSC_0); in ipu_dp_csc_init()
195 flow->out_cs = out; in ipu_dp_setup_channel()
197 if (flow->foreground.in_cs == flow->background.in_cs) { in ipu_dp_setup_channel()
203 flow->foreground.in_cs, flow->out_cs, in ipu_dp_setup_channel()
207 flow->foreground.in_cs == flow->out_cs) in ipu_dp_setup_channel()
324 if (flow & 1) in ipu_dp_get()
325 dp = &priv->flow[flow >> 1].foreground; in ipu_dp_get()
[all …]
/drivers/net/phy/mscc/
A Dmscc_macsec.c467 if ((flow->bank == MACSEC_INGR && flow->rx_sa && !flow->rx_sa->active) || in vsc8584_macsec_flow_enable()
468 (flow->bank == MACSEC_EGR && flow->tx_sa && !flow->tx_sa->active)) in vsc8584_macsec_flow_enable()
577 flow->rx_sa->next_pn : flow->tx_sa->next_pn); in vsc8584_macsec_transformation()
613 flow = kzalloc(sizeof(*flow), GFP_KERNEL); in vsc8584_macsec_alloc_flow()
614 if (!flow) in vsc8584_macsec_alloc_flow()
624 return flow; in vsc8584_macsec_alloc_flow()
635 kfree(flow); in vsc8584_macsec_free_flow()
651 if (IS_ERR(flow)) in vsc8584_macsec_default_flows()
667 if (IS_ERR(flow)) in vsc8584_macsec_default_flows()
820 if (flow->bank == MACSEC_INGR && flow->rx_sa && in vsc8584_macsec_del_rxsc()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/en/
A Dtc_priv.h131 struct mlx5e_tc_flow *flow,
151 set_bit(flag, &flow->flags); in __flow_flag_set()
154 #define flow_flag_set(flow, flag) __flow_flag_set(flow, MLX5E_TC_FLOW_FLAG_##flag) argument
164 __flow_flag_test_and_set(flow, \
171 clear_bit(flag, &flow->flags); in __flow_flag_clear()
174 #define flow_flag_clear(flow, flag) __flow_flag_clear(flow, \ argument
179 bool ret = test_bit(flag, &flow->flags); in __flow_flag_test()
186 #define flow_flag_test(flow, flag) __flow_flag_test(flow, \ argument
190 struct mlx5e_tc_flow *flow);
193 struct mlx5e_tc_flow *flow,
[all …]
A Dtc_tun_encap.c183 if (!mlx5e_is_offloaded_flow(flow) || !flow_flag_test(flow, SLOW)) in mlx5e_tc_encap_flows_add()
205 rule = mlx5e_tc_offload_fdb_rules(esw, flow, spec, flow->attr); in mlx5e_tc_encap_flows_add()
215 flow->rule[0] = rule; in mlx5e_tc_encap_flows_add()
261 mlx5e_tc_unoffload_fdb_rules(esw, flow, flow->attr); in mlx5e_tc_encap_flows_del()
263 flow->rule[0] = rule; in mlx5e_tc_encap_flows_del()
1105 mlx5e_detach_encap(flow->priv, flow, attr, out_index); in mlx5e_tc_tun_encap_dests_unset()
1337 flow->decap_route = r; in mlx5e_attach_decap_route()
1475 mlx5e_tc_unoffload_fdb_rules(esw, flow, flow->attr); in mlx5e_invalidate_encap()
1556 rule = mlx5e_tc_offload_fdb_rules(esw, flow, spec, flow->attr); in mlx5e_reoffload_encap()
1626 mlx5e_tc_unoffload_fdb_rules(esw, flow, flow->attr); in mlx5e_unoffload_flow_list()
[all …]
A Dtc_tun_encap.h10 struct mlx5e_tc_flow *flow,
15 struct mlx5e_tc_flow *flow,
23 struct mlx5e_tc_flow *flow,
26 struct mlx5e_tc_flow *flow);
29 struct mlx5e_tc_flow *flow);
31 struct mlx5e_tc_flow *flow);
34 struct mlx5e_tc_flow *flow,
39 struct mlx5e_tc_flow *flow,
44 int mlx5e_tc_set_attr_rx_tun(struct mlx5e_tc_flow *flow,
/drivers/net/ethernet/broadcom/bnxt/
A Dbnxt_tc.c705 is_wildcard(&flow->l4_mask, sizeof(flow->l4_mask))) { in bnxt_hwrm_cfa_flow_alloc()
1057 if (bits_set(&flow->l2_key.smac, sizeof(flow->l2_key.smac)) && in bnxt_tc_can_offload()
1058 !is_exactmatch(flow->l2_mask.smac, sizeof(flow->l2_mask.smac))) { in bnxt_tc_can_offload()
1062 if (bits_set(&flow->l2_key.dmac, sizeof(flow->l2_key.dmac)) && in bnxt_tc_can_offload()
1063 !is_exactmatch(&flow->l2_mask.dmac, sizeof(flow->l2_mask.dmac))) { in bnxt_tc_can_offload()
1283 &flow.daddr); in bnxt_tc_resolve_tunnel_hdrs()
1544 flow = &new_node->flow; in bnxt_tc_add_flow()
1551 bnxt_tc_set_flow_dir(bp, flow, flow->src_fid); in bnxt_tc_add_flow()
1642 flow = &flow_node->flow; in bnxt_tc_get_flow_stats()
1773 struct bnxt_tc_flow *flow = &flow_node->flow; in bnxt_tc_flow_stats_batch_update() local
[all …]
/drivers/infiniband/hw/hfi1/
A Dtid_rdma.c1117 flow->length = flow->req->seg_len - length; in kern_find_pages()
1342 flow->tid_entry[flow->tidcnt++] = in kern_program_rcv_group()
1347 flow->req->qp, flow->tidcnt - 1, in kern_program_rcv_group()
1348 flow->tid_entry[flow->tidcnt - 1]); in kern_program_rcv_group()
1416 trace_hfi1_tid_flow_alloc(flow->req->qp, flow->req->setup_head, flow); in kern_program_rcvarray()
1523 flow->flow_state.lpsn = flow->flow_state.spsn + flow->npkts - 1; in hfi1_kern_exp_rcv_setup()
1722 req_addr = &flow->tid_entry[flow->tid_idx]; in hfi1_build_tid_rdma_read_packet()
1724 (flow->tidcnt - flow->tid_idx); in hfi1_build_tid_rdma_read_packet()
1952 flow->flow_state.ib_lpsn = flow->flow_state.ib_spsn + flow->npkts - 1; in tid_rdma_rcv_read_request()
2363 last_pkt = (flow->sent >= flow->length); in hfi1_build_tid_rdma_read_resp()
[all …]
A Dtrace_tid.h561 TP_ARGS(qp, index, flow)
567 TP_ARGS(qp, index, flow)
573 TP_ARGS(qp, index, flow)
579 TP_ARGS(qp, index, flow)
585 TP_ARGS(qp, index, flow)
591 TP_ARGS(qp, index, flow)
597 TP_ARGS(qp, index, flow)
603 TP_ARGS(qp, index, flow)
609 TP_ARGS(qp, index, flow)
615 TP_ARGS(qp, index, flow)
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/
A Den_tc.c607 if (!flow || !refcount_inc_not_zero(&flow->refcnt)) in mlx5e_flow_get()
609 return flow; in mlx5e_flow_get()
644 return mlx5e_is_eswitch_flow(flow) ? flow->attr->esw_attr->counter_dev : flow->priv->mdev; in get_flow_counter_dev()
1873 err = mlx5e_tc_attach_mod_hdr(flow->priv, flow, attr); in post_process_attr()
4017 struct mlx5e_tc_flow *flow = parse_state->flow; in parse_tc_actions() local
4415 mlx5e_tc_detach_mod_hdr(flow->priv, flow, attr); in mlx5_free_flow_attr_actions()
4448 flow = kzalloc(sizeof(*flow), GFP_KERNEL); in mlx5e_alloc_flow()
4830 if (flow) { in mlx5e_configure_flower()
4847 if (flow) in mlx5e_configure_flower()
4896 if (!flow || !same_flow_direction(flow, flags)) { in mlx5e_delete_flower()
[all …]
/drivers/infiniband/hw/usnic/
A Dusnic_fwd.c213 flow = kzalloc(sizeof(*flow), GFP_ATOMIC); in usnic_fwd_alloc_flow()
214 if (!flow) in usnic_fwd_alloc_flow()
263 return flow; in usnic_fwd_alloc_flow()
265 kfree(flow); in usnic_fwd_alloc_flow()
274 a0 = flow->flow_id; in usnic_fwd_dealloc_flow()
276 status = usnic_fwd_devcmd(flow->ufdev, flow->vnic_idx, in usnic_fwd_dealloc_flow()
281 flow->flow_id, flow->vnic_idx, in usnic_fwd_dealloc_flow()
285 flow->ufdev->name, flow->vnic_idx, in usnic_fwd_dealloc_flow()
296 flow->ufdev->name, flow->vnic_idx, in usnic_fwd_dealloc_flow()
297 flow->flow_id); in usnic_fwd_dealloc_flow()
[all …]
A Dusnic_ib_qp_grp.c89 default_flow->flow->flow_id); in usnic_ib_qp_grp_dump_rows()
211 struct usnic_fwd_flow *flow; in create_roce_custom_flow() local
229 if (IS_ERR_OR_NULL(flow)) { in create_roce_custom_flow()
230 err = flow ? PTR_ERR(flow) : -EFAULT; in create_roce_custom_flow()
240 qp_flow->flow = flow; in create_roce_custom_flow()
247 usnic_fwd_dealloc_flow(flow); in create_roce_custom_flow()
271 struct usnic_fwd_flow *flow; in create_udp_flow() local
302 if (IS_ERR_OR_NULL(flow)) { in create_udp_flow()
303 err = flow ? PTR_ERR(flow) : -EFAULT; in create_udp_flow()
313 qp_flow->flow = flow; in create_udp_flow()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/en/tc/act/
A Dgoto.c10 struct mlx5e_tc_flow *flow, in validate_goto_chain() argument
16 bool is_esw = mlx5e_is_eswitch_flow(flow); in validate_goto_chain()
17 bool ft_flow = mlx5e_is_ft_flow(flow); in validate_goto_chain()
66 struct mlx5e_tc_flow *flow = parse_state->flow; in tc_act_can_offload_goto() local
68 if (validate_goto_chain(flow->priv, flow, attr, act, extack)) in tc_act_can_offload_goto()
93 struct mlx5e_tc_flow *flow = parse_state->flow; in tc_act_post_parse_goto() local
112 if (!mlx5e_is_eswitch_flow(flow) && parse_attr->mirred_ifindex[0]) { in tc_act_post_parse_goto()
A Dact.c75 struct mlx5e_tc_flow *flow, in mlx5e_tc_act_init_parse_state() argument
80 parse_state->flow = flow; in mlx5e_tc_act_init_parse_state()
96 priv = parse_state->flow->priv; in mlx5e_tc_act_post_parse()
118 mlx5e_tc_act_set_next_post_act(struct mlx5e_tc_flow *flow, in mlx5e_tc_act_set_next_post_act() argument
122 struct mlx5_core_dev *mdev = flow->priv->mdev; in mlx5e_tc_act_set_next_post_act()
/drivers/dma/ti/
A Dk3-udma-glue.c721 if (flow->ringrxfdq) in k3_udma_glue_release_rx_flow()
724 if (flow->ringrx) in k3_udma_glue_release_rx_flow()
728 flow->udma_rflow = NULL; in k3_udma_glue_release_rx_flow()
752 if (flow->udma_rflow_id != xudma_rflow_get_id(flow->udma_rflow)) { in k3_udma_glue_cfg_rx_flow()
769 &flow->ringrxfdq, in k3_udma_glue_cfg_rx_flow()
770 &flow->ringrx); in k3_udma_glue_cfg_rx_flow()
860 flow->udma_rflow = NULL; in k3_udma_glue_cfg_rx_flow()
1533 flow->udma_rflow_id); in k3_udma_glue_rx_get_irq()
1535 flow->virq = k3_ringacc_get_ring_irq_num(flow->ringrx); in k3_udma_glue_rx_get_irq()
1538 if (!flow->virq) in k3_udma_glue_rx_get_irq()
[all …]
/drivers/net/ethernet/marvell/octeontx2/nic/
A Dotx2_flows.c1045 flow->vf = vf; in otx2_add_flow_msg()
1133 if (!flow) { in otx2_add_flow()
1134 flow = kzalloc(sizeof(*flow), GFP_KERNEL); in otx2_add_flow()
1135 if (!flow) in otx2_add_flow()
1138 flow->entry = flow_cfg->flow_ent[flow->location]; in otx2_add_flow()
1164 kfree(flow); in otx2_add_flow()
1198 kfree(flow); in otx2_add_flow()
1208 if (flow->is_vf) in otx2_add_flow()
1277 if (!flow) in otx2_remove_flow()
1288 flow->entry); in otx2_remove_flow()
[all …]
/drivers/net/ethernet/marvell/mvpp2/
A Dmvpp2_cls.c474 mvpp2_prs_add_flow(priv, flow->flow_id, flow->prs_ri.ri, in mvpp2_cls_flow_prs_init()
485 le.lkpid = flow->flow_id; in mvpp2_cls_flow_lkp_init()
726 return &cls_flows[flow]; in mvpp2_cls_flow_get()
750 if (!flow) in mvpp2_port_rss_hash_opts_set()
838 if (!flow) in mvpp2_port_rss_hash_opts_get()
858 if (!flow) in mvpp2_cls_port_init_flows()
1189 if (!flow) in mvpp2_port_cls_rfs_rule_remove()
1222 if (!flow) in mvpp2_port_flt_rfs_rule_insert()
1245 struct flow_rule *flow = rule->flow; in mvpp2_cls_c2_build_match() local
1314 struct flow_rule *flow = rule->flow; in mvpp2_cls_rfs_parse_rule() local
[all …]
/drivers/net/ethernet/netronome/nfp/flower/
A Doffload.c710 u8 *mask = flow->mask_data; in nfp_flower_populate_merge_match()
1135 u8 *ext = flow->unmasked_data; in nfp_flower_validate_pre_tun_rule()
1137 u8 *mask = flow->mask_data; in nfp_flower_validate_pre_tun_rule()
1315 if (flow->common.chain_index) in offload_pre_check()
1344 extack = flow->common.extack; in nfp_flower_add_offload()
1348 if (is_pre_ct_flow(flow)) in nfp_flower_add_offload()
1351 if (is_post_ct_flow(flow)) in nfp_flower_add_offload()
1354 if (!offload_pre_check(flow)) in nfp_flower_add_offload()
1469 origin = link->sub_flow.flow; in nfp_flower_remove_merge_flow()
1561 extack = flow->common.extack; in nfp_flower_del_offload()
[all …]
A Dtunnel_conf.c336 if (neigh->flow) in nfp_tun_mutual_link()
353 neigh->flow = predt; in nfp_tun_mutual_link()
430 if (neigh->flow) in nfp_tun_cleanup_nn_entries()
450 neigh->flow = NULL; in nfp_tun_unlink_and_update_nn_entries()
505 nn_entry->flow = NULL; in nfp_tun_write_neigh()
576 if (nn_entry->flow) in nfp_tun_write_neigh()
604 nn_entry->flow = NULL; in nfp_tun_write_neigh()
754 struct flowi4 flow = {}; in nfp_tunnel_request_route_v4() local
767 flow.daddr = payload->ipv4_addr; in nfp_tunnel_request_route_v4()
768 flow.flowi4_proto = IPPROTO_UDP; in nfp_tunnel_request_route_v4()
[all …]
/drivers/net/xen-netback/
A Dhash.c152 struct flow_keys flow; in xenvif_set_skb_hash() local
181 memset(&flow, 0, sizeof(flow)); in xenvif_set_skb_hash()
182 if (!skb_flow_dissect_flow_keys(skb, &flow, 0)) in xenvif_set_skb_hash()
194 memcpy(&data[0], &flow.addrs.v4addrs.src, 4); in xenvif_set_skb_hash()
195 memcpy(&data[4], &flow.addrs.v4addrs.dst, 4); in xenvif_set_skb_hash()
196 memcpy(&data[8], &flow.ports.src, 2); in xenvif_set_skb_hash()
197 memcpy(&data[10], &flow.ports.dst, 2); in xenvif_set_skb_hash()
204 memcpy(&data[0], &flow.addrs.v4addrs.src, 4); in xenvif_set_skb_hash()
205 memcpy(&data[4], &flow.addrs.v4addrs.dst, 4); in xenvif_set_skb_hash()
220 memcpy(&data[32], &flow.ports.src, 2); in xenvif_set_skb_hash()
[all …]
/drivers/net/ethernet/intel/ice/
A Dice_ethtool_fdir.c43 switch (flow) { in ice_fltr_to_ethtool_flow()
300 int status, flow; in ice_fdir_rem_adq_chnl() local
305 for (flow = 0; flow < ICE_FLTR_PTYPE_MAX; flow++) { in ice_fdir_rem_adq_chnl()
424 int flow; in ice_fdir_release_flows() local
427 for (flow = 0; flow < ICE_FLTR_PTYPE_MAX; flow++) in ice_fdir_release_flows()
437 int flow; in ice_fdir_replay_flows() local
439 for (flow = 0; flow < ICE_FLTR_PTYPE_MAX; flow++) { in ice_fdir_replay_flows()
442 if (!hw->fdir_prof[flow] || !hw->fdir_prof[flow]->cnt) in ice_fdir_replay_flows()
467 flow); in ice_fdir_replay_flows()
679 flow); in ice_fdir_set_hw_fltr_rule()
[all …]
A Dice_virtchnl_fdir.c223 enum ice_fltr_ptype flow; in ice_vc_fdir_free_prof_all() local
228 for (flow = ICE_FLTR_PTYPE_NONF_NONE; flow < ICE_FLTR_PTYPE_MAX; flow++) in ice_vc_fdir_free_prof_all()
340 switch (flow) { in ice_vc_fdir_set_flow_hdr()
403 flow, vf->vf_id); in ice_vc_fdir_set_flow_hdr()
463 flow, vf->vf_id); in ice_vc_fdir_set_flow_hdr()
534 flow < ICE_FLTR_PTYPE_MAX; flow++) { in ice_vc_fdir_rem_prof_all()
549 flow < ICE_FLTR_PTYPE_MAX; flow++) { in ice_vc_fdir_reset_cnt_all()
685 flow, vf->vf_id); in ice_vc_fdir_write_flow_prof()
694 flow, vf->vf_id); in ice_vc_fdir_write_flow_prof()
704 flow, vf->vf_id); in ice_vc_fdir_write_flow_prof()
[all …]
/drivers/crypto/amlogic/
A Damlogic-gxl-cipher.c22 return atomic_inc_return(&mc->flow) % MAXFLOW; in get_engine_number()
92 int flow = rctx->flow; in meson_cipher() local
110 op->keylen, flow); in meson_cipher()
114 mc->chanlist[flow].stat_req++; in meson_cipher()
164 desc = &mc->chanlist[flow].tl[tloffset]; in meson_cipher()
207 desc = &mc->chanlist[flow].tl[tloffset]; in meson_cipher()
227 mc->chanlist[flow].status = 0; in meson_cipher()
228 writel(mc->chanlist[flow].t_phy | 2, mc->base + (flow << 2)); in meson_cipher()
231 if (mc->chanlist[flow].status == 0) { in meson_cipher()
287 rctx->flow = e; in meson_skdecrypt()
[all …]
A Damlogic-gxl-core.c28 int flow; in meson_irq_handler() local
31 for (flow = 0; flow < MAXFLOW; flow++) { in meson_irq_handler()
32 if (mc->irqs[flow] == irq) { in meson_irq_handler()
33 p = readl(mc->base + ((0x04 + flow) << 2)); in meson_irq_handler()
35 writel_relaxed(0xF, mc->base + ((0x4 + flow) << 2)); in meson_irq_handler()
36 mc->chanlist[flow].status = 1; in meson_irq_handler()
37 complete(&mc->chanlist[flow].complete); in meson_irq_handler()
40 dev_err(mc->dev, "%s %d Got irq for flow %d but ctrl is empty\n", __func__, irq, flow); in meson_irq_handler()

Completed in 97 milliseconds

123456789