| /drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
| A D | flowring.c | 85 hash = flow->hash; in brcmf_flowring_lookup() 130 hash = flow->hash; in brcmf_flowring_create() 363 flow = kzalloc(sizeof(*flow), GFP_KERNEL); in brcmf_flowring_attach() 364 if (flow) { in brcmf_flowring_attach() 365 flow->dev = dev; in brcmf_flowring_attach() 372 flow->rings = kcalloc(nrofrings, sizeof(*flow->rings), in brcmf_flowring_attach() 375 kfree(flow); in brcmf_flowring_attach() 376 flow = NULL; in brcmf_flowring_attach() 380 return flow; in brcmf_flowring_attach() 404 kfree(flow); in brcmf_flowring_detach() [all …]
|
| A D | flowring.h | 50 u32 brcmf_flowring_lookup(struct brcmf_flowring *flow, u8 da[ETH_ALEN], 52 u32 brcmf_flowring_create(struct brcmf_flowring *flow, u8 da[ETH_ALEN], 54 void brcmf_flowring_delete(struct brcmf_flowring *flow, u16 flowid); 55 void brcmf_flowring_open(struct brcmf_flowring *flow, u16 flowid); 56 u8 brcmf_flowring_tid(struct brcmf_flowring *flow, u16 flowid); 57 u32 brcmf_flowring_enqueue(struct brcmf_flowring *flow, u16 flowid, 60 void brcmf_flowring_reinsert(struct brcmf_flowring *flow, u16 flowid, 62 u32 brcmf_flowring_qlen(struct brcmf_flowring *flow, u16 flowid); 63 u8 brcmf_flowring_ifidx_get(struct brcmf_flowring *flow, u16 flowid); 65 void brcmf_flowring_detach(struct brcmf_flowring *flow); [all …]
|
| /drivers/gpu/ipu-v3/ |
| A D | ipu-dp.c | 46 u32 flow; member 151 flow->base + DP_CSC_0); in ipu_dp_csc_init() 153 flow->base + DP_CSC_1); in ipu_dp_csc_init() 161 flow->base + DP_CSC_0); in ipu_dp_csc_init() 195 flow->out_cs = out; in ipu_dp_setup_channel() 197 if (flow->foreground.in_cs == flow->background.in_cs) { in ipu_dp_setup_channel() 203 flow->foreground.in_cs, flow->out_cs, in ipu_dp_setup_channel() 207 flow->foreground.in_cs == flow->out_cs) in ipu_dp_setup_channel() 324 if (flow & 1) in ipu_dp_get() 325 dp = &priv->flow[flow >> 1].foreground; in ipu_dp_get() [all …]
|
| /drivers/net/phy/mscc/ |
| A D | mscc_macsec.c | 467 if ((flow->bank == MACSEC_INGR && flow->rx_sa && !flow->rx_sa->active) || in vsc8584_macsec_flow_enable() 468 (flow->bank == MACSEC_EGR && flow->tx_sa && !flow->tx_sa->active)) in vsc8584_macsec_flow_enable() 577 flow->rx_sa->next_pn : flow->tx_sa->next_pn); in vsc8584_macsec_transformation() 613 flow = kzalloc(sizeof(*flow), GFP_KERNEL); in vsc8584_macsec_alloc_flow() 614 if (!flow) in vsc8584_macsec_alloc_flow() 624 return flow; in vsc8584_macsec_alloc_flow() 635 kfree(flow); in vsc8584_macsec_free_flow() 651 if (IS_ERR(flow)) in vsc8584_macsec_default_flows() 667 if (IS_ERR(flow)) in vsc8584_macsec_default_flows() 820 if (flow->bank == MACSEC_INGR && flow->rx_sa && in vsc8584_macsec_del_rxsc() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/ |
| A D | tc_priv.h | 131 struct mlx5e_tc_flow *flow, 151 set_bit(flag, &flow->flags); in __flow_flag_set() 154 #define flow_flag_set(flow, flag) __flow_flag_set(flow, MLX5E_TC_FLOW_FLAG_##flag) argument 164 __flow_flag_test_and_set(flow, \ 171 clear_bit(flag, &flow->flags); in __flow_flag_clear() 174 #define flow_flag_clear(flow, flag) __flow_flag_clear(flow, \ argument 179 bool ret = test_bit(flag, &flow->flags); in __flow_flag_test() 186 #define flow_flag_test(flow, flag) __flow_flag_test(flow, \ argument 190 struct mlx5e_tc_flow *flow); 193 struct mlx5e_tc_flow *flow, [all …]
|
| A D | tc_tun_encap.c | 183 if (!mlx5e_is_offloaded_flow(flow) || !flow_flag_test(flow, SLOW)) in mlx5e_tc_encap_flows_add() 205 rule = mlx5e_tc_offload_fdb_rules(esw, flow, spec, flow->attr); in mlx5e_tc_encap_flows_add() 215 flow->rule[0] = rule; in mlx5e_tc_encap_flows_add() 261 mlx5e_tc_unoffload_fdb_rules(esw, flow, flow->attr); in mlx5e_tc_encap_flows_del() 263 flow->rule[0] = rule; in mlx5e_tc_encap_flows_del() 1105 mlx5e_detach_encap(flow->priv, flow, attr, out_index); in mlx5e_tc_tun_encap_dests_unset() 1337 flow->decap_route = r; in mlx5e_attach_decap_route() 1475 mlx5e_tc_unoffload_fdb_rules(esw, flow, flow->attr); in mlx5e_invalidate_encap() 1556 rule = mlx5e_tc_offload_fdb_rules(esw, flow, spec, flow->attr); in mlx5e_reoffload_encap() 1626 mlx5e_tc_unoffload_fdb_rules(esw, flow, flow->attr); in mlx5e_unoffload_flow_list() [all …]
|
| A D | tc_tun_encap.h | 10 struct mlx5e_tc_flow *flow, 15 struct mlx5e_tc_flow *flow, 23 struct mlx5e_tc_flow *flow, 26 struct mlx5e_tc_flow *flow); 29 struct mlx5e_tc_flow *flow); 31 struct mlx5e_tc_flow *flow); 34 struct mlx5e_tc_flow *flow, 39 struct mlx5e_tc_flow *flow, 44 int mlx5e_tc_set_attr_rx_tun(struct mlx5e_tc_flow *flow,
|
| /drivers/net/ethernet/broadcom/bnxt/ |
| A D | bnxt_tc.c | 705 is_wildcard(&flow->l4_mask, sizeof(flow->l4_mask))) { in bnxt_hwrm_cfa_flow_alloc() 1057 if (bits_set(&flow->l2_key.smac, sizeof(flow->l2_key.smac)) && in bnxt_tc_can_offload() 1058 !is_exactmatch(flow->l2_mask.smac, sizeof(flow->l2_mask.smac))) { in bnxt_tc_can_offload() 1062 if (bits_set(&flow->l2_key.dmac, sizeof(flow->l2_key.dmac)) && in bnxt_tc_can_offload() 1063 !is_exactmatch(&flow->l2_mask.dmac, sizeof(flow->l2_mask.dmac))) { in bnxt_tc_can_offload() 1283 &flow.daddr); in bnxt_tc_resolve_tunnel_hdrs() 1544 flow = &new_node->flow; in bnxt_tc_add_flow() 1551 bnxt_tc_set_flow_dir(bp, flow, flow->src_fid); in bnxt_tc_add_flow() 1642 flow = &flow_node->flow; in bnxt_tc_get_flow_stats() 1773 struct bnxt_tc_flow *flow = &flow_node->flow; in bnxt_tc_flow_stats_batch_update() local [all …]
|
| /drivers/infiniband/hw/hfi1/ |
| A D | tid_rdma.c | 1117 flow->length = flow->req->seg_len - length; in kern_find_pages() 1342 flow->tid_entry[flow->tidcnt++] = in kern_program_rcv_group() 1347 flow->req->qp, flow->tidcnt - 1, in kern_program_rcv_group() 1348 flow->tid_entry[flow->tidcnt - 1]); in kern_program_rcv_group() 1416 trace_hfi1_tid_flow_alloc(flow->req->qp, flow->req->setup_head, flow); in kern_program_rcvarray() 1523 flow->flow_state.lpsn = flow->flow_state.spsn + flow->npkts - 1; in hfi1_kern_exp_rcv_setup() 1722 req_addr = &flow->tid_entry[flow->tid_idx]; in hfi1_build_tid_rdma_read_packet() 1724 (flow->tidcnt - flow->tid_idx); in hfi1_build_tid_rdma_read_packet() 1952 flow->flow_state.ib_lpsn = flow->flow_state.ib_spsn + flow->npkts - 1; in tid_rdma_rcv_read_request() 2363 last_pkt = (flow->sent >= flow->length); in hfi1_build_tid_rdma_read_resp() [all …]
|
| A D | trace_tid.h | 561 TP_ARGS(qp, index, flow) 567 TP_ARGS(qp, index, flow) 573 TP_ARGS(qp, index, flow) 579 TP_ARGS(qp, index, flow) 585 TP_ARGS(qp, index, flow) 591 TP_ARGS(qp, index, flow) 597 TP_ARGS(qp, index, flow) 603 TP_ARGS(qp, index, flow) 609 TP_ARGS(qp, index, flow) 615 TP_ARGS(qp, index, flow) [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | en_tc.c | 607 if (!flow || !refcount_inc_not_zero(&flow->refcnt)) in mlx5e_flow_get() 609 return flow; in mlx5e_flow_get() 644 return mlx5e_is_eswitch_flow(flow) ? flow->attr->esw_attr->counter_dev : flow->priv->mdev; in get_flow_counter_dev() 1873 err = mlx5e_tc_attach_mod_hdr(flow->priv, flow, attr); in post_process_attr() 4017 struct mlx5e_tc_flow *flow = parse_state->flow; in parse_tc_actions() local 4415 mlx5e_tc_detach_mod_hdr(flow->priv, flow, attr); in mlx5_free_flow_attr_actions() 4448 flow = kzalloc(sizeof(*flow), GFP_KERNEL); in mlx5e_alloc_flow() 4830 if (flow) { in mlx5e_configure_flower() 4847 if (flow) in mlx5e_configure_flower() 4896 if (!flow || !same_flow_direction(flow, flags)) { in mlx5e_delete_flower() [all …]
|
| /drivers/infiniband/hw/usnic/ |
| A D | usnic_fwd.c | 213 flow = kzalloc(sizeof(*flow), GFP_ATOMIC); in usnic_fwd_alloc_flow() 214 if (!flow) in usnic_fwd_alloc_flow() 263 return flow; in usnic_fwd_alloc_flow() 265 kfree(flow); in usnic_fwd_alloc_flow() 274 a0 = flow->flow_id; in usnic_fwd_dealloc_flow() 276 status = usnic_fwd_devcmd(flow->ufdev, flow->vnic_idx, in usnic_fwd_dealloc_flow() 281 flow->flow_id, flow->vnic_idx, in usnic_fwd_dealloc_flow() 285 flow->ufdev->name, flow->vnic_idx, in usnic_fwd_dealloc_flow() 296 flow->ufdev->name, flow->vnic_idx, in usnic_fwd_dealloc_flow() 297 flow->flow_id); in usnic_fwd_dealloc_flow() [all …]
|
| A D | usnic_ib_qp_grp.c | 89 default_flow->flow->flow_id); in usnic_ib_qp_grp_dump_rows() 211 struct usnic_fwd_flow *flow; in create_roce_custom_flow() local 229 if (IS_ERR_OR_NULL(flow)) { in create_roce_custom_flow() 230 err = flow ? PTR_ERR(flow) : -EFAULT; in create_roce_custom_flow() 240 qp_flow->flow = flow; in create_roce_custom_flow() 247 usnic_fwd_dealloc_flow(flow); in create_roce_custom_flow() 271 struct usnic_fwd_flow *flow; in create_udp_flow() local 302 if (IS_ERR_OR_NULL(flow)) { in create_udp_flow() 303 err = flow ? PTR_ERR(flow) : -EFAULT; in create_udp_flow() 313 qp_flow->flow = flow; in create_udp_flow() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/tc/act/ |
| A D | goto.c | 10 struct mlx5e_tc_flow *flow, in validate_goto_chain() argument 16 bool is_esw = mlx5e_is_eswitch_flow(flow); in validate_goto_chain() 17 bool ft_flow = mlx5e_is_ft_flow(flow); in validate_goto_chain() 66 struct mlx5e_tc_flow *flow = parse_state->flow; in tc_act_can_offload_goto() local 68 if (validate_goto_chain(flow->priv, flow, attr, act, extack)) in tc_act_can_offload_goto() 93 struct mlx5e_tc_flow *flow = parse_state->flow; in tc_act_post_parse_goto() local 112 if (!mlx5e_is_eswitch_flow(flow) && parse_attr->mirred_ifindex[0]) { in tc_act_post_parse_goto()
|
| A D | act.c | 75 struct mlx5e_tc_flow *flow, in mlx5e_tc_act_init_parse_state() argument 80 parse_state->flow = flow; in mlx5e_tc_act_init_parse_state() 96 priv = parse_state->flow->priv; in mlx5e_tc_act_post_parse() 118 mlx5e_tc_act_set_next_post_act(struct mlx5e_tc_flow *flow, in mlx5e_tc_act_set_next_post_act() argument 122 struct mlx5_core_dev *mdev = flow->priv->mdev; in mlx5e_tc_act_set_next_post_act()
|
| /drivers/dma/ti/ |
| A D | k3-udma-glue.c | 721 if (flow->ringrxfdq) in k3_udma_glue_release_rx_flow() 724 if (flow->ringrx) in k3_udma_glue_release_rx_flow() 728 flow->udma_rflow = NULL; in k3_udma_glue_release_rx_flow() 752 if (flow->udma_rflow_id != xudma_rflow_get_id(flow->udma_rflow)) { in k3_udma_glue_cfg_rx_flow() 769 &flow->ringrxfdq, in k3_udma_glue_cfg_rx_flow() 770 &flow->ringrx); in k3_udma_glue_cfg_rx_flow() 860 flow->udma_rflow = NULL; in k3_udma_glue_cfg_rx_flow() 1533 flow->udma_rflow_id); in k3_udma_glue_rx_get_irq() 1535 flow->virq = k3_ringacc_get_ring_irq_num(flow->ringrx); in k3_udma_glue_rx_get_irq() 1538 if (!flow->virq) in k3_udma_glue_rx_get_irq() [all …]
|
| /drivers/net/ethernet/marvell/octeontx2/nic/ |
| A D | otx2_flows.c | 1045 flow->vf = vf; in otx2_add_flow_msg() 1133 if (!flow) { in otx2_add_flow() 1134 flow = kzalloc(sizeof(*flow), GFP_KERNEL); in otx2_add_flow() 1135 if (!flow) in otx2_add_flow() 1138 flow->entry = flow_cfg->flow_ent[flow->location]; in otx2_add_flow() 1164 kfree(flow); in otx2_add_flow() 1198 kfree(flow); in otx2_add_flow() 1208 if (flow->is_vf) in otx2_add_flow() 1277 if (!flow) in otx2_remove_flow() 1288 flow->entry); in otx2_remove_flow() [all …]
|
| /drivers/net/ethernet/marvell/mvpp2/ |
| A D | mvpp2_cls.c | 474 mvpp2_prs_add_flow(priv, flow->flow_id, flow->prs_ri.ri, in mvpp2_cls_flow_prs_init() 485 le.lkpid = flow->flow_id; in mvpp2_cls_flow_lkp_init() 726 return &cls_flows[flow]; in mvpp2_cls_flow_get() 750 if (!flow) in mvpp2_port_rss_hash_opts_set() 838 if (!flow) in mvpp2_port_rss_hash_opts_get() 858 if (!flow) in mvpp2_cls_port_init_flows() 1189 if (!flow) in mvpp2_port_cls_rfs_rule_remove() 1222 if (!flow) in mvpp2_port_flt_rfs_rule_insert() 1245 struct flow_rule *flow = rule->flow; in mvpp2_cls_c2_build_match() local 1314 struct flow_rule *flow = rule->flow; in mvpp2_cls_rfs_parse_rule() local [all …]
|
| /drivers/net/ethernet/netronome/nfp/flower/ |
| A D | offload.c | 710 u8 *mask = flow->mask_data; in nfp_flower_populate_merge_match() 1135 u8 *ext = flow->unmasked_data; in nfp_flower_validate_pre_tun_rule() 1137 u8 *mask = flow->mask_data; in nfp_flower_validate_pre_tun_rule() 1315 if (flow->common.chain_index) in offload_pre_check() 1344 extack = flow->common.extack; in nfp_flower_add_offload() 1348 if (is_pre_ct_flow(flow)) in nfp_flower_add_offload() 1351 if (is_post_ct_flow(flow)) in nfp_flower_add_offload() 1354 if (!offload_pre_check(flow)) in nfp_flower_add_offload() 1469 origin = link->sub_flow.flow; in nfp_flower_remove_merge_flow() 1561 extack = flow->common.extack; in nfp_flower_del_offload() [all …]
|
| A D | tunnel_conf.c | 336 if (neigh->flow) in nfp_tun_mutual_link() 353 neigh->flow = predt; in nfp_tun_mutual_link() 430 if (neigh->flow) in nfp_tun_cleanup_nn_entries() 450 neigh->flow = NULL; in nfp_tun_unlink_and_update_nn_entries() 505 nn_entry->flow = NULL; in nfp_tun_write_neigh() 576 if (nn_entry->flow) in nfp_tun_write_neigh() 604 nn_entry->flow = NULL; in nfp_tun_write_neigh() 754 struct flowi4 flow = {}; in nfp_tunnel_request_route_v4() local 767 flow.daddr = payload->ipv4_addr; in nfp_tunnel_request_route_v4() 768 flow.flowi4_proto = IPPROTO_UDP; in nfp_tunnel_request_route_v4() [all …]
|
| /drivers/net/xen-netback/ |
| A D | hash.c | 152 struct flow_keys flow; in xenvif_set_skb_hash() local 181 memset(&flow, 0, sizeof(flow)); in xenvif_set_skb_hash() 182 if (!skb_flow_dissect_flow_keys(skb, &flow, 0)) in xenvif_set_skb_hash() 194 memcpy(&data[0], &flow.addrs.v4addrs.src, 4); in xenvif_set_skb_hash() 195 memcpy(&data[4], &flow.addrs.v4addrs.dst, 4); in xenvif_set_skb_hash() 196 memcpy(&data[8], &flow.ports.src, 2); in xenvif_set_skb_hash() 197 memcpy(&data[10], &flow.ports.dst, 2); in xenvif_set_skb_hash() 204 memcpy(&data[0], &flow.addrs.v4addrs.src, 4); in xenvif_set_skb_hash() 205 memcpy(&data[4], &flow.addrs.v4addrs.dst, 4); in xenvif_set_skb_hash() 220 memcpy(&data[32], &flow.ports.src, 2); in xenvif_set_skb_hash() [all …]
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_ethtool_fdir.c | 43 switch (flow) { in ice_fltr_to_ethtool_flow() 300 int status, flow; in ice_fdir_rem_adq_chnl() local 305 for (flow = 0; flow < ICE_FLTR_PTYPE_MAX; flow++) { in ice_fdir_rem_adq_chnl() 424 int flow; in ice_fdir_release_flows() local 427 for (flow = 0; flow < ICE_FLTR_PTYPE_MAX; flow++) in ice_fdir_release_flows() 437 int flow; in ice_fdir_replay_flows() local 439 for (flow = 0; flow < ICE_FLTR_PTYPE_MAX; flow++) { in ice_fdir_replay_flows() 442 if (!hw->fdir_prof[flow] || !hw->fdir_prof[flow]->cnt) in ice_fdir_replay_flows() 467 flow); in ice_fdir_replay_flows() 679 flow); in ice_fdir_set_hw_fltr_rule() [all …]
|
| A D | ice_virtchnl_fdir.c | 223 enum ice_fltr_ptype flow; in ice_vc_fdir_free_prof_all() local 228 for (flow = ICE_FLTR_PTYPE_NONF_NONE; flow < ICE_FLTR_PTYPE_MAX; flow++) in ice_vc_fdir_free_prof_all() 340 switch (flow) { in ice_vc_fdir_set_flow_hdr() 403 flow, vf->vf_id); in ice_vc_fdir_set_flow_hdr() 463 flow, vf->vf_id); in ice_vc_fdir_set_flow_hdr() 534 flow < ICE_FLTR_PTYPE_MAX; flow++) { in ice_vc_fdir_rem_prof_all() 549 flow < ICE_FLTR_PTYPE_MAX; flow++) { in ice_vc_fdir_reset_cnt_all() 685 flow, vf->vf_id); in ice_vc_fdir_write_flow_prof() 694 flow, vf->vf_id); in ice_vc_fdir_write_flow_prof() 704 flow, vf->vf_id); in ice_vc_fdir_write_flow_prof() [all …]
|
| /drivers/crypto/amlogic/ |
| A D | amlogic-gxl-cipher.c | 22 return atomic_inc_return(&mc->flow) % MAXFLOW; in get_engine_number() 92 int flow = rctx->flow; in meson_cipher() local 110 op->keylen, flow); in meson_cipher() 114 mc->chanlist[flow].stat_req++; in meson_cipher() 164 desc = &mc->chanlist[flow].tl[tloffset]; in meson_cipher() 207 desc = &mc->chanlist[flow].tl[tloffset]; in meson_cipher() 227 mc->chanlist[flow].status = 0; in meson_cipher() 228 writel(mc->chanlist[flow].t_phy | 2, mc->base + (flow << 2)); in meson_cipher() 231 if (mc->chanlist[flow].status == 0) { in meson_cipher() 287 rctx->flow = e; in meson_skdecrypt() [all …]
|
| A D | amlogic-gxl-core.c | 28 int flow; in meson_irq_handler() local 31 for (flow = 0; flow < MAXFLOW; flow++) { in meson_irq_handler() 32 if (mc->irqs[flow] == irq) { in meson_irq_handler() 33 p = readl(mc->base + ((0x04 + flow) << 2)); in meson_irq_handler() 35 writel_relaxed(0xF, mc->base + ((0x4 + flow) << 2)); in meson_irq_handler() 36 mc->chanlist[flow].status = 1; in meson_irq_handler() 37 complete(&mc->chanlist[flow].complete); in meson_irq_handler() 40 dev_err(mc->dev, "%s %d Got irq for flow %d but ctrl is empty\n", __func__, irq, flow); in meson_irq_handler()
|