| /drivers/net/ethernet/intel/ice/ |
| A D | ice_eswitch.c | 202 xa_erase(&pf->eswitch.reprs, repr->id); in ice_eswitch_update_repr() 307 pf->eswitch.uplink_vsi = uplink_vsi; in ice_eswitch_enable_switchdev() 315 pf->eswitch.is_running = true; in ice_eswitch_enable_switchdev() 333 pf->eswitch.is_running = false; in ice_eswitch_disable_switchdev() 361 xa_destroy(&pf->eswitch.reprs); in ice_eswitch_mode_set() 374 xa_init(&pf->eswitch.reprs); in ice_eswitch_mode_set() 462 if (xa_empty(&pf->eswitch.reprs)) { in ice_eswitch_attach() 493 if (xa_empty(&pf->eswitch.reprs)) in ice_eswitch_attach() 562 if (xa_empty(&pf->eswitch.reprs)) in ice_eswitch_detach() 568 if (xa_empty(&pf->eswitch.reprs)) { in ice_eswitch_detach() [all …]
|
| A D | ice_eswitch_br.c | 959 struct ice_vsi *vsi = pf->eswitch.uplink_vsi; in ice_eswitch_br_uplink_port_init() 1197 struct ice_esw_br_offloads *br_offloads = pf->eswitch.br_offloads; in ice_eswitch_br_offloads_dealloc() 1206 pf->eswitch.br_offloads = NULL; in ice_eswitch_br_offloads_dealloc() 1217 if (pf->eswitch.br_offloads) in ice_eswitch_br_offloads_alloc() 1224 pf->eswitch.br_offloads = br_offloads; in ice_eswitch_br_offloads_alloc() 1235 br_offloads = pf->eswitch.br_offloads; in ice_eswitch_br_offloads_deinit()
|
| /drivers/infiniband/hw/mlx5/ |
| A D | ib_rep.c | 18 ibdev = mlx5_eswitch_uplink_get_proto_dev(dev->priv.eswitch, REP_IB); in mlx5_ib_set_vport_rep() 110 mlx5_ib_get_rep_netdev(lag_master->priv.eswitch, in mlx5_ib_vport_rep_load() 182 esw = peer_mdev->priv.eswitch; in mlx5_ib_vport_rep_unload() 203 esw = peer_mdev->priv.eswitch; in mlx5_ib_register_peer_vport_reps() 218 struct mlx5_eswitch *esw = dev->mdev->priv.eswitch; in create_flow_rule_vport_sq() 239 esw = mdev->priv.eswitch; in mlx5r_rep_probe() 250 esw = mdev->priv.eswitch; in mlx5r_rep_remove()
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/ |
| A D | tc_tun_encap.c | 155 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_encap_flows_add() 225 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_encap_flows_del() 419 esw = priv->mdev->priv.eswitch; in mlx5e_tc_update_neigh_used_value() 493 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_encap_put() 506 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_encap_put_locked() 519 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_decap_put() 539 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_detach_encap() 569 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_detach_decap() 658 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_encap_get() 1044 esw = priv->mdev->priv.eswitch; in mlx5e_tc_tun_encap_dests_set() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/rep/ |
| A D | bond.c | 229 err = mlx5_esw_acl_egress_vport_bond(priv->mdev->priv.eswitch, in mlx5e_rep_changelowerstate_event() 263 mlx5e_rep_bond_enslave(priv->mdev->priv.eswitch, netdev, lag_dev); in mlx5e_rep_changeupper_event() 265 mlx5e_rep_bond_unslave(priv->mdev->priv.eswitch, netdev, lag_dev); in mlx5e_rep_changeupper_event() 288 rpriv = mlx5_eswitch_get_uplink_priv(priv->mdev->priv.eswitch, REP_ETH); in mlx5e_rep_esw_bond_netevent() 315 if (!mlx5_esw_acl_egress_fwd2vport_supported(priv->mdev->priv.eswitch)) in mlx5e_rep_bond_init() 343 if (!mlx5_esw_acl_egress_fwd2vport_supported(priv->mdev->priv.eswitch) || in mlx5e_rep_bond_cleanup()
|
| A D | tc.c | 95 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_rep_update_flows() 212 esw = priv->mdev->priv.eswitch; in mlx5e_rep_setup_ft_cb() 392 esw = mpriv->mdev->priv.eswitch; in mlx5e_rep_indr_setup_ft_cb() 449 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_rep_check_indr_block_supported() 553 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_rep_indr_replace_act() 591 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_rep_indr_destroy_act() 613 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_rep_indr_stats_act() 722 esw = priv->mdev->priv.eswitch; in mlx5e_rep_tc_receive()
|
| A D | bridge.c | 26 return esw == priv->mdev->priv.eswitch; in mlx5_esw_bridge_dev_same_esw() 76 if (!priv->mdev->priv.eswitch->br_offloads) in mlx5_esw_bridge_rep_vport_num_vhca_id_get() 523 mdev->priv.eswitch; in mlx5e_rep_bridge_init() 583 mdev->priv.eswitch; in mlx5e_rep_bridge_cleanup()
|
| /drivers/net/ethernet/mellanox/mlx5/core/lag/ |
| A D | mpesw.c | 20 esw = dev->priv.eswitch; in mlx5_mpesw_metadata_cleanup() 41 esw = dev->priv.eswitch; in mlx5_mpesw_metadata_set() 101 err = mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch); in enable_mpesw() 115 mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch); in enable_mpesw()
|
| A D | lag.c | 656 master_esw = dev0->priv.eswitch; in mlx5_lag_create_single_fdb() 658 struct mlx5_eswitch *slave_esw = ldev->pf[i].dev->priv.eswitch; in mlx5_lag_create_single_fdb() 669 ldev->pf[j].dev->priv.eswitch); in mlx5_lag_create_single_fdb() 791 master_esw = dev0->priv.eswitch; in mlx5_deactivate_lag() 799 ldev->pf[i].dev->priv.eswitch); in mlx5_deactivate_lag() 930 mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch); in mlx5_disable_lag() 948 mlx5_eswitch_get_npeers(dev->priv.eswitch) == in mlx5_lag_shared_fdb_supported() 956 mlx5_eswitch_vport_match_metadata_enabled(dev->priv.eswitch) && in mlx5_lag_shared_fdb_supported() 957 mlx5_esw_offloads_devcom_is_ready(dev->priv.eswitch) && in mlx5_lag_shared_fdb_supported() 1036 mlx5_eswitch_reload_ib_reps(ldev->pf[i].dev->priv.eswitch); in mlx5_do_bond() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/tc/ |
| A D | int_port.c | 154 struct mlx5_eswitch *esw = priv->dev->priv.eswitch; in mlx5e_int_port_add() 230 struct mlx5_eswitch *esw = priv->dev->priv.eswitch; in mlx5e_int_port_remove() 308 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_int_port_init() 360 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_int_port_init_rep_rx() 380 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_int_port_cleanup_rep_rx()
|
| /drivers/net/ethernet/mellanox/mlx5/core/sf/ |
| A D | devlink.c | 243 struct mlx5_eswitch *esw = dev->priv.eswitch; in mlx5_sf_add() 297 return dev->priv.eswitch && MLX5_ESWITCH_MANAGER(dev) && in mlx5_sf_table_supported() 359 struct mlx5_eswitch *esw = table->dev->priv.eswitch; in mlx5_sf_del() 484 err = mlx5_esw_event_notifier_register(dev->priv.eswitch, &table->esw_nb); in mlx5_sf_table_init() 499 mlx5_esw_event_notifier_unregister(dev->priv.eswitch, &table->esw_nb); in mlx5_sf_table_init() 516 mlx5_esw_event_notifier_unregister(dev->priv.eswitch, &table->esw_nb); in mlx5_sf_table_cleanup()
|
| /drivers/net/ethernet/mellanox/mlx5/core/esw/ |
| A D | qos.c | 1009 esw_assert_qos_lock_held(vport->dev->priv.eswitch); in esw_qos_vport_tc_enable() 1108 esw_assert_qos_lock_held(vport->dev->priv.eswitch); in esw_qos_vport_enable() 1133 struct mlx5_eswitch *esw = vport->dev->priv.eswitch; in mlx5_esw_qos_vport_enable() 1162 struct mlx5_eswitch *esw = vport->dev->priv.eswitch; in mlx5_esw_qos_vport_disable() 1185 esw_assert_qos_lock_held(vport->dev->priv.eswitch); in mlx5_esw_qos_set_vport_max_rate() 1200 esw_assert_qos_lock_held(vport->dev->priv.eswitch); in mlx5_esw_qos_set_vport_min_rate() 1270 esw_assert_qos_lock_held(vport->dev->priv.eswitch); in esw_qos_vport_update() 1659 vport->dev->priv.eswitch; in esw_qos_vport_validate_unsupported_tc_bw() 1699 esw = vport->dev->priv.eswitch; in mlx5_esw_devlink_rate_leaf_tx_share_set() 1720 esw = vport->dev->priv.eswitch; in mlx5_esw_devlink_rate_leaf_tx_max_set() [all …]
|
| A D | ipsec_fs.c | 38 dest->ft = mlx5_chains_get_table(esw_chains(ipsec->mdev->priv.eswitch), 0, 1, 0); in mlx5_esw_ipsec_rx_status_pass_dest_get() 158 struct mlx5_eswitch *esw = mdev->priv.eswitch; in mlx5_esw_ipsec_restore_dest_uplink()
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | eswitch.c | 92 return dev->priv.eswitch; in __mlx5_devlink_eswitch_get() 728 struct mlx5_eswitch *esw = dev->priv.eswitch; in esw_vport_change_handle_locked() 1881 dev->priv.eswitch = esw; in mlx5_eswitch_init() 1921 dev->priv.eswitch = NULL; in mlx5_eswitch_init() 1951 esw->dev->priv.eswitch = NULL; in mlx5_eswitch_cleanup() 2212 struct mlx5_eswitch *esw = dev->priv.eswitch; in mlx5_eswitch_mode() 2223 esw = dev->priv.eswitch; in mlx5_eswitch_get_encap_mode() 2256 struct mlx5_eswitch *esw = mdev->priv.eswitch; in mlx5_esw_hold() 2366 esw = dev->priv.eswitch; in mlx5_eswitch_get_total_vports() 2385 struct mlx5_eswitch *esw = dev->priv.eswitch; in mlx5_eswitch_block_ipsec() [all …]
|
| A D | en_tc.c | 327 struct mlx5_eswitch *esw = dev->priv.eswitch; in mlx5e_get_flow_meters() 1626 esw = flow->priv->mdev->priv.eswitch; in add_unready_flow() 1641 esw = flow->priv->mdev->priv.eswitch; in remove_unready_flow() 1681 esw = out_priv->mdev->priv.eswitch; in mlx5e_tc_query_route_vport() 2264 esw = priv->mdev->priv.eswitch; in mlx5e_get_flow_tunnel_id() 2347 esw = flow->priv->mdev->priv.eswitch; in mlx5e_put_flow_tunnel_id() 5026 esw = priv->mdev->priv.eswitch; in apply_police_params() 5400 esw = priv->mdev->priv.eswitch; in mlx5e_tc_esw_init() 5481 esw = priv->mdev->priv.eswitch; in mlx5e_tc_esw_cleanup() 5818 esw = priv->mdev->priv.eswitch; in mlx5e_get_priv_obj_mapping() [all …]
|
| A D | en_rep.c | 205 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS() 573 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_add_sqs_fwd_rules() 628 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_remove_sqs_fwd_rules() 638 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_rep_add_meta_tunnel_rule() 1041 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_create_rep_vport_rx_rule() 1582 dl_port = mlx5_esw_offloads_devlink_port(dev->priv.eswitch, in mlx5e_vport_vf_rep_load() 1769 esw = mdev->priv.eswitch; in mlx5e_rep_probe() 1780 esw = mdev->priv.eswitch; in mlx5e_rep_remove()
|
| A D | sriov.c | 78 err = mlx5_eswitch_enable(dev->priv.eswitch, num_vfs); in mlx5_device_enable_sriov() 150 mlx5_eswitch_disable_sriov(dev->priv.eswitch, clear_vf); in mlx5_device_disable_sriov()
|
| A D | rdma.c | 65 esw = dev->priv.eswitch; in mlx5_rdma_enable_roce_steering()
|
| A D | eswitch_offloads.c | 757 esw_attr->in_mdev->priv.eswitch, in mlx5_eswitch_add_offloaded_rule() 850 esw_attr->in_mdev->priv.eswitch, in mlx5_eswitch_add_fwd_rule() 2444 struct mlx5_eswitch *esw = dev->priv.eswitch; in esw_port_metadata_set() 2724 dest.vport.num = slave->priv.eswitch->manager_vport; in __esw_set_master_egress_rule() 2822 struct mlx5_eswitch *esw = master->priv.eswitch; in esw_set_master_egress_rule() 2864 vport = mlx5_eswitch_get_vport(dev->priv.eswitch, in esw_unset_master_egress_rule() 2865 dev->priv.eswitch->manager_vport); in esw_unset_master_egress_rule() 3714 struct mlx5_eswitch *esw = dev->priv.eswitch; in mlx5_eswitch_block_mode() 3732 struct mlx5_eswitch *esw = dev->priv.eswitch; in mlx5_eswitch_unblock_mode() 3943 struct mlx5_eswitch *esw = dev->priv.eswitch; in mlx5_eswitch_block_encap() [all …]
|
| /drivers/net/ethernet/mellanox/mlx5/core/en/tc/act/ |
| A D | trap.c | 15 attr->dest_ft = mlx5_eswitch_get_slow_fdb(priv->mdev->priv.eswitch); in tc_act_parse_trap()
|
| A D | mirred.c | 28 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in verify_uplink_forwarding() 223 esw = priv->mdev->priv.eswitch; in parse_mirred()
|
| A D | goto.c | 24 esw = priv->mdev->priv.eswitch; in validate_goto_chain()
|
| A D | vlan.c | 182 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in tc_act_post_parse_vlan()
|
| /drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| A D | fs_chains.c | 127 if (!chains->dev->priv.eswitch || in mlx5_chains_get_prio_range() 128 chains->dev->priv.eswitch->mode != MLX5_ESWITCH_OFFLOADS) in mlx5_chains_get_prio_range() 220 struct mlx5_eswitch *esw = chain->chains->dev->priv.eswitch; in create_chain_restore()
|
| /drivers/net/ethernet/broadcom/ |
| A D | Kconfig | 235 offload for eswitch. This option enables SR-IOV switchdev eswitch
|