| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
| A D | dr_domain.c | 69 ret = mlx5_core_alloc_pd(dmn->mdev, &dmn->pdn); in dr_domain_init_resources() 75 dmn->uar = mlx5_get_uars_page(dmn->mdev); in dr_domain_init_resources() 109 mlx5_put_uars_page(dmn->mdev, dmn->uar); in dr_domain_init_resources() 111 mlx5_core_dealloc_pd(dmn->mdev, dmn->pdn); in dr_domain_init_resources() 118 mlx5dr_send_ring_free(dmn, dmn->send_ring); in dr_domain_uninit_resources() 121 mlx5_put_uars_page(dmn->mdev, dmn->uar); in dr_domain_uninit_resources() 122 mlx5_core_dealloc_pd(dmn->mdev, dmn->pdn); in dr_domain_uninit_resources() 366 dmn = kzalloc(sizeof(*dmn), GFP_KERNEL); in mlx5dr_domain_create() 367 if (!dmn) in mlx5dr_domain_create() 399 return dmn; in mlx5dr_domain_create() [all …]
|
| A D | dr_fw.c | 21 ft_attr.level = dmn->info.caps.max_ft_level - 1; in mlx5dr_fw_create_recalc_cs_ft() 24 ret = mlx5dr_cmd_create_flow_table(dmn->mdev, in mlx5dr_fw_create_recalc_cs_ft() 33 ret = mlx5dr_cmd_create_empty_flow_group(dmn->mdev, in mlx5dr_fw_create_recalc_cs_ft() 54 ret = mlx5dr_cmd_set_fte_modify_and_vport(dmn->mdev, in mlx5dr_fw_create_recalc_cs_ft() 73 mlx5dr_cmd_destroy_flow_group(dmn->mdev, in mlx5dr_fw_create_recalc_cs_ft() 86 mlx5dr_cmd_del_flow_table_entry(dmn->mdev, in mlx5dr_fw_destroy_recalc_cs_ft() 90 mlx5dr_cmd_destroy_flow_group(dmn->mdev, in mlx5dr_fw_destroy_recalc_cs_ft() 94 mlx5dr_cmd_destroy_flow_table(dmn->mdev, in mlx5dr_fw_destroy_recalc_cs_ft() 155 mlx5dr_cmd_destroy_flow_table(dmn->mdev, *tbl_id, in mlx5dr_fw_create_md_tbl() 164 mlx5dr_cmd_destroy_flow_group(dmn->mdev, in mlx5dr_fw_destroy_md_tbl() [all …]
|
| A D | dr_action.c | 556 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in mlx5dr_actions_build_ste_arr() local 588 if (dest_tbl->tbl->dmn != dmn) { in mlx5dr_actions_build_ste_arr() 818 action->dest_tbl->fw_tbl.dmn = dmn; in mlx5dr_action_create_dest_table_num() 932 action->dest_tbl->fw_tbl.dmn = dmn; in mlx5dr_action_create_mult_dest_tbl() 963 action->dest_tbl->fw_tbl.dmn = dmn; in mlx5dr_action_create_dest_flow_fw_table() 1013 action->sampler->dmn = dmn; in mlx5dr_action_create_flow_sampler() 1230 action->reformat->dmn = dmn; in mlx5dr_action_create_packet_reformat() 1438 struct mlx5dr_domain *dmn = action->rewrite->dmn; in dr_action_modify_check_set_field_limitation() local 1469 struct mlx5dr_domain *dmn = action->rewrite->dmn; in dr_action_modify_check_add_field_limitation() local 1733 action->rewrite->dmn = dmn; in mlx5dr_action_create_modify_header() [all …]
|
| A D | dr_matcher.c | 124 struct mlx5dr_domain *dmn) in dr_mask_is_tnl_vxlan_gpe() argument 169 struct mlx5dr_domain *dmn) in dr_mask_is_tnl_gtpu() argument 396 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_matcher_set_ste_builders() local 719 struct mlx5dr_domain *dmn = tbl->dmn; in dr_matcher_add_to_tbl() local 783 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_matcher_uninit() local 785 switch (dmn->type) { in dr_matcher_uninit() 804 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_matcher_set_all_ste_builders() local 822 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_matcher_init_nic() local 880 struct mlx5dr_domain *dmn = tbl->dmn; in dr_matcher_init() local 905 switch (dmn->type) { in dr_matcher_init() [all …]
|
| A D | dr_table.c | 17 mlx5dr_domain_lock(tbl->dmn); in mlx5dr_table_set_miss_action() 81 mlx5dr_domain_unlock(tbl->dmn); in mlx5dr_table_set_miss_action() 98 mlx5dr_domain_lock(tbl->dmn); in dr_table_uninit() 100 switch (tbl->dmn->type) { in dr_table_uninit() 115 mlx5dr_domain_unlock(tbl->dmn); in dr_table_uninit() 180 mlx5dr_domain_lock(tbl->dmn); in dr_table_init() 182 switch (tbl->dmn->type) { in dr_table_init() 204 mlx5dr_domain_unlock(tbl->dmn); in dr_table_init() 250 refcount_inc(&dmn->refcount); in mlx5dr_table_create() 256 tbl->dmn = dmn; in mlx5dr_table_create() [all …]
|
| A D | dr_send.c | 928 dmn->send_ring->cq = dr_create_cq(dmn->mdev, dmn->uar, cq_size); in mlx5dr_send_ring_alloc() 946 dmn->send_ring->qp = dr_create_rc_qp(dmn->mdev, &init_attr); in mlx5dr_send_ring_alloc() 953 dmn->send_ring->cq->qp = dmn->send_ring->qp; in mlx5dr_send_ring_alloc() 959 dmn->send_ring->signal_th = dmn->info.max_send_wr / in mlx5dr_send_ring_alloc() 981 dmn->send_ring->mr = dr_reg_mr(dmn->mdev, in mlx5dr_send_ring_alloc() 982 dmn->pdn, dmn->send_ring->buf, size); in mlx5dr_send_ring_alloc() 988 dmn->send_ring->sync_mr = dr_reg_mr(dmn->mdev, in mlx5dr_send_ring_alloc() 989 dmn->pdn, dmn->send_ring->sync_buff, in mlx5dr_send_ring_alloc() 999 dr_dereg_mr(dmn->mdev, dmn->send_ring->mr); in mlx5dr_send_ring_alloc() 1003 dr_destroy_qp(dmn->mdev, dmn->send_ring->qp); in mlx5dr_send_ring_alloc() [all …]
|
| A D | dr_rule.c | 45 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_create_collision_htbl() local 181 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_rehash_handle_collision() local 238 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_rehash_copy_ste() local 377 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_rehash_htbl() local 494 struct mlx5dr_domain *dmn = rule->matcher->tbl->dmn; in dr_rule_rehash() local 515 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_handle_collision() local 695 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_handle_action_stes() local 761 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_handle_empty_entry() local 814 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_handle_ste_branch() local 1070 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in dr_rule_create_rule_nic() local [all …]
|
| A D | dr_icm_pool.c | 12 struct mlx5dr_domain *dmn; member 29 struct mlx5dr_domain *dmn; member 64 struct mlx5_core_dev *mdev = pool->dmn->mdev; in dr_icm_pool_mr_create() 74 icm_mr->dmn = pool->dmn; in dr_icm_pool_mr_create() 98 err = dr_icm_create_dm_mkey(mdev, pool->dmn->pdn, in dr_icm_pool_mr_create() 130 struct mlx5_core_dev *mdev = icm_mr->dmn->mdev; in dr_icm_pool_mr_destroy() 266 mlx5dr_err(pool->dmn, in dr_icm_chunk_create() 299 err = mlx5dr_cmd_sync_steering(pool->dmn->mdev); in dr_icm_pool_sync_all_buddy_pools() 341 mlx5dr_err(pool->dmn, in dr_icm_handle_buddies_get_mem() 351 mlx5dr_err(pool->dmn, in dr_icm_handle_buddies_get_mem() [all …]
|
| A D | dr_types.h | 24 #define mlx5dr_err(dmn, arg...) mlx5_core_err((dmn)->mdev, ##arg) argument 25 #define mlx5dr_info(dmn, arg...) mlx5_core_info((dmn)->mdev, ##arg) argument 26 #define mlx5dr_dbg(dmn, arg...) mlx5_core_dbg((dmn)->mdev, ##arg) argument 211 struct mlx5dr_domain *dmn; member 472 struct mlx5dr_domain *dmn, 855 struct mlx5dr_domain *dmn; member 901 struct mlx5dr_domain *dmn; member 912 struct mlx5dr_domain *dmn; member 920 struct mlx5dr_domain *dmn; member 931 struct mlx5dr_domain *dmn; member [all …]
|
| A D | dr_ste.c | 294 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in mlx5dr_ste_free() local 408 dmn->info.caps.gvmi, in mlx5dr_ste_htbl_init_and_postsend() 424 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in mlx5dr_ste_create_next_htbl() local 520 struct mlx5dr_domain *dmn, in mlx5dr_ste_set_actions_tx() argument 531 struct mlx5dr_domain *dmn, in mlx5dr_ste_set_actions_rx() argument 612 mlx5dr_err(dmn, in mlx5dr_ste_build_pre_check() 618 mlx5dr_err(dmn, in mlx5dr_ste_build_pre_check() 634 struct mlx5dr_domain *dmn = matcher->tbl->dmn; in mlx5dr_ste_build_ste_arr() local 649 dmn->info.caps.gvmi); in mlx5dr_ste_build_ste_arr() 1240 struct mlx5dr_domain *dmn, in mlx5dr_ste_build_src_gvmi_qpn() argument [all …]
|
| A D | mlx5dr.h | 50 void mlx5dr_domain_set_peer(struct mlx5dr_domain *dmn, 81 mlx5dr_action_create_dest_table_num(struct mlx5dr_domain *dmn, u32 table_num); 96 mlx5dr_action_create_mult_dest_tbl(struct mlx5dr_domain *dmn, 106 mlx5dr_action_create_flow_sampler(struct mlx5dr_domain *dmn, u32 sampler_id); 112 mlx5dr_action_create_packet_reformat(struct mlx5dr_domain *dmn,
|
| A D | dr_ste_v0.c | 408 dr_ste_v0_set_actions_tx(struct mlx5dr_domain *dmn, in dr_ste_v0_set_actions_tx() argument 465 if (MLX5_CAP_GEN(dmn->mdev, prio_tag_required)) in dr_ste_v0_set_actions_tx() 477 dr_ste_v0_set_actions_rx(struct mlx5dr_domain *dmn, in dr_ste_v0_set_actions_rx() argument 1647 struct mlx5dr_domain *dmn = sb->dmn; in dr_ste_v0_build_src_gvmi_qpn_tag() local 1656 if (misc->source_eswitch_owner_vhca_id == dmn->info.caps.gvmi) in dr_ste_v0_build_src_gvmi_qpn_tag() 1657 vport_dmn = dmn; in dr_ste_v0_build_src_gvmi_qpn_tag() 1658 else if (dmn->peer_dmn && (misc->source_eswitch_owner_vhca_id == in dr_ste_v0_build_src_gvmi_qpn_tag() 1659 dmn->peer_dmn->info.caps.gvmi)) in dr_ste_v0_build_src_gvmi_qpn_tag() 1660 vport_dmn = dmn->peer_dmn; in dr_ste_v0_build_src_gvmi_qpn_tag() 1666 vport_dmn = dmn; in dr_ste_v0_build_src_gvmi_qpn_tag() [all …]
|
| A D | dr_ste.h | 160 void (*set_actions_rx)(struct mlx5dr_domain *dmn, 165 void (*set_actions_tx)(struct mlx5dr_domain *dmn,
|
| A D | dr_ste_v1.c | 513 static void dr_ste_v1_set_actions_tx(struct mlx5dr_domain *dmn, in dr_ste_v1_set_actions_tx() argument 633 static void dr_ste_v1_set_actions_rx(struct mlx5dr_domain *dmn, in dr_ste_v1_set_actions_rx() argument 1780 struct mlx5dr_domain *dmn = sb->dmn; in dr_ste_v1_build_src_gvmi_qpn_tag() local 1788 if (misc->source_eswitch_owner_vhca_id == dmn->info.caps.gvmi) in dr_ste_v1_build_src_gvmi_qpn_tag() 1789 vport_dmn = dmn; in dr_ste_v1_build_src_gvmi_qpn_tag() 1790 else if (dmn->peer_dmn && (misc->source_eswitch_owner_vhca_id == in dr_ste_v1_build_src_gvmi_qpn_tag() 1791 dmn->peer_dmn->info.caps.gvmi)) in dr_ste_v1_build_src_gvmi_qpn_tag() 1792 vport_dmn = dmn->peer_dmn; in dr_ste_v1_build_src_gvmi_qpn_tag() 1798 vport_dmn = dmn; in dr_ste_v1_build_src_gvmi_qpn_tag() 1806 mlx5dr_err(dmn, "Vport 0x%x is disabled or invalid\n", in dr_ste_v1_build_src_gvmi_qpn_tag()
|
| /linux/drivers/powercap/ |
| A D | intel_rapl_common.c | 1080 int dmn, prim; in rapl_update_domain_data() local 1083 for (dmn = 0; dmn < rp->nr_domains; dmn++) { in rapl_update_domain_data() 1085 rp->domains[dmn].name); in rapl_update_domain_data() 1088 if (!rapl_read_data_raw(&rp->domains[dmn], prim, in rapl_update_domain_data() 1090 rp->domains[dmn].rdd.primitives[prim] = val; in rapl_update_domain_data()
|