| /drivers/net/ethernet/chelsio/cxgb4/ |
| A D | smt.c | 65 static struct smt_entry *find_or_alloc_smte(struct smt_data *s, u8 *smac) in find_or_alloc_smte() argument 79 if (memcmp(e->src_mac, smac, ETH_ALEN) == 0) in find_or_alloc_smte() 210 u8 *smac) in t4_smt_alloc_switching() argument 216 e = find_or_alloc_smte(s, smac); in t4_smt_alloc_switching() 223 memcpy(e->src_mac, smac, ETH_ALEN); in t4_smt_alloc_switching() 242 struct smt_entry *cxgb4_smt_alloc_switching(struct net_device *dev, u8 *smac) in cxgb4_smt_alloc_switching() argument 246 return t4_smt_alloc_switching(adap, 0x0, smac); in cxgb4_smt_alloc_switching()
|
| A D | smt.h | 73 struct smt_entry *cxgb4_smt_alloc_switching(struct net_device *dev, u8 *smac);
|
| A D | cxgb4_tc_flower.c | 49 PEDIT_FIELDS(ETH_, SMAC_15_0, 2, smac, 0), 50 PEDIT_FIELDS(ETH_, SMAC_47_16, 4, smac, 2),
|
| /drivers/net/ethernet/marvell/prestera/ |
| A D | prestera_main.c | 292 if (smac.valid) { in prestera_pcs_get_state() 293 state->link = smac.oper ? 1 : 0; in prestera_pcs_get_state() 296 state->speed = smac.speed; in prestera_pcs_get_state() 297 state->duplex = smac.duplex; in prestera_pcs_get_state() 811 memset(&smac, 0, sizeof(smac)); in prestera_port_handle_event() 812 smac.valid = true; in prestera_port_handle_event() 813 smac.oper = pevt->data.mac.oper; in prestera_port_handle_event() 814 if (smac.oper) { in prestera_port_handle_event() 815 smac.mode = pevt->data.mac.mode; in prestera_port_handle_event() 818 smac.fc = pevt->data.mac.fc; in prestera_port_handle_event() [all …]
|
| /drivers/infiniband/core/ |
| A D | lag.c | 19 u8 smac[ETH_ALEN]; in rdma_build_skb() local 73 rdma_read_gid_l2_fields(ah_attr->grh.sgid_attr, NULL, smac); in rdma_build_skb() 74 memcpy(eth->h_source, smac, ETH_ALEN); in rdma_build_skb()
|
| A D | cache.c | 1365 u16 *vlan_id, u8 *smac) in rdma_read_gid_l2_fields() argument 1378 if (smac) in rdma_read_gid_l2_fields() 1379 ether_addr_copy(smac, ndev->dev_addr); in rdma_read_gid_l2_fields()
|
| /drivers/infiniband/hw/mlx4/ |
| A D | qp.c | 1420 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) { in destroy_qp_common() 1422 qp->pri.smac = 0; in destroy_qp_common() 1425 if (qp->alt.smac) { in destroy_qp_common() 1427 qp->alt.smac = 0; in destroy_qp_common() 1873 smac_info->smac != smac) { in _mlx4_set_path() 2303 u8 smac[ETH_ALEN]; in __mlx4_ib_modify_qp() local 2602 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) { in __mlx4_ib_modify_qp() 2604 qp->pri.smac = 0; in __mlx4_ib_modify_qp() 2609 qp->alt.smac = 0; in __mlx4_ib_modify_qp() 2636 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) in __mlx4_ib_modify_qp() [all …]
|
| A D | cq.c | 603 memcpy(&(wc->smac[0]), (char *)&hdr->tun.mac_31_0, 4); in use_tunnel_data() 604 memcpy(&(wc->smac[4]), (char *)&hdr->tun.slid_mac_47_32, 2); in use_tunnel_data() 870 memcpy(wc->smac, cqe->smac, ETH_ALEN); in mlx4_ib_poll_one()
|
| /drivers/net/ethernet/microchip/vcap/ |
| A D | vcap_tc.c | 32 struct vcap_u48_key smac, dmac; in vcap_tc_flower_handler_ethaddr_usage() local 38 vcap_netbytes_copy(smac.value, match.key->src, ETH_ALEN); in vcap_tc_flower_handler_ethaddr_usage() 39 vcap_netbytes_copy(smac.mask, match.mask->src, ETH_ALEN); in vcap_tc_flower_handler_ethaddr_usage() 40 err = vcap_rule_add_key_u48(st->vrule, smac_key, &smac); in vcap_tc_flower_handler_ethaddr_usage()
|
| /drivers/net/ethernet/broadcom/bnxt/ |
| A D | bnxt_tc.c | 408 ether_addr_copy(flow->l2_key.smac, match.key->src); in bnxt_tc_parse_flow() 409 ether_addr_copy(flow->l2_mask.smac, match.mask->src); in bnxt_tc_parse_flow() 691 memcpy(req->smac, flow->l2_key.smac, ETH_ALEN); in bnxt_hwrm_cfa_flow_alloc() 758 memcpy(&req->l2_rewrite_smac, &req->smac, ETH_ALEN); in bnxt_hwrm_cfa_flow_alloc() 766 memcpy(&req->l2_rewrite_smac, &req->smac, ETH_ALEN); in bnxt_hwrm_cfa_flow_alloc() 901 ether_addr_copy(encap->src_mac_addr, l2_info->smac); in hwrm_cfa_encap_record_alloc() 1057 if (bits_set(&flow->l2_key.smac, sizeof(flow->l2_key.smac)) && in bnxt_tc_can_offload() 1058 !is_exactmatch(flow->l2_mask.smac, sizeof(flow->l2_mask.smac))) { in bnxt_tc_can_offload() 1291 ether_addr_copy(l2_info->smac, dst_dev->dev_addr); in bnxt_tc_resolve_tunnel_hdrs() 1346 ether_addr_copy(decap_l2_info->dmac, l2_info.smac); in bnxt_tc_get_decap_handle()
|
| A D | bnxt_tc.h | 21 u8 smac[ETH_ALEN]; member
|
| /drivers/infiniband/hw/irdma/ |
| A D | puda.h | 42 u8 smac[ETH_ALEN]; member 78 u8 smac[ETH_ALEN]; member
|
| A D | puda.c | 296 info->smac[0] = (u8)((qword6 >> 40) & 0xFF); in irdma_puda_poll_info() 297 info->smac[1] = (u8)((qword6 >> 32) & 0xFF); in irdma_puda_poll_info() 298 info->smac[2] = (u8)((qword6 >> 24) & 0xFF); in irdma_puda_poll_info() 299 info->smac[3] = (u8)((qword6 >> 16) & 0xFF); in irdma_puda_poll_info() 300 info->smac[4] = (u8)((qword6 >> 8) & 0xFF); in irdma_puda_poll_info() 301 info->smac[5] = (u8)(qword6 & 0xFF); in irdma_puda_poll_info()
|
| /drivers/net/ethernet/mscc/ |
| A D | ocelot_vcap.c | 396 etype->smac.value, etype->smac.mask); in is2_entry_set() 418 llc->smac.value, llc->smac.mask); in is2_entry_set() 434 snap->smac.value, snap->smac.mask); in is2_entry_set() 445 arp->smac.value, arp->smac.mask); in is2_entry_set() 710 etype->smac.value, etype->smac.mask); in is1_entry_set()
|
| /drivers/infiniband/hw/bnxt_re/ |
| A D | qplib_sp.c | 311 struct bnxt_qplib_gid *gid, const u8 *smac, in bnxt_qplib_add_sgid() argument 373 req.src_mac[0] = cpu_to_be16(((u16 *)smac)[0]); in bnxt_qplib_add_sgid() 374 req.src_mac[1] = cpu_to_be16(((u16 *)smac)[1]); in bnxt_qplib_add_sgid() 375 req.src_mac[2] = cpu_to_be16(((u16 *)smac)[2]); in bnxt_qplib_add_sgid() 402 const u8 *smac) in bnxt_qplib_update_sgid() argument 428 req.src_mac[0] = cpu_to_be16(((u16 *)smac)[0]); in bnxt_qplib_update_sgid() 429 req.src_mac[1] = cpu_to_be16(((u16 *)smac)[1]); in bnxt_qplib_update_sgid() 430 req.src_mac[2] = cpu_to_be16(((u16 *)smac)[2]); in bnxt_qplib_update_sgid()
|
| A D | qplib_fp.h | 299 u8 smac[6]; member 423 u8 smac[6]; member
|
| A D | qplib_sp.h | 329 const u8 *smac);
|
| A D | ib_verbs.c | 1213 ether_addr_copy(qp->qplib_qp.smac, rdev->netdev->dev_addr); in bnxt_re_create_shadow_qp() 1447 ether_addr_copy(qplqp->smac, rdev->netdev->dev_addr); in bnxt_re_init_qp_attr() 2136 &qp->qplib_qp.smac[0]); in bnxt_re_modify_qp() 2408 ether_addr_copy(qp->qp1_hdr.eth.smac_h, qp->qplib_qp.smac); in bnxt_re_build_qp1_send_v2() 3706 memcpy(wc->smac, orig_cqe->smac, ETH_ALEN); in bnxt_re_process_res_shadow_qp_wc() 3746 memcpy(wc->smac, cqe->smac, ETH_ALEN); in bnxt_re_process_res_ud_wc() 3854 memcpy(wc->smac, cqe->smac, ETH_ALEN); in bnxt_re_poll_cq()
|
| /drivers/infiniband/hw/qedr/ |
| A D | qedr_roce_cm.c | 112 *((u32 *)&qp->rqe_wr_id[qp->rq.gsi_cons].smac[0]) = in qedr_ll2_complete_rx_packet() 114 *((u16 *)&qp->rqe_wr_id[qp->rq.gsi_cons].smac[4]) = in qedr_ll2_complete_rx_packet() 693 ether_addr_copy(wc[i].smac, qp->rqe_wr_id[qp->rq.cons].smac); in qedr_gsi_poll_cq()
|
| /drivers/net/ethernet/mellanox/mlxsw/ |
| A D | spectrum_span.h | 33 unsigned char smac[ETH_ALEN]; member
|
| A D | spectrum_span.c | 400 memcpy(sparmsp->smac, edev->dev_addr, ETH_ALEN); in mlxsw_sp_span_entry_tunnel_parms_common() 497 sparms.ttl, sparms.smac, in mlxsw_sp_span_entry_gretap4_configure() 601 mlxsw_reg_mpat_eth_rspan_l3_ipv6_pack(mpat_pl, sparms.ttl, sparms.smac, in mlxsw_sp_span_entry_gretap6_configure()
|
| /drivers/net/ethernet/mellanox/mlx5/core/diag/ |
| A D | fs_tracepoint.c | 68 DECLARE_MASK_VAL(u64, smac) = { in print_lyr_2_4_hdrs() 81 PRINT_MASKED_VALP(smac, u8 *, p, "%pM"); in print_lyr_2_4_hdrs()
|
| /drivers/infiniband/hw/ocrdma/ |
| A D | ocrdma_ah.c | 119 memcpy(ð.smac[0], &dev->nic_info.mac_addr[0], ETH_ALEN); in set_av_attr()
|
| /drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
| A D | ipsec.h | 114 u8 smac[ETH_ALEN]; member
|
| A D | ipsec.c | 283 dst = attrs->smac; in mlx5e_ipsec_init_macs() 303 src = attrs->smac; in mlx5e_ipsec_init_macs() 681 ether_addr_copy(attrs->smac, data->addr); in mlx5e_ipsec_handle_netdev_event()
|