| /linux/tools/testing/selftests/bpf/prog_tests/ |
| A D | flow_dissector.c | 123 .nhoff = ETH_HLEN, 144 .nhoff = ETH_HLEN, 215 .nhoff = ETH_HLEN, 241 .nhoff = ETH_HLEN, 264 .nhoff = ETH_HLEN, 290 .nhoff = ETH_HLEN, 313 .nhoff = ETH_HLEN, 337 .nhoff = ETH_HLEN, 360 .nhoff = ETH_HLEN, 388 .nhoff = ETH_HLEN, [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| A D | test_tc_tunnel.c | 30 #define L2_PAD_SZ (sizeof(struct vxlanhdr) + ETH_HLEN) 120 if (bpf_skb_load_bytes(skb, ETH_HLEN, &iph6_inner, in __encap_ipv4() 137 if (bpf_skb_load_bytes(skb, ETH_HLEN, &iph_inner, in __encap_ipv4() 148 if (bpf_skb_load_bytes(skb, ETH_HLEN + tcp_off, in __encap_ipv4() 166 l2_len = ETH_HLEN; in __encap_ipv4() 218 if (bpf_skb_load_bytes(skb, 0, l2_hdr, ETH_HLEN)) in __encap_ipv4() 274 if (bpf_skb_load_bytes(skb, ETH_HLEN, &iph_inner, in __encap_ipv6() 297 l2_len = ETH_HLEN; in __encap_ipv6() 639 olen += ETH_HLEN; in decap_internal() 652 olen += ETH_HLEN; in decap_internal() [all …]
|
| A D | test_check_mtu.c | 71 delta = GLOBAL_USER_MTU - (data_len - ETH_HLEN) + 1; in xdp_exceed_mtu() 96 delta = -((data_len - ETH_HLEN) + 1); in xdp_minus_delta() 123 __u32 mtu_len = data_len - ETH_HLEN; in xdp_input_len() 186 delta = GLOBAL_USER_MTU - (skb_len - ETH_HLEN) + 1; in tc_exceed_mtu() 213 delta = GLOBAL_USER_MTU - (data_len - ETH_HLEN) + 1; in tc_exceed_mtu_da() 236 delta = -((skb_len - ETH_HLEN) + 1); in tc_minus_delta()
|
| A D | decap_sanity.c | 42 if (bpf_skb_load_bytes(skb, ETH_HLEN, &ip6h, sizeof(ip6h))) in decap_sanity() 48 if (bpf_skb_load_bytes(skb, ETH_HLEN + sizeof(ip6h), &udph, sizeof(udph))) in decap_sanity() 56 err = bpf_skb_adjust_room(skb, -(s32)(ETH_HLEN + sizeof(ip6h) + sizeof(udph)), in decap_sanity()
|
| A D | crypto_sanity.c | 29 if (bpf_skb_load_bytes(skb, ETH_HLEN, &ip6h, sizeof(ip6h))) in skb_dynptr_validate() 35 if (bpf_skb_load_bytes(skb, ETH_HLEN + sizeof(ip6h), &udph, sizeof(udph))) in skb_dynptr_validate() 41 offset = ETH_HLEN + sizeof(ip6h) + sizeof(udph); in skb_dynptr_validate()
|
| /linux/samples/bpf/ |
| A D | tcbpf1_kern.c | 21 #define IP_CSUM_OFF (ETH_HLEN + offsetof(struct iphdr, check)) 22 #define TOS_OFF (ETH_HLEN + offsetof(struct iphdr, tos)) 32 #define TCP_CSUM_OFF (ETH_HLEN + sizeof(struct iphdr) + offsetof(struct tcphdr, check)) 33 #define IP_SRC_OFF (ETH_HLEN + offsetof(struct iphdr, saddr)) 46 #define TCP_DPORT_OFF (ETH_HLEN + sizeof(struct iphdr) + offsetof(struct tcphdr, dest)) 58 __u8 proto = load_byte(skb, ETH_HLEN + offsetof(struct iphdr, protocol)); in bpf_prog1()
|
| A D | parse_ldabs.c | 30 __u64 troff = ETH_HLEN + sizeof(struct iphdr); in handle_ingress() 34 if (load_byte(skb, ETH_HLEN + offsetof(struct iphdr, protocol)) != IPPROTO_UDP || in handle_ingress() 35 load_byte(skb, ETH_HLEN) != 0x45) in handle_ingress() 37 if (ip_is_fragment(skb, ETH_HLEN)) in handle_ingress()
|
| /linux/drivers/net/wireless/ath/ath10k/ |
| A D | wow.c | 84 u8 hdr_8023_pattern[ETH_HLEN] = {}; in ath10k_wow_convert_8023_to_80211() 85 u8 hdr_8023_bit_mask[ETH_HLEN] = {}; in ath10k_wow_convert_8023_to_80211() 107 old->pattern, ETH_HLEN - old->pkt_offset); in ath10k_wow_convert_8023_to_80211() 109 old->mask, ETH_HLEN - old->pkt_offset); in ath10k_wow_convert_8023_to_80211() 139 if (total_len > ETH_HLEN) in ath10k_wow_convert_8023_to_80211() 159 if (total_len > ETH_HLEN) { in ath10k_wow_convert_8023_to_80211() 163 total_len - ETH_HLEN); in ath10k_wow_convert_8023_to_80211() 166 total_len - ETH_HLEN); in ath10k_wow_convert_8023_to_80211() 168 new->pattern_len += total_len - ETH_HLEN; in ath10k_wow_convert_8023_to_80211() 343 if (patterns[i].pkt_offset < ETH_HLEN) { in ath10k_vif_wow_set_wakeups() [all …]
|
| /linux/tools/testing/selftests/net/ |
| A D | gro.c | 121 ipproto_off = ETH_HLEN + next_off; in setup_sock_filter() 318 fill_networklayer(buf + ETH_HLEN, payload_len); in create_packet() 361 static char buf[ETH_HLEN + IP_MAXPACKET]; in send_data_pkts() 598 iph = (struct iphdr *)(optpkt + ETH_HLEN); in add_ipv4_ts_option() 640 iph1 = (struct iphdr *)(buf1 + ETH_HLEN); in send_flush_id_case() 641 iph2 = (struct iphdr *)(buf2 + ETH_HLEN); in send_flush_id_case() 642 iph3 = (struct iphdr *)(buf3 + ETH_HLEN); in send_flush_id_case() 780 fill_networklayer(buf + ETH_HLEN, PAYLOAD_LEN); in send_fragment4() 840 buf[ETH_HLEN + 1] ^= 0x2; // ECN set to 10 in send_changed_ECN() 844 buf[ETH_HLEN + 1] ^= 0x20; // ECN set to 10 in send_changed_ECN() [all …]
|
| A D | psock_snd.c | 50 #define TEST_SZ (sizeof(struct virtio_net_hdr) + ETH_HLEN + ETH_MAX_MTU + 1) 80 vh->hdr_len = ETH_HLEN + sizeof(struct iphdr) + sizeof(struct udphdr); in build_vnet_header() 84 vh->csum_start = ETH_HLEN + sizeof(struct iphdr); in build_vnet_header() 106 uint16_t *tag = header + ETH_HLEN; in build_eth_header() 110 return ETH_HLEN + 4; in build_eth_header() 114 return ETH_HLEN; in build_eth_header() 193 buf += ETH_HLEN; in do_send() 194 len -= ETH_HLEN; in do_send()
|
| /linux/net/core/ |
| A D | timestamping.c | 54 if (skb_headroom(skb) < ETH_HLEN) in skb_defer_rx_timestamp() 57 __skb_push(skb, ETH_HLEN); in skb_defer_rx_timestamp() 61 __skb_pull(skb, ETH_HLEN); in skb_defer_rx_timestamp()
|
| /linux/net/dsa/ |
| A D | tag.c | 39 if (skb_headroom(skb) < ETH_HLEN) in dsa_skb_defer_rx_timestamp() 42 __skb_push(skb, ETH_HLEN); in dsa_skb_defer_rx_timestamp() 46 __skb_pull(skb, ETH_HLEN); in dsa_skb_defer_rx_timestamp() 93 skb_push(skb, ETH_HLEN); in dsa_switch_rcv() 118 dev_sw_netstats_rx_add(skb->dev, skb->len + ETH_HLEN); in dsa_switch_rcv()
|
| A D | tag_ocelot.c | 84 skb_push(skb, ETH_HLEN); in ocelot_rcv() 100 skb_pull(skb, ETH_HLEN); in ocelot_rcv() 147 skb_push_rcsum(skb, ETH_HLEN); in ocelot_rcv() 149 skb_pull_rcsum(skb, ETH_HLEN); in ocelot_rcv()
|
| A D | tag_lan9303.c | 95 skb_push_rcsum(skb, ETH_HLEN); in lan9303_rcv() 97 skb_pull_rcsum(skb, ETH_HLEN); in lan9303_rcv()
|
| /linux/drivers/net/wireless/ath/ath11k/ |
| A D | wow.c | 154 u8 hdr_8023_pattern[ETH_HLEN] = {}; in ath11k_wow_convert_8023_to_80211() 155 u8 hdr_8023_bit_mask[ETH_HLEN] = {}; in ath11k_wow_convert_8023_to_80211() 177 old->pattern, ETH_HLEN - old->pkt_offset); in ath11k_wow_convert_8023_to_80211() 179 old->mask, ETH_HLEN - old->pkt_offset); in ath11k_wow_convert_8023_to_80211() 209 if (total_len > ETH_HLEN) in ath11k_wow_convert_8023_to_80211() 229 if (total_len > ETH_HLEN) { in ath11k_wow_convert_8023_to_80211() 233 total_len - ETH_HLEN); in ath11k_wow_convert_8023_to_80211() 236 total_len - ETH_HLEN); in ath11k_wow_convert_8023_to_80211() 238 new->pattern_len += total_len - ETH_HLEN; in ath11k_wow_convert_8023_to_80211() 415 if (patterns[i].pkt_offset < ETH_HLEN) { in ath11k_vif_wow_set_wakeups() [all …]
|
| /linux/drivers/net/ |
| A D | ntb_netdev.c | 130 skb = netdev_alloc_skb(ndev, ndev->mtu + ETH_HLEN); in ntb_netdev_rx_handler() 138 rc = ntb_transport_rx_enqueue(qp, skb, skb->data, ndev->mtu + ETH_HLEN); in ntb_netdev_rx_handler() 255 skb = netdev_alloc_skb(ndev, ndev->mtu + ETH_HLEN); in ntb_netdev_open() 262 ndev->mtu + ETH_HLEN); in ntb_netdev_open() 305 if (new_mtu > ntb_transport_max_size(dev->qp) - ETH_HLEN) in ntb_netdev_change_mtu() 323 skb = netdev_alloc_skb(ndev, new_mtu + ETH_HLEN); in ntb_netdev_change_mtu() 330 new_mtu + ETH_HLEN); in ntb_netdev_change_mtu() 446 ndev->mtu = ntb_transport_max_size(dev->qp) - ETH_HLEN; in ntb_netdev_probe()
|
| /linux/net/ethernet/ |
| A D | eth.c | 83 struct ethhdr *eth = skb_push(skb, ETH_HLEN); in eth_header() 100 return ETH_HLEN; in eth_header() 109 return ETH_HLEN; in eth_header() 112 return -ETH_HLEN; in eth_header() 235 smp_store_release(&hh->hh_len, ETH_HLEN); in eth_header_cache() 349 dev->hard_header_len = ETH_HLEN; in ether_setup() 350 dev->min_header_len = ETH_HLEN; in ether_setup()
|
| /linux/net/batman-adv/ |
| A D | fragmentation.c | 279 memmove(skb_out->data - ETH_HLEN, skb_mac_header(skb_out), ETH_HLEN); in batadv_frag_merge_packets() 280 skb_set_mac_header(skb_out, -ETH_HLEN); in batadv_frag_merge_packets() 371 skb->len + ETH_HLEN); in batadv_frag_skb_fwd() 522 skb_fragment->len + ETH_HLEN); in batadv_frag_send_packet() 535 ret = skb_cow_head(skb, ETH_HLEN + header_size); in batadv_frag_send_packet() 545 skb->len + ETH_HLEN); in batadv_frag_send_packet()
|
| A D | network-coding.c | 1172 count = skb_src->len + ETH_HLEN; in batadv_nc_code_packets() 1173 count += skb_dest->len + ETH_HLEN; in batadv_nc_code_packets() 1179 count = skb_src->len + ETH_HLEN; in batadv_nc_code_packets() 1180 count += skb_dest->len + ETH_HLEN; in batadv_nc_code_packets() 1188 skb_src->len + ETH_HLEN); in batadv_nc_code_packets() 1191 skb_dest->len + ETH_HLEN); in batadv_nc_code_packets() 1197 skb_src->len + ETH_HLEN); in batadv_nc_code_packets() 1200 skb_dest->len + ETH_HLEN); in batadv_nc_code_packets() 1383 skb_push(skb, ETH_HLEN); in batadv_nc_skb_store_before_coding() 1620 skb_push(skb, ETH_HLEN); in batadv_nc_skb_store_sniffed_unicast() [all …]
|
| A D | routing.c | 239 if (skb_cow(skb, ETH_HLEN) < 0) in batadv_recv_my_icmp_packet() 305 if (skb_cow(skb, ETH_HLEN) < 0) in batadv_recv_icmp_ttl_exceeded() 377 if (skb_cow(skb, ETH_HLEN) < 0) in batadv_recv_icmp_packet() 405 if (skb_cow(skb, ETH_HLEN) < 0) in batadv_recv_icmp_packet() 682 if (skb_cow(skb, ETH_HLEN) < 0) in batadv_route_unicast_packet() 714 len + ETH_HLEN); in batadv_route_unicast_packet() 798 if (!pskb_may_pull(skb, hdr_len + ETH_HLEN)) in batadv_check_unicast_ttvn() 1300 if (skb_cow(skb, ETH_HLEN) < 0) in batadv_recv_mcast_packet() 1322 skb->len + ETH_HLEN); in batadv_recv_mcast_packet() 1327 if (ret == NET_RX_SUCCESS && (skb->len - hdr_size >= ETH_HLEN)) { in batadv_recv_mcast_packet()
|
| /linux/drivers/net/ethernet/broadcom/asp2/ |
| A D | bcmasp.c | 342 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset + 1, in bcmasp_netfilt_tcpip4_wr() 345 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset + 12, in bcmasp_netfilt_tcpip4_wr() 348 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset + 16, in bcmasp_netfilt_tcpip4_wr() 351 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset + 20, in bcmasp_netfilt_tcpip4_wr() 354 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset + 22, in bcmasp_netfilt_tcpip4_wr() 373 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset, in bcmasp_netfilt_tcpip6_wr() 375 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset + 8, in bcmasp_netfilt_tcpip6_wr() 378 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset + 24, in bcmasp_netfilt_tcpip6_wr() 381 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset + 40, in bcmasp_netfilt_tcpip6_wr() 465 bcmasp_netfilt_wr_m_wake(priv, nfilt, ETH_HLEN + offset, in bcmasp_netfilt_wr_to_hw() [all …]
|
| /linux/net/bridge/ |
| A D | br_forward.c | 35 skb_push(skb, ETH_HLEN); in br_dev_queue_push_xmit() 103 skb_push(skb, ETH_HLEN); in __br_forward() 270 __skb_push(skb, ETH_HLEN); in maybe_deliver_addr() 272 __skb_pull(skb, ETH_HLEN); in maybe_deliver_addr() 279 __skb_pull(skb, ETH_HLEN); in maybe_deliver_addr()
|
| /linux/drivers/net/ethernet/qualcomm/ |
| A D | qca_7k_common.h | 34 #define QCAFRM_MIN_MTU (ETH_ZLEN - ETH_HLEN) 38 #define QCAFRM_MIN_LEN (QCAFRM_MIN_MTU + ETH_HLEN)
|
| /linux/drivers/staging/gdm724x/ |
| A D | gdm_lte.c | 81 len = skb->len + ETH_HLEN; in gdm_lte_rx() 125 mac_header_len = ETH_HLEN; in gdm_lte_emulate_arp() 167 skb_pull(skb_out, ETH_HLEN); in gdm_lte_emulate_arp() 249 mac_header_len = ETH_HLEN; in gdm_lte_emulate_ndp() 328 skb_pull(skb_out, ETH_HLEN); in gdm_lte_emulate_ndp() 358 network_data = skb->data + ETH_HLEN; in gdm_lte_tx_nic_type() 441 data_buf = skb->data + (VLAN_ETH_HLEN - ETH_HLEN); in gdm_lte_tx() 442 data_len = skb->len - (VLAN_ETH_HLEN - ETH_HLEN); in gdm_lte_tx() 628 mac_header_len = ETH_HLEN; in gdm_lte_netif_rx() 670 skb_pull(skb, ETH_HLEN); in gdm_lte_netif_rx()
|
| /linux/net/bridge/netfilter/ |
| A D | nf_conntrack_bridge.c | 334 skb_copy_from_linear_data_offset(skb, -ETH_HLEN, data->mac, ETH_HLEN); in nf_ct_bridge_frag_save() 370 err = skb_cow_head(skb, ETH_HLEN); in nf_ct_bridge_frag_restore() 380 skb_copy_to_linear_data_offset(skb, -ETH_HLEN, data->mac, ETH_HLEN); in nf_ct_bridge_frag_restore()
|