| /net/ax25/ |
| A D | ax25_ip.c | 49 unsigned char *buff; in ax25_hard_header() local 62 buff[6] &= ~AX25_CBIT; in ax25_hard_header() 63 buff[6] &= ~AX25_EBIT; in ax25_hard_header() 64 buff[6] |= AX25_SSSID_SPARE; in ax25_hard_header() 65 buff += AX25_ADDR_LEN; in ax25_hard_header() 72 buff[6] &= ~AX25_CBIT; in ax25_hard_header() 73 buff[6] |= AX25_EBIT; in ax25_hard_header() 75 buff += AX25_ADDR_LEN; in ax25_hard_header() 82 *buff++ = AX25_P_IP; in ax25_hard_header() 85 *buff++ = AX25_P_ARP; in ax25_hard_header() [all …]
|
| /net/netrom/ |
| A D | nr_dev.c | 72 buff[6] &= ~AX25_CBIT; in nr_header() 73 buff[6] &= ~AX25_EBIT; in nr_header() 75 buff += AX25_ADDR_LEN; in nr_header() 79 buff[6] &= ~AX25_CBIT; in nr_header() 80 buff[6] |= AX25_EBIT; in nr_header() 82 buff += AX25_ADDR_LEN; in nr_header() 86 *buff++ = NR_PROTO_IP; in nr_header() 87 *buff++ = NR_PROTO_IP; in nr_header() 88 *buff++ = 0; in nr_header() 89 *buff++ = 0; in nr_header() [all …]
|
| /net/ipv4/ |
| A D | proc.c | 418 unsigned long buff[TCPUDP_MIB_MAX]; in snmp_seq_show_tcp_udp() local 434 seq_printf(seq, " %ld", buff[i]); in snmp_seq_show_tcp_udp() 436 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp() 448 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp() 460 seq_printf(seq, " %lu", buff[i]); in snmp_seq_show_tcp_udp() 486 unsigned long *buff; in netstat_seq_show() local 496 if (buff) { in netstat_seq_show() 500 seq_printf(seq, " %lu", buff[i]); in netstat_seq_show() 512 if (buff) { in netstat_seq_show() 513 u64 *buff64 = (u64 *)buff; in netstat_seq_show() [all …]
|
| A D | tcp_output.c | 1611 struct sk_buff *buff; in tcp_fragment() local 1641 if (!buff) in tcp_fragment() 1649 buff->truesize += nlen; in tcp_fragment() 1683 tcp_skb_pcount(buff); in tcp_fragment() 2168 struct sk_buff *buff; in tso_fragment() local 2175 if (unlikely(!buff)) in tso_fragment() 4082 struct sk_buff *buff; in tcp_connect() local 4145 if (unlikely(!buff)) in tcp_connect() 4170 if (unlikely(buff)) { in tcp_connect() 4249 struct sk_buff *buff; in __tcp_send_ack() local [all …]
|
| A D | icmp.c | 1052 char buff[IFNAMSIZ]; in icmp_build_probe() local 1092 memset(buff, 0, sizeof(buff)); in icmp_build_probe() 1093 memcpy(buff, &iio->ident.name, ident_len); in icmp_build_probe() 1094 dev = dev_get_by_name(net, buff); in icmp_build_probe()
|
| /net/rose/ |
| A D | rose_dev.c | 39 unsigned char *buff = skb_push(skb, ROSE_MIN_LEN + 2); in rose_header() local 42 memcpy(buff + 7, daddr, dev->addr_len); in rose_header() 44 *buff++ = ROSE_GFI | ROSE_Q_BIT; in rose_header() 45 *buff++ = 0x00; in rose_header() 46 *buff++ = ROSE_DATA; in rose_header() 47 *buff++ = 0x7F; in rose_header() 48 *buff++ = AX25_P_IP; in rose_header()
|
| /net/atm/ |
| A D | mpoa_proc.c | 51 static ssize_t proc_mpc_write(struct file *file, const char __user *buff, 54 static int parse_qos(const char *buff); 206 static ssize_t proc_mpc_write(struct file *file, const char __user *buff, in proc_mpc_write() argument 223 if (get_user(*p, buff++)) { in proc_mpc_write() 242 static int parse_qos(const char *buff) in parse_qos() argument 254 if (sscanf(buff, "del %hhu.%hhu.%hhu.%hhu", in parse_qos() 260 if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=tx", in parse_qos() 264 } else if (sscanf(buff, "add %hhu.%hhu.%hhu.%hhu tx=%d,%d rx=%d,%d", in parse_qos()
|
| A D | lec.c | 132 char *buff; in lec_handle_bridge() local 140 buff = skb->data + skb->dev->hard_header_len; in lec_handle_bridge() 141 if (*buff++ == 0x42 && *buff++ == 0x42 && *buff++ == 0x03) { in lec_handle_bridge() 152 buff += 4; in lec_handle_bridge() 153 mesg->content.normal.flag = *buff & 0x01; in lec_handle_bridge()
|
| A D | mpc.c | 496 char *buff; in send_via_shortcut() local 507 buff = skb->data + mpc->dev->hard_header_len; in send_via_shortcut() 508 iph = (struct iphdr *)buff; in send_via_shortcut()
|
| /net/netfilter/ipvs/ |
| A D | ip_vs_sync.c | 570 if (buff) { in ip_vs_sync_conn_v0() 573 if (buff->head + len > buff->end || !m->nr_conns) { in ip_vs_sync_conn_v0() 576 buff = NULL; in ip_vs_sync_conn_v0() 579 if (!buff) { in ip_vs_sync_conn_v0() 581 if (!buff) { in ip_vs_sync_conn_v0() 689 if (buff) { in ip_vs_sync_conn() 693 if (buff->head + len + pad > buff->end || m->reserved) { in ip_vs_sync_conn() 696 buff = NULL; in ip_vs_sync_conn() 701 if (!buff) { in ip_vs_sync_conn() 703 if (!buff) { in ip_vs_sync_conn() [all …]
|
| /net/xfrm/ |
| A D | xfrm_proc.c | 53 unsigned long buff[LINUX_MIB_XFRMMAX]; in xfrm_statistics_seq_show() local 57 memset(buff, 0, sizeof(unsigned long) * LINUX_MIB_XFRMMAX); in xfrm_statistics_seq_show() 60 snmp_get_cpu_field_batch(buff, xfrm_mib_list, in xfrm_statistics_seq_show() 64 buff[i]); in xfrm_statistics_seq_show()
|
| /net/netfilter/ |
| A D | nft_exthdr.c | 192 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_eval() local 199 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_eval() 236 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_set_eval() local 242 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_set_eval() 316 u8 buff[sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE]; in nft_exthdr_tcp_strip_eval() local 322 tcph = nft_tcp_header_pointer(pkt, sizeof(buff), buff, &tcphdr_len); in nft_exthdr_tcp_strip_eval()
|
| A D | nf_conntrack_proto_tcp.c | 338 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_options() local 346 length, buff); in tcp_options() 394 unsigned char buff[(15 * 4) - sizeof(struct tcphdr)]; in tcp_sack() local 403 length, buff); in tcp_sack()
|
| /net/ipv6/ |
| A D | proc.c | 187 unsigned long buff[SNMP_MIB_MAX]; in snmp6_seq_show_item() local 191 memset(buff, 0, sizeof(unsigned long) * SNMP_MIB_MAX); in snmp6_seq_show_item() 193 snmp_get_cpu_field_batch(buff, itemlist, pcpumib); in snmp6_seq_show_item() 196 itemlist[i].name, buff[i]); in snmp6_seq_show_item()
|
| A D | tcp_ipv6.c | 877 struct sk_buff *buff; in tcp_v6_send_response() local 898 buff = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC); in tcp_v6_send_response() 899 if (!buff) in tcp_v6_send_response() 902 skb_reserve(buff, MAX_TCP_HEADER); in tcp_v6_send_response() 904 t1 = skb_push(buff, tot_len); in tcp_v6_send_response() 905 skb_reset_transport_header(buff); in tcp_v6_send_response() 959 buff->ip_summed = CHECKSUM_PARTIAL; in tcp_v6_send_response() 975 skb_set_owner_edemux(buff, (struct sock *)sk); in tcp_v6_send_response() 985 skb_set_hash(buff, txhash, PKT_HASH_TYPE_L4); in tcp_v6_send_response() 1002 skb_dst_set(buff, dst); in tcp_v6_send_response() [all …]
|
| A D | ndisc.c | 1660 struct sk_buff *buff; in ndisc_send_redirect() local 1740 buff = ndisc_alloc_skb(dev, sizeof(*msg) + optlen); in ndisc_send_redirect() 1741 if (!buff) in ndisc_send_redirect() 1744 msg = skb_put(buff, sizeof(*msg)); in ndisc_send_redirect() 1758 ndisc_fill_redirect_addr_option(buff, ha, ops_data); in ndisc_send_redirect() 1765 ndisc_fill_redirect_hdr_option(buff, skb, rd_len); in ndisc_send_redirect() 1767 skb_dst_set(buff, dst); in ndisc_send_redirect() 1768 ndisc_send_skb(buff, &ipv6_hdr(skb)->saddr, &saddr_buf); in ndisc_send_redirect()
|
| A D | addrconf.c | 5764 u64 buff[IPSTATS_MIB_MAX]; in __snmp6_fill_stats64() local 5769 memset(buff, 0, sizeof(buff)); in __snmp6_fill_stats64() 5770 buff[0] = IPSTATS_MIB_MAX; in __snmp6_fill_stats64() 5774 buff[i] += snmp_get_cpu_field64(mib, c, i, syncpoff); in __snmp6_fill_stats64() 5777 memcpy(stats, buff, IPSTATS_MIB_MAX * sizeof(u64)); in __snmp6_fill_stats64()
|
| /net/sctp/ |
| A D | proc.c | 61 unsigned long buff[SCTP_MIB_MAX]; in sctp_snmp_seq_show() local 65 memset(buff, 0, sizeof(unsigned long) * SCTP_MIB_MAX); in sctp_snmp_seq_show() 67 snmp_get_cpu_field_batch(buff, sctp_snmp_list, in sctp_snmp_seq_show() 71 buff[i]); in sctp_snmp_seq_show()
|
| /net/xdp/ |
| A D | xsk_buff_pool.c | 648 struct xdp_buff *buff; in xp_alloc_slow() local 650 buff = xp_alloc(pool); in xp_alloc_slow() 651 if (unlikely(!buff)) in xp_alloc_slow() 653 *xdp = buff; in xp_alloc_slow()
|
| /net/ |
| A D | socket.c | 2193 int __sys_sendto(int fd, void __user *buff, size_t len, unsigned int flags, in __sys_sendto() argument 2201 err = import_ubuf(ITER_SOURCE, buff, len, &msg.msg_iter); in __sys_sendto() 2231 SYSCALL_DEFINE6(sendto, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE6() argument 2235 return __sys_sendto(fd, buff, len, flags, addr, addr_len); in SYSCALL_DEFINE6() 2242 SYSCALL_DEFINE4(send, int, fd, void __user *, buff, size_t, len, in SYSCALL_DEFINE4() argument 2245 return __sys_sendto(fd, buff, len, flags, NULL, 0); in SYSCALL_DEFINE4()
|
| /net/packet/ |
| A D | af_packet.c | 197 static void packet_increment_head(struct packet_ring_buffer *buff); 1245 static void packet_increment_head(struct packet_ring_buffer *buff) in packet_increment_head() argument 1247 buff->head = buff->head != buff->frame_max ? buff->head+1 : 0; in packet_increment_head()
|
| /net/core/ |
| A D | neighbour.c | 1254 struct sk_buff *buff; in __neigh_event_send() local 1256 buff = __skb_dequeue(&neigh->arp_queue); in __neigh_event_send() 1257 if (!buff) in __neigh_event_send() 1259 neigh->arp_queue_len_bytes -= buff->truesize; in __neigh_event_send() 1260 kfree_skb_reason(buff, SKB_DROP_REASON_NEIGH_QUEUEFULL); in __neigh_event_send()
|