/linux-6.3-rc2/net/netrom/ |
A D | nr_out.c | 34 struct sk_buff *skbn; in nr_output() local 49 skb_reserve(skbn, frontlen); in nr_output() 62 skbn->data[4] |= NR_MORE_FLAG; in nr_output() 99 struct sk_buff *skb, *skbn; in nr_send_nak_frame() local 108 skbn->data[2] = nr->va; in nr_send_nak_frame() 109 skbn->data[3] = nr->vr; in nr_send_nak_frame() 112 skbn->data[4] |= NR_CHOKE_FLAG; in nr_send_nak_frame() 114 nr_transmit_buffer(sk, skbn); in nr_send_nak_frame() 125 struct sk_buff *skb, *skbn; in nr_kick() local 161 skb_set_owner_w(skbn, sk); in nr_kick() [all …]
|
A D | nr_in.c | 31 struct sk_buff *skbo, *skbn = skb; in nr_queue_rx_frame() local 48 if ((skbn = alloc_skb(nr->fraglen, GFP_ATOMIC)) == NULL) in nr_queue_rx_frame() 51 skb_reset_transport_header(skbn); in nr_queue_rx_frame() 55 skb_put(skbn, skbo->len), in nr_queue_rx_frame() 63 return sock_queue_rcv_skb(sk, skbn); in nr_queue_rx_frame() 150 struct sk_buff *skbn; in nr_state3_machine() local 233 ns = skbn->data[17]; in nr_state3_machine() 239 skb_queue_tail(&temp_queue, skbn); in nr_state3_machine() 242 skb_queue_tail(&temp_queue, skbn); in nr_state3_machine() 244 kfree_skb(skbn); in nr_state3_machine() [all …]
|
A D | nr_loopback.c | 32 struct sk_buff *skbn; in nr_loopback_queue() local 34 if ((skbn = alloc_skb(skb->len, GFP_ATOMIC)) != NULL) { in nr_loopback_queue() 35 skb_copy_from_linear_data(skb, skb_put(skbn, skb->len), skb->len); in nr_loopback_queue() 36 skb_reset_transport_header(skbn); in nr_loopback_queue() 38 skb_queue_tail(&loopback_queue, skbn); in nr_loopback_queue()
|
A D | nr_subr.c | 213 struct sk_buff *skbn; in __nr_transmit_reply() local 219 if ((skbn = alloc_skb(len, GFP_ATOMIC)) == NULL) in __nr_transmit_reply() 222 skb_reserve(skbn, 0); in __nr_transmit_reply() 224 dptr = skb_put(skbn, NR_NETWORK_LEN + NR_TRANSPORT_LEN); in __nr_transmit_reply() 255 if (!nr_route_frame(skbn, NULL)) in __nr_transmit_reply() 256 kfree_skb(skbn); in __nr_transmit_reply()
|
A D | nr_route.c | 760 struct sk_buff *skbn; in nr_route_frame() local 813 if ((skbn=skb_copy_expand(skb, dev->hard_header_len, 0, GFP_ATOMIC)) == NULL) { in nr_route_frame() 820 skb=skbn; in nr_route_frame()
|
/linux-6.3-rc2/net/x25/ |
A D | x25_out.c | 49 struct sk_buff *skbn; in x25_output() local 67 skbn = sock_alloc_send_skb(sk, frontlen + max_len, in x25_output() 70 if (!skbn) { in x25_output() 81 skb_reserve(skbn, frontlen); in x25_output() 90 skb_push(skbn, header_len); in x25_output() 95 skbn->data[3] |= X25_EXT_M_BIT; in x25_output() 97 skbn->data[2] |= X25_STD_M_BIT; in x25_output() 100 skb_queue_tail(&sk->sk_write_queue, skbn); in x25_output() 138 struct sk_buff *skb, *skbn; in x25_kick() local 185 skb_set_owner_w(skbn, sk); in x25_kick() [all …]
|
A D | x25_forward.c | 23 struct sk_buff *skbn; in x25_forward_call() local 72 if ( (skbn = skb_clone(skb, GFP_ATOMIC)) == NULL){ in x25_forward_call() 75 x25_transmit_link(skbn, neigh_new); in x25_forward_call() 95 struct sk_buff *skbn; in x25_forward_data() local 115 if ( (skbn = pskb_copy(skb, GFP_ATOMIC)) == NULL){ in x25_forward_data() 119 x25_transmit_link(skbn, nb); in x25_forward_data()
|
A D | x25_in.c | 34 struct sk_buff *skbo, *skbn = skb; in x25_queue_rx_frame() local 47 if ((skbn = alloc_skb(len, GFP_ATOMIC)) == NULL){ in x25_queue_rx_frame() 54 skb_reset_transport_header(skbn); in x25_queue_rx_frame() 57 skb_copy_from_linear_data(skbo, skb_put(skbn, skbo->len), in x25_queue_rx_frame() 66 skb_put(skbn, skbo->len), in x25_queue_rx_frame() 74 skb_set_owner_r(skbn, sk); in x25_queue_rx_frame() 75 skb_queue_tail(&sk->sk_receive_queue, skbn); in x25_queue_rx_frame()
|
A D | x25_link.c | 67 struct sk_buff *skbn; in x25_link_control() local 125 while ((skbn = skb_dequeue(&nb->queue)) != NULL) in x25_link_control() 126 x25_send_frame(skbn, nb); in x25_link_control()
|
/linux-6.3-rc2/net/ax25/ |
A D | ax25_out.c | 118 struct sk_buff *skbn; in ax25_output() local 151 skb_set_owner_w(skbn, skb->sk); in ax25_output() 158 skb_reserve(skbn, frontlen + 2); in ax25_output() 159 skb_set_network_header(skbn, in ax25_output() 162 p = skb_push(skbn, 2); in ax25_output() 172 skb_reserve(skbn, frontlen + 1); in ax25_output() 173 skb_set_network_header(skbn, in ax25_output() 176 p = skb_push(skbn, 1); in ax25_output() 243 struct sk_buff *skb, *skbn; in ax25_kick() local 285 skb_set_owner_w(skbn, skb->sk); in ax25_kick() [all …]
|
A D | ax25_in.c | 36 struct sk_buff *skbn, *skbo; in ax25_rx_fragment() local 49 skbn = alloc_skb(AX25_MAX_HEADER_LEN + in ax25_rx_fragment() 52 if (!skbn) { in ax25_rx_fragment() 59 skbn->dev = ax25->ax25_dev->dev; in ax25_rx_fragment() 60 skb_reset_network_header(skbn); in ax25_rx_fragment() 61 skb_reset_transport_header(skbn); in ax25_rx_fragment() 66 skb_put(skbn, skbo->len), in ax25_rx_fragment() 73 if (ax25_rx_iframe(ax25, skbn) == 0) in ax25_rx_fragment() 74 kfree_skb(skbn); in ax25_rx_fragment() 117 if (skbn != NULL) { in ax25_rx_iframe() [all …]
|
/linux-6.3-rc2/net/rose/ |
A D | rose_out.c | 48 struct sk_buff *skb, *skbn; in rose_kick() local 76 if ((skbn = skb_clone(skb, GFP_ATOMIC)) == NULL) { in rose_kick() 81 skb_set_owner_w(skbn, sk); in rose_kick() 86 rose_send_iframe(sk, skbn); in rose_kick()
|
A D | rose_loopback.c | 36 struct sk_buff *skbn = NULL; in rose_loopback_queue() local 39 skbn = skb_clone(skb, GFP_ATOMIC); in rose_loopback_queue() 41 if (skbn) { in rose_loopback_queue() 43 skb_queue_tail(&loopback_queue, skbn); in rose_loopback_queue()
|
A D | rose_link.c | 141 struct sk_buff *skbn; in rose_link_rx_restart() local 167 while ((skbn = skb_dequeue(&neigh->queue)) != NULL) in rose_link_rx_restart() 168 if (!rose_send_frame(skbn, neigh)) in rose_link_rx_restart() 169 kfree_skb(skbn); in rose_link_rx_restart()
|
A D | af_rose.c | 1183 struct sk_buff *skbn; in rose_sendmsg() local 1194 if ((skbn = sock_alloc_send_skb(sk, frontlen + ROSE_PACLEN, 0, &err)) == NULL) { in rose_sendmsg() 1199 skbn->sk = sk; in rose_sendmsg() 1200 skbn->free = 1; in rose_sendmsg() 1201 skbn->arp = 1; in rose_sendmsg() 1203 skb_reserve(skbn, frontlen); in rose_sendmsg() 1208 skb_copy_from_linear_data(skb, skb_put(skbn, lg), lg); in rose_sendmsg() 1212 skb_push(skbn, ROSE_MIN_LEN); in rose_sendmsg() 1213 skb_copy_to_linear_data(skbn, header, ROSE_MIN_LEN); in rose_sendmsg() 1216 skbn->data[2] |= M_BIT; in rose_sendmsg() [all …]
|
/linux-6.3-rc2/net/lapb/ |
A D | lapb_out.c | 68 struct sk_buff *skb, *skbn; in lapb_kick() local 85 skbn = skb_copy(skb, GFP_ATOMIC); in lapb_kick() 86 if (!skbn) { in lapb_kick() 92 skb_set_owner_w(skbn, skb->sk); in lapb_kick() 97 lapb_send_iframe(lapb, skbn, LAPB_POLLOFF); in lapb_kick()
|
/linux-6.3-rc2/drivers/net/wwan/ |
A D | mhi_wwan_mbim.c | 296 struct sk_buff *skbn; in mhi_mbim_rx() local 307 skbn = netdev_alloc_skb(link->ndev, dgram_len); in mhi_mbim_rx() 308 if (!skbn) in mhi_mbim_rx() 311 skb_put(skbn, dgram_len); in mhi_mbim_rx() 312 skb_copy_bits(skb, dgram_offset, skbn->data, dgram_len); in mhi_mbim_rx() 314 switch (skbn->data[0] & 0xf0) { in mhi_mbim_rx() 316 skbn->protocol = htons(ETH_P_IP); in mhi_mbim_rx() 319 skbn->protocol = htons(ETH_P_IPV6); in mhi_mbim_rx() 324 dev_kfree_skb_any(skbn); in mhi_mbim_rx() 333 u64_stats_add(&link->rx_bytes, skbn->len); in mhi_mbim_rx() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/freescale/fs_enet/ |
A D | fs_enet-main.c | 91 struct sk_buff *skb, *skbn; in fs_enet_napi() local 228 skbn = fep->rx_skbuff[curidx]; in fs_enet_napi() 242 if (skbn != NULL) { in fs_enet_napi() 245 skbn->data, pkt_len); in fs_enet_napi() 246 swap(skb, skbn); in fs_enet_napi() 255 if (skbn) { in fs_enet_napi() 258 skb_align(skbn, ENET_RX_ALIGN); in fs_enet_napi() 266 skbn->data, in fs_enet_napi() 273 if (skbn != NULL) { in fs_enet_napi() 280 skbn = skb; in fs_enet_napi() [all …]
|
/linux-6.3-rc2/drivers/net/ethernet/qualcomm/rmnet/ |
A D | rmnet_map_data.c | 348 struct sk_buff *skbn; in rmnet_map_deaggregate() local 385 skbn = alloc_skb(packet_len + RMNET_MAP_DEAGGR_SPACING, GFP_ATOMIC); in rmnet_map_deaggregate() 386 if (!skbn) in rmnet_map_deaggregate() 389 skb_reserve(skbn, RMNET_MAP_DEAGGR_HEADROOM); in rmnet_map_deaggregate() 390 skb_put(skbn, packet_len); in rmnet_map_deaggregate() 391 memcpy(skbn->data, skb->data, packet_len); in rmnet_map_deaggregate() 394 return skbn; in rmnet_map_deaggregate()
|
A D | rmnet_handlers.c | 112 struct sk_buff *skbn; in rmnet_map_ingress_handler() local 124 while ((skbn = rmnet_map_deaggregate(skb, port)) != NULL) in rmnet_map_ingress_handler() 125 __rmnet_map_ingress_handler(skbn, port); in rmnet_map_ingress_handler()
|
/linux-6.3-rc2/drivers/net/usb/ |
A D | qmi_wwan.c | 167 struct sk_buff *skbn; in qmimux_rx_fixup() local 191 skbn = netdev_alloc_skb(net, pkt_len + LL_MAX_HEADER); in qmimux_rx_fixup() 192 if (!skbn) in qmimux_rx_fixup() 197 skbn->protocol = htons(ETH_P_IP); in qmimux_rx_fixup() 200 skbn->protocol = htons(ETH_P_IPV6); in qmimux_rx_fixup() 207 skb_reserve(skbn, LL_MAX_HEADER); in qmimux_rx_fixup() 208 skb_put_data(skbn, skb->data + offset + qmimux_hdr_sz, pkt_len); in qmimux_rx_fixup() 209 if (netif_rx(skbn) != NET_RX_SUCCESS) { in qmimux_rx_fixup()
|
/linux-6.3-rc2/net/qrtr/ |
A D | af_qrtr.c | 874 struct sk_buff *skbn; in qrtr_bcast_enqueue() local 878 skbn = skb_clone(skb, GFP_KERNEL); in qrtr_bcast_enqueue() 879 if (!skbn) in qrtr_bcast_enqueue() 881 skb_set_owner_w(skbn, skb->sk); in qrtr_bcast_enqueue() 882 qrtr_node_enqueue(node, skbn, type, from, to); in qrtr_bcast_enqueue()
|