Lines Matching refs:skb
285 struct sk_buff *skb, in prep_msg() argument
292 nr_frags = skb_shinfo(skb)->nr_frags; in prep_msg()
294 if (skb_linearize(skb) != 0) in prep_msg()
299 vp->form_header(iov[iov_index].iov_base, skb, vp); in prep_msg()
302 iov[iov_index].iov_base = skb->data; in prep_msg()
304 iov[iov_index].iov_len = skb->len - skb->data_len; in prep_msg()
307 iov[iov_index].iov_len = skb->len; in prep_msg()
310 skb_frag = &skb_shinfo(skb)->frags[frag]; in prep_msg()
325 static int vector_enqueue(struct vector_queue *qi, struct sk_buff *skb) in vector_enqueue() argument
336 if (skb) in vector_enqueue()
337 packet_len = skb->len; in vector_enqueue()
341 *(qi->skbuff_vector + qi->tail) = skb; in vector_enqueue()
345 skb, in vector_enqueue()
361 if (skb != NULL) { in vector_enqueue()
362 packet_len = skb->len; in vector_enqueue()
363 dev_consume_skb_any(skb); in vector_enqueue()
372 struct sk_buff *skb; in consume_vector_skbs() local
377 skb = *(qi->skbuff_vector + skb_index); in consume_vector_skbs()
381 bytes_compl += skb->len; in consume_vector_skbs()
383 dev_consume_skb_any(skb); in consume_vector_skbs()
838 struct sk_buff *skb; in vector_legacy_rx() local
853 skb = prep_skb(vp, &hdr); in vector_legacy_rx()
855 if (skb == NULL) { in vector_legacy_rx()
871 if (skb != NULL) { in vector_legacy_rx()
875 vp->header_rxbuffer, skb, vp); in vector_legacy_rx()
877 dev_kfree_skb_irq(skb); in vector_legacy_rx()
884 skb->ip_summed = CHECKSUM_UNNECESSARY; in vector_legacy_rx()
887 pskb_trim(skb, pkt_len - vp->rx_header_size); in vector_legacy_rx()
888 skb->protocol = eth_type_trans(skb, skb->dev); in vector_legacy_rx()
889 vp->dev->stats.rx_bytes += skb->len; in vector_legacy_rx()
891 napi_gro_receive(&vp->napi, skb); in vector_legacy_rx()
893 dev_kfree_skb_irq(skb); in vector_legacy_rx()
906 static int writev_tx(struct vector_private *vp, struct sk_buff *skb) in writev_tx() argument
912 iov_count = prep_msg(vp, skb, (struct iovec *) &iov); in writev_tx()
930 vp->dev->stats.tx_bytes += skb->len; in writev_tx()
935 consume_skb(skb); in writev_tx()
939 consume_skb(skb); in writev_tx()
954 struct sk_buff *skb; in vector_mmsg_rx() local
987 skb = (*skbuff_vector); in vector_mmsg_rx()
992 skb, in vector_mmsg_rx()
1001 dev_kfree_skb_irq(skb); in vector_mmsg_rx()
1007 skb->ip_summed = CHECKSUM_UNNECESSARY; in vector_mmsg_rx()
1010 pskb_trim(skb, in vector_mmsg_rx()
1012 skb->protocol = eth_type_trans(skb, skb->dev); in vector_mmsg_rx()
1017 vp->dev->stats.rx_bytes += skb->len; in vector_mmsg_rx()
1019 napi_gro_receive(&vp->napi, skb); in vector_mmsg_rx()
1025 if (skb != NULL) in vector_mmsg_rx()
1026 dev_kfree_skb_irq(skb); in vector_mmsg_rx()
1042 static int vector_net_start_xmit(struct sk_buff *skb, struct net_device *dev) in vector_net_start_xmit() argument
1055 writev_tx(vp, skb); in vector_net_start_xmit()
1063 netdev_sent_queue(vp->dev, skb->len); in vector_net_start_xmit()
1064 queue_depth = vector_enqueue(vp->tx_queue, skb); in vector_net_start_xmit()