Searched refs:skbuff (Results 1 – 10 of 10) sorted by relevance
237 struct sk_buff *skbuff; member1064 np->lack_rxbuf->skbuff = skb; in allocate_rx_buffers()1244 np->rx_ring[i].skbuff = NULL; in init_ring()1261 np->rx_ring[i].skbuff = skb; in init_ring()1328 next->skbuff = skb; in start_tx()1386 if (cur->skbuff) { in reset_tx_descriptors()1390 cur->skbuff = NULL; in reset_tx_descriptors()1415 if (cur->skbuff) in reset_rx_descriptors()1533 np->cur_tx->skbuff->len, in intr_handler()1536 np->cur_tx->skbuff = NULL; in intr_handler()[all …]
181 if (!pool->skbuff) { in ibmveth_alloc_buffer_pool()247 if (!pool->skbuff[index]) { in ibmveth_replenish_buffer_pool()268 pool->skbuff[index] = skb; in ibmveth_replenish_buffer_pool()343 if (pool->skbuff[index]) { in ibmveth_replenish_buffer_pool()351 pool->skbuff[index] = NULL; in ibmveth_replenish_buffer_pool()430 pool->skbuff[i] = NULL; in ibmveth_free_buffer_pool()440 if (pool->skbuff) { in ibmveth_free_buffer_pool()441 kfree(pool->skbuff); in ibmveth_free_buffer_pool()442 pool->skbuff = NULL; in ibmveth_free_buffer_pool()2211 pool->skbuff[0] = NULL; in ibmveth_remove_buffer_from_pool_test()[all …]
142 struct sk_buff **skbuff; member
515 sq->skbuff = kcalloc(q_len, sizeof(u64), GFP_KERNEL); in nicvf_init_snd_queue()516 if (!sq->skbuff) in nicvf_init_snd_queue()594 skb = (struct sk_buff *)sq->skbuff[sq->head]; in nicvf_free_snd_queue()622 kfree(sq->skbuff); in nicvf_free_snd_queue()1195 skb = (struct sk_buff *)sq->skbuff[sq->head]; in nicvf_sq_free_used_descs()1352 sq->skbuff[qentry] = (u64)skb; in nicvf_sq_add_hdr_subdesc()1442 sq->skbuff[qentry] = (u64)skb; in nicvf_sq_add_cqe_subdesc()1535 sq->skbuff[hdr_qentry] = (u64)NULL; in nicvf_sq_append_tso()1541 sq->skbuff[hdr_qentry] = (u64)skb; in nicvf_sq_append_tso()
272 u64 *skbuff; member
693 skb = (struct sk_buff *)sq->skbuff[cqe_tx->sqe_ptr]; in nicvf_snd_pkt_handler()717 sq->skbuff[cqe_tx->sqe_ptr] = (u64)NULL; in nicvf_snd_pkt_handler()
222 skb = ring->skbuff[entry]; in rtase_tx_clear_range()228 ring->skbuff[entry] = NULL; in rtase_tx_clear_range()287 if (ring->skbuff[entry]) { in tx_handler()289 bytes_compl += ring->skbuff[entry]->len; in tx_handler()290 napi_consume_skb(ring->skbuff[entry], budget); in tx_handler()291 ring->skbuff[entry] = NULL; in tx_handler()325 memset(ring->skbuff, 0x0, sizeof(ring->skbuff)); in rtase_tx_desc_init()1288 ring->skbuff[entry] = skb; in rtase_xmit_frags()1372 ring->skbuff[entry] = skb; in rtase_start_xmit()1422 ring->skbuff[entry] = NULL; in rtase_start_xmit()
293 struct sk_buff *skbuff[RTASE_NUM_DESC]; member
1456 struct sk_buff *skbuff = NULL; in ath6kl_rx() local1485 while ((skbuff = skb_dequeue(&conn->psq))) { in ath6kl_rx()1487 ath6kl_data_tx(skbuff, vif->ndev); in ath6kl_rx()1492 while ((skbuff = skb_dequeue(&conn->apsdq))) { in ath6kl_rx()1494 ath6kl_data_tx(skbuff, vif->ndev); in ath6kl_rx()
580 void *skbuff, in liquidio_push_packet() argument589 struct sk_buff *skb = (struct sk_buff *)skbuff; in liquidio_push_packet()
Completed in 36 milliseconds