Lines Matching refs:buffer
39 tx_queue->buffer = kcalloc(entries, sizeof(*tx_queue->buffer), in efx_probe_tx_queue()
41 if (!tx_queue->buffer) in efx_probe_tx_queue()
63 kfree(tx_queue->buffer); in efx_probe_tx_queue()
64 tx_queue->buffer = NULL; in efx_probe_tx_queue()
105 struct efx_tx_buffer *buffer; in efx_fini_tx_queue() local
112 if (!tx_queue->buffer) in efx_fini_tx_queue()
121 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_fini_tx_queue()
122 efx_dequeue_buffer(tx_queue, buffer, &pkts_compl, &bytes_compl, in efx_fini_tx_queue()
136 if (!tx_queue->buffer) in efx_remove_tx_queue()
151 kfree(tx_queue->buffer); in efx_remove_tx_queue()
152 tx_queue->buffer = NULL; in efx_remove_tx_queue()
157 struct efx_tx_buffer *buffer, in efx_dequeue_buffer() argument
164 if (buffer->unmap_len) { in efx_dequeue_buffer()
166 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in efx_dequeue_buffer()
168 if (buffer->flags & EFX_TX_BUF_MAP_SINGLE) in efx_dequeue_buffer()
169 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, in efx_dequeue_buffer()
172 dma_unmap_page(dma_dev, unmap_addr, buffer->unmap_len, in efx_dequeue_buffer()
174 buffer->unmap_len = 0; in efx_dequeue_buffer()
177 if (buffer->flags & EFX_TX_BUF_SKB) { in efx_dequeue_buffer()
178 struct sk_buff *skb = (struct sk_buff *)buffer->skb; in efx_dequeue_buffer()
180 if (unlikely(buffer->flags & EFX_TX_BUF_EFV)) { in efx_dequeue_buffer()
201 dev_consume_skb_any((struct sk_buff *)buffer->skb); in efx_dequeue_buffer()
205 } else if (buffer->flags & EFX_TX_BUF_XDP) { in efx_dequeue_buffer()
206 xdp_return_frame_rx_napi(buffer->xdpf); in efx_dequeue_buffer()
210 (*xdp_bytes) += buffer->xdpf->len; in efx_dequeue_buffer()
213 buffer->len = 0; in efx_dequeue_buffer()
214 buffer->flags = 0; in efx_dequeue_buffer()
237 struct efx_tx_buffer *buffer = &tx_queue->buffer[read_ptr]; in efx_dequeue_buffers() local
239 if (!efx_tx_buffer_in_use(buffer)) { in efx_dequeue_buffers()
247 efx_dequeue_buffer(tx_queue, buffer, pkts_compl, bytes_compl, in efx_dequeue_buffers()
314 struct efx_tx_buffer *buffer; in efx_enqueue_unwind() local
321 buffer = __efx_tx_queue_get_insert_buffer(tx_queue); in efx_enqueue_unwind()
322 efx_dequeue_buffer(tx_queue, buffer, &pkts_compl, &bytes_compl, in efx_enqueue_unwind()
332 struct efx_tx_buffer *buffer; in efx_tx_map_chunk() local
337 buffer = efx_tx_queue_get_insert_buffer(tx_queue); in efx_tx_map_chunk()
344 buffer->len = dma_len; in efx_tx_map_chunk()
345 buffer->dma_addr = dma_addr; in efx_tx_map_chunk()
346 buffer->flags = EFX_TX_BUF_CONT; in efx_tx_map_chunk()
352 return buffer; in efx_tx_map_chunk()
408 struct efx_tx_buffer *buffer; in efx_tx_map_data() local
411 buffer = efx_tx_map_chunk(tx_queue, dma_addr, len); in efx_tx_map_data()
416 buffer->flags = EFX_TX_BUF_CONT | dma_flags; in efx_tx_map_data()
417 buffer->unmap_len = unmap_len; in efx_tx_map_data()
418 buffer->dma_offset = buffer->dma_addr - unmap_addr; in efx_tx_map_data()
424 buffer->skb = skb; in efx_tx_map_data()
425 buffer->flags = EFX_TX_BUF_SKB | dma_flags; in efx_tx_map_data()