Searched refs:nb_entries (Results 1 – 4 of 4) sorted by relevance
/linux-6.3-rc2/net/xdp/ |
A D | xsk_queue.h | 217 u32 cached_cons = q->cached_cons, nb_entries = 0; in xskq_cons_read_desc_batch() local 220 while (cached_cons != q->cached_prod && nb_entries < max) { in xskq_cons_read_desc_batch() 224 descs[nb_entries] = ring->desc[idx]; in xskq_cons_read_desc_batch() 225 if (unlikely(!xskq_cons_is_valid_desc(q, &descs[nb_entries], pool))) { in xskq_cons_read_desc_batch() 231 nb_entries++; in xskq_cons_read_desc_batch() 237 return nb_entries; in xskq_cons_read_desc_batch() 357 u32 nb_entries) in xskq_prod_write_addr_batch() argument 364 for (i = 0; i < nb_entries; i++) in xskq_prod_write_addr_batch() 406 static inline void xskq_prod_submit_n(struct xsk_queue *q, u32 nb_entries) in xskq_prod_submit_n() argument 408 __xskq_prod_submit(q, q->ring->producer + nb_entries); in xskq_prod_submit_n()
|
A D | xsk_buff_pool.c | 545 u32 i, cached_cons, nb_entries; in xp_alloc_new_from_fq() local 552 nb_entries = max; in xp_alloc_new_from_fq() 565 nb_entries--; in xp_alloc_new_from_fq() 583 return nb_entries; in xp_alloc_new_from_fq() 586 static u32 xp_alloc_reused(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 nb_entries) in xp_alloc_reused() argument 591 nb_entries = min_t(u32, nb_entries, pool->free_list_cnt); in xp_alloc_reused() 593 i = nb_entries; in xp_alloc_reused() 601 pool->free_list_cnt -= nb_entries; in xp_alloc_reused() 603 return nb_entries; in xp_alloc_reused()
|
A D | xsk.c | 297 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) in xsk_tx_completed() argument 299 xskq_prod_submit_n(pool->cq, nb_entries); in xsk_tx_completed()
|
/linux-6.3-rc2/include/net/ |
A D | xdp_sock_drv.h | 17 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries); 147 static inline void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) in xsk_tx_completed() argument
|
Completed in 9 milliseconds