Lines Matching refs:cached_prod
43 u32 cached_prod; member
130 if (q->cached_cons != q->cached_prod) { in xskq_cons_read_addr_unchecked()
194 return q->cached_cons != q->cached_prod; in xskq_has_descs()
212 if (q->cached_cons != q->cached_prod) { in xskq_cons_read_desc()
247 while (cached_cons != q->cached_prod && nb_entries < max) { in xskq_cons_read_desc_batch()
287 q->cached_prod = smp_load_acquire(&q->ring->producer); /* C, matches B */ in __xskq_cons_peek()
298 u32 entries = q->cached_prod - q->cached_cons; in xskq_cons_nb_entries()
304 entries = q->cached_prod - q->cached_cons; in xskq_cons_nb_entries()
311 if (q->cached_prod == q->cached_cons) in xskq_cons_peek_addr_unchecked()
320 if (q->cached_prod == q->cached_cons) in xskq_cons_peek_desc()
349 u32 free_entries = q->nentries - (q->cached_prod - q->cached_cons); in xskq_prod_nb_free()
356 free_entries = q->nentries - (q->cached_prod - q->cached_cons); in xskq_prod_nb_free()
368 q->cached_prod -= cnt; in xskq_prod_cancel_n()
377 q->cached_prod++; in xskq_prod_reserve()
389 ring->desc[q->cached_prod++ & q->ring_mask] = addr; in xskq_prod_reserve_addr()
397 u32 i, cached_prod; in xskq_prod_write_addr_batch() local
400 cached_prod = q->cached_prod; in xskq_prod_write_addr_batch()
402 ring->desc[cached_prod++ & q->ring_mask] = descs[i].addr; in xskq_prod_write_addr_batch()
403 q->cached_prod = cached_prod; in xskq_prod_write_addr_batch()
416 idx = q->cached_prod++ & q->ring_mask; in xskq_prod_reserve_desc()
431 __xskq_prod_submit(q, q->cached_prod); in xskq_prod_submit()