Lines Matching refs:batch_size
256 sock_opt = xsk->batch_size; in enable_busy_poll()
506 ifobj->xsk_arr[j].batch_size = DEFAULT_BATCH_SIZE; in __test_spec_init()
1155 static int complete_pkts(struct xsk_socket_info *xsk, int batch_size) in complete_pkts() argument
1167 rcvd = xsk_ring_cons__peek(&xsk->umem->cq, batch_size, &idx); in complete_pkts()
1221 rcvd = xsk_ring_cons__peek(&xsk->rx, xsk->batch_size, &idx_rx); in __receive_pkts()
1373 if (pkts_in_flight >= (int)((umem_size(umem) - xsk->batch_size * buffer_len) / in __send_pkts()
1384 while (xsk_ring_prod__reserve(&xsk->tx, xsk->batch_size, &idx) < xsk->batch_size) { in __send_pkts()
1404 complete_pkts(xsk, xsk->batch_size); in __send_pkts()
1407 for (i = 0; i < xsk->batch_size; i++) { in __send_pkts()
1415 if (nb_frags > xsk->batch_size - i) { in __send_pkts()
1417 xsk_ring_prod__cancel(&xsk->tx, xsk->batch_size - i); in __send_pkts()
1505 complete_pkts(xsk, xsk->batch_size); in wait_for_tx_completion()
2547 test->ifobj_tx->xsk->batch_size = 1; in testapp_hw_sw_min_ring_size()
2548 test->ifobj_rx->xsk->batch_size = 1; in testapp_hw_sw_min_ring_size()
2554 test->ifobj_tx->xsk->batch_size = DEFAULT_BATCH_SIZE - 1; in testapp_hw_sw_min_ring_size()
2555 test->ifobj_rx->xsk->batch_size = DEFAULT_BATCH_SIZE - 1; in testapp_hw_sw_min_ring_size()
2571 test->ifobj_tx->xsk->batch_size = XSK_RING_PROD__DEFAULT_NUM_DESCS; in testapp_hw_sw_max_ring_size()
2572 test->ifobj_rx->xsk->batch_size = XSK_RING_PROD__DEFAULT_NUM_DESCS; in testapp_hw_sw_max_ring_size()
2581 test->ifobj_tx->xsk->batch_size = test->ifobj_tx->ring.tx_max_pending - 8; in testapp_hw_sw_max_ring_size()
2582 test->ifobj_rx->xsk->batch_size = test->ifobj_tx->ring.tx_max_pending - 8; in testapp_hw_sw_max_ring_size()
2663 test->ifobj_tx->xsk->batch_size = nr_packets; in testapp_tx_queue_consumer()