Home
last modified time | relevance | path

Searched refs:backlog (Results 1 – 25 of 183) sorted by relevance

12345678

/linux-6.3-rc2/drivers/crypto/qat/qat_common/
A Dqat_algs_send.c24 void qat_alg_send_backlog(struct qat_instance_backlog *backlog) in qat_alg_send_backlog() argument
28 spin_lock_bh(&backlog->lock); in qat_alg_send_backlog()
29 list_for_each_entry_safe(req, tmp, &backlog->list, list) { in qat_alg_send_backlog()
40 spin_unlock_bh(&backlog->lock); in qat_alg_send_backlog()
44 struct qat_instance_backlog *backlog) in qat_alg_backlog_req() argument
48 spin_lock_bh(&backlog->lock); in qat_alg_backlog_req()
49 list_add_tail(&req->list, &backlog->list); in qat_alg_backlog_req()
50 spin_unlock_bh(&backlog->lock); in qat_alg_backlog_req()
55 struct qat_instance_backlog *backlog = req->backlog; in qat_alg_send_message_maybacklog() local
60 if (!list_empty(&backlog->list)) in qat_alg_send_message_maybacklog()
[all …]
A Dqat_algs_send.h19 struct qat_instance_backlog *backlog; member
23 void qat_alg_send_backlog(struct qat_instance_backlog *backlog);
/linux-6.3-rc2/drivers/crypto/ccp/
A Dccp-crypto-main.c55 struct list_head *backlog; member
95 *backlog = NULL; in ccp_crypto_cmd_complete()
114 if (req_queue.backlog != &req_queue.cmds) { in ccp_crypto_cmd_complete()
119 *backlog = container_of(req_queue.backlog, in ccp_crypto_cmd_complete()
121 req_queue.backlog = req_queue.backlog->next; in ccp_crypto_cmd_complete()
160 if (backlog) { in ccp_crypto_complete()
161 backlog->ret = -EINPROGRESS; in ccp_crypto_complete()
192 if (backlog) { in ccp_crypto_complete()
193 backlog->ret = -EINPROGRESS; in ccp_crypto_complete()
242 req_queue.backlog = &crypto_cmd->entry; in ccp_crypto_enqueue_cmd()
[all …]
A Dccp-dev.c315 list_add_tail(&cmd->entry, &ccp->backlog); in ccp_enqueue_cmd()
378 struct ccp_cmd *backlog = NULL; in ccp_dequeue_cmd() local
403 if (!list_empty(&ccp->backlog)) { in ccp_dequeue_cmd()
404 backlog = list_first_entry(&ccp->backlog, struct ccp_cmd, in ccp_dequeue_cmd()
406 list_del(&backlog->entry); in ccp_dequeue_cmd()
411 if (backlog) { in ccp_dequeue_cmd()
412 INIT_WORK(&backlog->work, ccp_do_cmd_backlog); in ccp_dequeue_cmd()
413 schedule_work(&backlog->work); in ccp_dequeue_cmd()
488 INIT_LIST_HEAD(&ccp->backlog); in ccp_alloc_struct()
/linux-6.3-rc2/tools/perf/ui/tui/
A Dhelpline.c33 static int backlog; in tui_helpline__show() local
36 ret = vscnprintf(ui_helpline__last_msg + backlog, in tui_helpline__show()
37 sizeof(ui_helpline__last_msg) - backlog, format, ap); in tui_helpline__show()
38 backlog += ret; in tui_helpline__show()
42 if (ui_helpline__last_msg[backlog - 1] == '\n') { in tui_helpline__show()
45 backlog = 0; in tui_helpline__show()
/linux-6.3-rc2/tools/perf/ui/gtk/
A Dhelpline.c32 static int backlog; in gtk_helpline_show() local
34 ret = vscnprintf(ui_helpline__current + backlog, in gtk_helpline_show()
35 sizeof(ui_helpline__current) - backlog, fmt, ap); in gtk_helpline_show()
36 backlog += ret; in gtk_helpline_show()
40 if (ptr && (ptr - ui_helpline__current) <= backlog) { in gtk_helpline_show()
43 backlog = 0; in gtk_helpline_show()
/linux-6.3-rc2/tools/testing/selftests/net/forwarding/
A Dsch_red.sh164 qdisc_stats_get $swp3 11: .backlog
258 local backlog
266 backlog=$(build_backlog $((2 * limit / 3)) udp)
269 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected == 0."
276 backlog=$(build_backlog $((3 * limit / 2)) tcp tos=0x01)
332 local backlog
342 backlog=$(build_backlog $((2 * limit / 3)) tcp tos=0x01)
350 backlog=$(build_backlog $((3 * limit / 2)) tcp tos=0x01)
363 local backlog
418 backlog=$(build_backlog $((2 * limit / 3)) tcp tos=0x01)
[all …]
/linux-6.3-rc2/net/sched/
A Dsch_pie.c34 struct pie_vars *vars, u32 backlog, u32 packet_size) in pie_drop_early() argument
54 if (backlog < 2 * mtu) in pie_drop_early()
96 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, in pie_qdisc_enqueue()
206 struct pie_vars *vars, u32 backlog) in pie_process_dequeue() argument
222 if (backlog == 0) in pie_process_dequeue()
274 if (backlog < QUEUE_THRESHOLD) { in pie_process_dequeue()
298 u32 backlog) in pie_calculate_probability() argument
313 qdelay = (backlog << PIE_SCALE) / vars->avg_dq_rate; in pie_calculate_probability()
324 if (qdelay == 0 && backlog != 0) in pie_calculate_probability()
400 vars->backlog_old = backlog; in pie_calculate_probability()
[all …]
A Dsch_gred.c118 return sch->qstats.backlog; in gred_backlog()
120 return q->backlog; in gred_backlog()
248 q->backlog += qdisc_pkt_len(skb); in gred_enqueue()
276 q->backlog -= qdisc_pkt_len(skb); in gred_dequeue()
279 if (!sch->qstats.backlog) in gred_dequeue()
282 if (!q->backlog) in gred_dequeue()
307 q->backlog = 0; in gred_reset()
346 opt->set.tab[i].backlog = &q->backlog; in gred_offload()
386 table->tab[i]->backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats()
391 sch->qstats.backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats()
[all …]
A Dsch_fq_pie.c47 u32 backlog; member
163 sel_flow->backlog, skb->len)) { in fq_pie_qdisc_enqueue()
182 sch->qstats.backlog += pkt_len; in fq_pie_qdisc_enqueue()
190 sel_flow->backlog = 0; in fq_pie_qdisc_enqueue()
193 sel_flow->backlog += pkt_len; in fq_pie_qdisc_enqueue()
255 sch->qstats.backlog -= pkt_len; in fq_pie_qdisc_dequeue()
271 flow->backlog -= pkt_len; in fq_pie_qdisc_dequeue()
273 pie_process_dequeue(skb, &q->p_params, &flow->vars, flow->backlog); in fq_pie_qdisc_dequeue()
381 q->flows[idx].backlog); in fq_pie_timer()
A Dsch_sfq.c109 unsigned int backlog; member
308 slot->backlog -= len; in sfq_drop()
381 slot->backlog); in sfq_enqueue()
432 sch->qstats.backlog -= delta; in sfq_enqueue()
433 slot->backlog -= delta; in sfq_enqueue()
443 slot->backlog += qdisc_pkt_len(skb); in sfq_enqueue()
504 slot->backlog -= qdisc_pkt_len(skb); in sfq_dequeue()
556 slot->backlog = 0; in sfq_rehash()
587 slot->backlog); in sfq_rehash()
588 slot->backlog += qdisc_pkt_len(skb); in sfq_rehash()
[all …]
A Dsch_skbprio.c85 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue()
108 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue()
117 q->qstats[lp].backlog -= qdisc_pkt_len(to_drop); in skbprio_enqueue()
152 q->qstats[q->highest_prio].backlog -= qdisc_pkt_len(skb); in skbprio_dequeue()
/linux-6.3-rc2/tools/testing/selftests/drivers/net/mlxsw/
A Dsch_red_core.sh316 qdisc_stats_get $swp3 $(get_qdisc_handle $vlan) .backlog
419 local backlog
427 backlog=$(build_backlog $vlan $((2 * limit / 3)) udp)
430 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected == 0."
437 backlog=$(build_backlog $vlan $((3 * limit / 2)) tcp tos=0x01)
514 local backlog
524 backlog=$(build_backlog $vlan $((2 * limit / 3)) tcp tos=0x01)
532 backlog=$(build_backlog $vlan $((3 * limit / 2)) tcp tos=0x01)
536 local diff=$((limit - backlog))
539 check_err $? "backlog $backlog / $limit expected <= 10% distance"
[all …]
/linux-6.3-rc2/include/net/
A Dfq_impl.h24 flow->backlog -= bytes; in __fq_adjust_removal()
25 fq->backlog -= packets; in __fq_adjust_removal()
28 if (flow->backlog) in __fq_adjust_removal()
176 cur_len = cur->backlog; in fq_find_fattest_flow()
185 unsigned int cur_len = tin->default_flow.backlog; in fq_find_fattest_flow()
210 if (!flow->backlog) { in fq_tin_enqueue()
220 flow->backlog += skb->len; in fq_tin_enqueue()
224 fq->backlog++; in fq_tin_enqueue()
235 while (fq->backlog > fq->limit || oom) { in fq_tin_enqueue()
308 WARN_ON_ONCE(flow->backlog); in fq_flow_reset()
A Dcodel_impl.h111 u32 *backlog, in codel_should_drop() argument
129 *backlog <= params->mtu) { in codel_should_drop()
147 u32 *backlog, in codel_dequeue() argument
166 skb_len_func, skb_time_func, backlog, now); in codel_dequeue()
202 backlog, now)) { in codel_dequeue()
227 skb_time_func, backlog, now); in codel_dequeue()
A Dpie.h127 struct pie_vars *vars, u32 backlog, u32 packet_size);
130 struct pie_vars *vars, u32 backlog);
133 u32 backlog);
A Dsch_generic.h843 sch->qstats.backlog -= qdisc_pkt_len(skb); in qdisc_qstats_backlog_dec()
855 sch->qstats.backlog += qdisc_pkt_len(skb); in qdisc_qstats_backlog_inc()
917 __u32 *backlog) in qdisc_qstats_qlen_backlog() argument
923 *backlog = qstats.backlog; in qdisc_qstats_qlen_backlog()
928 __u32 qlen, backlog; in qdisc_tree_flush_backlog() local
930 qdisc_qstats_qlen_backlog(sch, &qlen, &backlog); in qdisc_tree_flush_backlog()
931 qdisc_tree_reduce_backlog(sch, qlen, backlog); in qdisc_tree_flush_backlog()
936 __u32 qlen, backlog; in qdisc_purge_queue() local
938 qdisc_qstats_qlen_backlog(sch, &qlen, &backlog); in qdisc_purge_queue()
940 qdisc_tree_reduce_backlog(sch, qlen, backlog); in qdisc_purge_queue()
[all …]
A Dfq.h35 u32 backlog; member
79 u32 backlog; member
A Dred.h339 unsigned int backlog) in red_calc_qavg_no_idle_time() argument
350 return v->qavg + (backlog - (v->qavg >> p->Wlog)); in red_calc_qavg_no_idle_time()
355 unsigned int backlog) in red_calc_qavg() argument
358 return red_calc_qavg_no_idle_time(p, v, backlog); in red_calc_qavg()
/linux-6.3-rc2/drivers/net/ethernet/mellanox/mlxsw/
A Dspectrum_qdisc.c79 u64 backlog; member
200 tmp->stats_base.backlog -= mlxsw_sp_qdisc->stats_base.backlog; in mlxsw_sp_qdisc_reduce_parent_backlog()
569 backlog -= stats_base->backlog; in mlxsw_sp_qdisc_update_stats()
575 stats_base->backlog += backlog; in mlxsw_sp_qdisc_update_stats()
751 u64 backlog; in mlxsw_sp_qdisc_leaf_unoffload() local
755 qstats->backlog -= backlog; in mlxsw_sp_qdisc_leaf_unoffload()
1298 u64 backlog; in mlxsw_sp_qdisc_walk_cb_clean_stats() local
1301 backlog = mlxsw_sp_qdisc->stats_base.backlog; in mlxsw_sp_qdisc_walk_cb_clean_stats()
1305 mlxsw_sp_qdisc->stats_base.backlog = backlog; in mlxsw_sp_qdisc_walk_cb_clean_stats()
1422 u64 backlog; in __mlxsw_sp_qdisc_ets_unoffload() local
[all …]
/linux-6.3-rc2/net/core/
A Dgen_stats.c349 qstats->backlog += qcpu->backlog; in gnet_stats_add_queue_cpu()
364 qstats->backlog += q->backlog; in gnet_stats_add_queue()
399 d->tc_stats.backlog = qstats.backlog; in gnet_stats_copy_queue()
/linux-6.3-rc2/net/sunrpc/
A Dstats.c158 ktime_t backlog, execute, now; in rpc_count_iostats_metrics() local
174 backlog = 0; in rpc_count_iostats_metrics()
176 backlog = ktime_sub(req->rq_xtime, task->tk_start); in rpc_count_iostats_metrics()
177 op_metrics->om_queue = ktime_add(op_metrics->om_queue, backlog); in rpc_count_iostats_metrics()
189 trace_rpc_stats_latency(req->rq_task, backlog, req->rq_rtt, execute); in rpc_count_iostats_metrics()
/linux-6.3-rc2/net/rxrpc/
A Dcall_accept.c163 struct rxrpc_backlog *b = rx->backlog; in rxrpc_service_prealloc()
169 rx->backlog = b; in rxrpc_service_prealloc()
180 struct rxrpc_backlog *b = rx->backlog; in rxrpc_discard_prealloc()
186 rx->backlog = NULL; in rxrpc_discard_prealloc()
250 struct rxrpc_backlog *b = rx->backlog; in rxrpc_alloc_incoming_call()
438 struct rxrpc_backlog *b = rx->backlog; in rxrpc_user_charge_accept()
470 struct rxrpc_backlog *b = rx->backlog; in rxrpc_kernel_charge_accept()
/linux-6.3-rc2/drivers/crypto/qce/
A Dcore.c83 struct crypto_async_request *async_req, *backlog; in qce_handle_queue() local
98 backlog = crypto_get_backlog(&qce->queue); in qce_handle_queue()
108 if (backlog) { in qce_handle_queue()
110 crypto_request_complete(backlog, -EINPROGRESS); in qce_handle_queue()
/linux-6.3-rc2/drivers/net/ipvlan/
A Dipvlan_core.c240 spin_lock_bh(&port->backlog.lock); in ipvlan_process_multicast()
241 skb_queue_splice_tail_init(&port->backlog, &list); in ipvlan_process_multicast()
242 spin_unlock_bh(&port->backlog.lock); in ipvlan_process_multicast()
550 spin_lock(&port->backlog.lock); in ipvlan_multicast_enqueue()
551 if (skb_queue_len(&port->backlog) < IPVLAN_QBACKLOG_LIMIT) { in ipvlan_multicast_enqueue()
554 __skb_queue_tail(&port->backlog, skb); in ipvlan_multicast_enqueue()
555 spin_unlock(&port->backlog.lock); in ipvlan_multicast_enqueue()
558 spin_unlock(&port->backlog.lock); in ipvlan_multicast_enqueue()

Completed in 64 milliseconds

12345678