| /linux/drivers/gpu/drm/scheduler/ |
| A D | sched_main.c | 208 rq->sched = sched; in drm_sched_rq_init() 375 queue_work(sched->submit_wq, &sched->work_run_job); in drm_sched_run_job_queue() 413 struct drm_gpu_scheduler *sched = s_fence->sched; in drm_sched_job_done() local 451 mod_delayed_work(sched->timeout_wq, &sched->work_tdr, sched->timeout); in drm_sched_start_timeout() 545 struct drm_gpu_scheduler *sched = s_job->sched; in drm_sched_job_begin() local 618 if (bad && bad->sched == sched) in drm_sched_stop() 840 sched = entity->rq->sched; in drm_sched_job_arm() 842 job->sched = sched; in drm_sched_job_arm() 1256 sched->score = score ? score : &sched->_score; in drm_sched_init() 1300 drm_sched_rq_init(sched, sched->sched_rq[i]); in drm_sched_init() [all …]
|
| A D | sched_entity.c | 194 job->sched->ops->free_job(job); in drm_sched_entity_kill_jobs_work() 285 struct drm_gpu_scheduler *sched; in drm_sched_entity_flush() local 292 sched = entity->rq->sched; in drm_sched_entity_flush() 300 sched->job_scheduled, in drm_sched_entity_flush() 411 struct drm_gpu_scheduler *sched = entity->rq->sched; in drm_sched_entity_add_dependency_cb() local 427 if (!fence->error && s_fence && s_fence->sched == sched && in drm_sched_entity_add_dependency_cb() 470 if (job->sched->ops->prepare_job) in drm_sched_job_dependency() 533 struct drm_gpu_scheduler *sched; in drm_sched_entity_select_rq() local 560 rq = sched ? sched->sched_rq[entity->priority] : NULL; in drm_sched_entity_select_rq() 615 sched = rq->sched; in drm_sched_entity_push_job() [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| A D | xe_gpu_scheduler.c | 11 queue_work(sched->base.submit_wq, &sched->work_process_msg); in xe_sched_process_msg_queue() 18 xe_sched_msg_lock(sched); in xe_sched_process_msg_queue_if_ready() 22 xe_sched_msg_unlock(sched); in xe_sched_process_msg_queue_if_ready() 30 xe_sched_msg_lock(sched); in xe_sched_get_msg() 35 xe_sched_msg_unlock(sched); in xe_sched_get_msg() 66 sched->ops = xe_ops; in xe_sched_init() 67 INIT_LIST_HEAD(&sched->msgs); in xe_sched_init() 84 queue_work(sched->base.submit_wq, &sched->work_process_msg); in xe_sched_submission_start() 95 drm_sched_resume_timeout(&sched->base, sched->base.timeout); in xe_sched_submission_resume_tdr() 101 xe_sched_msg_lock(sched); in xe_sched_add_msg() [all …]
|
| A D | xe_gpu_scheduler.h | 12 int xe_sched_init(struct xe_gpu_scheduler *sched, 20 void xe_sched_fini(struct xe_gpu_scheduler *sched); 34 spin_lock(&sched->base.job_list_lock); in xe_sched_msg_lock() 39 spin_unlock(&sched->base.job_list_lock); in xe_sched_msg_unlock() 44 drm_sched_stop(&sched->base, NULL); in xe_sched_stop() 49 drm_sched_tdr_queue_imm(&sched->base); in xe_sched_tdr_queue_imm() 54 drm_sched_resubmit_jobs(&sched->base); in xe_sched_resubmit_jobs() 66 spin_lock(&sched->base.job_list_lock); in xe_sched_add_pending_job() 68 spin_unlock(&sched->base.job_list_lock); in xe_sched_add_pending_job() 80 struct xe_gpu_scheduler *sched) in xe_sched_entity_init() argument [all …]
|
| /linux/tools/perf/ |
| A D | builtin-sched.c | 475 sched->tasks = realloc(sched->tasks, sched->nr_tasks * sizeof(struct task_desc *)); in register_pid() 617 struct perf_sched *sched = parms->sched; in thread_func() local 674 parms->sched = sched; in create_tasks() 741 sched->runavg_cpu_usage = sched->cpu_usage; in wait_for_tasks() 742 …sched->runavg_cpu_usage = (sched->runavg_cpu_usage * (sched->replay_repeat - 1) + sched->cpu_usage… in wait_for_tasks() 781 sched->run_avg = (sched->run_avg * (sched->replay_repeat - 1) + delta) / sched->replay_repeat; in run_one_test() 2202 if (sched->show_wakeups && !sched->show_next) in timehist_print_sample() 2750 if (sched->show_prio || sched->prio_str) in timehist_sched_change_event() 3023 totals.sched = sched; in timehist_print_summary() 3306 sched->summary = sched->summary_only; in perf_sched__timehist() [all …]
|
| /linux/net/mptcp/ |
| A D | sched.c | 46 ret = sched; in mptcp_sched_find() 111 if (!sched) in mptcp_init_sched() 114 if (!bpf_try_module_get(sched, sched->owner)) in mptcp_init_sched() 117 msk->sched = sched; in mptcp_init_sched() 128 struct mptcp_sched_ops *sched = msk->sched; in mptcp_release_sched() local 130 if (!sched) in mptcp_release_sched() 133 msk->sched = NULL; in mptcp_release_sched() 134 if (sched->release) in mptcp_release_sched() 137 bpf_module_put(sched, sched->owner); in mptcp_release_sched() 170 if (msk->sched == &mptcp_sched_default || !msk->sched) in mptcp_sched_get_send() [all …]
|
| /linux/tools/testing/selftests/ftrace/test.d/trigger/ |
| A D | trigger-filter.tc | 14 echo 'traceoff if child_pid == 0' > events/sched/sched_process_fork/trigger 23 ! echo 'traceoff if a' > events/sched/sched_process_fork/trigger 24 ! echo 'traceoff if common_pid=0' > events/sched/sched_process_fork/trigger 25 ! echo 'traceoff if common_pid==b' > events/sched/sched_process_fork/trigger 26 echo 'traceoff if common_pid == 0' > events/sched/sched_process_fork/trigger 27 echo '!traceoff' > events/sched/sched_process_fork/trigger 29 echo 'traceoff if common_pid <= 0' > events/sched/sched_process_fork/trigger 30 echo '!traceoff' > events/sched/sched_process_fork/trigger 32 echo '!traceoff' > events/sched/sched_process_fork/trigger 34 echo '!traceoff' > events/sched/sched_process_fork/trigger [all …]
|
| A D | trigger-eventonoff.tc | 12 FEATURE=`grep enable_event events/sched/sched_process_fork/trigger` 19 echo 0 > events/sched/sched_switch/enable 20 echo 'enable_event:sched:sched_switch' > events/sched/sched_process_fork/trigger 22 if [ `cat events/sched/sched_switch/enable` != '1*' ]; then 29 echo 1 > events/sched/sched_switch/enable 30 echo 'disable_event:sched:sched_switch' > events/sched/sched_process_fork/trigger 32 if [ `cat events/sched/sched_switch/enable` != '0*' ]; then 39 ! echo 'enable_event:nogroup:noevent' > events/sched/sched_process_fork/trigger 40 ! echo 'disable_event+1' > events/sched/sched_process_fork/trigger 41 echo 'enable_event:sched:sched_switch' > events/sched/sched_process_fork/trigger [all …]
|
| A D | trigger-multihist.tc | 14 echo 'hist:keys=parent_pid:vals=child_pid' > events/sched/sched_process_fork/trigger 15 echo 'hist:keys=parent_comm:vals=child_pid' >> events/sched/sched_process_fork/trigger 17 grep parent_pid events/sched/sched_process_fork/hist > /dev/null || \ 19 grep child events/sched/sched_process_fork/hist > /dev/null || \ 22 grep "parent_comm: $COMM" events/sched/sched_process_fork/hist > /dev/null || \ 29 echo 'hist:name=test_hist:keys=common_pid' > events/sched/sched_process_fork/trigger 31 grep test_hist events/sched/sched_process_fork/hist > /dev/null || \ 36 echo 'hist:name=test_hist:keys=common_pid' > events/sched/sched_process_exit/trigger 38 grep test_hist events/sched/sched_process_exit/hist > /dev/null || \ 41 diffs=`diff events/sched/sched_process_exit/hist events/sched/sched_process_fork/hist | wc -l`
|
| A D | trigger-traceonoff.tc | 13 echo 'traceoff' > events/sched/sched_process_fork/trigger 23 echo 'traceon' > events/sched/sched_process_fork/trigger 32 ! echo 'traceoff:badparam' > events/sched/sched_process_fork/trigger 33 ! echo 'traceoff+0' > events/sched/sched_process_fork/trigger 34 echo 'traceon' > events/sched/sched_process_fork/trigger 35 ! echo 'traceon' > events/sched/sched_process_fork/trigger 36 ! echo 'traceoff' > events/sched/sched_process_fork/trigger
|
| A D | trigger-hist.tc | 14 echo 'hist:keys=parent_pid:vals=child_pid' > events/sched/sched_process_fork/trigger 16 grep parent_pid events/sched/sched_process_fork/hist > /dev/null || \ 18 grep child events/sched/sched_process_fork/hist > /dev/null || \ 25 echo 'hist:keys=parent_pid,child_pid' > events/sched/sched_process_fork/trigger 27 grep '^{ parent_pid:.*, child_pid:.*}' events/sched/sched_process_fork/hist > /dev/null || \ 34 echo 'hist:keys=parent_comm' > events/sched/sched_process_fork/trigger 37 grep "parent_comm: $COMM" events/sched/sched_process_fork/hist > /dev/null || \ 62 echo 'hist:keys=parent_pid,child_pid:sort=child_pid.ascending' > events/sched/sched_process_fork/tr… 73 events/sched/sched_process_fork/hist | cut -d: -f2 ` ||
|
| A D | trigger-snapshot.tc | 11 FEATURE=`grep snapshot events/sched/sched_process_fork/trigger` 19 echo 1 > events/sched/sched_process_fork/enable 21 echo 'snapshot:1' > events/sched/sched_process_fork/trigger 28 echo 0 > events/sched/sched_process_fork/enable 32 ! echo "snapshot+1" > events/sched/sched_process_fork/trigger 33 echo "snapshot" > events/sched/sched_process_fork/trigger 34 ! echo "snapshot" > events/sched/sched_process_fork/trigger
|
| A D | trigger-stacktrace.tc | 11 FEATURE=`grep stacktrace events/sched/sched_process_fork/trigger` 20 echo 'stacktrace' > events/sched/sched_process_fork/trigger 29 ! echo "stacktrace:foo" > events/sched/sched_process_fork/trigger 30 echo "stacktrace" > events/sched/sched_process_fork/trigger 31 ! echo "stacktrace" > events/sched/sched_process_fork/trigger
|
| /linux/Documentation/scheduler/ |
| A D | index.rst | 11 sched-arch 12 sched-bwc 13 sched-deadline 15 sched-eevdf 16 sched-domains 17 sched-capacity 18 sched-energy 22 sched-rt-group 23 sched-stats 24 sched-ext [all …]
|
| /linux/drivers/gpu/drm/nouveau/ |
| A D | nouveau_sched.c | 29 struct nouveau_sched *sched = args->sched; in nouveau_job_init() local 36 job->sched = sched; in nouveau_job_init() 121 struct nouveau_sched *sched = job->sched; in nouveau_job_done() local 276 struct nouveau_sched *sched = job->sched; in nouveau_job_submit() local 371 struct drm_gpu_scheduler *sched = sched_job->sched; in nouveau_sched_timedout_job() local 451 if (sched->wq) in nouveau_sched_init() 463 sched = kzalloc(sizeof(*sched), GFP_KERNEL); in nouveau_sched_create() 464 if (!sched) in nouveau_sched_create() 469 kfree(sched); in nouveau_sched_create() 494 if (sched->wq) in nouveau_sched_fini() [all …]
|
| /linux/drivers/slimbus/ |
| A D | sched.c | 29 struct slim_sched *sched = &ctrl->sched; in slim_ctrl_clk_pause() local 38 mutex_lock(&sched->m_reconf); in slim_ctrl_clk_pause() 41 mutex_unlock(&sched->m_reconf); in slim_ctrl_clk_pause() 52 mutex_unlock(&sched->m_reconf); in slim_ctrl_clk_pause() 67 mutex_unlock(&sched->m_reconf); in slim_ctrl_clk_pause() 74 mutex_unlock(&sched->m_reconf); in slim_ctrl_clk_pause() 83 mutex_unlock(&sched->m_reconf); in slim_ctrl_clk_pause() 112 sched->clk_state = SLIM_CLK_ACTIVE; in slim_ctrl_clk_pause() 114 sched->clk_state = SLIM_CLK_PAUSED; in slim_ctrl_clk_pause() 115 complete(&sched->pause_comp); in slim_ctrl_clk_pause() [all …]
|
| /linux/drivers/gpu/drm/panthor/ |
| A D | panthor_sched.c | 693 queue_work((sched)->wq, &(sched)->wname ## _work); \ 2283 u64 resched_target = sched->last_tick + sched->tick_period; in tick_ctx_update_resched_target() 2285 if (time_before64(sched->resched_target, sched->last_tick) || in tick_ctx_update_resched_target() 2289 return sched->resched_target - sched->last_tick; in tick_ctx_update_resched_target() 2517 if (sched->used_csg_slot_count < sched->csg_slot_count) in group_schedule_locked() 2527 sched->resched_target = sched->last_tick + sched->tick_period; in group_schedule_locked() 2528 if (sched->used_csg_slot_count == sched->csg_slot_count && in group_schedule_locked() 3481 if (!sched || !sched->csg_slot_count) in panthor_sched_fini() 3486 if (sched->wq) in panthor_sched_fini() 3510 if (!sched) in panthor_sched_init() [all …]
|
| /linux/net/netfilter/ipvs/ |
| A D | ip_vs_sched.c | 70 if (sched->done_service) in ip_vs_unbind_scheduler() 71 sched->done_service(svc); in ip_vs_unbind_scheduler() 81 struct ip_vs_scheduler *sched; in ip_vs_sched_getbyname() local 91 if (sched->module && !try_module_get(sched->module)) { in ip_vs_sched_getbyname() 100 return sched; in ip_vs_sched_getbyname() 102 module_put(sched->module); in ip_vs_sched_getbyname() 115 struct ip_vs_scheduler *sched; in ip_vs_scheduler_get() local 125 if (sched == NULL) { in ip_vs_scheduler_get() 130 return sched; in ip_vs_scheduler_get() 146 char *sched_name = sched ? sched->name : "none"; in ip_vs_scheduler_err() [all …]
|
| /linux/Documentation/translations/zh_CN/scheduler/ |
| A D | index.rst | 22 sched-arch 23 sched-bwc 24 sched-design-CFS 25 sched-domains 26 sched-capacity 27 sched-energy 29 sched-nice-design 30 sched-stats 31 sched-debug 35 sched-deadline [all …]
|
| /linux/crypto/ |
| A D | fcrypt.c | 54 __be32 sched[ROUNDS]; member 223 #define F_ENCRYPT(R, L, sched) \ argument 226 u.l = sched ^ R; \ 242 F_ENCRYPT(X.r, X.l, ctx->sched[0x0]); in fcrypt_encrypt() 243 F_ENCRYPT(X.l, X.r, ctx->sched[0x1]); in fcrypt_encrypt() 244 F_ENCRYPT(X.r, X.l, ctx->sched[0x2]); in fcrypt_encrypt() 245 F_ENCRYPT(X.l, X.r, ctx->sched[0x3]); in fcrypt_encrypt() 246 F_ENCRYPT(X.r, X.l, ctx->sched[0x4]); in fcrypt_encrypt() 247 F_ENCRYPT(X.l, X.r, ctx->sched[0x5]); in fcrypt_encrypt() 343 ctx->sched[0xf] = cpu_to_be32(k); in fcrypt_setkey() [all …]
|
| /linux/drivers/net/wireless/ath/ath9k/ |
| A D | channel.c | 451 if (sc->sched.extend_absence) in ath_chanctx_set_periodic_noa() 460 sc->sched.extend_absence) in ath_chanctx_set_periodic_noa() 535 if (sc->sched.offchannel_pending && !sc->sched.wait_switch) { in ath_chanctx_event() 558 if (sc->sched.mgd_prepare_tx) in ath_chanctx_event() 665 sc->sched.beacon_pending) in ath_chanctx_event() 666 sc->sched.beacon_miss++; in ath_chanctx_event() 680 sc->sched.beacon_miss = 0; in ath_chanctx_event() 732 sc->sched.beacon_miss = 0; in ath_chanctx_event() 780 if (sc->sched.beacon_pending) in ath_chanctx_beacon_sent_ev() 814 sc->sched.wait_switch = true; in ath_chanctx_switch() [all …]
|
| /linux/net/sctp/ |
| A D | stream_sched.c | 119 sctp_sched_ops[sched] = sched_ops; in sctp_sched_ops_register() 137 sched->unsched_all(stream); in sctp_sched_free_sched() 142 sched->free_sid(stream, i); in sctp_sched_free_sched() 149 enum sctp_sched_type sched) in sctp_sched_set_sched() argument 157 if (sched > SCTP_SS_MAX) in sctp_sched_set_sched() 160 n = sctp_sched_ops[sched]; in sctp_sched_set_sched() 167 asoc->outqueue.sched = n; in sctp_sched_set_sched() 200 if (asoc->outqueue.sched == sctp_sched_ops[i]) in sctp_sched_get_sched() 253 q->sched->dequeue_done(q, ch); in sctp_sched_dequeue_done() 270 return sched->init_sid(stream, sid, gfp); in sctp_sched_init_sid() [all …]
|
| /linux/include/drm/ |
| A D | gpu_scheduler.h | 258 struct drm_gpu_scheduler *sched; member 302 struct drm_gpu_scheduler *sched; member 340 struct drm_gpu_scheduler *sched; member 545 int drm_sched_init(struct drm_gpu_scheduler *sched, 552 void drm_sched_fini(struct drm_gpu_scheduler *sched); 577 void drm_sched_wakeup(struct drm_gpu_scheduler *sched); 578 bool drm_sched_wqueue_ready(struct drm_gpu_scheduler *sched); 579 void drm_sched_wqueue_stop(struct drm_gpu_scheduler *sched); 580 void drm_sched_wqueue_start(struct drm_gpu_scheduler *sched); 582 void drm_sched_start(struct drm_gpu_scheduler *sched); [all …]
|
| /linux/tools/perf/Documentation/ |
| A D | perf-sched.txt | 1 perf-sched(1) 15 There are several variants of 'perf sched': 24 perf sched record -- sleep 1 25 perf sched latency 62 perf sched record -- sleep 1 63 perf sched timehist 66 time (time between sched-out and next sched-in events for the task), the 101 OPTIONS for 'perf sched latency' 117 OPTIONS for 'perf sched map' 143 OPTIONS for 'perf sched timehist' [all …]
|
| /linux/tools/perf/tests/shell/ |
| A D | lock_contention.sh | 47 perf lock record -o ${perfdata} -- perf bench sched messaging > /dev/null 2>&1 67 perf lock con -a -b -E 1 -q -- perf bench sched messaging > /dev/null 2> ${result} 78 perf lock record -o- -- perf bench sched messaging 2> /dev/null | \ 102 perf lock con -a -b -t -E 1 -q -- perf bench sched messaging > /dev/null 2> ${result} 125 perf lock con -a -b -l -E 1 -q -- perf bench sched messaging > /dev/null 2> ${result} 143 perf lock con -a -b -g -E 1 -q -- perf bench sched messaging > /dev/null 2> ${result} 165 perf lock con -a -b -Y spinlock -q -- perf bench sched messaging > /dev/null 2> ${result} 197 perf lock con -a -b -L tasklist_lock -q -- perf bench sched messaging > /dev/null 2> ${result} 225 perf lock con -a -b -S unix_stream -E 1 -q -- perf bench sched messaging > /dev/null 2> ${result} 269 perf lock con -a -b -g -E 1 -F wait_total -q -- perf bench sched messaging > /dev/null 2> ${result} [all …]
|