| /linux/arch/x86/events/amd/ |
| A D | brs.c | 384 void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in amd_pmu_brs_sched_task() argument 397 if (sched_in) in amd_pmu_brs_sched_task()
|
| A D | lbr.c | 374 void amd_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in amd_pmu_lbr_sched_task() argument 383 if (cpuc->lbr_users && sched_in) in amd_pmu_lbr_sched_task()
|
| /linux/tools/perf/ |
| A D | builtin-sched.c | 1128 struct thread *sched_out, *sched_in; in latency_switch_event() local 1149 if (sched_out == NULL || sched_in == NULL) in latency_switch_event() 1185 thread__put(sched_in); in latency_switch_event() 1628 struct thread *sched_in, *sched_out; in map_switch_event() local 1674 tr = thread__get_runtime(sched_in); in map_switch_event() 1676 thread__put(sched_in); in map_switch_event() 1683 str = thread__comm_str(sched_in); in map_switch_event() 1717 str = thread__comm_str(sched_in); in map_switch_event() 1743 if (thread__has_color(sched_in)) in map_switch_event() 1747 tr->shortname, thread__comm_str(sched_in), thread__tid(sched_in)); in map_switch_event() [all …]
|
| /linux/arch/x86/events/ |
| A D | perf_event.h | 857 bool sched_in); 1382 void amd_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1436 void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1461 static inline void amd_pmu_brs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in amd_pmu_brs_sched_task() argument 1627 void intel_pmu_pebs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in); 1642 void intel_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in);
|
| A D | core.c | 2629 static void x86_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in x86_pmu_sched_task() argument 2631 static_call_cond(x86_pmu_sched_task)(pmu_ctx, sched_in); in x86_pmu_sched_task()
|
| /linux/include/linux/ |
| A D | preempt.h | 338 void (*sched_in)(struct preempt_notifier *notifier, int cpu); member
|
| A D | perf_event.h | 477 bool sched_in);
|
| /linux/arch/s390/kernel/ |
| A D | perf_pai_ext.c | 545 static void paiext_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in paiext_sched_task() argument 550 if (!sched_in) in paiext_sched_task()
|
| A D | perf_pai_crypto.c | 521 static void paicrypt_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in paicrypt_sched_task() argument 526 if (!sched_in) in paicrypt_sched_task()
|
| /linux/arch/x86/events/intel/ |
| A D | lbr.c | 542 void intel_pmu_lbr_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in intel_pmu_lbr_sched_task() argument 557 if (sched_in) in intel_pmu_lbr_sched_task() 570 if (sched_in) in intel_pmu_lbr_sched_task()
|
| A D | ds.c | 1243 void intel_pmu_pebs_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in intel_pmu_pebs_sched_task() argument 1247 if (!sched_in && pebs_needs_sched_cb(cpuc)) in intel_pmu_pebs_sched_task()
|
| A D | core.c | 5112 bool sched_in) in intel_pmu_sched_task() argument 5114 intel_pmu_pebs_sched_task(pmu_ctx, sched_in); in intel_pmu_sched_task() 5115 intel_pmu_lbr_sched_task(pmu_ctx, sched_in); in intel_pmu_sched_task()
|
| /linux/arch/powerpc/perf/ |
| A D | core-book3s.c | 135 static void power_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) {} in power_pmu_sched_task() argument 447 static void power_pmu_sched_task(struct perf_event_pmu_context *pmu_ctx, bool sched_in) in power_pmu_sched_task() argument 452 if (sched_in) in power_pmu_sched_task()
|
| /linux/kernel/events/ |
| A D | core.c | 3561 static void perf_ctx_sched_task_cb(struct perf_event_context *ctx, bool sched_in) in perf_ctx_sched_task_cb() argument 3570 pmu_ctx->pmu->sched_task(pmu_ctx, sched_in); in perf_ctx_sched_task_cb() 3705 static void __perf_pmu_sched_task(struct perf_cpu_pmu_context *cpc, bool sched_in) in __perf_pmu_sched_task() argument 3719 pmu->sched_task(cpc->task_epc, sched_in); in __perf_pmu_sched_task() 3727 bool sched_in) in perf_pmu_sched_task() argument 3737 __perf_pmu_sched_task(cpc, sched_in); in perf_pmu_sched_task() 3741 struct task_struct *next_prev, bool sched_in); 9234 struct task_struct *next_prev, bool sched_in) in perf_event_switch() argument 9246 .misc = sched_in ? 0 : PERF_RECORD_MISC_SWITCH_OUT, in perf_event_switch() 9254 if (!sched_in && task_is_runnable(task)) { in perf_event_switch()
|
| /linux/virt/kvm/ |
| A D | kvm_main.c | 6498 kvm_preempt_ops.sched_in = kvm_sched_in; in kvm_init()
|
| /linux/kernel/sched/ |
| A D | core.c | 4892 notifier->ops->sched_in(notifier, raw_smp_processor_id()); in __fire_sched_in_preempt_notifiers()
|