| /kernel/trace/rv/monitors/sssw/ |
| A D | sssw.c | 31 static void handle_sched_switch(void *data, bool preempt, in handle_sched_switch() argument 36 if (preempt) in handle_sched_switch()
|
| /kernel/trace/rv/monitors/nrp/ |
| A D | nrp.c | 78 static void handle_schedule_entry(void *data, bool preempt) in handle_schedule_entry() argument 80 if (preempt) in handle_schedule_entry()
|
| /kernel/trace/rv/monitors/sts/ |
| A D | sts.c | 77 static void handle_sched_switch(void *data, bool preempt, in handle_sched_switch() argument 85 static void handle_schedule_entry(void *data, bool preempt) in handle_schedule_entry() argument
|
| /kernel/rcu/ |
| A D | tree_plugin.h | 324 void rcu_note_context_switch(bool preempt) in rcu_note_context_switch() argument 332 …WARN_ONCE(!preempt && rcu_preempt_depth() > 0, "Voluntary context switch within RCU read-side crit… in rcu_note_context_switch() 371 rcu_tasks_qs(current, preempt); in rcu_note_context_switch() 998 void rcu_note_context_switch(bool preempt) in rcu_note_context_switch() argument 1009 rcu_tasks_qs(current, preempt); in rcu_note_context_switch()
|
| A D | Kconfig.debug | 63 Note that PREEMPT_COUNT must be enabled if the preempt-disabled
|
| A D | Kconfig | 269 and where the "x" is "p" for RCU-preempt (PREEMPTION kernels)
|
| /kernel/trace/rv/monitors/snep/ |
| A D | Kconfig | 11 Monitor to ensure schedule does not enable preempt.
|
| A D | snep.c | 33 static void handle_schedule_entry(void *data, bool preempt) in handle_schedule_entry() argument
|
| /kernel/trace/rv/monitors/sco/ |
| A D | sco.c | 27 static void handle_schedule_entry(void *data, bool preempt) in handle_schedule_entry() argument
|
| /kernel/trace/rv/monitors/snroc/ |
| A D | snroc.c | 27 static void handle_sched_switch(void *data, bool preempt, in handle_sched_switch() argument
|
| /kernel/trace/rv/monitors/wwnr/ |
| A D | wwnr.c | 21 static void handle_switch(void *data, bool preempt, struct task_struct *p, in handle_switch() argument
|
| /kernel/trace/rv/monitors/scpd/ |
| A D | scpd.c | 33 static void handle_schedule_entry(void *data, bool preempt) in handle_schedule_entry() argument
|
| /kernel/ |
| A D | Kconfig.preempt | 49 low priority process to voluntarily preempt itself even if it 83 eager to preempt SCHED_NORMAL tasks in an attempt to
|
| /kernel/sched/ |
| A D | core.c | 2399 guard(preempt)(); in migrate_disable() local 2431 guard(preempt)(); in migrate_enable() local 3497 guard(preempt)(); in kick_process() local 4180 guard(preempt)(); in try_to_wake_up() local 6378 guard(preempt)(); in sched_core_balance() local 6823 bool preempt = sched_mode > SM_NONE; in __schedule() local 6838 schedule_debug(prev, preempt); in __schedule() 6846 rcu_note_context_switch(preempt); in __schedule() 6876 preempt = sched_mode == SM_PREEMPT; in __schedule() 6889 } else if (!preempt && prev_state) { in __schedule() [all …]
|
| A D | syscalls.c | 1421 int __sched yield_to(struct task_struct *p, bool preempt) in yield_to() argument 1459 if (preempt && rq != p_rq) in yield_to()
|
| A D | ext.c | 2050 bool preempt = false; in dispatch_enqueue() local 2055 preempt = true; in dispatch_enqueue() 2058 if (preempt || sched_class_above(&ext_sched_class, in dispatch_enqueue()
|
| A D | fair.c | 8702 goto preempt; in check_preempt_wakeup_fair() 8726 goto preempt; in check_preempt_wakeup_fair() 8733 preempt: in check_preempt_wakeup_fair()
|
| /kernel/trace/ |
| A D | trace_sched_switch.c | 25 probe_sched_switch(void *ignore, bool preempt, in probe_sched_switch() argument
|
| A D | Kconfig | 159 Create preempt/irq toggle tracepoints if needed, so that other parts 412 enabled. This option and the preempt-off timing option can be 605 of the overall system. This is enabled by default when the preempt 1173 tristate "Test module to create a preempt / IRQ disable delay thread to test latency tracers" 1177 tracers by executing a preempt or irq disable section with a user
|
| A D | trace_sched_wakeup.c | 444 probe_wakeup_sched_switch(void *ignore, bool preempt, in probe_wakeup_sched_switch() argument
|
| A D | fgraph.c | 1078 ftrace_graph_probe_sched_switch(void *ignore, bool preempt, in ftrace_graph_probe_sched_switch() argument
|
| A D | trace_osnoise.c | 1253 trace_sched_switch_callback(void *data, bool preempt, in trace_sched_switch_callback() argument
|
| A D | trace_events.c | 1063 event_filter_pid_sched_switch_probe_pre(void *data, bool preempt, in event_filter_pid_sched_switch_probe_pre() argument 1089 event_filter_pid_sched_switch_probe_post(void *data, bool preempt, in event_filter_pid_sched_switch_probe_post() argument
|
| A D | trace.c | 2656 scoped_guard(preempt,) { in trace_buffered_event_enable() 3146 guard(preempt)(); in ftrace_trace_userstack() local
|
| A D | ftrace.c | 8145 ftrace_filter_pid_sched_switch_probe(void *data, bool preempt, in ftrace_filter_pid_sched_switch_probe() argument
|