Lines Matching refs:next

1015 	if (unlikely(cmpxchg_relaxed(&node->next, NULL, WAKE_Q_TAIL)))  in __wake_q_add()
1022 head->lastp = &node->next; in __wake_q_add()
1075 node = node->next; in wake_up_q()
1077 WRITE_ONCE(task->wake_q.next, NULL); in wake_up_q()
4940 struct task_struct *next) in __fire_sched_out_preempt_notifiers() argument
4945 notifier->ops->sched_out(notifier, next); in __fire_sched_out_preempt_notifiers()
4950 struct task_struct *next) in fire_sched_out_preempt_notifiers() argument
4953 __fire_sched_out_preempt_notifiers(curr, next); in fire_sched_out_preempt_notifiers()
4964 struct task_struct *next) in fire_sched_out_preempt_notifiers() argument
4970 static inline void prepare_task(struct task_struct *next) in prepare_task() argument
4979 WRITE_ONCE(next->on_cpu, 1); in prepare_task()
5001 struct balance_callback *next; in do_balance_callbacks() local
5007 next = head->next; in do_balance_callbacks()
5008 head->next = NULL; in do_balance_callbacks()
5009 head = next; in do_balance_callbacks()
5029 .next = NULL,
5080 prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf) in prepare_lock_switch() argument
5092 rq_lockp(rq)->owner = next; in prepare_lock_switch()
5113 # define prepare_arch_switch(next) do { } while (0) argument
5151 struct task_struct *next) in prepare_task_switch() argument
5154 sched_info_switch(rq, prev, next); in prepare_task_switch()
5155 perf_event_task_sched_out(prev, next); in prepare_task_switch()
5157 fire_sched_out_preempt_notifiers(prev, next); in prepare_task_switch()
5159 prepare_task(next); in prepare_task_switch()
5160 prepare_arch_switch(next); in prepare_task_switch()
5302 struct task_struct *next, struct rq_flags *rf) in context_switch() argument
5304 prepare_task_switch(rq, prev, next); in context_switch()
5323 if (!next->mm) { // to kernel in context_switch()
5324 enter_lazy_tlb(prev->active_mm, next); in context_switch()
5326 next->active_mm = prev->active_mm; in context_switch()
5332 membarrier_switch_mm(rq, prev->active_mm, next->mm); in context_switch()
5341 switch_mm_irqs_off(prev->active_mm, next->mm, next); in context_switch()
5342 lru_gen_use_mm(next->mm); in context_switch()
5352 switch_mm_cid(rq, prev, next); in context_switch()
5354 prepare_lock_switch(rq, next, rf); in context_switch()
5357 switch_to(prev, next, prev); in context_switch()
6076 struct task_struct *next, *p, *max = NULL; in pick_next_task() local
6116 next = rq->core_pick; in pick_next_task()
6161 next = pick_task(rq); in pick_next_task()
6162 if (!next->core_cookie) { in pick_next_task()
6170 task_vruntime_update(rq, next, false); in pick_next_task()
6237 next = rq->core_pick; in pick_next_task()
6241 WARN_ON_ONCE(!next); in pick_next_task()
6284 WARN_ON_ONCE(!cookie_match(next, rq_i->core_pick)); in pick_next_task()
6296 put_prev_set_next_task(rq, prev, next); in pick_next_task()
6297 if (rq->core->core_forceidle_count && next == rq->idle) in pick_next_task()
6300 return next; in pick_next_task()
6326 goto next; in try_steal_cookie()
6329 goto next; in try_steal_cookie()
6332 goto next; in try_steal_cookie()
6340 goto next; in try_steal_cookie()
6348 next: in try_steal_cookie()
6818 struct task_struct *prev, *next; in __schedule() local
6886 next = prev; in __schedule()
6902 next = pick_next_task(rq, rq->donor, &rf); in __schedule()
6903 rq_set_donor(rq, next); in __schedule()
6904 if (unlikely(task_is_blocked(next))) { in __schedule()
6905 next = find_proxy_task(rq, next, &rf); in __schedule()
6906 if (!next) in __schedule()
6908 if (next == rq->idle) in __schedule()
6917 is_switch = prev != next; in __schedule()
6924 RCU_INIT_POINTER(rq->curr, next); in __schedule()
6926 if (!task_current_donor(rq, next)) in __schedule()
6927 proxy_tag_curr(rq, next); in __schedule()
6954 psi_account_irqtime(rq, prev, next); in __schedule()
6955 psi_sched_switch(prev, next, !task_on_rq_queued(prev) || in __schedule()
6958 trace_sched_switch(preempt, prev, next, prev_state); in __schedule()
6961 rq = context_switch(rq, prev, next, &rf); in __schedule()
6964 if (!task_current_donor(rq, next)) in __schedule()
6965 proxy_tag_curr(rq, next); in __schedule()
10793 work->next = work; /* Prevent double-add */ in task_mm_cid_work()
10837 t->cid_work.next = &t->cid_work; /* Protect against double add */ in init_sched_mm_cid()
10847 work->next != work) in task_tick_mm_cid()