Searched refs:cpu_of (Results 1 – 8 of 8) sorted by relevance
/linux-6.3-rc2/kernel/sched/ |
A D | pelt.h | 118 delta = cap_scale(delta, arch_scale_cpu_capacity(cpu_of(rq))); in update_rq_clock_pelt() 119 delta = cap_scale(delta, arch_scale_freq_capacity(cpu_of(rq))); in update_rq_clock_pelt()
|
A D | pelt.c | 439 running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq))); in update_irq_load_avg() 440 running = cap_scale(running, arch_scale_cpu_capacity(cpu_of(rq))); in update_irq_load_avg()
|
A D | core_sched.c | 242 const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq)); in __sched_core_account_forceidle()
|
A D | sched.h | 1182 static inline int cpu_of(struct rq *rq) in cpu_of() function 1273 for_each_cpu(cpu, cpu_smt_mask(cpu_of(rq))) { in sched_core_cookie_match() 2402 int cpu = cpu_of(rq); in sched_update_tick_dependency() 2487 if (!cpu_active(cpu_of(rq))) in hrtick_enabled() 2879 cpu_of(rq))); in cpufreq_update_util() 3081 rq_util = cpu_util_cfs(cpu_of(rq)) + cpu_util_rt(rq); in uclamp_rq_is_capped()
|
A D | fair.c | 366 int cpu = cpu_of(rq); in list_add_leaf_cfs_rq() 4584 if (task_fits_cpu(p, cpu_of(rq))) { in update_misfit_status() 5277 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up() 5335 se = cfs_rq->tg->se[cpu_of(rq_of(cfs_rq))]; in throttle_cfs_rq() 5401 se = cfs_rq->tg->se[cpu_of(rq)]; in unthrottle_cfs_rq() 5590 if (cpu_of(rq) != this_cpu || in distribute_cfs_runtime() 6747 int core = cpu_of(rq); in __update_idle_core() 8817 int cpu = cpu_of(rq); in __update_blocked_fair() 10775 __cpumask_clear_cpu(cpu_of(busiest), cpus); in load_balance() 10835 stop_one_cpu_nowait(cpu_of(busiest), in load_balance() [all …]
|
A D | core.c | 710 steal = paravirt_steal_clock(cpu_of(rq)); in update_rq_clock_task() 745 delta = sched_clock_cpu(cpu_of(rq)) - rq->clock; in update_rq_clock() 772 WARN_ON_ONCE(cpu_of(rq) != smp_processor_id()); in hrtick() 1031 cpu = cpu_of(rq); in resched_curr() 1158 int cpu = cpu_of(rq); in nohz_csd_func() 2869 stop_one_cpu_nowait(cpu_of(rq), migration_cpu_stop, in affine_move_task() 3809 if (WARN_ON_ONCE(task_cpu(p) != cpu_of(rq))) in sched_ttwu_pending() 3810 set_task_cpu(p, cpu_of(rq)); in sched_ttwu_pending() 6009 cpu = cpu_of(rq); in pick_next_task() 6293 int cpu = cpu_of(rq); in sched_core_balance() [all …]
|
A D | rt.c | 562 (rt_rq = iter->rt_rq[cpu_of(rq)]);) 581 int cpu = cpu_of(rq); in sched_rt_rq_enqueue() 599 int cpu = cpu_of(rq_of_rt_rq(rt_rq)); in sched_rt_rq_dequeue() 2553 if (p->prio < rq->curr->prio && cpu_online(cpu_of(rq))) in switched_to_rt()
|
A D | deadline.c | 1312 int cpu = cpu_of(rq); in update_curr_dl() 2506 src_dl_b = dl_bw_of(cpu_of(rq)); in set_cpus_allowed_dl()
|
Completed in 68 milliseconds