Searched refs:cfs_rq (Results 1 – 8 of 8) sorted by relevance
/linux-6.3-rc2/kernel/sched/ |
A D | fair.c | 3933 struct cfs_rq *cfs_rq; in migrate_se_pelt_lag() local 5277 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up() local 5295 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_throttle_down() local 5551 struct cfs_rq *cfs_rq; in distribute_cfs_runtime() local 6005 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in update_runtime_enabled() local 6023 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in unthrottle_offline_cfs_rqs() local 6196 struct cfs_rq *cfs_rq; in enqueue_task_fair() local 6289 struct cfs_rq *cfs_rq; in dequeue_task_fair() local 6396 struct cfs_rq *cfs_rq; in cpu_load_without() local 6419 struct cfs_rq *cfs_rq; in cpu_runnable_without() local [all …]
|
A D | pelt.h | 5 int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se); 6 int __update_load_avg_cfs_rq(u64 now, struct cfs_rq *cfs_rq); 156 static inline void update_idle_cfs_rq_clock_pelt(struct cfs_rq *cfs_rq) in update_idle_cfs_rq_clock_pelt() argument 169 static inline u64 cfs_rq_clock_pelt(struct cfs_rq *cfs_rq) in cfs_rq_clock_pelt() argument 172 return cfs_rq->throttled_clock_pelt - cfs_rq->throttled_clock_pelt_time; in cfs_rq_clock_pelt() 174 return rq_clock_pelt(rq_of(cfs_rq)) - cfs_rq->throttled_clock_pelt_time; in cfs_rq_clock_pelt() 177 static inline void update_idle_cfs_rq_clock_pelt(struct cfs_rq *cfs_rq) { } in update_idle_cfs_rq_clock_pelt() argument 178 static inline u64 cfs_rq_clock_pelt(struct cfs_rq *cfs_rq) in cfs_rq_clock_pelt() argument 180 return rq_clock_pelt(rq_of(cfs_rq)); in cfs_rq_clock_pelt() 187 update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq) in update_cfs_rq_load_avg() argument [all …]
|
A D | pelt.c | 306 int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se) in __update_load_avg_se() argument 309 cfs_rq->curr == se)) { in __update_load_avg_se() 320 int __update_load_avg_cfs_rq(u64 now, struct cfs_rq *cfs_rq) in __update_load_avg_cfs_rq() argument 322 if (___update_load_sum(now, &cfs_rq->avg, in __update_load_avg_cfs_rq() 323 scale_load_down(cfs_rq->load.weight), in __update_load_avg_cfs_rq() 324 cfs_rq->h_nr_running, in __update_load_avg_cfs_rq() 325 cfs_rq->curr != NULL)) { in __update_load_avg_cfs_rq() 327 ___update_load_avg(&cfs_rq->avg, 1); in __update_load_avg_cfs_rq() 328 trace_pelt_cfs_tp(cfs_rq); in __update_load_avg_cfs_rq()
|
A D | debug.c | 581 void print_cfs_rq(struct seq_file *m, int cpu, struct cfs_rq *cfs_rq) in print_cfs_rq() argument 621 cfs_rq->nr_spread_over); in print_cfs_rq() 625 cfs_rq->idle_nr_running); in print_cfs_rq() 627 cfs_rq->idle_h_nr_running); in print_cfs_rq() 631 cfs_rq->avg.load_avg); in print_cfs_rq() 633 cfs_rq->avg.runnable_avg); in print_cfs_rq() 635 cfs_rq->avg.util_avg); in print_cfs_rq() 639 cfs_rq->removed.load_avg); in print_cfs_rq() 641 cfs_rq->removed.util_avg); in print_cfs_rq() 653 cfs_rq->throttled); in print_cfs_rq() [all …]
|
A D | sched.h | 337 struct cfs_rq; 376 struct cfs_rq **cfs_rq; member 467 extern void unthrottle_cfs_rq(struct cfs_rq *cfs_rq); 493 struct cfs_rq *prev, struct cfs_rq *next); 496 struct cfs_rq *prev, struct cfs_rq *next) { } in set_task_rq_fair() 1169 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of() argument 1176 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of() argument 1954 set_task_rq_fair(&p->se, p->se.cfs_rq, tg->cfs_rq[cpu]); in set_task_rq() 1955 p->se.cfs_rq = tg->cfs_rq[cpu]; in set_task_rq() 2782 extern void init_cfs_rq(struct cfs_rq *cfs_rq); [all …]
|
A D | core.c | 4437 p->se.cfs_rq = NULL; in __sched_fork() 5458 struct sched_entity *curr = (&p->se)->cfs_rq->curr; in prefetch_curr_exec_start() 9808 root_task_group.cfs_rq = (struct cfs_rq **)ptr; in sched_init() 10751 struct cfs_rq *cfs_rq = tg->cfs_rq[i]; in tg_set_cfs_bandwidth() local 10752 struct rq *rq = cfs_rq->rq; in tg_set_cfs_bandwidth() 10756 cfs_rq->runtime_enabled = runtime_enabled; in tg_set_cfs_bandwidth() 10757 cfs_rq->runtime_remaining = 0; in tg_set_cfs_bandwidth() 10759 if (cfs_rq->throttled) in tg_set_cfs_bandwidth() 10760 unthrottle_cfs_rq(cfs_rq); in tg_set_cfs_bandwidth()
|
/linux-6.3-rc2/include/trace/events/ |
A D | sched.h | 695 TP_PROTO(struct cfs_rq *cfs_rq), 696 TP_ARGS(cfs_rq)); 727 TP_PROTO(struct cfs_rq *cfs_rq), 728 TP_ARGS(cfs_rq));
|
/linux-6.3-rc2/include/linux/ |
A D | sched.h | 49 struct cfs_rq; 565 struct cfs_rq *cfs_rq; member 567 struct cfs_rq *my_q;
|
Completed in 60 milliseconds