Lines Matching refs:sa
106 accumulate_sum(u64 delta, struct sched_avg *sa, in accumulate_sum() argument
112 delta += sa->period_contrib; in accumulate_sum()
119 sa->load_sum = decay_load(sa->load_sum, periods); in accumulate_sum()
120 sa->runnable_sum = in accumulate_sum()
121 decay_load(sa->runnable_sum, periods); in accumulate_sum()
122 sa->util_sum = decay_load((u64)(sa->util_sum), periods); in accumulate_sum()
140 1024 - sa->period_contrib, delta); in accumulate_sum()
143 sa->period_contrib = delta; in accumulate_sum()
146 sa->load_sum += load * contrib; in accumulate_sum()
148 sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum()
150 sa->util_sum += contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum()
184 ___update_load_sum(u64 now, struct sched_avg *sa, in ___update_load_sum() argument
189 delta = now - sa->last_update_time; in ___update_load_sum()
195 sa->last_update_time = now; in ___update_load_sum()
207 sa->last_update_time += delta << 10; in ___update_load_sum()
230 if (!accumulate_sum(delta, sa, load, runnable, running)) in ___update_load_sum()
261 ___update_load_avg(struct sched_avg *sa, unsigned long load) in ___update_load_avg() argument
263 u32 divider = get_pelt_divider(sa); in ___update_load_avg()
268 sa->load_avg = div_u64(load * sa->load_sum, divider); in ___update_load_avg()
269 sa->runnable_avg = div_u64(sa->runnable_sum, divider); in ___update_load_avg()
270 WRITE_ONCE(sa->util_avg, sa->util_sum / divider); in ___update_load_avg()