Lines Matching refs:scurr

3250                struct csched2_vcpu *scurr,  in runq_candidate()  argument
3261 if ( unlikely(is_idle_vcpu(scurr->vcpu)) ) in runq_candidate()
3263 snext = scurr; in runq_candidate()
3267 yield = __test_and_clear_bit(__CSFLAG_vcpu_yield, &scurr->flags); in runq_candidate()
3278 if ( !yield && prv->ratelimit_us && vcpu_runnable(scurr->vcpu) && in runq_candidate()
3279 (now - scurr->vcpu->runstate.state_entry_time) < in runq_candidate()
3288 d.dom = scurr->vcpu->domain->domain_id; in runq_candidate()
3289 d.vcpu = scurr->vcpu->vcpu_id; in runq_candidate()
3290 d.runtime = now - scurr->vcpu->runstate.state_entry_time; in runq_candidate()
3295 return scurr; in runq_candidate()
3299 if ( has_soft_affinity(scurr->vcpu, scurr->vcpu->cpu_hard_affinity) ) in runq_candidate()
3301 affinity_balance_cpumask(scurr->vcpu, BALANCE_SOFT_AFFINITY, in runq_candidate()
3305 cpumask_t *online = cpupool_domain_cpumask(scurr->vcpu->domain); in runq_candidate()
3333 if ( vcpu_runnable(scurr->vcpu) && !soft_aff_preempt ) in runq_candidate()
3334 snext = scurr; in runq_candidate()
3440 struct csched2_vcpu * const scurr = csched2_vcpu(current); in csched2_schedule() local
3456 BUG_ON(!is_idle_vcpu(scurr->vcpu) && scurr->rqd != rqd); in csched2_schedule()
3485 burn_credits(rqd, scurr, now); in csched2_schedule()
3492 if ( unlikely(scurr->budget <= 0) ) in csched2_schedule()
3493 vcpu_grab_budget(scurr); in csched2_schedule()
3515 __clear_bit(__CSFLAG_vcpu_yield, &scurr->flags); in csched2_schedule()
3520 snext = runq_candidate(rqd, scurr, cpu, now, &skipped_vcpus); in csched2_schedule()
3524 if ( snext != scurr in csched2_schedule()
3525 && !is_idle_vcpu(scurr->vcpu) in csched2_schedule()
3527 __set_bit(__CSFLAG_delayed_runq_add, &scurr->flags); in csched2_schedule()
3535 if ( snext != scurr ) in csched2_schedule()