Lines Matching refs:new_mask
2202 .new_mask = cpumask_of(rq->cpu), in migrate_disable_switch()
2238 .new_mask = &p->cpus_mask, in migrate_enable()
2539 p->cpus_ptr = ctx->new_mask; in set_cpus_allowed_common()
2543 cpumask_copy(&p->cpus_mask, ctx->new_mask); in set_cpus_allowed_common()
2544 p->nr_cpus_allowed = cpumask_weight(ctx->new_mask); in set_cpus_allowed_common()
2602 void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) in do_set_cpus_allowed() argument
2605 .new_mask = new_mask, in do_set_cpus_allowed()
2941 if (!kthread && !cpumask_subset(ctx->new_mask, cpu_allowed_mask)) { in __set_cpus_allowed_ptr_locked()
2956 if (cpumask_equal(&p->cpus_mask, ctx->new_mask)) { in __set_cpus_allowed_ptr_locked()
2964 !cpumask_test_cpu(task_cpu(p), ctx->new_mask))) { in __set_cpus_allowed_ptr_locked()
2975 dest_cpu = cpumask_any_and_distribute(cpu_valid_mask, ctx->new_mask); in __set_cpus_allowed_ptr_locked()
3013 cpumask_and(rq->scratch_mask, ctx->new_mask, p->user_cpus_ptr)) in __set_cpus_allowed_ptr()
3014 ctx->new_mask = rq->scratch_mask; in __set_cpus_allowed_ptr()
3019 int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) in set_cpus_allowed_ptr() argument
3022 .new_mask = new_mask, in set_cpus_allowed_ptr()
3040 struct cpumask *new_mask, in restrict_cpus_allowed_ptr() argument
3044 .new_mask = new_mask, in restrict_cpus_allowed_ptr()
3063 if (!cpumask_and(new_mask, task_user_cpus(p), subset_mask)) { in restrict_cpus_allowed_ptr()
3083 cpumask_var_t new_mask; in force_compatible_cpus_allowed_ptr() local
3086 alloc_cpumask_var(&new_mask, GFP_KERNEL); in force_compatible_cpus_allowed_ptr()
3094 if (!cpumask_available(new_mask)) in force_compatible_cpus_allowed_ptr()
3097 if (!restrict_cpus_allowed_ptr(p, new_mask, override_mask)) in force_compatible_cpus_allowed_ptr()
3104 cpuset_cpus_allowed(p, new_mask); in force_compatible_cpus_allowed_ptr()
3105 override_mask = new_mask; in force_compatible_cpus_allowed_ptr()
3117 free_cpumask_var(new_mask); in force_compatible_cpus_allowed_ptr()
3133 .new_mask = task_user_cpus(p), in relax_compatible_cpus_allowed_ptr()
3623 return set_cpus_allowed_ptr(p, ctx->new_mask); in __set_cpus_allowed_ptr()
8213 cpumask_var_t cpus_allowed, new_mask; in __sched_setaffinity() local
8218 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { in __sched_setaffinity()
8224 cpumask_and(new_mask, ctx->new_mask, cpus_allowed); in __sched_setaffinity()
8226 ctx->new_mask = new_mask; in __sched_setaffinity()
8229 retval = dl_task_check_affinity(p, new_mask); in __sched_setaffinity()
8238 if (!cpumask_subset(new_mask, cpus_allowed)) { in __sched_setaffinity()
8243 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity()
8254 bool empty = !cpumask_and(new_mask, new_mask, in __sched_setaffinity()
8258 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity()
8265 free_cpumask_var(new_mask); in __sched_setaffinity()
8322 .new_mask = in_mask, in sched_setaffinity()
8336 struct cpumask *new_mask) in get_user_cpu_mask() argument
8339 cpumask_clear(new_mask); in get_user_cpu_mask()
8343 return copy_from_user(new_mask, user_mask_ptr, len) ? -EFAULT : 0; in get_user_cpu_mask()
8357 cpumask_var_t new_mask; in SYSCALL_DEFINE3() local
8360 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) in SYSCALL_DEFINE3()
8363 retval = get_user_cpu_mask(user_mask_ptr, len, new_mask); in SYSCALL_DEFINE3()
8365 retval = sched_setaffinity(pid, new_mask); in SYSCALL_DEFINE3()
8366 free_cpumask_var(new_mask); in SYSCALL_DEFINE3()
9112 .new_mask = cpumask_of(cpu), in init_idle()