| /linux/kernel/irq/ |
| A D | ipi.c | 35 if (!cpumask_subset(dest, cpu_possible_mask)) { in irq_reserve_ipi() 135 if (!ipimask || WARN_ON(!cpumask_subset(dest, ipimask))) in irq_destroy_ipi() 207 if (!cpumask_subset(dest, ipimask)) in ipi_send_verify()
|
| A D | cpuhotplug.c | 199 if (cpumask_subset(irq_data_get_effective_affinity_mask(data), hk_mask)) in hk_should_isolate()
|
| /linux/drivers/base/ |
| A D | arch_topology.c | 36 return cpumask_subset(cpus, &scale_freq_counters_mask); in supports_scale_freq_counters() 735 if (cpumask_subset(&cpu_topology[cpu].core_sibling, core_mask)) { in cpu_coregroup_mask() 741 if (cpumask_subset(&cpu_topology[cpu].llc_sibling, core_mask)) in cpu_coregroup_mask() 751 cpumask_subset(core_mask, &cpu_topology[cpu].cluster_sibling)) in cpu_coregroup_mask() 763 if (cpumask_subset(cpu_coregroup_mask(cpu), in cpu_clustergroup_mask()
|
| /linux/arch/powerpc/include/asm/ |
| A D | tlb.h | 55 return cpumask_subset(mm_cpumask(mm), in mm_is_core_local()
|
| /linux/kernel/cgroup/ |
| A D | cpuset.c | 609 if (!cpumask_empty(acpus) && cpumask_subset(acpus, xcpus)) in validate_change() 1019 if (!cpumask_subset(cs->effective_cpus, in rebuild_sched_domains_locked() 1180 return (cpumask_subset(parent->effective_cpus, xcpus) && in tasks_nocpu_error() 1370 cpumask_subset(top_cpuset.effective_cpus, tmp->new_cpus)) in remote_partition_enable() 1402 WARN_ON_ONCE(!cpumask_subset(tmp->new_cpus, subpartitions_cpus)); in remote_partition_disable() 1456 cpumask_subset(top_cpuset.effective_cpus, tmp->addmask))) in remote_cpus_update() 1737 cpumask_subset(xcpus, parent->effective_xcpus)) { in update_parent_effective_cpumask() 1898 if (!cpumask_subset(child->effective_xcpus, in compute_partition_effective_cpumask() 1902 cpumask_subset(new_ecpus, child->effective_xcpus)) in compute_partition_effective_cpumask() 2204 if (!cpumask_subset(trialcs->cpus_allowed, in update_cpumask() [all …]
|
| A D | cpuset-v1.c | 341 return cpumask_subset(p->cpus_allowed, q->cpus_allowed) && in is_cpuset_subset()
|
| /linux/arch/mips/kernel/ |
| A D | mips-mt-fpaff.c | 133 if (!cpumask_subset(effective_mask, cpus_allowed)) { in mipsmt_sys_sched_setaffinity()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | irq_affinity.c | 100 if (!cpumask_subset(iter_mask, req_mask)) in irq_pool_find_least_loaded()
|
| /linux/arch/arm64/kernel/ |
| A D | topology.c | 197 if (unlikely(cpumask_subset(cpus, amu_fie_cpus))) in amu_fie_setup()
|
| /linux/arch/powerpc/platforms/pseries/ |
| A D | hotplug-cpu.c | 167 BUG_ON(!cpumask_subset(cpu_present_mask, cpu_possible_mask)); in find_cpu_id_range() 190 if (cpumask_subset(*cpu_mask, candidate_mask)) in find_cpu_id_range()
|
| /linux/kernel/bpf/ |
| A D | cpumask.c | 335 return cpumask_subset(src1, src2); in bpf_cpumask_subset()
|
| /linux/drivers/iommu/ |
| A D | hyperv-iommu.c | 47 if (!cpumask_subset(mask, &ioapic_max_cpumask)) in hyperv_ir_set_affinity()
|
| /linux/kernel/sched/ |
| A D | syscalls.c | 664 if (!cpumask_subset(span, p->cpus_ptr) || in __sched_setscheduler() 1187 if (!cpumask_subset(task_rq(p)->rd->span, mask)) in dl_task_check_affinity() 1222 if (!cpumask_subset(new_mask, cpus_allowed)) { in __sched_setaffinity()
|
| A D | topology.c | 126 !cpumask_subset(groupmask, sched_domain_span(sd->parent))) in sched_domain_debug_one() 1023 !cpumask_subset(sched_domain_span(sibling->child), in find_descended_sibling() 1105 !cpumask_subset(sched_domain_span(sibling->child), span)) in build_overlap_sched_groups() 2338 if (!cpumask_subset(sched_domain_span(child), in build_sched_domain()
|
| /linux/arch/powerpc/sysdev/xics/ |
| A D | xics-common.c | 285 if (!cpumask_subset(cpu_possible_mask, cpumask)) { in xics_get_irq_server()
|
| /linux/arch/sparc/kernel/ |
| A D | leon_kernel.c | 111 if (cpumask_subset(cpu_online_mask, affinity) || cpu >= nr_cpu_ids) in irq_choose_cpu()
|
| /linux/kernel/ |
| A D | taskstats.c | 302 if (!cpumask_subset(mask, cpu_possible_mask)) in add_del_listener()
|
| /linux/kernel/time/ |
| A D | tick-broadcast.c | 741 if (WARN_ON_ONCE(!cpumask_subset(tmpmask, cpu_online_mask))) in tick_handle_oneshot_broadcast()
|
| /linux/include/linux/ |
| A D | cpumask.h | 739 bool cpumask_subset(const struct cpumask *src1p, const struct cpumask *src2p) in cpumask_subset() function
|
| /linux/arch/x86/kernel/cpu/resctrl/ |
| A D | pseudo_lock.c | 1529 if (!cpumask_subset(current->cpus_ptr, &plr->d->hdr.cpu_mask)) { in pseudo_lock_dev_mmap()
|
| /linux/arch/x86/kernel/apic/ |
| A D | vector.c | 422 if (!cpumask_subset(irq_data_get_effective_affinity_mask(irqd), in activate_reserved()
|
| /linux/arch/powerpc/kernel/ |
| A D | smp.c | 1433 !cpumask_subset(submask_fn(cpu), cpu_l2_cache_mask(cpu))) { in update_mask_by_l2()
|
| /linux/kernel/rcu/ |
| A D | tree_nocb.h | 1327 if (!cpumask_subset(rcu_nocb_mask, cpu_possible_mask)) { in rcu_init_nohz()
|
| /linux/io_uring/ |
| A D | io-wq.c | 1338 if (cpumask_subset(mask, allowed_mask)) in io_wq_cpu_affinity()
|
| /linux/arch/mips/cavium-octeon/ |
| A D | octeon-irq.c | 2522 if (!cpumask_subset(dest, cpumask_of_node(cd->ciu_node))) in octeon_irq_ciu3_set_affinity()
|