/xen/tools/debugger/kdd/ |
A D | kdd-xen.c | 172 if ((void *) cpu + sizeof (*cpu) <= g->hvm_buf + g->hvm_sz) in get_cpu() 330 if (!cpu) in kdd_get_regs() 348 if (!cpu) in kdd_set_regs() 364 c->cr0 = cpu->cr0; in kdd_get_ctrl_x86_32() 365 c->cr2 = cpu->cr2; in kdd_get_ctrl_x86_32() 366 c->cr3 = cpu->cr3; in kdd_get_ctrl_x86_32() 367 c->cr4 = cpu->cr4; in kdd_get_ctrl_x86_32() 368 c->dr0 = cpu->dr0; in kdd_get_ctrl_x86_32() 412 if (!cpu) in kdd_get_ctrl() 432 if (!cpu) in kdd_wrmsr() [all …]
|
/xen/xen/include/xen/ |
A D | cpumask.h | 85 return cpu; in cpumask_check() 231 cpu = cpumask_next(cpu, srcp)) in cpumask_last() 232 pcpu = cpu; in cpumask_last() 268 return cpu; in cpumask_any() 376 (cpu) = cpumask_next(cpu, mask)) 379 for ((cpu) = 0; (cpu) < cpumask_test_cpu(0, mask); ++(cpu)) 446 #define cpu_online(cpu) cpumask_test_cpu(cpu, &cpu_online_map) argument 448 #define cpu_present(cpu) cpumask_test_cpu(cpu, &cpu_present_map) argument 453 #define cpu_online(cpu) ((cpu) == 0) argument 454 #define cpu_possible(cpu) ((cpu) == 0) argument [all …]
|
A D | irq_cpustat.h | 22 #define __IRQ_STAT(cpu, member) (irq_stat[cpu].member) argument 25 #define softirq_pending(cpu) __IRQ_STAT((cpu), __softirq_pending) argument 26 #define local_irq_count(cpu) __IRQ_STAT((cpu), __local_irq_count) argument 27 #define mwait_wakeup(cpu) __IRQ_STAT((cpu), __mwait_wakeup) argument
|
/xen/xen/arch/x86/ |
A D | smpboot.c | 244 cpumask_set_cpu(cpu, socket_cpumask[cpu_to_socket(cpu)]); in set_cpu_sibling_map() 245 cpumask_set_cpu(cpu, per_cpu(cpu_core_mask, cpu)); in set_cpu_sibling_map() 246 cpumask_set_cpu(cpu, per_cpu(cpu_sibling_mask, cpu)); in set_cpu_sibling_map() 305 else if ( (i != cpu) && !c[cpu].booted_cores ) in set_cpu_sibling_map() 551 int cpu; in alloc_cpu_id() local 555 return (cpu < nr_cpu_ids) ? cpu : -ENODEV; in alloc_cpu_id() 1048 (stack_base[cpu] = cpu_alloc_stack(cpu)) == NULL ) in cpu_smpboot_alloc() 1090 stub_page = alloc_stub_page(cpu, &per_cpu(stubs.mfn, cpu)); in cpu_smpboot_alloc() 1247 cpumask_clear_cpu(cpu, socket_cpumask[cpu_to_socket(cpu)]); in remove_siblinginfo() 1305 int cpu = -1; in cpu_add() local [all …]
|
A D | percpu.c | 20 unsigned int cpu; in percpu_init_areas() local 22 for ( cpu = 1; cpu < NR_CPUS; cpu++ ) in percpu_init_areas() 26 static int init_percpu_area(unsigned int cpu) in init_percpu_area() argument 43 unsigned int cpu; member 51 unsigned int cpu = info->cpu; in _free_percpu_area() local 58 static void free_percpu_area(unsigned int cpu) in free_percpu_area() argument 62 info->cpu = cpu; in free_percpu_area() 69 unsigned int cpu = (unsigned long)hcpu; in cpu_percpu_callback() local 75 rc = init_percpu_area(cpu); in cpu_percpu_callback() 81 free_percpu_area(cpu); in cpu_percpu_callback() [all …]
|
/xen/xen/common/ |
A D | cpu.c | 117 if ( (cpu >= nr_cpu_ids) || (cpu == 0) ) in cpu_down() 121 if ( !cpu_online(cpu) ) in cpu_down() 133 __cpu_die(cpu); in cpu_down() 134 err = cpu_online(cpu); in cpu_down() 150 int cpu_up(unsigned int cpu) in cpu_up() argument 158 if ( (cpu >= nr_cpu_ids) || !cpu_present(cpu) ) in cpu_up() 162 if ( cpu_online(cpu) ) in cpu_up() 169 err = __cpu_up(cpu); in cpu_up() 196 int cpu, error = 0; in disable_nonboot_cpus() local 206 if ( cpu == 0 ) in disable_nonboot_cpus() [all …]
|
A D | core_parking.c | 59 unsigned int cpu = -1; in core_parking_performance() local 71 for_each_cpu(cpu, &cpu_online_map) in core_parking_performance() 73 if ( cpu == 0 ) in core_parking_performance() 112 return cpu; in core_parking_performance() 117 unsigned int cpu = -1; in core_parking_power() local 131 if ( cpu == 0 ) in core_parking_power() 170 return cpu; in core_parking_power() 176 unsigned int cpu; in core_parking_helper() local 186 ret = cpu_down(cpu); in core_parking_helper() 200 ret = cpu_up(cpu); in core_parking_helper() [all …]
|
A D | softirq.c | 30 unsigned int i, cpu; in __do_softirq() local 42 cpu = smp_processor_id(); in __do_softirq() 45 rcu_check_callbacks(cpu); in __do_softirq() 48 || cpu_is_offline(cpu) ) in __do_softirq() 52 clear_bit(i, &softirq_pending(cpu)); in __do_softirq() 96 for_each_cpu(cpu, mask) in cpumask_raise_softirq() 98 cpu != this_cpu && in cpumask_raise_softirq() 111 || (cpu == this_cpu) in cpu_raise_softirq() 116 smp_send_event_check_cpu(cpu); in cpu_raise_softirq() 132 for_each_cpu ( cpu, mask ) in cpu_raise_softirq_batch_finish() [all …]
|
A D | tasklet.c | 35 unsigned int cpu = t->scheduled_on; in tasklet_enqueue() local 62 t->scheduled_on = cpu; in tasklet_schedule_on_cpu() 109 unsigned int cpu = smp_processor_id(); in do_tasklet() local 122 ASSERT(tasklet_work_to_do(cpu)); in do_tasklet() 126 do_tasklet_work(cpu, list); in do_tasklet() 140 unsigned int cpu = smp_processor_id(); in tasklet_softirq_action() local 145 do_tasklet_work(cpu, list); in tasklet_softirq_action() 193 BUG_ON(t->scheduled_on != cpu); in migrate_tasklets_from_cpu() 221 unsigned int cpu = (unsigned long)hcpu; in cpu_callback() local 231 migrate_tasklets_from_cpu(cpu, &per_cpu(tasklet_list, cpu)); in cpu_callback() [all …]
|
A D | perfc.c | 40 unsigned int k, cpu; in perfc_printall() local 48 for_each_online_cpu ( cpu ) in perfc_printall() 56 for_each_online_cpu ( cpu ) in perfc_printall() 60 printk(" CPU%02u[%10"PRIperfc"u]", cpu, per_cpu(perfcounters, cpu)[j]); in perfc_printall() 68 for_each_online_cpu ( cpu ) in perfc_printall() 84 for_each_online_cpu ( cpu ) in perfc_printall() 94 for_each_online_cpu ( cpu ) in perfc_printall() 130 unsigned int cpu; in perfc_reset() local 135 for_each_online_cpu ( cpu ) in perfc_reset() 142 for_each_online_cpu ( cpu ) in perfc_reset() [all …]
|
/xen/xen/arch/arm/arm64/ |
A D | smpboot.c | 14 int (*prepare_cpu)(int cpu); 20 static int __init smp_spin_table_cpu_up(int cpu) in smp_spin_table_cpu_up() argument 24 if (!cpu_release_addr[cpu]) in smp_spin_table_cpu_up() 57 static int __init smp_psci_init(int cpu) in smp_psci_init() argument 87 smp_spin_table_init(cpu, dn); in dt_arch_cpu_init() 89 return smp_psci_init(cpu); in dt_arch_cpu_init() 102 return dt_arch_cpu_init(cpu, dn); in arch_cpu_init() 105 return smp_psci_init(cpu); in arch_cpu_init() 108 int arch_cpu_up(int cpu) in arch_cpu_up() argument 112 if ( !smp_enable_ops[cpu].prepare_cpu ) in arch_cpu_up() [all …]
|
/xen/xen/arch/arm/ |
A D | percpu.c | 14 unsigned int cpu; in percpu_init_areas() local 15 for ( cpu = 1; cpu < NR_CPUS; cpu++ ) in percpu_init_areas() 19 static int init_percpu_area(unsigned int cpu) in init_percpu_area() argument 27 __per_cpu_offset[cpu] = p - __per_cpu_start; in init_percpu_area() 32 unsigned int cpu; member 40 unsigned int cpu = info->cpu; in _free_percpu_area() local 46 static void free_percpu_area(unsigned int cpu) in free_percpu_area() argument 49 info->cpu = cpu; in free_percpu_area() 56 unsigned int cpu = (unsigned long)hcpu; in cpu_percpu_callback() local 62 rc = init_percpu_area(cpu); in cpu_percpu_callback() [all …]
|
A D | smpboot.c | 99 cpumask_set_cpu(cpu, per_cpu(cpu_sibling_mask, cpu)); in setup_cpu_sibling_map() 129 struct dt_device_node *cpu; in dt_smp_init_cpus() local 160 dt_node_full_name(cpu), dt_n_size_cells(cpu)); in dt_smp_init_cpus() 473 int __init cpu_up_send_sgi(int cpu) in cpu_up_send_sgi() argument 484 int __cpu_up(unsigned int cpu) in __cpu_up() argument 491 rc = prepare_secondary_mm(cpu); in __cpu_up() 501 init_data.cpuid = cpu; in __cpu_up() 506 rc = arch_cpu_up(cpu); in __cpu_up() 545 if ( !cpu_online(cpu) ) in __cpu_up() 555 void __cpu_die(unsigned int cpu) in __cpu_die() argument [all …]
|
/xen/xen/arch/x86/genapic/ |
A D | x2apic.c | 43 for_each_online_cpu ( cpu ) in init_apic_ldr_x2apic_cluster() 45 if ( this_cpu == cpu ) in init_apic_ldr_x2apic_cluster() 54 BUG_ON(cpu >= NR_CPUS); in init_apic_ldr_x2apic_cluster() 67 int cpu) in vector_allocation_cpumask_x2apic_cluster() argument 69 return per_cpu(cluster_cpus, cpu); in vector_allocation_cpumask_x2apic_cluster() 79 for_each_cpu ( cpu, cluster_cpus ) in cpu_mask_to_apicid_x2apic_cluster() 94 unsigned int cpu; in send_IPI_mask_x2apic_phys() local 112 for_each_cpu ( cpu, cpumask ) in send_IPI_mask_x2apic_phys() 114 if ( !cpu_online(cpu) || (cpu == smp_processor_id()) ) in send_IPI_mask_x2apic_phys() 128 unsigned int cpu = smp_processor_id(); in send_IPI_mask_x2apic_cluster() local [all …]
|
/xen/xen/common/sched/ |
A D | null.c | 127 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in unit_check_affinity() 130 return cpumask_test_cpu(cpu, cpumask_scratch_cpu(cpu)); in unit_check_affinity() 289 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), cpus); in pick_res() 299 && cpumask_test_cpu(cpu, cpumask_scratch_cpu(cpu))) ) in pick_res() 306 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in pick_res() 367 .cpu = cpu, in unit_assign() 400 .cpu = cpu, in unit_deassign() 466 unsigned int cpu; in null_unit_insert() local 532 unsigned int cpu; in null_unit_remove() local 635 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in null_unit_wake() [all …]
|
A D | private.h | 78 return get_sched_res(cpu)->curr; in curr_on_cpu() 156 ASSERT(cpu < nr_cpu_ids); in sched_set_res() 157 v->processor = cpu; in sched_set_res() 158 cpu = cpumask_next(cpu, res->cpus); in sched_set_res() 263 sched_lock(pcpu, unsigned int cpu, cpu, ) 265 sched_lock(pcpu, unsigned int cpu, cpu, _irq) in sched_lock() argument 267 sched_unlock(pcpu, unsigned int cpu, cpu, ) in sched_lock() 269 sched_unlock(pcpu, unsigned int cpu, cpu, _irq) in sched_lock() 275 sched_lock(pcpu, unsigned int cpu, cpu, _irqsave, *flags) in sched_lock() 278 sched_unlock(pcpu, unsigned int cpu, cpu, _irqrestore, flags) in sched_lock() [all …]
|
A D | cpupool.c | 111 unsigned int cpu; in cpupool_check_granularity() local 117 for_each_online_cpu ( cpu ) in cpupool_check_granularity() 555 cpupool_moving_cpu = cpu; in cpupool_unassign_cpu_start() 613 if ( (!c->cpupool_id && !cpu) || !cpu_online(cpu) ) in cpupool_unassign_cpu() 807 sched_rm_cpu(cpu); in cpupool_cpu_remove_forced() 866 unsigned int cpu; in cpupool_do_sysctl() local 869 cpu = op->cpu; in cpupool_do_sysctl() 913 unsigned int cpu; in cpupool_do_sysctl() local 919 cpu = op->cpu; in cpupool_do_sysctl() 922 ret = (cpu < nr_cpu_ids) ? cpupool_unassign_cpu(c, cpu) : -EINVAL; in cpupool_do_sysctl() [all …]
|
A D | credit.c | 389 ASSERT(cpumask_cycle(cpu, unit->cpu_hard_affinity) == cpu); in __runq_tickle() 496 trace_time(TRC_CSCHED_TICKLE, sizeof(cpu), &cpu); in __runq_tickle() 601 prv->master = cpu; in init_pdata() 608 prv->balance_bias[cpu_to_node(cpu)] = cpu; in init_pdata() 610 init_timer(&spc->ticker, csched_tick, (void *)(unsigned long)cpu, cpu); in init_pdata() 789 if ( sched_unit_master(unit) == cpu && is_runq_idle(cpu) ) in _csched_cpu_pick() 806 cpu = cpumask_cycle(cpu, cpus); in _csched_cpu_pick() 863 spc->idle_bias = cpu; in _csched_cpu_pick() 867 return cpu; in _csched_cpu_pick() 1864 uint16_t cpu; in csched_schedule() member [all …]
|
A D | credit2.c | 740 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in get_fallback_cpu() 747 if ( likely(cpumask_test_cpu(cpu, cpumask_scratch_cpu(cpu))) ) in get_fallback_cpu() 789 cpu = cpumask_first(cpumask_scratch_cpu(cpu)); in get_fallback_cpu() 909 cpumask_or(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in cpu_runqueue_siblings_match() 1636 ASSERT(cpumask_cycle(cpu, unit->cpu_hard_affinity) == cpu); in runq_tickle() 1638 ipid = cpu; in runq_tickle() 2507 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in csched2_res_pick() 2509 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in csched2_res_pick() 2523 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in csched2_res_pick() 2655 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in migrate() [all …]
|
/xen/xen/drivers/cpufreq/ |
A D | cpufreq_misc_governors.c | 33 unsigned int cpu; in cpufreq_governor_userspace() local 36 unlikely(!cpu_online(cpu = policy->cpu))) in cpufreq_governor_userspace() 41 if (!per_cpu(cpu_set_freq, cpu)) in cpufreq_governor_userspace() 42 per_cpu(cpu_set_freq, cpu) = policy->cur; in cpufreq_governor_userspace() 45 per_cpu(cpu_set_freq, cpu) = 0; in cpufreq_governor_userspace() 48 if (policy->max < per_cpu(cpu_set_freq, cpu)) in cpufreq_governor_userspace() 71 if (!cpu_online(cpu) || !(policy = per_cpu(cpufreq_cpu_policy, cpu))) in write_userspace_scaling_setspeed() 74 per_cpu(cpu_set_freq, cpu) = freq; in write_userspace_scaling_setspeed() 97 unsigned int cpu = (unsigned long)hcpu; in cpufreq_userspace_cpu_callback() local 121 unsigned int cpu; in cpufreq_gov_userspace_init() local [all …]
|
A D | cpufreq.c | 176 if (!cpu_online(cpu) || !(data = per_cpu(cpufreq_cpu_policy, cpu)) || in cpufreq_limit_change() 177 !processor_pminfo[cpu]) in cpufreq_limit_change() 206 if ( !processor_pminfo[cpu] || !cpu_online(cpu) ) in cpufreq_add_cpu() 278 policy->cpu = cpu; in cpufreq_add_cpu() 333 cpufreq_statistic_exit(cpu); in cpufreq_add_cpu() 364 if ( !processor_pminfo[cpu] || !cpu_online(cpu) ) in cpufreq_del_cpu() 462 int ret = 0, cpu; in set_px_pminfo() local 466 cpu = get_cpu_id(acpi_id); in set_px_pminfo() 467 if ( cpu < 0 || !perf ) in set_px_pminfo() 474 acpi_id, cpu); in set_px_pminfo() [all …]
|
A D | utility.c | 50 total_idle_ns = get_cpu_idle_time(cpu); in cpufreq_residency_update() 72 pxpt = per_cpu(cpufreq_statistic_data, cpu); in cpufreq_statistic_update() 82 cpufreq_residency_update(cpu, from); in cpufreq_statistic_update() 89 int cpufreq_statistic_init(unsigned int cpu) in cpufreq_statistic_init() argument 103 pxpt = per_cpu(cpufreq_statistic_data, cpu); in cpufreq_statistic_init() 147 void cpufreq_statistic_exit(unsigned int cpu) in cpufreq_statistic_exit() argument 168 void cpufreq_statistic_reset(unsigned int cpu) in cpufreq_statistic_reset() argument 252 if (!cpu_online(policy->cpu)) in cpufreq_frequency_table_verify() 302 if (!cpu_online(policy->cpu)) in cpufreq_frequency_table_target() 379 if (!cpu_online(cpu) || !(policy = per_cpu(cpufreq_cpu_policy, cpu))) in cpufreq_driver_getavg() [all …]
|
/xen/xen/arch/x86/acpi/cpufreq/ |
A D | hwp.c | 72 #define hwp_err(cpu, fmt, args...) \ argument 251 policy->cpu, hwp_req.raw); in hwp_write_request() 261 unsigned int cpu = policy->cpu; in hwp_cpufreq_target() local 420 unsigned int cpu = policy->cpu; in hwp_cpufreq_cpu_init() local 429 per_cpu(hwp_drv_data, cpu) = data; in hwp_cpufreq_cpu_init() 436 per_cpu(hwp_drv_data, cpu) = NULL; in hwp_cpufreq_cpu_init() 490 policy->cpu); in hwp_set_misc_turbo() 504 policy->cpu, msr); in hwp_set_misc_turbo() 513 return per_cpu(hwp_drv_data, cpu)->ret; in hwp_cpufreq_update() 526 int get_hwp_para(unsigned int cpu, in get_hwp_para() argument [all …]
|
/xen/xen/include/acpi/cpufreq/ |
A D | processor_perf.h | 12 unsigned int get_measured_perf(unsigned int cpu, unsigned int flag); 13 void cpufreq_residency_update(unsigned int cpu, uint8_t state); 14 void cpufreq_statistic_update(unsigned int cpu, uint8_t from, uint8_t to); 15 int cpufreq_statistic_init(unsigned int cpu); 16 void cpufreq_statistic_exit(unsigned int cpu); 17 void cpufreq_statistic_reset(unsigned int cpu); 19 int cpufreq_limit_change(unsigned int cpu); 21 int cpufreq_add_cpu(unsigned int cpu); 22 int cpufreq_del_cpu(unsigned int cpu); 62 int cpufreq_cpu_init(unsigned int cpu);
|
/xen/ |
A D | config.sub | 939 cpu=i386 1017 cpu=bfin 1037 cpu=mt 1040 cpu=m68k 1093 cpu=arm 1105 cpu=`echo "$cpu" | sed 's/^xscale/arm/'` 1125 cpu=neo 1129 cpu=nse 1133 cpu=nsr 1137 cpu=nsv [all …]
|