/xen-4.10.0-shim-comet/xen/arch/arm/ |
A D | processor.c | 21 static const struct processor *processor = NULL; variable 31 processor = procinfo->processor; in processor_setup() 36 if ( !processor || !processor->vcpu_initialise ) in processor_vcpu_initialise() 39 processor->vcpu_initialise(v); in processor_vcpu_initialise()
|
A D | gic-v2.c | 1063 struct acpi_madt_generic_interrupt *processor = in gic_acpi_parse_madt_cpu() local 1066 if ( BAD_MADT_ENTRY(processor, end) ) in gic_acpi_parse_madt_cpu() 1072 cbase = processor->base_address; in gic_acpi_parse_madt_cpu() 1074 hbase = processor->gich_base_address; in gic_acpi_parse_madt_cpu() 1075 vbase = processor->gicv_base_address; in gic_acpi_parse_madt_cpu() 1076 gicv2_info.maintenance_irq = processor->vgic_interrupt; in gic_acpi_parse_madt_cpu() 1078 if ( processor->flags & ACPI_MADT_VGIC_IRQ_MODE ) in gic_acpi_parse_madt_cpu() 1087 if ( cbase != processor->base_address in gic_acpi_parse_madt_cpu() 1088 || hbase != processor->gich_base_address in gic_acpi_parse_madt_cpu() 1089 || vbase != processor->gicv_base_address in gic_acpi_parse_madt_cpu() [all …]
|
A D | vtimer.c | 110 init_timer(&t->timer, phys_timer_expired, t, v->processor); in vcpu_vtimer_init() 119 init_timer(&t->timer, virt_timer_expired, t, v->processor); in vcpu_vtimer_init() 161 migrate_timer(&v->arch.virt_timer.timer, v->processor); in virt_timer_restore() 162 migrate_timer(&v->arch.phys_timer.timer, v->processor); in virt_timer_restore()
|
A D | gic-v3.c | 1448 struct acpi_madt_generic_interrupt *processor = in gic_acpi_parse_madt_cpu() local 1451 if ( BAD_MADT_ENTRY(processor, end) ) in gic_acpi_parse_madt_cpu() 1457 cbase = processor->base_address; in gic_acpi_parse_madt_cpu() 1458 vbase = processor->gicv_base_address; in gic_acpi_parse_madt_cpu() 1459 gicv3_info.maintenance_irq = processor->vgic_interrupt; in gic_acpi_parse_madt_cpu() 1461 if ( processor->flags & ACPI_MADT_VGIC_IRQ_MODE ) in gic_acpi_parse_madt_cpu() 1470 if ( cbase != processor->base_address in gic_acpi_parse_madt_cpu() 1471 || vbase != processor->gicv_base_address in gic_acpi_parse_madt_cpu() 1501 struct acpi_madt_generic_interrupt *processor; in gic_acpi_parse_cpu_redistributor() local 1504 processor = (struct acpi_madt_generic_interrupt *)header; in gic_acpi_parse_cpu_redistributor() [all …]
|
A D | Kconfig | 96 If a Cortex-A53 processor is executing a store or prefetch for 97 write instruction at the same time as a processor in another 119 If the processor is executing a load and store exclusive sequence at 120 the same time as a processor in another cluster is executing a cache
|
A D | vgic.c | 277 irq_set_affinity(p->desc, cpumask_of(new->processor)); in vgic_migrate_irq() 285 irq_set_affinity(p->desc, cpumask_of(new->processor)); in vgic_migrate_irq() 301 const cpumask_t *cpu_mask = cpumask_of(v->processor); in arch_move_irqs() 402 irq_set_affinity(p->desc, cpumask_of(v_target->processor)); in vgic_enable_irqs() 567 smp_send_event_check_mask(cpumask_of(v->processor)); in vgic_vcpu_inject_irq()
|
/xen-4.10.0-shim-comet/xen/arch/x86/acpi/ |
A D | boot.c | 84 struct acpi_madt_local_x2apic *processor = in acpi_parse_x2apic() local 88 if (BAD_MADT_ENTRY(processor, end)) in acpi_parse_x2apic() 94 if (processor->local_apic_id >= MAX_APICS || in acpi_parse_x2apic() 95 processor->uid >= MAX_MADT_ENTRIES) { in acpi_parse_x2apic() 100 processor->local_apic_id, processor->uid); in acpi_parse_x2apic() 108 x86_acpiid_to_apicid[processor->uid] = in acpi_parse_x2apic() 109 processor->local_apic_id; in acpi_parse_x2apic() 128 struct acpi_madt_local_apic *processor = in acpi_parse_lapic() local 132 if (BAD_MADT_ENTRY(processor, end)) in acpi_parse_lapic() 139 x86_acpiid_to_apicid[processor->processor_id] = processor->id; in acpi_parse_lapic() [all …]
|
/xen-4.10.0-shim-comet/xen/arch/x86/pv/ |
A D | traps.c | 162 unsigned int processor; /* physical cpu to inject trap */ member 181 if ( (cpu != st->processor) || in nmi_mce_softirq() 182 (st->processor != st->vcpu->processor) ) in nmi_mce_softirq() 189 vcpu_set_hard_affinity(st->vcpu, cpumask_of(st->processor)); in nmi_mce_softirq() 227 st->processor = v->processor; in pv_raise_interrupt() 247 st->processor = v->processor; in pv_raise_interrupt()
|
/xen-4.10.0-shim-comet/xen/arch/arm/acpi/ |
A D | boot.c | 51 acpi_map_gic_cpu_interface(struct acpi_madt_generic_interrupt *processor) in acpi_map_gic_cpu_interface() argument 55 u64 mpidr = processor->arm_mpidr & MPIDR_HWID_MASK; in acpi_map_gic_cpu_interface() 56 bool enabled = processor->flags & ACPI_MADT_ENABLED; in acpi_map_gic_cpu_interface() 130 struct acpi_madt_generic_interrupt *processor = in acpi_parse_gic_cpu_interface() local 133 if ( BAD_MADT_ENTRY(processor, end) ) in acpi_parse_gic_cpu_interface() 137 acpi_map_gic_cpu_interface(processor); in acpi_parse_gic_cpu_interface()
|
/xen-4.10.0-shim-comet/xen/include/asm-arm/ |
A D | procinfo.h | 26 struct processor { struct 35 struct processor *processor; argument
|
/xen-4.10.0-shim-comet/xen/include/xen/ |
A D | sched-if.h | 92 sched_lock(vcpu, const struct vcpu *v, v->processor, ) 94 sched_lock(vcpu, const struct vcpu *v, v->processor, _irq) in sched_lock() 96 sched_unlock(vcpu, const struct vcpu *v, v->processor, ) in sched_lock() 98 sched_unlock(vcpu, const struct vcpu *v, v->processor, _irq) in sched_lock() 104 sched_lock(vcpu, const struct vcpu *v, v->processor, _irqsave, *flags) in sched_lock() 107 sched_unlock(vcpu, const struct vcpu *v, v->processor, _irqrestore, flags) in sched_lock()
|
A D | numa.h | 15 #define vcpu_to_node(v) (cpu_to_node((v)->processor))
|
/xen-4.10.0-shim-comet/xen/common/ |
A D | sched_null.c | 295 unsigned int cpu = v->processor, new_cpu; in pick_cpu() 366 v->processor = cpu; in vcpu_assign() 450 cpu = v->processor = pick_cpu(prv, v); in null_vcpu_insert() 497 unsigned int cpu = v->processor; in _vcpu_remove() 552 ASSERT(per_cpu(npc, v->processor).vcpu == v); in null_vcpu_remove() 567 if ( unlikely(curr_on_cpu(v->processor) == v) ) in null_vcpu_wake() 586 cpu_raise_softirq(v->processor, SCHEDULE_SOFTIRQ); in null_vcpu_wake() 594 if ( curr_on_cpu(v->processor) == v ) in null_vcpu_sleep() 614 if ( v->processor == new_cpu ) in null_vcpu_migrate() 625 d.cpu = v->processor; in null_vcpu_migrate() [all …]
|
A D | schedule.c | 264 v->processor = processor; in sched_init_vcpu() 274 v, v->processor); in sched_init_vcpu() 276 v, v->processor); in sched_init_vcpu() 278 v, v->processor); in sched_init_vcpu() 373 v->processor = new_p; in sched_move_domain() 520 unsigned int old_cpu = v->processor; in vcpu_move_locked() 540 v->processor = new_cpu; in vcpu_move_locked() 563 ASSERT(new_cpu != v->processor); in vcpu_move_nosched() 577 old_cpu = new_cpu = v->processor; in vcpu_migrate() 590 old_cpu = v->processor; in vcpu_migrate() [all …]
|
A D | sched_credit2.c | 695 int cpu = v->processor; in get_fallback_cpu() 1378 if ( new->vcpu->processor != cpu ) in tickle_score() 1446 unsigned processor; in runq_tickle() member 1451 d.processor = new->vcpu->processor; in runq_tickle() 1628 svc_cpu = svc->vcpu->processor; in reset_credit() 2074 unsigned int cpu = vc->processor; in csched2_vcpu_wake() 2448 int cpu = svc->vcpu->processor; in migrate() 2517 int cpu = svc->vcpu->processor; in vcpu_is_migrateable() 2744 vc->processor = new_cpu; in csched2_vcpu_migrate() 2765 vc->processor = new_cpu; in csched2_vcpu_migrate() [all …]
|
A D | sched_arinc653.c | 501 if ( per_cpu(schedule_data, vc->processor).curr == vc ) in a653sched_vcpu_sleep() 502 cpu_raise_softirq(vc->processor, SCHEDULE_SOFTIRQ); in a653sched_vcpu_sleep() 517 cpu_raise_softirq(vc->processor, SCHEDULE_SOFTIRQ); in a653sched_vcpu_wake() 608 && (new_task->processor != cpu) ) in a653sched_do_schedule() 646 if ( cpumask_test_cpu(vc->processor, online) in a653sched_pick_cpu() 648 cpu = vc->processor; in a653sched_pick_cpu()
|
A D | sched_rt.c | 332 mask = cpumask_scratch_cpu(svc->vcpu->processor); in rt_dump_vcpu() 343 svc->vcpu->processor, in rt_dump_vcpu() 655 cpu = cpumask_test_cpu(vc->processor, &cpus) in rt_cpu_pick() 656 ? vc->processor in rt_cpu_pick() 657 : cpumask_cycle(vc->processor, &cpus); in rt_cpu_pick() 936 vc->processor = rt_cpu_pick(ops, vc); in rt_vcpu_insert() 1163 if ( snext->vcpu->processor != cpu ) in rt_schedule() 1165 snext->vcpu->processor = cpu; in rt_schedule() 1187 if ( curr_on_cpu(vc->processor) == vc ) in rt_vcpu_sleep() 1188 cpu_raise_softirq(vc->processor, SCHEDULE_SOFTIRQ); in rt_vcpu_sleep() [all …]
|
A D | sched_credit.c | 275 unsigned int cpu = svc->vcpu->processor; in __runq_insert() 307 inc_nr_runnable(svc->vcpu->processor); in runq_insert() 320 dec_nr_runnable(svc->vcpu->processor); in runq_remove() 351 unsigned int cpu = new->vcpu->processor; in __runq_tickle() 729 int cpu = vc->processor; in _csched_cpu_pick() 766 ? vc->processor in _csched_cpu_pick() 943 ASSERT( current->processor == cpu ); in csched_vcpu_acct() 1091 unsigned int cpu = vc->processor; in csched_vcpu_sleep() 1671 vc->processor = cpu; in csched_runq_steal() 1697 BUG_ON( cpu != snext->vcpu->processor ); in csched_load_balance() [all …]
|
A D | keyhandler.c | 218 tasklet_schedule_on_cpu(&dump_hwdom_tasklet, v->processor); in dump_hwdom_action() 240 tasklet_schedule_on_cpu(&dump_hwdom_tasklet, v->processor); in dump_hwdom_registers() 339 v->vcpu_id, v->processor, in dump_domains()
|
/xen-4.10.0-shim-comet/xen/arch/x86/ |
A D | mpparse.c | 481 struct mpc_config_processor processor; in construct_default_ISA_mptable() local 496 processor.mpc_type = MP_PROCESSOR; in construct_default_ISA_mptable() 499 processor.mpc_cpuflag = CPU_ENABLED; in construct_default_ISA_mptable() 500 processor.mpc_cpufeature = (boot_cpu_data.x86 << 8) | in construct_default_ISA_mptable() 503 processor.mpc_featureflag = in construct_default_ISA_mptable() 505 processor.mpc_reserved[0] = 0; in construct_default_ISA_mptable() 506 processor.mpc_reserved[1] = 0; in construct_default_ISA_mptable() 508 processor.mpc_apicid = i; in construct_default_ISA_mptable() 509 MP_processor_info(&processor); in construct_default_ISA_mptable() 786 struct mpc_config_processor processor = { in mp_register_lapic() local [all …]
|
/xen-4.10.0-shim-comet/tools/libacpi/ |
A D | build.c | 218 struct acpi_20_srat_processor *processor; in construct_srat() local 224 size = sizeof(*srat) + sizeof(*processor) * config->hvminfo->nr_vcpus + in construct_srat() 241 processor = (struct acpi_20_srat_processor *)(srat + 1); in construct_srat() 244 processor->type = ACPI_PROCESSOR_AFFINITY; in construct_srat() 245 processor->length = sizeof(*processor); in construct_srat() 246 processor->domain = config->numa.vcpu_to_vnode[i]; in construct_srat() 247 processor->apic_id = config->lapic_id(i); in construct_srat() 248 processor->flags = ACPI_LOCAL_APIC_AFFIN_ENABLED; in construct_srat() 249 processor++; in construct_srat() 252 memory = (struct acpi_20_srat_memory *)processor; in construct_srat()
|
/xen-4.10.0-shim-comet/xen/arch/arm/arm32/ |
A D | proc-caxx.c | 33 const struct processor caxx_processor = {
|
/xen-4.10.0-shim-comet/docs/misc/ |
A D | amd-ucode-container.txt | 16 processor families 10h, 11h, 12h, 14h, and 15h. 19 * For AMD processor families 15h and later, there is a separate container file 22 sure you have the latest container file for your AMD processor family.
|
/xen-4.10.0-shim-comet/docs/man/ |
A D | xen-tscmode.pod.7 | 28 per second per processor, we call this a "high-TSC-frequency" 121 processor that increases monotonically. Historically, TSC incremented 122 every processor cycle, but on recent processors, it increases 123 at a constant rate even if the processor changes frequency (for example, 124 to reduce processor power usage). TSC is known by x86 programmers 135 one processor, then was moved by the OS to another processor, then read 143 machines, the TSC was synchronized but if one processor were to enter 196 per second per processor), this performance degradation is not noticeable 223 When a pvrdtscp-modified app is running on a processor that is both TSC-safe 227 running on a TSC-unsafe processor or a processor that doesn't support
|
/xen-4.10.0-shim-comet/docs/misc/arm/ |
A D | early-printk.txt | 46 - seattle: printk with pl011 for AMD Seattle processor 49 - thunderx: printk with pl011 for Cavium ThunderX processor
|