Lines Matching refs:cycles
195 tsc_t cycles; member
276 long long cycles; member
282 unsigned long long cycles; member
2195 static inline double __cycles_percent(long long cycles, long long total) { in __cycles_percent() argument
2196 return (double)(cycles*100) / total; in __cycles_percent()
2201 return __cycles_percent(s->cycles, f->total_cycles); in __summary_percent()
2254 s->cycles += c; in update_cycles()
2257 s->interval.cycles += c; in update_cycles()
2261 e->cycles = 0; in clear_interval_cycles()
2270 avg = s->cycles / s->count; in print_cpu_affinity()
2297 avg = s->cycles / s->count; in print_cycle_percent_summary()
2299 seconds = ((double)s->cycles) / opt.cpu_hz; in print_cycle_percent_summary()
2301 percent = ((double)(s->cycles * 100)) / total; in print_cycle_percent_summary()
2333 avg = s->cycles / s->count; in print_cycle_summary()
2347 p, s->count, ((double)s->cycles)/opt.cpu_hz, in print_cycle_summary()
2351 p, s->count, ((double)s->cycles)/opt.cpu_hz, avg); in print_cycle_summary()
2370 ((double)(_s).cycles)/opt.cpu_hz, \
2372 (_s).count ? (_s).cycles / (_s).count:0, \
2378 ((double)(_s).cycles)/opt.cpu_hz, \
2380 (_s).count ? (_s).cycles / (_s).count:0); \
2392 void __interval_cycle_percent_output(struct interval_element *e, tsc_t cycles) { in __interval_cycle_percent_output() argument
2394 __cycles_percent(e->cycles, cycles)); in __interval_cycle_percent_output()
2399 __interval_cycle_percent_output(e, opt.interval.cycles); in interval_cycle_percent_output()
2524 if(a->total_time.interval.cycles < b->total_time.interval.cycles) in cr3_time_compare()
2526 else if(b->total_time.interval.cycles == a->total_time.interval.cycles) { in cr3_time_compare()
2563 if(p->total_time.interval.cycles > 0) { in interval_cr3_schedule_ordered_output()
2566 __cycles_percent(p->total_time.interval.cycles, in interval_cr3_schedule_ordered_output()
2567 opt.interval.cycles), in interval_cr3_schedule_ordered_output()
2597 __cycles_percent(p->total_time.interval.cycles, in interval_cr3_short_summary_output()
2598 opt.interval.cycles)); in interval_cr3_short_summary_output()
2602 p->total_time.interval.cycles); in interval_cr3_short_summary_output()
2806 unsigned long long cycles, int type, void * extra) { in update_eip() argument
2836 update_cycles(&p->summary, cycles); in update_eip()
2843 if(a->summary.cycles < b->summary.cycles) in eip_compare()
2845 else if(b->summary.cycles == a->summary.cycles) { in eip_compare()
2891 if ( p->summary.cycles ) in dump_eip()
3898 unsigned long long cycles = now - v->cr3.start_time; in cr3_switch() local
3901 update_cycles(&cur->total_time, cycles); in cr3_switch()
9129 while ( tsc - P.interval.start_tsc > opt.interval.cycles ) { in process_record_tsc()
9131 P.interval.start_tsc += opt.interval.cycles; in process_record_tsc()
9291 tsc_t cycles; in process_cpu_change() local
9295 cycles = (p->volume.buffer_first_tsc > p->volume.buffer_dom0_runstate_tsc) ? in process_cpu_change()
9299 += ri->tsc - cycles; in process_cpu_change()
9304 ri->tsc - cycles); in process_cpu_change()
9307 cycles = ri->tsc - p->volume.buffer_first_tsc; in process_cpu_change()
9308 cycles_to_time(cycles, &t); in process_cpu_change()
9310 t.s, t.ns, cycles); in process_cpu_change()
9313 cycles / p->volume.buffer_size); in process_cpu_change()
9316 cycles = ri->tsc - P.buffer_trace_virq_tsc; in process_cpu_change()
9317 cycles_to_time(cycles, &t); in process_cpu_change()
9319 t.s, t.ns, cycles); in process_cpu_change()
10541 opt.interval.cycles = ( opt.interval.msec * opt.cpu_hz ) / 1000 ; in cmd_parser()