Lines Matching refs:pc
95 struct perf_event_mmap_page *pc = map->base; in perf_mmap__empty() local
97 return perf_mmap__read_head(map) == map->prev && !pc->aux_size; in perf_mmap__empty()
482 struct perf_event_mmap_page *pc = map->base; in perf_mmap__read_self() local
486 if (!pc || !pc->cap_user_rdpmc) in perf_mmap__read_self()
490 seq = READ_ONCE(pc->lock); in perf_mmap__read_self()
493 count->ena = READ_ONCE(pc->time_enabled); in perf_mmap__read_self()
494 count->run = READ_ONCE(pc->time_running); in perf_mmap__read_self()
496 if (pc->cap_user_time && count->ena != count->run) { in perf_mmap__read_self()
498 time_mult = READ_ONCE(pc->time_mult); in perf_mmap__read_self()
499 time_shift = READ_ONCE(pc->time_shift); in perf_mmap__read_self()
500 time_offset = READ_ONCE(pc->time_offset); in perf_mmap__read_self()
502 if (pc->cap_user_time_short) { in perf_mmap__read_self()
503 time_cycles = READ_ONCE(pc->time_cycles); in perf_mmap__read_self()
504 time_mask = READ_ONCE(pc->time_mask); in perf_mmap__read_self()
508 idx = READ_ONCE(pc->index); in perf_mmap__read_self()
509 cnt = READ_ONCE(pc->offset); in perf_mmap__read_self()
510 if (pc->cap_user_rdpmc && idx) { in perf_mmap__read_self()
512 u16 width = READ_ONCE(pc->pmc_width); in perf_mmap__read_self()
521 } while (READ_ONCE(pc->lock) != seq); in perf_mmap__read_self()