Lines Matching refs:evsel

37 					  struct perf_evsel *evsel)  in __perf_evlist__propagate_maps()  argument
43 if (evsel->system_wide) { in __perf_evlist__propagate_maps()
44 perf_cpu_map__put(evsel->cpus); in __perf_evlist__propagate_maps()
45 evsel->cpus = perf_cpu_map__new(NULL); in __perf_evlist__propagate_maps()
46 } else if (!evsel->own_cpus || evlist->has_user_cpus || in __perf_evlist__propagate_maps()
47 (!evsel->requires_cpu && perf_cpu_map__empty(evlist->user_requested_cpus))) { in __perf_evlist__propagate_maps()
48 perf_cpu_map__put(evsel->cpus); in __perf_evlist__propagate_maps()
49 evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus); in __perf_evlist__propagate_maps()
50 } else if (evsel->cpus != evsel->own_cpus) { in __perf_evlist__propagate_maps()
51 perf_cpu_map__put(evsel->cpus); in __perf_evlist__propagate_maps()
52 evsel->cpus = perf_cpu_map__get(evsel->own_cpus); in __perf_evlist__propagate_maps()
55 if (evsel->system_wide) { in __perf_evlist__propagate_maps()
56 perf_thread_map__put(evsel->threads); in __perf_evlist__propagate_maps()
57 evsel->threads = perf_thread_map__new_dummy(); in __perf_evlist__propagate_maps()
59 perf_thread_map__put(evsel->threads); in __perf_evlist__propagate_maps()
60 evsel->threads = perf_thread_map__get(evlist->threads); in __perf_evlist__propagate_maps()
63 evlist->all_cpus = perf_cpu_map__merge(evlist->all_cpus, evsel->cpus); in __perf_evlist__propagate_maps()
68 struct perf_evsel *evsel; in perf_evlist__propagate_maps() local
72 perf_evlist__for_each_evsel(evlist, evsel) in perf_evlist__propagate_maps()
73 __perf_evlist__propagate_maps(evlist, evsel); in perf_evlist__propagate_maps()
77 struct perf_evsel *evsel) in perf_evlist__add() argument
79 evsel->idx = evlist->nr_entries; in perf_evlist__add()
80 list_add_tail(&evsel->node, &evlist->entries); in perf_evlist__add()
84 __perf_evlist__propagate_maps(evlist, evsel); in perf_evlist__add()
88 struct perf_evsel *evsel) in perf_evlist__remove() argument
90 list_del_init(&evsel->node); in perf_evlist__remove()
185 struct perf_evsel *evsel; in perf_evlist__open() local
188 perf_evlist__for_each_entry(evlist, evsel) { in perf_evlist__open()
189 err = perf_evsel__open(evsel, evsel->cpus, evsel->threads); in perf_evlist__open()
203 struct perf_evsel *evsel; in perf_evlist__close() local
205 perf_evlist__for_each_entry_reverse(evlist, evsel) in perf_evlist__close()
206 perf_evsel__close(evsel); in perf_evlist__close()
211 struct perf_evsel *evsel; in perf_evlist__enable() local
213 perf_evlist__for_each_entry(evlist, evsel) in perf_evlist__enable()
214 perf_evsel__enable(evsel); in perf_evlist__enable()
219 struct perf_evsel *evsel; in perf_evlist__disable() local
221 perf_evlist__for_each_entry(evlist, evsel) in perf_evlist__disable()
222 perf_evsel__disable(evsel); in perf_evlist__disable()
235 struct perf_evsel *evsel, in perf_evlist__id_hash() argument
239 struct perf_sample_id *sid = SID(evsel, cpu, thread); in perf_evlist__id_hash()
242 sid->evsel = evsel; in perf_evlist__id_hash()
256 struct perf_evsel *evsel, in perf_evlist__id_add() argument
259 perf_evlist__id_hash(evlist, evsel, cpu, thread, id); in perf_evlist__id_add()
260 evsel->id[evsel->ids++] = id; in perf_evlist__id_add()
264 struct perf_evsel *evsel, in perf_evlist__id_add_fd() argument
288 if (!(evsel->attr.read_format & PERF_FORMAT_ID) || in perf_evlist__id_add_fd()
292 if (evsel->attr.read_format & PERF_FORMAT_TOTAL_TIME_ENABLED) in perf_evlist__id_add_fd()
294 if (evsel->attr.read_format & PERF_FORMAT_TOTAL_TIME_RUNNING) in perf_evlist__id_add_fd()
300 perf_evlist__id_add(evlist, evsel, cpu, thread, id); in perf_evlist__id_add_fd()
309 struct perf_evsel *evsel; in perf_evlist__alloc_pollfd() local
311 perf_evlist__for_each_entry(evlist, evsel) { in perf_evlist__alloc_pollfd()
312 if (evsel->system_wide) in perf_evlist__alloc_pollfd()
385 static void perf_evsel__set_sid_idx(struct perf_evsel *evsel, int idx, int cpu, int thread) in perf_evsel__set_sid_idx() argument
387 struct perf_sample_id *sid = SID(evsel, cpu, thread); in perf_evsel__set_sid_idx()
390 sid->cpu = perf_cpu_map__cpu(evsel->cpus, cpu); in perf_evsel__set_sid_idx()
391 sid->tid = perf_thread_map__pid(evsel->threads, thread); in perf_evsel__set_sid_idx()
439 struct perf_evsel *evsel; in mmap_per_evsel() local
442 perf_evlist__for_each_entry(evlist, evsel) { in mmap_per_evsel()
443 bool overwrite = evsel->attr.write_backward; in mmap_per_evsel()
448 if (evsel->system_wide && thread) in mmap_per_evsel()
451 cpu = perf_cpu_map__idx(evsel->cpus, evlist_cpu); in mmap_per_evsel()
467 fd = FD(evsel, cpu, thread); in mmap_per_evsel()
488 ops->idx(evlist, evsel, mp, idx); in mmap_per_evsel()
510 flgs = evsel->system_wide ? fdarray_flag__nonfilterable : fdarray_flag__default; in mmap_per_evsel()
516 if (evsel->attr.read_format & PERF_FORMAT_ID) { in mmap_per_evsel()
517 if (perf_evlist__id_add_fd(evlist, evsel, cpu, thread, in mmap_per_evsel()
520 perf_evsel__set_sid_idx(evsel, idx, cpu, thread); in mmap_per_evsel()
622 struct perf_evsel *evsel; in perf_evlist__mmap_ops() local
631 perf_evlist__for_each_entry(evlist, evsel) { in perf_evlist__mmap_ops()
632 if ((evsel->attr.read_format & PERF_FORMAT_ID) && in perf_evlist__mmap_ops()
633 evsel->sample_id == NULL && in perf_evlist__mmap_ops()
634 perf_evsel__alloc_id(evsel, evsel->fd->max_x, evsel->fd->max_y) < 0) in perf_evlist__mmap_ops()
690 struct perf_evsel *first, *last, *evsel; in __perf_evlist__set_leader() local
697 __perf_evlist__for_each_entry(list, evsel) in __perf_evlist__set_leader()
698 evsel->leader = leader; in __perf_evlist__set_leader()