Lines Matching refs:evsel
22 void perf_evsel__init(struct perf_evsel *evsel, struct perf_event_attr *attr, in perf_evsel__init() argument
25 INIT_LIST_HEAD(&evsel->node); in perf_evsel__init()
26 evsel->attr = *attr; in perf_evsel__init()
27 evsel->idx = idx; in perf_evsel__init()
28 evsel->leader = evsel; in perf_evsel__init()
33 struct perf_evsel *evsel = zalloc(sizeof(*evsel)); in perf_evsel__new() local
35 if (evsel != NULL) in perf_evsel__new()
36 perf_evsel__init(evsel, attr, 0); in perf_evsel__new()
38 return evsel; in perf_evsel__new()
41 void perf_evsel__delete(struct perf_evsel *evsel) in perf_evsel__delete() argument
43 free(evsel); in perf_evsel__delete()
52 int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_fd() argument
54 evsel->fd = xyarray__new(ncpus, nthreads, sizeof(int)); in perf_evsel__alloc_fd()
56 if (evsel->fd) { in perf_evsel__alloc_fd()
61 int *fd = FD(evsel, idx, thread); in perf_evsel__alloc_fd()
69 return evsel->fd != NULL ? 0 : -ENOMEM; in perf_evsel__alloc_fd()
72 static int perf_evsel__alloc_mmap(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_mmap() argument
74 evsel->mmap = xyarray__new(ncpus, nthreads, sizeof(struct perf_mmap)); in perf_evsel__alloc_mmap()
76 return evsel->mmap != NULL ? 0 : -ENOMEM; in perf_evsel__alloc_mmap()
87 static int get_group_fd(struct perf_evsel *evsel, int cpu_map_idx, int thread, int *group_fd) in get_group_fd() argument
89 struct perf_evsel *leader = evsel->leader; in get_group_fd()
92 if (evsel == leader) { in get_group_fd()
113 int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus, in perf_evsel__open() argument
143 if (evsel->fd == NULL && in perf_evsel__open()
144 perf_evsel__alloc_fd(evsel, perf_cpu_map__nr(cpus), threads->nr) < 0) in perf_evsel__open()
151 evsel_fd = FD(evsel, idx, thread); in perf_evsel__open()
157 err = get_group_fd(evsel, idx, thread, &group_fd); in perf_evsel__open()
161 fd = sys_perf_event_open(&evsel->attr, in perf_evsel__open()
175 perf_evsel__close(evsel); in perf_evsel__open()
180 static void perf_evsel__close_fd_cpu(struct perf_evsel *evsel, int cpu_map_idx) in perf_evsel__close_fd_cpu() argument
184 for (thread = 0; thread < xyarray__max_y(evsel->fd); ++thread) { in perf_evsel__close_fd_cpu()
185 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__close_fd_cpu()
194 void perf_evsel__close_fd(struct perf_evsel *evsel) in perf_evsel__close_fd() argument
196 for (int idx = 0; idx < xyarray__max_x(evsel->fd); idx++) in perf_evsel__close_fd()
197 perf_evsel__close_fd_cpu(evsel, idx); in perf_evsel__close_fd()
200 void perf_evsel__free_fd(struct perf_evsel *evsel) in perf_evsel__free_fd() argument
202 xyarray__delete(evsel->fd); in perf_evsel__free_fd()
203 evsel->fd = NULL; in perf_evsel__free_fd()
206 void perf_evsel__close(struct perf_evsel *evsel) in perf_evsel__close() argument
208 if (evsel->fd == NULL) in perf_evsel__close()
211 perf_evsel__close_fd(evsel); in perf_evsel__close()
212 perf_evsel__free_fd(evsel); in perf_evsel__close()
215 void perf_evsel__close_cpu(struct perf_evsel *evsel, int cpu_map_idx) in perf_evsel__close_cpu() argument
217 if (evsel->fd == NULL) in perf_evsel__close_cpu()
220 perf_evsel__close_fd_cpu(evsel, cpu_map_idx); in perf_evsel__close_cpu()
223 void perf_evsel__munmap(struct perf_evsel *evsel) in perf_evsel__munmap() argument
227 if (evsel->fd == NULL || evsel->mmap == NULL) in perf_evsel__munmap()
230 for (idx = 0; idx < xyarray__max_x(evsel->fd); idx++) { in perf_evsel__munmap()
231 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__munmap()
232 int *fd = FD(evsel, idx, thread); in perf_evsel__munmap()
237 perf_mmap__munmap(MMAP(evsel, idx, thread)); in perf_evsel__munmap()
241 xyarray__delete(evsel->mmap); in perf_evsel__munmap()
242 evsel->mmap = NULL; in perf_evsel__munmap()
245 int perf_evsel__mmap(struct perf_evsel *evsel, int pages) in perf_evsel__mmap() argument
253 if (evsel->fd == NULL || evsel->mmap) in perf_evsel__mmap()
256 if (perf_evsel__alloc_mmap(evsel, xyarray__max_x(evsel->fd), xyarray__max_y(evsel->fd)) < 0) in perf_evsel__mmap()
259 for (idx = 0; idx < xyarray__max_x(evsel->fd); idx++) { in perf_evsel__mmap()
260 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__mmap()
261 int *fd = FD(evsel, idx, thread); in perf_evsel__mmap()
263 struct perf_cpu cpu = perf_cpu_map__cpu(evsel->cpus, idx); in perf_evsel__mmap()
268 map = MMAP(evsel, idx, thread); in perf_evsel__mmap()
273 perf_evsel__munmap(evsel); in perf_evsel__mmap()
282 void *perf_evsel__mmap_base(struct perf_evsel *evsel, int cpu_map_idx, int thread) in perf_evsel__mmap_base() argument
284 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__mmap_base()
286 if (fd == NULL || *fd < 0 || MMAP(evsel, cpu_map_idx, thread) == NULL) in perf_evsel__mmap_base()
289 return MMAP(evsel, cpu_map_idx, thread)->base; in perf_evsel__mmap_base()
292 int perf_evsel__read_size(struct perf_evsel *evsel) in perf_evsel__read_size() argument
294 u64 read_format = evsel->attr.read_format; in perf_evsel__read_size()
312 nr = evsel->nr_members; in perf_evsel__read_size()
321 static int perf_evsel__read_group(struct perf_evsel *evsel, int cpu_map_idx, in perf_evsel__read_group() argument
324 size_t size = perf_evsel__read_size(evsel); in perf_evsel__read_group()
325 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__read_group()
326 u64 read_format = evsel->attr.read_format; in perf_evsel__read_group()
366 static void perf_evsel__adjust_values(struct perf_evsel *evsel, u64 *buf, in perf_evsel__adjust_values() argument
369 u64 read_format = evsel->attr.read_format; in perf_evsel__adjust_values()
387 int perf_evsel__read(struct perf_evsel *evsel, int cpu_map_idx, int thread, in perf_evsel__read() argument
390 size_t size = perf_evsel__read_size(evsel); in perf_evsel__read()
391 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__read()
392 u64 read_format = evsel->attr.read_format; in perf_evsel__read()
401 return perf_evsel__read_group(evsel, cpu_map_idx, thread, count); in perf_evsel__read()
403 if (MMAP(evsel, cpu_map_idx, thread) && in perf_evsel__read()
405 !perf_mmap__read_self(MMAP(evsel, cpu_map_idx, thread), count)) in perf_evsel__read()
411 perf_evsel__adjust_values(evsel, buf.values, count); in perf_evsel__read()
415 static int perf_evsel__ioctl(struct perf_evsel *evsel, int ioc, void *arg, in perf_evsel__ioctl() argument
418 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__ioctl()
426 static int perf_evsel__run_ioctl(struct perf_evsel *evsel, in perf_evsel__run_ioctl() argument
432 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__run_ioctl()
433 int err = perf_evsel__ioctl(evsel, ioc, arg, cpu_map_idx, thread); in perf_evsel__run_ioctl()
442 int perf_evsel__enable_cpu(struct perf_evsel *evsel, int cpu_map_idx) in perf_evsel__enable_cpu() argument
444 return perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, cpu_map_idx); in perf_evsel__enable_cpu()
447 int perf_evsel__enable_thread(struct perf_evsel *evsel, int thread) in perf_evsel__enable_thread() argument
453 perf_cpu_map__for_each_cpu(cpu, idx, evsel->cpus) { in perf_evsel__enable_thread()
454 err = perf_evsel__ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, idx, thread); in perf_evsel__enable_thread()
462 int perf_evsel__enable(struct perf_evsel *evsel) in perf_evsel__enable() argument
467 for (i = 0; i < xyarray__max_x(evsel->fd) && !err; i++) in perf_evsel__enable()
468 err = perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, i); in perf_evsel__enable()
472 int perf_evsel__disable_cpu(struct perf_evsel *evsel, int cpu_map_idx) in perf_evsel__disable_cpu() argument
474 return perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_DISABLE, NULL, cpu_map_idx); in perf_evsel__disable_cpu()
477 int perf_evsel__disable(struct perf_evsel *evsel) in perf_evsel__disable() argument
482 for (i = 0; i < xyarray__max_x(evsel->fd) && !err; i++) in perf_evsel__disable()
483 err = perf_evsel__run_ioctl(evsel, PERF_EVENT_IOC_DISABLE, NULL, i); in perf_evsel__disable()
487 int perf_evsel__apply_filter(struct perf_evsel *evsel, const char *filter) in perf_evsel__apply_filter() argument
491 for (i = 0; i < perf_cpu_map__nr(evsel->cpus) && !err; i++) in perf_evsel__apply_filter()
492 err = perf_evsel__run_ioctl(evsel, in perf_evsel__apply_filter()
498 struct perf_cpu_map *perf_evsel__cpus(struct perf_evsel *evsel) in perf_evsel__cpus() argument
500 return evsel->cpus; in perf_evsel__cpus()
503 struct perf_thread_map *perf_evsel__threads(struct perf_evsel *evsel) in perf_evsel__threads() argument
505 return evsel->threads; in perf_evsel__threads()
508 struct perf_event_attr *perf_evsel__attr(struct perf_evsel *evsel) in perf_evsel__attr() argument
510 return &evsel->attr; in perf_evsel__attr()
513 int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_id() argument
518 evsel->sample_id = xyarray__new(ncpus, nthreads, sizeof(struct perf_sample_id)); in perf_evsel__alloc_id()
519 if (evsel->sample_id == NULL) in perf_evsel__alloc_id()
522 evsel->id = zalloc(ncpus * nthreads * sizeof(u64)); in perf_evsel__alloc_id()
523 if (evsel->id == NULL) { in perf_evsel__alloc_id()
524 xyarray__delete(evsel->sample_id); in perf_evsel__alloc_id()
525 evsel->sample_id = NULL; in perf_evsel__alloc_id()
532 void perf_evsel__free_id(struct perf_evsel *evsel) in perf_evsel__free_id() argument
534 xyarray__delete(evsel->sample_id); in perf_evsel__free_id()
535 evsel->sample_id = NULL; in perf_evsel__free_id()
536 zfree(&evsel->id); in perf_evsel__free_id()
537 evsel->ids = 0; in perf_evsel__free_id()