Lines Matching refs:evsel_list
103 static struct evlist *evsel_list; variable
105 .evlistp = &evsel_list,
248 evlist__reset_stats(evsel_list); in perf_stat__reset_stats()
321 int nthreads = perf_thread_map__nr(evsel_list->core.threads); in read_counter_cpu()
381 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in read_affinity_counters()
401 evlist__for_each_entry(evsel_list, counter) { in read_bpf_map_counters()
426 evlist__for_each_entry(evsel_list, counter) { in process_counters()
434 perf_stat_merge_counters(&stat_config, evsel_list); in process_counters()
435 perf_stat_process_percore(&stat_config, evsel_list); in process_counters()
445 evlist__reset_aggr_stats(evsel_list); in process_interval()
475 evlist__for_each_entry(evsel_list, evsel) { in enable_counters()
486 evlist__enable(evsel_list); in enable_counters()
501 evlist__for_each_entry(evsel_list, counter) in disable_counters()
504 evlist__disable(evsel_list); in disable_counters()
608 child_exited = !is_target_alive(&target, evsel_list->core.threads) ? 1 : 0; in dispatch_events()
614 if (!(evlist__poll(evsel_list, time_to_sleep) > 0)) { /* poll timeout or EINTR */ in dispatch_events()
619 process_evlist(evsel_list, interval); in dispatch_events()
662 evsel_list->core.threads && in stat_handle_error()
663 evsel_list->core.threads->err_thread != -1) { in stat_handle_error()
668 if (!thread_map__remove(evsel_list->core.threads, in stat_handle_error()
669 evsel_list->core.threads->err_thread)) { in stat_handle_error()
670 evsel_list->core.threads->err_thread = -1; in stat_handle_error()
711 …if (evlist__prepare_workload(evsel_list, &target, argv, is_pipe, workload_exec_failed_signal) < 0)… in __run_perf_stat()
715 child_pid = evsel_list->workload.pid; in __run_perf_stat()
718 if (!cpu_map__is_dummy(evsel_list->core.user_requested_cpus)) { in __run_perf_stat()
724 evlist__for_each_entry(evsel_list, counter) { in __run_perf_stat()
732 evlist__reset_aggr_stats(evsel_list); in __run_perf_stat()
734 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in __run_perf_stat()
762 evlist__reset_weak_group(evsel_list, counter, false); in __run_perf_stat()
790 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in __run_perf_stat()
799 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in __run_perf_stat()
825 evlist__for_each_entry(evsel_list, counter) { in __run_perf_stat()
836 evsel__store_ids(counter, evsel_list)) in __run_perf_stat()
840 if (evlist__apply_filters(evsel_list, &counter, &target)) { in __run_perf_stat()
853 err = perf_session__write_header(perf_stat.session, evsel_list, in __run_perf_stat()
860 err = perf_event__synthesize_stat_events(&stat_config, NULL, evsel_list, in __run_perf_stat()
876 evlist__start_workload(evsel_list); in __run_perf_stat()
891 if (interval || timeout || evlist__ctlfd_initialized(evsel_list)) in __run_perf_stat()
924 evlist__copy_prev_raw_counts(evsel_list); in __run_perf_stat()
925 evlist__reset_prev_raw_counts(evsel_list); in __run_perf_stat()
926 evlist__reset_aggr_stats(evsel_list); in __run_perf_stat()
946 evlist__close(evsel_list); in __run_perf_stat()
985 evlist__print_counters(evsel_list, &stat_config, &target, ts, argc, argv); in print_counters()
1475 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode()
1485 nr = perf_thread_map__nr(evsel_list->core.threads); in perf_stat_init_aggr_mode()
1504 if (!perf_cpu_map__is_any_cpu_or_is_empty(evsel_list->core.user_requested_cpus)) in perf_stat_init_aggr_mode()
1505 nr = perf_cpu_map__max(evsel_list->core.user_requested_cpus).cpu; in perf_stat_init_aggr_mode()
1788 int nr = perf_thread_map__nr(evsel_list->core.threads); in perf_stat_init_aggr_mode_file()
1806 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode_file()
1952 return metricgroup__parse_groups(evsel_list, pmu, "transaction", in add_default_attributes()
1986 return metricgroup__parse_groups(evsel_list, pmu, "smi", in add_default_attributes()
2020 if (metricgroup__parse_groups(evsel_list, in add_default_attributes()
2035 if (!evsel_list->core.nr_entries) { in add_default_attributes()
2040 if (evlist__add_default_attrs(evsel_list, default_attrs0) < 0) in add_default_attributes()
2043 if (evlist__add_default_attrs(evsel_list, frontend_attrs) < 0) in add_default_attributes()
2047 if (evlist__add_default_attrs(evsel_list, backend_attrs) < 0) in add_default_attributes()
2050 if (evlist__add_default_attrs(evsel_list, default_attrs1) < 0) in add_default_attributes()
2077 evlist__splice_list_tail(evsel_list, &metric_evlist->core.entries); in add_default_attributes()
2082 if (evlist__add_default_attrs(evsel_list, default_null_attrs) < 0) in add_default_attributes()
2092 if (evlist__add_default_attrs(evsel_list, detailed_attrs) < 0) in add_default_attributes()
2099 if (evlist__add_default_attrs(evsel_list, very_detailed_attrs) < 0) in add_default_attributes()
2106 return evlist__add_default_attrs(evsel_list, very_very_detailed_attrs); in add_default_attributes()
2154 session->evlist = evsel_list; in __cmd_record()
2223 perf_evlist__set_maps(&evsel_list->core, st->cpus, st->threads); in set_maps()
2225 if (evlist__alloc_stats(&stat_config, evsel_list, /*alloc_raw=*/true)) in set_maps()
2335 evlist__delete(evsel_list); in __cmd_report()
2336 evsel_list = session->evlist; in __cmd_report()
2365 evlist__for_each_entry(evsel_list, counter) { in setup_system_wide()
2372 if (evsel_list->core.nr_entries) in setup_system_wide()
2386 OPT_CALLBACK(0, "filter", &evsel_list, "filter", in cmd_stat()
2433 OPT_CALLBACK('G', "cgroup", &evsel_list, "name", in cmd_stat()
2486 OPT_CALLBACK('M', "metrics", &evsel_list, "metric/metric group list", in cmd_stat()
2505 OPT_CALLBACK(0, "cputype", &evsel_list, "hybrid cpu type", in cmd_stat()
2510 OPT_CALLBACK(0, "pfm-events", &evsel_list, "event", in cmd_stat()
2519 OPT_CALLBACK_OPTARG(0, "iostat", &evsel_list, &stat_config, "default", in cmd_stat()
2537 evsel_list = evlist__new(); in cmd_stat()
2538 if (evsel_list == NULL) in cmd_stat()
2717 status = iostat_prepare(evsel_list, &stat_config); in cmd_stat()
2721 iostat_list(evsel_list, &stat_config); in cmd_stat()
2724 iostat_list(evsel_list, &stat_config); in cmd_stat()
2747 int ret = metricgroup__parse_groups(evsel_list, pmu, metrics, in cmd_stat()
2774 if (evlist__expand_cgroup(evsel_list, stat_config.cgroup_list, in cmd_stat()
2782 evlist__warn_user_requested_cpus(evsel_list, target.cpu_list); in cmd_stat()
2784 if (evlist__create_maps(evsel_list, &target) < 0) { in cmd_stat()
2797 evlist__check_cpu_maps(evsel_list); in cmd_stat()
2804 thread_map__read_comms(evsel_list->core.threads); in cmd_stat()
2840 if (evlist__alloc_stats(&stat_config, evsel_list, interval)) in cmd_stat()
2867 if (evlist__initialize_ctlfd(evsel_list, stat_config.ctl_fd, stat_config.ctl_fd_ack)) in cmd_stat()
2871 evlist__first(evsel_list)->ignore_missing_thread = target.pid; in cmd_stat()
2879 evlist__reset_prev_raw_counts(evsel_list); in cmd_stat()
2890 evlist__copy_res_stats(&stat_config, evsel_list); in cmd_stat()
2894 evlist__finalize_ctlfd(evsel_list); in cmd_stat()
2926 perf_session__write_header(perf_stat.session, evsel_list, fd, true); in cmd_stat()
2929 evlist__close(evsel_list); in cmd_stat()
2934 evlist__free_stats(evsel_list); in cmd_stat()
2937 iostat_release(evsel_list); in cmd_stat()
2945 evlist__delete(evsel_list); in cmd_stat()