| /tools/perf/util/ |
| A D | arm64-frame-pointer-unwind-support.c | 13 struct entries { struct 31 struct entries *entries = arg; in add_entry() argument 33 entries->stack[entries->length++] = entry->ip; in add_entry() 40 struct entries entries = {}; in get_leaf_frame_caller_aarch64() local 63 ret = unwind__get_entries(add_entry, &entries, thread, sample, 2, true); in get_leaf_frame_caller_aarch64() 66 if (ret || entries.length != 2) in get_leaf_frame_caller_aarch64() 69 return callchain_param.order == ORDER_CALLER ? entries.stack[0] : entries.stack[1]; in get_leaf_frame_caller_aarch64()
|
| A D | mem2node.c | 50 struct phys_entry *entries, *tmp_entries; in mem2node__init() local 62 entries = zalloc(sizeof(*entries) * max); in mem2node__init() 63 if (!entries) in mem2node__init() 84 struct phys_entry *prev = &entries[j - 1]; in mem2node__init() 93 phys_entry__init(&entries[j++], start, bsize, n->node); in mem2node__init() 98 tmp_entries = realloc(entries, sizeof(*entries) * j); in mem2node__init() 101 entries = tmp_entries; in mem2node__init() 105 entries[i].node, entries[i].start, entries[i].end); in mem2node__init() 107 phys_entry__insert(&entries[i], &map->root); in mem2node__init() 110 map->entries = entries; in mem2node__init() [all …]
|
| A D | pstack.c | 18 void *entries[]; member 45 if (pstack->entries[i] == key) { in pstack__remove() 47 memmove(pstack->entries + i, in pstack__remove() 48 pstack->entries + i + 1, in pstack__remove() 63 pstack->entries[pstack->top++] = key; in pstack__push() 70 return pstack->entries[pstack->top - 1]; in pstack__peek()
|
| A D | rblist.c | 15 struct rb_node **p = &rblist->entries.rb_root.rb_node; in rblist__add_node() 40 rb_insert_color_cached(new_node, &rblist->entries, leftmost); in rblist__add_node() 48 rb_erase_cached(rb_node, &rblist->entries); in rblist__remove_node() 57 struct rb_node **p = &rblist->entries.rb_root.rb_node; in __rblist__findnew() 82 &rblist->entries, leftmost); in __rblist__findnew() 103 rblist->entries = RB_ROOT_CACHED; in rblist__init() 112 struct rb_node *pos, *next = rb_first_cached(&rblist->entries); in rblist__exit() 133 for (node = rb_first_cached(&rblist->entries); node; in rblist__entry()
|
| /tools/lib/api/fd/ |
| A D | array.c | 15 fda->entries = NULL; in fdarray__init() 27 struct pollfd *entries = realloc(fda->entries, size); in fdarray__grow() local 29 if (entries == NULL) in fdarray__grow() 34 free(entries); in fdarray__grow() 42 fda->entries = entries; in fdarray__grow() 65 free(fda->entries); in fdarray__exit() 84 fda->entries[fda->nr].fd = fd; in fdarray__add() 85 fda->entries[fda->nr].events = revents; in fdarray__add() 99 entry = &from->entries[pos]; in fdarray__dup_entry_from() 118 if (!fda->entries[fd].events) in fdarray__filter() [all …]
|
| /tools/perf/trace/beauty/ |
| A D | ioctl.c | 41 if (nr < strarray__ioctl_tty_cmd.nr_entries && strarray__ioctl_tty_cmd.entries[nr] != NULL) in ioctl__scnprintf_tty_cmd() 42 return scnprintf(bf, size, "%s", strarray__ioctl_tty_cmd.entries[nr]); in ioctl__scnprintf_tty_cmd() 52 if (nr < strarray__drm_ioctl_cmds.nr_entries && strarray__drm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_drm_cmd() 53 return scnprintf(bf, size, "DRM_%s", strarray__drm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_drm_cmd() 64 return scnprintf(bf, size, "SNDRV_PCM_%s", strarray__sndrv_pcm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_sndrv_pcm_cmd() 75 return scnprintf(bf, size, "SNDRV_CTL_%s", strarray__sndrv_ctl_ioctl_cmds.entries[nr]); in ioctl__scnprintf_sndrv_ctl_cmd() 86 return scnprintf(bf, size, "KVM_%s", strarray__kvm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_kvm_cmd() 98 if (nr < s->nr_entries && s->entries[nr] != NULL) in ioctl__scnprintf_vhost_virtio_cmd() 99 return scnprintf(bf, size, "VHOST_%s", s->entries[nr]); in ioctl__scnprintf_vhost_virtio_cmd() 110 return scnprintf(bf, size, "PERF_%s", strarray__perf_ioctl_cmds.entries[nr]); in ioctl__scnprintf_perf_cmd() [all …]
|
| A D | pkey_alloc.c | 17 const char *s = sa->entries[0]; in strarray__scnprintf_flags() 32 if (sa->entries[i] != NULL) in strarray__scnprintf_flags() 33 … += scnprintf(bf + printed, size - printed, "%s%s", show_prefix ? sa->prefix : "", sa->entries[i]); in strarray__scnprintf_flags()
|
| /tools/testing/selftests/bpf/progs/ |
| A D | get_branch_snapshot.c | 16 struct perf_branch_entry entries[ENTRY_CNT] = {}; variable 28 total_entries = bpf_get_branch_snapshot(entries, sizeof(entries), 0); in BPF_PROG() 34 if (gbs_in_range(entries[i].from) && gbs_in_range(entries[i].to)) in BPF_PROG()
|
| A D | bpf_iter_task_stack.c | 9 unsigned long entries[MAX_STACK_TRACE_DEPTH] = {}; variable 22 retlen = bpf_get_task_stack(task, entries, in dump_task_stack() 31 BPF_SEQ_PRINTF(seq, "[<0>] %pB\n", (void *)entries[i]); in dump_task_stack() 51 res = bpf_get_task_stack(task, entries, in get_task_user_stacks() 66 bpf_seq_write(seq, &entries, buf_sz); in get_task_user_stacks()
|
| A D | test_perf_branches.c | 24 __u64 entries[4 * 3] = {0}; in perf_branches() local 28 written_stack = bpf_read_branch_records(ctx, entries, sizeof(entries), 0); in perf_branches()
|
| /tools/perf/tests/ |
| A D | fdarray.c | 14 fda->entries[fd].fd = fda->nr - fd; in fdarray__init_revents() 15 fda->entries[fd].events = revents; in fdarray__init_revents() 16 fda->entries[fd].revents = revents; in fdarray__init_revents() 58 fda->entries[2].revents = POLLIN; in test__fdarray__filter() 70 fda->entries[0].revents = POLLIN; in test__fdarray__filter() 71 fda->entries[3].revents = POLLIN; in test__fdarray__filter() 103 if (fda->entries[_idx].fd != _fd) { \ in test__fdarray__add() 105 __LINE__, _idx, fda->entries[1].fd, _fd); \ in test__fdarray__add() 108 if (fda->entries[_idx].events != (_revents)) { \ in test__fdarray__add() 110 __LINE__, _idx, fda->entries[_idx].fd, _revents); \ in test__fdarray__add() [all …]
|
| /tools/build/feature/ |
| A D | test-backtrace.c | 8 int entries; in main() local 10 entries = backtrace(backtrace_fns, 10); in main() 11 backtrace_symbols_fd(backtrace_fns, entries, 1); in main()
|
| /tools/lib/perf/include/internal/ |
| A D | evlist.h | 18 struct list_head entries; member 79 __perf_evlist__for_each_entry(&(evlist)->entries, evsel) 95 __perf_evlist__for_each_entry_reverse(&(evlist)->entries, evsel) 113 __perf_evlist__for_each_entry_safe(&(evlist)->entries, tmp, evsel) 117 return list_entry(evlist->entries.next, struct perf_evsel, node); in perf_evlist__first() 122 return list_entry(evlist->entries.prev, struct perf_evsel, node); in perf_evlist__last()
|
| /tools/testing/selftests/kvm/lib/x86/ |
| A D | hyperv.c | 46 if (cpuid_sys->entries[i].function >= 0x40000000 && in vcpu_set_hv_cpuid() 47 cpuid_sys->entries[i].function < 0x40000100) in vcpu_set_hv_cpuid() 49 cpuid_full->entries[nent] = cpuid_sys->entries[i]; in vcpu_set_hv_cpuid() 53 memcpy(&cpuid_full->entries[nent], cpuid_hv->entries, in vcpu_set_hv_cpuid()
|
| /tools/testing/selftests/bpf/ |
| A D | xsk.h | 108 __u32 entries = r->cached_prod - r->cached_cons; in xsk_cons_nb_avail() local 110 if (entries == 0) { in xsk_cons_nb_avail() 112 entries = r->cached_prod - r->cached_cons; in xsk_cons_nb_avail() 115 return (entries > nb) ? nb : entries; in xsk_cons_nb_avail() 144 __u32 entries = xsk_cons_nb_avail(cons, nb); in xsk_ring_cons__peek() local 146 if (entries > 0) { in xsk_ring_cons__peek() 148 cons->cached_cons += entries; in xsk_ring_cons__peek() 151 return entries; in xsk_ring_cons__peek()
|
| A D | test_bpftool_synctypes.py | 49 entries = set() 56 entries.add(capture.group(1)) 57 return entries 84 entries = set() 91 entries |= {capture.group(1)} 92 return entries 105 entries = set() 110 entries.update(pattern.findall(line)) 113 return entries
|
| /tools/tracing/rtla/src/ |
| A D | timerlat_bpf.c | 25 bpf->rodata->entries = params->entries; in timerlat_bpf_init() 30 if (params->entries != 0) { in timerlat_bpf_init() 35 bpf_map__set_max_entries(bpf->maps.hist_irq, params->entries); in timerlat_bpf_init() 36 bpf_map__set_max_entries(bpf->maps.hist_thread, params->entries); in timerlat_bpf_init() 37 bpf_map__set_max_entries(bpf->maps.hist_user, params->entries); in timerlat_bpf_init()
|
| A D | timerlat.bpf.c | 46 const volatile int entries = 256; variable 77 if (entries == 0) in update_main_hist() 81 if (bucket >= entries) in update_main_hist() 98 if (bucket >= entries) in update_summary()
|
| A D | osnoise_hist.c | 31 int entries; member 61 *osnoise_alloc_histogram(int nr_cpus, int entries, int bucket_size) in osnoise_alloc_histogram() argument 70 data->entries = entries; in osnoise_alloc_histogram() 101 int entries = data->entries; in osnoise_hist_update_multiple() local 118 if (bucket < entries) in osnoise_hist_update_multiple() 121 hist[entries] += count; in osnoise_hist_update_multiple() 352 for (bucket = 0; bucket < data->entries; bucket++) { in osnoise_print_stats() 405 data->hist[cpu].samples[data->entries]); in osnoise_print_stats() 495 params->entries = 256; in osnoise_hist_parse_args() 591 params->entries = get_llong_from_str(optarg); in osnoise_hist_parse_args() [all …]
|
| A D | timerlat_hist.c | 46 int entries; member 92 data->entries = entries; in timerlat_alloc_histogram() 140 int entries = data->entries; in timerlat_hist_update() local 169 if (bucket < entries) in timerlat_hist_update() 172 hist[entries]++; in timerlat_hist_update() 209 for (i = 0; i < data->entries; i++) { in timerlat_hist_bpf_pull_data() 687 data->hist[cpu].irq[data->entries]); in timerlat_print_stats() 691 data->hist[cpu].thread[data->entries]); in timerlat_print_stats() 695 data->hist[cpu].user[data->entries]); in timerlat_print_stats() 809 params->entries = 256; in timerlat_hist_parse_args() [all …]
|
| /tools/testing/selftests/bpf/prog_tests/ |
| A D | fd_htab_lookup.c | 11 unsigned int entries; member 25 for (j = 0; j < ctx->entries; j++) { in htab_lookup_fn() 70 for (j = 0; j < ctx->entries; j++) { in htab_update_fn() 100 static int setup_htab(int fd, unsigned int entries) in setup_htab() argument 104 for (i = 0; i < entries; i++) { in setup_htab() 155 ctx.entries = 8; in test_fd_htab_lookup() 157 err = setup_htab(ctx.fd, ctx.entries); in test_fd_htab_lookup()
|
| /tools/testing/selftests/kvm/x86/ |
| A D | cpuid_test.c | 33 __cpuid(guest_cpuid->entries[i].function, in test_guest_cpuids() 34 guest_cpuid->entries[i].index, in test_guest_cpuids() 37 GUEST_ASSERT_EQ(eax, guest_cpuid->entries[i].eax); in test_guest_cpuids() 38 GUEST_ASSERT_EQ(ebx, guest_cpuid->entries[i].ebx); in test_guest_cpuids() 39 GUEST_ASSERT_EQ(ecx, guest_cpuid->entries[i].ecx); in test_guest_cpuids() 40 GUEST_ASSERT_EQ(edx, guest_cpuid->entries[i].edx); in test_guest_cpuids() 95 e1 = &cpuid1->entries[i]; in compare_cpuids() 96 e2 = &cpuid2->entries[i]; in compare_cpuids() 145 int size = sizeof(*cpuid) + cpuid->nent * sizeof(cpuid->entries[0]); in vcpu_alloc_cpuid()
|
| /tools/perf/pmu-events/ |
| A D | empty-pmu-events.c | 16 const struct compact_pmu_event *entries; member 128 .entries = pmu_events__common_software, 133 .entries = pmu_events__common_tool, 168 .entries = pmu_events__test_soc_cpu_default_core, 173 .entries = pmu_events__test_soc_cpu_hisi_sccl_ddrc, 178 .entries = pmu_events__test_soc_cpu_hisi_sccl_l3c, 183 .entries = pmu_events__test_soc_cpu_uncore_cbox, 188 .entries = pmu_events__test_soc_cpu_uncore_imc, 220 .entries = pmu_metrics__test_soc_cpu_default_core, 239 .entries = pmu_events__test_soc_sys_uncore_sys_ccn_pmu, [all …]
|
| /tools/lib/perf/ |
| A D | xyarray.c | 15 xy->entries = xlen * ylen; in xyarray__new() 25 size_t n = xy->entries * xy->entry_size; in xyarray__reset()
|
| /tools/perf/ui/browsers/ |
| A D | annotate-data.c | 34 struct list_head entries; member 108 parent_list = &browser->entries; in add_child_entries() 147 list_for_each_entry(entry, &browser->entries, node) in count_visible_entries() 159 INIT_LIST_HEAD(&browser->entries); in annotated_data_browser__collect_entries() 164 browser->b.entries = &browser->entries; in annotated_data_browser__collect_entries() 173 list_for_each_entry_safe(pos, tmp, &browser->entries, node) { in annotated_data_browser__delete_entries() 220 first = list_first_entry(&browser->entries, struct browser_entry, node); in browser__prev_entry() 243 last = list_last_entry(&browser->entries, struct browser_entry, node); in browser__next_entry() 273 entry = list_first_entry(&browser->entries, typeof(*entry), node); in browser__seek() 281 entry = list_last_entry(&browser->entries, typeof(*entry), node); in browser__seek() [all …]
|