| /linux/tools/testing/selftests/bpf/progs/ |
| A D | test_stacktrace_build_id.c | 7 #ifndef PERF_MAX_STACK_DEPTH 8 #define PERF_MAX_STACK_DEPTH 127 macro 25 typedef struct bpf_stack_build_id stack_trace_t[PERF_MAX_STACK_DEPTH]; 46 * PERF_MAX_STACK_DEPTH; in oncpu()
|
| A D | test_stacktrace_map.c | 7 #ifndef PERF_MAX_STACK_DEPTH 8 #define PERF_MAX_STACK_DEPTH 127 macro 25 typedef __u64 stack_trace_t[PERF_MAX_STACK_DEPTH]; 56 __u32 max_len = PERF_MAX_STACK_DEPTH * sizeof(__u64); in oncpu()
|
| A D | perf_event_stackmap.c | 6 #ifndef PERF_MAX_STACK_DEPTH 7 #define PERF_MAX_STACK_DEPTH 127 macro 10 typedef __u64 stack_trace_t[PERF_MAX_STACK_DEPTH];
|
| A D | strobemeta.h | 19 #define PERF_MAX_STACK_DEPTH 127 macro 222 __uint(value_size, sizeof(uint64_t) * PERF_MAX_STACK_DEPTH); 229 __uint(value_size, sizeof(uint64_t) * PERF_MAX_STACK_DEPTH);
|
| /linux/samples/bpf/ |
| A D | spintest.bpf.c | 12 #ifndef PERF_MAX_STACK_DEPTH 13 #define PERF_MAX_STACK_DEPTH 127 macro 32 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
|
| A D | offwaketime_user.c | 49 __u64 ip[PERF_MAX_STACK_DEPTH] = {}; in print_stack() 57 for (i = PERF_MAX_STACK_DEPTH - 1; i >= 0; i--) in print_stack() 64 for (i = 0; i < PERF_MAX_STACK_DEPTH; i++) in print_stack()
|
| A D | offwaketime.bpf.c | 13 #ifndef PERF_MAX_STACK_DEPTH 14 #define PERF_MAX_STACK_DEPTH 127 macro 56 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
|
| A D | trace_event_user.c | 63 __u64 ip[PERF_MAX_STACK_DEPTH] = {}; in print_stack() 71 for (i = PERF_MAX_STACK_DEPTH - 1; i >= 0; i--) in print_stack() 78 for (i = PERF_MAX_STACK_DEPTH - 1; i >= 0; i--) in print_stack()
|
| A D | trace_event_kern.c | 30 __uint(value_size, PERF_MAX_STACK_DEPTH * sizeof(u64));
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| A D | stacktrace_build_id.c | 13 struct bpf_stack_build_id id_offs[PERF_MAX_STACK_DEPTH]; in test_stacktrace_build_id() 71 for (i = 0; i < PERF_MAX_STACK_DEPTH; ++i) in test_stacktrace_build_id() 95 stack_trace_len = PERF_MAX_STACK_DEPTH * in test_stacktrace_build_id()
|
| A D | stacktrace_build_id_nmi.c | 17 struct bpf_stack_build_id id_offs[PERF_MAX_STACK_DEPTH]; in test_stacktrace_build_id_nmi() 99 for (i = 0; i < PERF_MAX_STACK_DEPTH; ++i) in test_stacktrace_build_id_nmi()
|
| A D | stacktrace_map.c | 65 stack_trace_len = PERF_MAX_STACK_DEPTH * sizeof(__u64); in test_stacktrace_map()
|
| /linux/kernel/events/ |
| A D | callchain.c | 23 int sysctl_perf_event_max_stack __read_mostly = PERF_MAX_STACK_DEPTH;
|
| /linux/tools/perf/util/ |
| A D | util.c | 45 int sysctl_perf_event_max_stack = PERF_MAX_STACK_DEPTH;
|
| A D | db-export.c | 232 sample, NULL, NULL, PERF_MAX_STACK_DEPTH); in call_path_from_sample()
|
| A D | python.c | 1357 unsigned int scripting_max_stack = PERF_MAX_STACK_DEPTH;
|
| /linux/tools/include/uapi/linux/ |
| A D | perf_event.h | 1243 #define PERF_MAX_STACK_DEPTH 127 macro
|
| /linux/include/uapi/linux/ |
| A D | perf_event.h | 1243 #define PERF_MAX_STACK_DEPTH 127 macro
|
| /linux/tools/perf/ |
| A D | builtin-report.c | 1293 .max_stack = PERF_MAX_STACK_DEPTH, in cmd_report() 1360 "Default: kernel.perf_event_max_stack or " __stringify(PERF_MAX_STACK_DEPTH)), in cmd_report()
|
| A D | builtin-annotate.c | 227 ret = hist_entry_iter__add(&iter, &a, PERF_MAX_STACK_DEPTH, ann); in process_branch_callback()
|
| A D | builtin-top.c | 1536 "Default: kernel.perf_event_max_stack or " __stringify(PERF_MAX_STACK_DEPTH)), in cmd_top()
|
| A D | builtin-diff.c | 439 if (hist_entry_iter__add(&iter, &al, PERF_MAX_STACK_DEPTH, in diff__process_sample_event()
|
| A D | builtin-script.c | 95 unsigned int scripting_max_stack = PERF_MAX_STACK_DEPTH; 4040 "Default: kernel.perf_event_max_stack or " __stringify(PERF_MAX_STACK_DEPTH)), in cmd_script()
|
| A D | builtin-inject.c | 944 sample__for_each_callchain_node(thread, evsel, sample, PERF_MAX_STACK_DEPTH, in perf_event__inject_buildid()
|
| /linux/tools/testing/selftests/bpf/ |
| A D | test_progs.c | 589 char val_buf[PERF_MAX_STACK_DEPTH * in compare_map_keys()
|