Searched refs:cpu (Results 1 – 9 of 9) sorted by relevance
| /samples/bpf/ |
| A D | map_perf_test_user.c | 82 static void test_hash_prealloc(int cpu) in test_hash_prealloc() argument 140 assert(cpu < MAX_NR_CPUS); in do_test_lru() 146 inner_lru_map_fds[cpu] = in do_test_lru() 152 if (inner_lru_map_fds[cpu] == -1) { in do_test_lru() 159 &inner_lru_map_fds[cpu], in do_test_lru() 199 cpu, test_name, in do_test_lru() 220 do_test_lru(LRU_HASH_LOOKUP, cpu); in test_lru_hash_lookup() 300 typedef void (*test_func)(int cpu); 331 static void loop(int cpu) in loop() argument 337 CPU_SET(cpu, &cpuset); in loop() [all …]
|
| A D | cpustat_kern.c | 107 u32 key, cpu, pstate_idx; in bpf_prog1() local 113 cpu = ctx->cpu_id; in bpf_prog1() 115 key = cpu * MAP_OFF_NUM + MAP_OFF_CSTATE_TIME; in bpf_prog1() 120 key = cpu * MAP_OFF_NUM + MAP_OFF_CSTATE_IDX; in bpf_prog1() 125 key = cpu * MAP_OFF_NUM + MAP_OFF_PSTATE_TIME; in bpf_prog1() 130 key = cpu * MAP_OFF_NUM + MAP_OFF_PSTATE_IDX; in bpf_prog1() 175 key = cpu * MAX_PSTATE_ENTRIES + pstate_idx; in bpf_prog1() 198 key = cpu * MAX_CSTATE_ENTRIES + prev_state; in bpf_prog1() 215 u32 key, cpu, pstate_idx; in bpf_prog2() local 218 cpu = ctx->cpu_id; in bpf_prog2() [all …]
|
| A D | tracex6_user.c | 25 static void check_on_cpu(int cpu, struct perf_event_attr *attr) in check_on_cpu() argument 34 CPU_SET(cpu, &set); in check_on_cpu() 39 fprintf(stderr, "sys_perf_event_open failed on CPU %d\n", cpu); in check_on_cpu() 46 bpf_map_get_next_key(map_fd[1], &cpu, NULL); in check_on_cpu() 48 if (bpf_map_lookup_elem(map_fd[1], &cpu, &value)) { in check_on_cpu() 49 fprintf(stderr, "Value missing for CPU %d\n", cpu); in check_on_cpu() 53 fprintf(stderr, "CPU %d: %llu\n", cpu, value); in check_on_cpu() 56 if (bpf_map_lookup_elem(map_fd[2], &cpu, &value2)) { in check_on_cpu() 57 fprintf(stderr, "Value2 missing for CPU %d\n", cpu); in check_on_cpu() 66 assert(bpf_map_delete_elem(map_fd[0], &cpu) == 0 || error); in check_on_cpu() [all …]
|
| A D | lathist_kern.c | 31 int cpu = bpf_get_smp_processor_id(); in bpf_prog1() local 32 u64 *ts = bpf_map_lookup_elem(&my_map, &cpu); in bpf_prog1() 75 int key, cpu; in bpf_prog2() local 78 cpu = bpf_get_smp_processor_id(); in bpf_prog2() 79 ts = bpf_map_lookup_elem(&my_map, &cpu); in bpf_prog2() 89 key = cpu * MAX_ENTRIES + delta; in bpf_prog2()
|
| A D | xdp_sample.bpf.c | 68 u32 cpu = bpf_get_smp_processor_id(); in xdp_redirect_collect_stat() local 78 idx = key * nr_cpus + cpu; in xdp_redirect_collect_stat() 131 u32 cpu = bpf_get_smp_processor_id(); in BPF_PROG() local 138 idx = to_cpu * nr_cpus + cpu; in BPF_PROG() 160 u32 cpu; in BPF_PROG() local 165 cpu = bpf_get_smp_processor_id(); in BPF_PROG() 184 u32 cpu = bpf_get_smp_processor_id(); in BPF_PROG() local 196 idx = key * nr_cpus + cpu; in BPF_PROG() 211 u32 cpu; in BPF_PROG() local 221 cpu = bpf_get_smp_processor_id(); in BPF_PROG() [all …]
|
| A D | xdp_sample_user.c | 93 struct datarec *cpu; member 335 if (!e->val.cpu) in map_entry_init() 461 if (!rec->rx_cnt.cpu) { in alloc_stats_record() 483 if (!rec->kthread.cpu) { in alloc_stats_record() 516 if (!rec->enq[i].cpu) { in alloc_stats_record() 536 free(rec->kthread.cpu); in alloc_stats_record() 541 free(rec->rx_cnt.cpu); in alloc_stats_record() 554 free(r->enq[i].cpu); in free_stats_record() 557 free(e->val.cpu); in free_stats_record() 563 free(r->kthread.cpu); in free_stats_record() [all …]
|
| A D | trace_event_kern.c | 44 u32 cpu = bpf_get_smp_processor_id(); in bpf_prog1() local 57 bpf_trace_printk(fmt, sizeof(fmt), cpu, ctx->sample_period, in bpf_prog1()
|
| A D | trace_output_user.c | 22 static void print_bpf_output(void *ctx, int cpu, void *data, __u32 size) in print_bpf_output() argument
|
| A D | map_perf_test.bpf.c | 216 int cpu = bpf_get_smp_processor_id(); in BPF_KSYSCALL() local 219 &cpu); in BPF_KSYSCALL()
|
Completed in 19 milliseconds