Lines Matching refs:i
243 for (int i = 2; i < XDP_REDIRECT_ERR_MAX; i++) in sample_print_help() local
244 printf("\t\t\t %s: %s\n", xdp_redirect_err_names[i], in sample_print_help()
245 xdp_redirect_err_help[i - 1]); in sample_print_help()
297 int i; in sample_usage() local
303 for (i = 0; long_options[i].name != 0; i++) { in sample_usage()
304 printf(" --%-15s", long_options[i].name); in sample_usage()
305 if (long_options[i].flag != NULL) in sample_usage()
307 *long_options[i].flag); in sample_usage()
309 printf("\t short-option: -%c", long_options[i].val); in sample_usage()
350 int i; in map_collect_percpu() local
356 for (i = 0; i < nr_cpus; i++) { in map_collect_percpu()
357 rec->cpu[i].processed = READ_ONCE(values[i].processed); in map_collect_percpu()
358 rec->cpu[i].dropped = READ_ONCE(values[i].dropped); in map_collect_percpu()
359 rec->cpu[i].issue = READ_ONCE(values[i].issue); in map_collect_percpu()
360 rec->cpu[i].xdp_pass = READ_ONCE(values[i].xdp_pass); in map_collect_percpu()
361 rec->cpu[i].xdp_drop = READ_ONCE(values[i].xdp_drop); in map_collect_percpu()
362 rec->cpu[i].xdp_redirect = READ_ONCE(values[i].xdp_redirect); in map_collect_percpu()
364 sum_processed += rec->cpu[i].processed; in map_collect_percpu()
365 sum_dropped += rec->cpu[i].dropped; in map_collect_percpu()
366 sum_issue += rec->cpu[i].issue; in map_collect_percpu()
367 sum_xdp_pass += rec->cpu[i].xdp_pass; in map_collect_percpu()
368 sum_xdp_drop += rec->cpu[i].xdp_drop; in map_collect_percpu()
369 sum_xdp_redirect += rec->cpu[i].xdp_redirect; in map_collect_percpu()
387 int i, ret; in map_collect_percpu_devmap() local
409 for (i = 0; i < count; i++) { in map_collect_percpu_devmap()
411 __u64 pair = keys[i]; in map_collect_percpu_devmap()
414 arr = &values[i * nr_cpus]; in map_collect_percpu_devmap()
451 int i; in alloc_stats_record() local
468 for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++) { in alloc_stats_record()
469 rec->redir_err[i].cpu = alloc_record_per_cpu(); in alloc_stats_record()
470 if (!rec->redir_err[i].cpu) { in alloc_stats_record()
474 xdp_redirect_err_names[i]); in alloc_stats_record()
475 while (i--) in alloc_stats_record()
476 free(rec->redir_err[i].cpu); in alloc_stats_record()
490 for (i = 0; i < XDP_ACTION_MAX; i++) { in alloc_stats_record()
491 rec->exception[i].cpu = alloc_record_per_cpu(); in alloc_stats_record()
492 if (!rec->exception[i].cpu) { in alloc_stats_record()
496 action2str(i)); in alloc_stats_record()
497 while (i--) in alloc_stats_record()
498 free(rec->exception[i].cpu); in alloc_stats_record()
514 for (i = 0; i < sample_n_cpus; i++) { in alloc_stats_record()
515 rec->enq[i].cpu = alloc_record_per_cpu(); in alloc_stats_record()
516 if (!rec->enq[i].cpu) { in alloc_stats_record()
520 i); in alloc_stats_record()
521 while (i--) in alloc_stats_record()
522 free(rec->enq[i].cpu); in alloc_stats_record()
533 for (i = 0; i < XDP_ACTION_MAX; i++) in alloc_stats_record()
534 free(rec->exception[i].cpu); in alloc_stats_record()
538 for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++) in alloc_stats_record()
539 free(rec->redir_err[i].cpu); in alloc_stats_record()
551 int i; in free_stats_record() local
553 for (i = 0; i < sample_n_cpus; i++) in free_stats_record()
554 free(r->enq[i].cpu); in free_stats_record()
555 hash_for_each_safe(r->xmit_map, i, tmp, e, node) { in free_stats_record()
561 for (i = 0; i < XDP_ACTION_MAX; i++) in free_stats_record()
562 free(r->exception[i].cpu); in free_stats_record()
564 for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++) in free_stats_record()
565 free(r->redir_err[i].cpu); in free_stats_record()
654 int i; in stats_get_rx_cnt() local
660 for (i = 0; i < nr_cpus; i++) { in stats_get_rx_cnt()
661 struct datarec *r = &rec->cpu[i]; in stats_get_rx_cnt()
662 struct datarec *p = &prev->cpu[i]; in stats_get_rx_cnt()
671 snprintf(str, sizeof(str), "cpu:%d", i); in stats_get_rx_cnt()
697 int i, to_cpu; in stats_get_cpumap_enqueue() local
723 for (i = 0; i < nr_cpus; i++) { in stats_get_cpumap_enqueue()
724 struct datarec *r = &rec->cpu[i]; in stats_get_cpumap_enqueue()
725 struct datarec *p = &prev->cpu[i]; in stats_get_cpumap_enqueue()
734 snprintf(str, sizeof(str), "cpu:%d->%d", i, to_cpu); in stats_get_cpumap_enqueue()
752 int i; in stats_get_cpumap_remote() local
767 for (i = 0; i < nr_cpus; i++) { in stats_get_cpumap_remote()
768 struct datarec *r = &rec->cpu[i]; in stats_get_cpumap_remote()
769 struct datarec *p = &prev->cpu[i]; in stats_get_cpumap_remote()
776 snprintf(str, sizeof(str), "cpu:%d", i); in stats_get_cpumap_remote()
790 int i; in stats_get_cpumap_kthread() local
804 for (i = 0; i < nr_cpus; i++) { in stats_get_cpumap_kthread()
805 struct datarec *r = &rec->cpu[i]; in stats_get_cpumap_kthread()
806 struct datarec *p = &prev->cpu[i]; in stats_get_cpumap_kthread()
815 snprintf(str, sizeof(str), "cpu:%d", i); in stats_get_cpumap_kthread()
829 int i; in stats_get_redirect_cnt() local
834 for (i = 0; i < nr_cpus; i++) { in stats_get_redirect_cnt()
835 struct datarec *r = &rec->cpu[i]; in stats_get_redirect_cnt()
836 struct datarec *p = &prev->cpu[i]; in stats_get_redirect_cnt()
843 snprintf(str, sizeof(str), "cpu:%d", i); in stats_get_redirect_cnt()
861 int rec_i, i; in stats_get_redirect_err_cnt() local
880 for (i = 0; i < nr_cpus; i++) { in stats_get_redirect_err_cnt()
881 struct datarec *r = &rec->cpu[i]; in stats_get_redirect_err_cnt()
882 struct datarec *p = &prev->cpu[i]; in stats_get_redirect_err_cnt()
889 snprintf(str, sizeof(str), "cpu:%d", i); in stats_get_redirect_err_cnt()
910 int rec_i, i; in stats_get_exception_cnt() local
925 for (i = 0; i < nr_cpus; i++) { in stats_get_exception_cnt()
926 struct datarec *r = &rec->cpu[i]; in stats_get_exception_cnt()
927 struct datarec *p = &prev->cpu[i]; in stats_get_exception_cnt()
935 snprintf(str, sizeof(str), "cpu:%d", i); in stats_get_exception_cnt()
956 int i; in stats_get_devmap_xmit() local
961 for (i = 0; i < nr_cpus; i++) { in stats_get_devmap_xmit()
962 struct datarec *r = &rec->cpu[i]; in stats_get_devmap_xmit()
963 struct datarec *p = &prev->cpu[i]; in stats_get_devmap_xmit()
973 snprintf(str, sizeof(str), "cpu:%d", i); in stats_get_devmap_xmit()
1022 int i; in stats_get_devmap_xmit_multi() local
1076 for (i = 0; i < nr_cpus; i++) { in stats_get_devmap_xmit_multi()
1077 struct datarec *rc = &r->cpu[i]; in stats_get_devmap_xmit_multi()
1081 pc = p == &beg ? &p_beg : &p->cpu[i]; in stats_get_devmap_xmit_multi()
1090 snprintf(str, sizeof(str), "cpu:%d", i); in stats_get_devmap_xmit_multi()
1199 for (int i = 0; i < MAP_DEVMAP_XMIT_MULTI; i++) { in sample_setup_maps() local
1200 sample_map[i] = maps[i]; in sample_setup_maps()
1202 switch (i) { in sample_setup_maps()
1206 sample_map_count[i] = sample_n_cpus; in sample_setup_maps()
1209 sample_map_count[i] = in sample_setup_maps()
1213 sample_map_count[i] = XDP_ACTION_MAX * sample_n_cpus; in sample_setup_maps()
1215 sample_map_count[i] = sample_n_cpus * sample_n_cpus; in sample_setup_maps()
1220 if (bpf_map__set_max_entries(sample_map[i], sample_map_count[i]) < 0) in sample_setup_maps()
1229 for (int i = 0; i < MAP_DEVMAP_XMIT_MULTI; i++) { in sample_setup_maps_mappings() local
1230 size_t size = sample_map_count[i] * sizeof(struct datarec); in sample_setup_maps_mappings()
1232 sample_mmap[i] = mmap(NULL, size, PROT_READ | PROT_WRITE, in sample_setup_maps_mappings()
1233 MAP_SHARED, bpf_map__fd(sample_map[i]), 0); in sample_setup_maps_mappings()
1234 if (sample_mmap[i] == MAP_FAILED) in sample_setup_maps_mappings()
1367 for (int i = 0; i < NUM_MAP; i++) { in sample_exit() local
1368 size = sample_map_count[i] * sizeof(**sample_mmap); in sample_exit()
1369 munmap(sample_mmap[i], size); in sample_exit()
1372 int i = sample_xdp_cnt, ifindex, xdp_flags; in sample_exit() local
1375 prog_id = sample_xdp_progs[i].prog_id; in sample_exit()
1376 ifindex = sample_xdp_progs[i].ifindex; in sample_exit()
1377 xdp_flags = sample_xdp_progs[i].flags; in sample_exit()
1388 int i; in sample_stats_collect() local
1397 for (i = 1; i < XDP_REDIRECT_ERR_MAX; i++) in sample_stats_collect()
1398 map_collect_percpu(&sample_mmap[MAP_REDIRECT_ERR][i * sample_n_cpus], in sample_stats_collect()
1399 &rec->redir_err[i]); in sample_stats_collect()
1403 for (i = 0; i < sample_n_cpus; i++) in sample_stats_collect()
1404 map_collect_percpu(&sample_mmap[MAP_CPUMAP_ENQUEUE][i * sample_n_cpus], in sample_stats_collect()
1405 &rec->enq[i]); in sample_stats_collect()
1412 for (i = 0; i < XDP_ACTION_MAX; i++) in sample_stats_collect()
1413 map_collect_percpu(&sample_mmap[MAP_EXCEPTION][i * sample_n_cpus], in sample_stats_collect()
1414 &rec->exception[i]); in sample_stats_collect()