| /tools/perf/util/ |
| A D | bpf-trace-summary.c | 190 nodes = reallocarray(data->nodes, data->nr_nodes + 1, sizeof(*nodes)); in update_thread_stats() 191 if (nodes == NULL) in update_thread_stats() 194 data->nodes = nodes; in update_thread_stats() 195 nodes = &data->nodes[data->nr_nodes++]; in update_thread_stats() 207 qsort(data->nodes, data->nr_nodes, sizeof(*data->nodes), nodecmp); in print_thread_stat() 243 data->nodes = zalloc(sizeof(*data->nodes)); in update_total_stats() 323 nodes = reallocarray(data->nodes, data->nr_nodes + 1, sizeof(*nodes)); in update_cgroup_stats() 324 if (nodes == NULL) in update_cgroup_stats() 327 data->nodes = nodes; in update_cgroup_stats() 328 nodes = &data->nodes[data->nr_nodes++]; in update_cgroup_stats() [all …]
|
| A D | cputopo.h | 46 struct numa_topology_node nodes[]; member 56 struct hybrid_topology_node nodes[]; member
|
| A D | cputopo.c | 406 tp = zalloc(sizeof(*tp) + sizeof(tp->nodes[0])*nr); in numa_topology__new() 413 if (load_numa_node(&tp->nodes[i], perf_cpu_map__cpu(node_map, i).cpu)) { in numa_topology__new() 432 zfree(&tp->nodes[i].cpus); in numa_topology__delete() 480 tp = zalloc(sizeof(*tp) + sizeof(tp->nodes[0]) * nr); in hybrid_topology__new() 486 if (load_hybrid_node(&tp->nodes[i], pmu)) { in hybrid_topology__new() 501 zfree(&tp->nodes[i].pmu_name); in hybrid_topology__delete() 502 zfree(&tp->nodes[i].cpus); in hybrid_topology__delete()
|
| A D | mem2node.c | 49 struct memory_node *n, *nodes = &env->memory_nodes[0]; in mem2node__init() local 58 n = &nodes[i]; in mem2node__init() 69 n = &nodes[i]; in mem2node__init()
|
| A D | header.c | 1416 reallocarray(nodes, cnt + 4, sizeof(*nodes)); in build_mem_topology() 1435 qsort(nodes, cnt, sizeof(nodes[0]), memory_node__sort); in build_mem_topology() 2704 nodes = zalloc(sizeof(*nodes) * nr); in process_numa_topology() 2705 if (!nodes) in process_numa_topology() 2735 free(nodes); in process_numa_topology() 2984 nodes = zalloc(sizeof(*nodes) * nr); in process_mem_topology() 2985 if (!nodes) in process_mem_topology() 3013 free(nodes); in process_mem_topology() 3074 nodes = zalloc(sizeof(*nodes) * nr); in process_hybrid_topology() 3075 if (!nodes) in process_hybrid_topology() [all …]
|
| /tools/perf/tests/ |
| A D | mem2node.c | 50 struct memory_node nodes[3]; in test__mem2node() local 52 .memory_nodes = (struct memory_node *) &nodes[0], in test__mem2node() 53 .nr_memory_nodes = ARRAY_SIZE(nodes), in test__mem2node() 58 for (i = 0; i < ARRAY_SIZE(nodes); i++) { in test__mem2node() 59 nodes[i].node = test_nodes[i].node; in test__mem2node() 60 nodes[i].size = 10; in test__mem2node() 63 (nodes[i].set = get_bitmap(test_nodes[i].map, 10))); in test__mem2node() 75 for (i = 0; i < ARRAY_SIZE(nodes); i++) in test__mem2node() 76 zfree(&nodes[i].set); in test__mem2node()
|
| /tools/testing/selftests/bpf/progs/ |
| A D | rbtree.c | 134 struct node_data *nodes[3][2] = {{NULL, NULL}, {NULL, NULL}, {NULL, NULL}}; in rbtree_add_and_remove_array() local 141 nodes[i][j] = bpf_obj_new(typeof(*nodes[i][j])); in rbtree_add_and_remove_array() 142 if (!nodes[i][j]) in rbtree_add_and_remove_array() 144 nodes[i][j]->key = i * 2 + j; in rbtree_add_and_remove_array() 151 bpf_rbtree_add(&groot_array[i], &nodes[i][j]->node, less); in rbtree_add_and_remove_array() 153 bpf_rbtree_add(&groot_array_one[0], &nodes[2][j]->node, less); in rbtree_add_and_remove_array() 154 res1 = bpf_rbtree_remove(&groot_array[0], &nodes[0][0]->node); in rbtree_add_and_remove_array() 155 res2 = bpf_rbtree_remove(&groot_array[1], &nodes[1][0]->node); in rbtree_add_and_remove_array() 156 res3 = bpf_rbtree_remove(&groot_array_one[0], &nodes[2][0]->node); in rbtree_add_and_remove_array() 182 if (nodes[i][j]) in rbtree_add_and_remove_array() [all …]
|
| /tools/testing/selftests/kvm/lib/ |
| A D | lru_gen_util.c | 117 stats->nodes[ctx->current_node_idx].node = found_node_id; in memcg_stats_handle_in_memcg() 149 node_stats = &stats->nodes[ctx->current_node_idx]; in memcg_stats_handle_in_node() 187 pr_debug("\tnode %d\n", stats->nodes[node].node); in print_memcg_stats() 188 for (gen = 0; gen < stats->nodes[node].nr_gens; ++gen) { in print_memcg_stats() 190 &stats->nodes[node].gens[gen]; in print_memcg_stats() 252 const struct node_stats *node_stats = &stats->nodes[node]; in lru_gen_sum_memcg_stats_for_gen() 314 for (gen = 0; gen < stats->nodes[node].nr_gens; ++gen) { in lru_gen_do_aging() 315 int this_gen = stats->nodes[node].gens[gen].gen; in lru_gen_do_aging() 320 run_aging_impl(stats->memcg_id, stats->nodes[node].node, in lru_gen_do_aging() 338 for (gen_idx = 0; gen_idx < stats->nodes[node].nr_gens; in lru_gen_find_generation() [all …]
|
| /tools/testing/selftests/net/ |
| A D | reuseport_bpf_numa.c | 235 int *rcv_fd, nodes; in main() local 240 nodes = numa_max_node() + 1; in main() 242 rcv_fd = calloc(nodes, sizeof(int)); in main() 247 test(rcv_fd, nodes, AF_INET, SOCK_DGRAM); in main() 250 test(rcv_fd, nodes, AF_INET6, SOCK_DGRAM); in main() 253 test(rcv_fd, nodes, AF_INET, SOCK_STREAM); in main() 256 test(rcv_fd, nodes, AF_INET6, SOCK_STREAM); in main()
|
| /tools/testing/selftests/kvm/x86/ |
| A D | xapic_ipi_test.c | 256 int nodes = 0; in do_migrations() local 284 nodemasks[nodes] = nodemask & bit; in do_migrations() 285 nodes++; in do_migrations() 289 TEST_ASSERT(nodes > 1, in do_migrations() 292 fprintf(stderr, "Migrating amongst %d nodes found\n", nodes); in do_migrations() 329 if (to == nodes) in do_migrations()
|
| /tools/perf/bench/ |
| A D | numa.c | 980 int nodes; in count_process_nodes() local 985 for (nodes = 0; nodes < g->p.nr_nodes; nodes++) in count_process_nodes() 1005 nodes = 0; in count_process_nodes() 1011 return nodes; in count_process_nodes() 1057 if (!nodes) { in calc_convergence_compression() 1079 int *nodes; in calc_convergence() local 1094 BUG_ON(!nodes); in calc_convergence() 1096 nodes[node] = 0; in calc_convergence() 1113 nodes[node]++; in calc_convergence() 1137 free(nodes); in calc_convergence() [all …]
|
| /tools/testing/selftests/kvm/include/ |
| A D | lru_gen_util.h | 40 struct node_stats nodes[MAX_NR_NODES]; member
|
| /tools/testing/selftests/bpf/ |
| A D | verify_sig_setup.sh | 41 openssl req -new -nodes -utf8 -sha256 -days 36500 \
|
| /tools/testing/selftests/zram/ |
| A D | README | 11 Statistics for individual zram devices are exported through sysfs nodes at
|
| /tools/testing/selftests/devices/probe/boards/ |
| A D | google,spherion.yaml | 17 # allows the matching to include parent nodes as well to make it unique.
|
| /tools/testing/selftests/mm/ |
| A D | run_vmtests.sh | 57 ksm tests that do not require >=2 NUMA nodes 59 ksm tests that require >=2 NUMA nodes
|
| /tools/perf/ |
| A D | builtin-c2c.c | 92 unsigned long **nodes; member 1189 bitmap_and(set, c2c_he->cpuset, c2c.nodes[node], c2c.cpus_cnt); in node_entry() 2273 unsigned long **nodes; in setup_nodes() local 2289 nodes = zalloc(sizeof(unsigned long *) * c2c.nodes_cnt); in setup_nodes() 2290 if (!nodes) in setup_nodes() 2293 c2c.nodes = nodes; in setup_nodes() 2312 nodes[node] = set; in setup_nodes()
|
| A D | builtin-record.c | 3925 spec[s] = topo->nodes[s].cpus; in record__init_thread_numa_masks()
|
| /tools/perf/Documentation/ |
| A D | perf.data-file-format.txt | 205 } nodes[nr]; /* Variable length records */ 303 u64 count; // number of nodes 321 # memory nodes (nr 1, block size 0x8000000):
|
| A D | perf-c2c.txt | 285 - nodes participating on the access (see NODE INFO section) 289 The 'Node' field displays nodes that accesses given cacheline
|
| A D | perf-stat.txt | 338 Aggregate counts per NUMA nodes for system-wide mode measurements. This 339 is a useful mode to detect imbalance between NUMA nodes. To enable this
|
| /tools/arch/x86/kcpuid/ |
| A D | cpuid.csv | 952 0x8000001e, 0, ecx, 10:8, nnodes_per_socket , #nodes in invoking logical CPU's …
|
| /tools/testing/radix-tree/ |
| A D | maple.c | 97 struct maple_node *nodes[100]; in check_new_node() local 260 nodes[j] = mas_pop_node(&mas); in check_new_node() 266 mas_push_node(&mas, nodes[j]); in check_new_node()
|