| /tools/testing/selftests/rcutorture/bin/ |
| A D | cpus2use.sh | 17 ncpus=`grep '^processor' /proc/cpuinfo | wc -l` 21 awk -v ncpus=$ncpus '{ print ncpus * ($7 + $NF) / 100 }'` 24 idlecpus=$ncpus 26 awk -v ncpus=$ncpus -v idlecpus=$idlecpus < /dev/null '
|
| A D | kvm-build.sh | 44 ncpus="`getconf _NPROCESSORS_ONLN`" 45 make -j$((2 * ncpus)) $TORTURE_KMAKE_ARG > $resdir/Make.out 2>&1
|
| A D | kvm.sh | 360 awk < $T/cfgcpu.sort > $T/cfgcpu.pack -v ncpus=$cpus ' 468 if (ja[1] == -1 && ncpus == 0) 471 njitter = ncpus; 486 if (cpusr[jn] > ncpus && ncpus != 0) 554 nc = ncpus; 560 if (ncpus == 0) { 569 nc = ncpus; 576 if (ncpus != 0) 584 -v ncpus=$cpus \
|
| /tools/perf/util/ |
| A D | counts.c | 10 struct perf_counts *perf_counts__new(int ncpus, int nthreads) in perf_counts__new() argument 17 values = xyarray__new(ncpus, nthreads, sizeof(struct perf_counts_values)); in perf_counts__new() 25 values = xyarray__new(ncpus, nthreads, sizeof(bool)); in perf_counts__new()
|
| A D | bpf_ftrace.c | 23 int i, ncpus = 1, ntasks = 1; in perf_ftrace__latency_prepare_bpf() local 60 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus); in perf_ftrace__latency_prepare_bpf() 61 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in perf_ftrace__latency_prepare_bpf() 87 for (i = 0; i < ncpus; i++) { in perf_ftrace__latency_prepare_bpf() 172 int ncpus = cpu__max_cpu().cpu; in perf_ftrace__latency_read_bpf() local 176 hist = calloc(ncpus, sizeof(*hist)); in perf_ftrace__latency_read_bpf() 187 for (i = 0; i < ncpus; i++) in perf_ftrace__latency_read_bpf()
|
| A D | counts.h | 37 struct perf_counts *perf_counts__new(int ncpus, int nthreads);
|
| A D | bpf_off_cpu.c | 151 int ncpus = 1, ntasks = 1, ncgrps = 1; in off_cpu_prepare() local 168 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in off_cpu_prepare() 169 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in off_cpu_prepare() 239 for (i = 0; i < ncpus; i++) { in off_cpu_prepare()
|
| A D | cputopo.c | 262 long ncpus; in cpu_topology__new() local 267 ncpus = cpu__max_present_cpu().cpu; in cpu_topology__new() 276 nr = (u32)(ncpus & UINT_MAX); in cpu_topology__new()
|
| A D | bpf_lock_contention.c | 183 int ncpus = 1, ntasks = 1, ntypes = 1, naddrs = 1, ncgrps = 1, nslabs = 1; in lock_contention_prepare() local 219 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in lock_contention_prepare() 285 bpf_map__set_max_entries(skel->maps.cpu_filter, ncpus); in lock_contention_prepare() 323 for (i = 0; i < ncpus; i++) { in lock_contention_prepare()
|
| A D | data-convert-bt.c | 1342 int ncpus; in setup_streams() local 1348 ncpus = env->nr_cpus_avail ?: MAX_CPUS; in setup_streams() 1350 stream = zalloc(sizeof(*stream) * ncpus); in setup_streams() 1357 cw->stream_cnt = ncpus; in setup_streams()
|
| A D | stat.c | 465 int ncpus = evsel__nr_cpus(counter); in process_counter_maps() local 469 for (idx = 0; idx < ncpus; idx++) { in process_counter_maps()
|
| A D | python.c | 914 int nthreads, ncpus; in evsel__ensure_counts() local 920 ncpus = perf_cpu_map__nr(evsel->core.cpus); in evsel__ensure_counts() 922 evsel->counts = perf_counts__new(ncpus, nthreads); in evsel__ensure_counts() 926 evsel->prev_raw_counts = perf_counts__new(ncpus, nthreads); in evsel__ensure_counts()
|
| /tools/lib/perf/include/internal/ |
| A D | evsel.h | 137 int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads); 143 int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads);
|
| /tools/lib/perf/ |
| A D | evsel.c | 65 int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_fd() argument 67 evsel->fd = xyarray__new(ncpus, nthreads, sizeof(int)); in perf_evsel__alloc_fd() 72 for (idx = 0; idx < ncpus; idx++) { in perf_evsel__alloc_fd() 85 static int perf_evsel__alloc_mmap(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_mmap() argument 87 evsel->mmap = xyarray__new(ncpus, nthreads, sizeof(struct perf_mmap)); in perf_evsel__alloc_mmap() 526 int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_id() argument 528 if (ncpus == 0 || nthreads == 0) in perf_evsel__alloc_id() 531 evsel->sample_id = xyarray__new(ncpus, nthreads, sizeof(struct perf_sample_id)); in perf_evsel__alloc_id() 535 evsel->id = zalloc(ncpus * nthreads * sizeof(u64)); in perf_evsel__alloc_id()
|
| /tools/testing/selftests/powerpc/ |
| A D | utils.c | 419 int ncpus, cpu = -1; in pick_online_cpu() local 423 ncpus = get_nprocs_conf(); in pick_online_cpu() 424 size = CPU_ALLOC_SIZE(ncpus); in pick_online_cpu() 425 mask = CPU_ALLOC(ncpus); in pick_online_cpu() 439 for (cpu = 8; cpu < ncpus; cpu += 8) in pick_online_cpu() 444 for (cpu = ncpus - 1; cpu >= 0; cpu--) in pick_online_cpu()
|
| /tools/testing/selftests/powerpc/benchmarks/ |
| A D | context_switch.c | 108 int pid, ncpus; in start_process_on() local 121 ncpus = get_nprocs(); in start_process_on() 122 size = CPU_ALLOC_SIZE(ncpus); in start_process_on() 123 cpuset = CPU_ALLOC(ncpus); in start_process_on()
|
| /tools/testing/selftests/powerpc/ptrace/ |
| A D | perf-hwbreak.c | 147 int i, ncpus, cpu, ret = 0; in perf_systemwide_event_open() local 162 ncpus = get_nprocs_conf(); in perf_systemwide_event_open() 163 size = CPU_ALLOC_SIZE(ncpus); in perf_systemwide_event_open() 164 mask = CPU_ALLOC(ncpus); in perf_systemwide_event_open() 178 for (i = 0, cpu = 0; i < nprocs && cpu < ncpus; cpu++) { in perf_systemwide_event_open()
|
| /tools/testing/selftests/cgroup/ |
| A D | test_cpu.c | 376 static pid_t weight_hog_ncpus(const struct cpu_hogger *child, int ncpus) in weight_hog_ncpus() argument 380 .nprocs = ncpus, in weight_hog_ncpus()
|
| /tools/perf/ |
| A D | builtin-sched.c | 2093 u32 ncpus = sched->max_cpu.cpu + 1; in timehist_header() local 2100 for (i = 0, j = 0; i < ncpus; ++i) { in timehist_header() 2129 printf(" %*s ", ncpus, ""); in timehist_header() 2149 printf(" %.*s ", ncpus, graph_dotted_line); in timehist_header()
|