| /tools/perf/tests/ |
| A D | Build | 3 perf-test-y += builtin-test.o 6 perf-test-y += dso-data.o 16 perf-test-y += pmu.o 27 perf-test-y += wp.o 40 perf-test-y += mem.o 47 perf-test-y += sdt.o 54 perf-test-y += time-utils-test.o 57 perf-test-y += demangle-java-test.o 58 perf-test-y += demangle-ocaml-test.o 59 perf-test-y += demangle-rust-v0-test.o [all …]
|
| A D | expr.c | 92 ret = test(ctx, "1+1", 2); in test__expr() 93 ret |= test(ctx, "FOO+BAR", 3); in test__expr() 94 ret |= test(ctx, "(BAR/2)%2", 1); in test__expr() 95 ret |= test(ctx, "1 - -4", 5); in test__expr() 97 ret |= test(ctx, "1-1 | 1", 1); in test__expr() 98 ret |= test(ctx, "1-1 & 1", 0); in test__expr() 107 ret |= test(ctx, ".1 + 2.", 2.1); in test__expr() 110 ret |= test(ctx, "1.1 < 2.2", 1); in test__expr() 111 ret |= test(ctx, "2.2 > 1.1", 1); in test__expr() 112 ret |= test(ctx, "1.1 < 1.1", 0); in test__expr() [all …]
|
| A D | unit_number__scnprintf.c | 15 } test[] = { in test__unit_number__scnprint() local 25 while (test[i].str) { in test__unit_number__scnprint() 28 unit_number__scnprintf(buf, sizeof(buf), test[i].n); in test__unit_number__scnprint() 31 test[i].n, test[i].str, buf); in test__unit_number__scnprint() 33 if (strcmp(test[i].str, buf)) in test__unit_number__scnprint()
|
| /tools/testing/scatterlist/ |
| A D | main.c | 8 struct test { struct 30 static void fail(struct test *test, struct sg_table *st, const char *cond) in fail() argument 37 test->size, test->max_seg, test->expected_segments, st->nents, in fail() 55 struct test *test, tests[] = { in main() local 86 for (i = 0, test = tests; test->expected_segments; test++, i++) { in main() 87 int left_pages = test->pfn_app ? test->num_pages : 0; in main() 92 set_pages(pages, test->pfn, test->num_pages); in main() 96 &append, pages, test->num_pages, 0, test->size, in main() 101 test->size, test->max_seg, GFP_KERNEL); in main() 109 set_pages(pages, test->pfn_app, test->num_pages); in main() [all …]
|
| /tools/testing/selftests/sched_ext/ |
| A D | runner.c | 45 printf("TEST: %s\n", test->name); in print_test_preamble() 85 if (test->setup) { in run_test() 91 status = test->run(context); in run_test() 93 if (test->cleanup) in run_test() 94 test->cleanup(context); in run_test() 101 if (!test) { in test_valid() 106 if (!test->name) { in test_valid() 112 if (!test->description) { in test_valid() 117 if (!test->run) { in test_valid() 162 printf("%s\n", test->name); in main() [all …]
|
| /tools/testing/kunit/ |
| A D | kunit_parser.py | 504 message = test.name 661 if test.ok_status() or test.counts.total() < 100: 681 counts = test.counts 682 status = test.status 751 test = Test() 761 test.name = "main" 810 if test.name != "" and not peek_test_name_match(lines, test): 836 return test 852 test = Test() 860 test.status = test.counts.get_status() [all …]
|
| /tools/testing/selftests/bpf/ |
| A D | xskxceiver.c | 548 memset(test, 0, sizeof(*test)); in test_spec_init() 570 __test_spec_init(test, test->ifobj_tx, test->ifobj_rx); in test_spec_reset() 1908 xsk_reattach_xdp(ifobj_rx, test->xdp_prog_rx, test->xskmap_rx, test->mode); in xsk_attach_xdp_progs() 1914 xsk_reattach_xdp(ifobj_tx, test->xdp_prog_tx, test->xskmap_tx, test->mode); in xsk_attach_xdp_progs() 1935 err = test_spec_set_mtu(test, test->mtu); in __testapp_validate_traffic() 1971 if (test->total_steps == test->current_step || test->fail) { in __testapp_validate_traffic() 2057 swap_directions(&test->ifobj_rx, &test->ifobj_tx); in testapp_bidirectional() 2058 res = __testapp_validate_traffic(test, test->ifobj_rx, test->ifobj_tx); in testapp_bidirectional() 2060 swap_directions(&test->ifobj_rx, &test->ifobj_tx); in testapp_bidirectional() 2072 test->ifobj_tx->xsk = &test->ifobj_tx->xsk_arr[1]; in swap_xsk_resources() [all …]
|
| /tools/testing/selftests/powerpc/vphn/ |
| A D | test-vphn.c | 21 static struct test { struct 366 static int test_one(struct test *test) in test_one() argument 371 vphn_unpack_associativity(test->input, output); in test_one() 374 if (len != test->expected[0]) { in test_one() 375 printf("expected %d elements, got %d\n", test->expected[0], in test_one() 382 if (val != test->expected[i]) { in test_one() 384 test->expected[i]); in test_one() 394 static struct test *test; in test_vphn() local 396 for (test = all_tests; test->descr; test++) { in test_vphn() 399 ret = test_one(test); in test_vphn() [all …]
|
| /tools/testing/selftests/exec/ |
| A D | binfmt_script.py | 117 test(name="too-big", size=SIZE+80, good=False) 119 test(name="exact", size=SIZE, good=False) 141 test(name="test.pl", size=439, leading=" ", 145 test(name="one-under", size=SIZE-1) 147 test(name="two-under", size=SIZE-2) 149 test(name="exact-trunc-whitespace", size=SIZE, arg=" ") 151 test(name="exact-trunc-arg", size=SIZE, arg=" f") 153 test(name="one-under-full-arg", size=SIZE-1, arg=" f") 155 test(name="one-under-no-nl", size=SIZE-1, newline="") 159 test(name="one-under-trunc-arg", size=SIZE-1, arg=" ") [all …]
|
| /tools/perf/tests/shell/attr/ |
| A D | README | 1 The struct perf_event_attr test (attr tests) support 28 the test definition, executes it and checks results. 32 Directory containing all attr test definitions. 35 perf record kill (test-record-basic) 44 perf record -c 123 kill (test-record-count) 45 perf record -d kill (test-record-data) 46 perf record -F 100 kill (test-record-freq) 49 perf record --call-graph dwarf kill (test-record-graph-dwarf) 61 perf record -R kill (test-record-raw) 65 perf stat -e cycles kill (test-stat-basic) [all …]
|
| /tools/perf/arch/x86/tests/ |
| A D | Build | 1 perf-test-$(CONFIG_DWARF_UNWIND) += regs_load.o 2 perf-test-$(CONFIG_DWARF_UNWIND) += dwarf-unwind.o 4 perf-test-y += arch-tests.o 5 perf-test-y += hybrid.o 6 perf-test-$(CONFIG_AUXTRACE) += intel-pt-test.o 8 perf-test-$(CONFIG_AUXTRACE) += insn-x86.o 10 perf-test-$(CONFIG_X86_64) += bp-modify.o 11 perf-test-y += amd-ibs-via-core-pmu.o 12 perf-test-y += amd-ibs-period.o 13 perf-test-y += topdown.o [all …]
|
| /tools/testing/selftests/bpf/prog_tests/ |
| A D | flow_dissector_classification.c | 213 if (test->cfg_l3_extra) in build_packet() 263 test->cfg_l3_inner); in build_packet() 278 &test->out_saddr6, &test->out_daddr6, in build_packet() 317 &test->extra_saddr6, &test->extra_daddr6, in build_packet() 333 if (test->cfg_l3_extra) in setup_tx() 344 if (test->cfg_l3_extra) { in setup_tx() 452 fdr = setup_rx(test); in run_test() 453 fdt = setup_tx(test); in run_test() 459 len = build_packet(test, in run_test() 780 test = &tests_input[i]; in test_flow_dissector_classification() [all …]
|
| A D | section_names.c | 214 rc = libbpf_prog_type_by_name(test->sec_name, &prog_type, in test_prog_type_by_name() 217 CHECK(rc != test->expected_load.rc, "check_code", in test_prog_type_by_name() 218 "prog: unexpected rc=%d for %s\n", rc, test->sec_name); in test_prog_type_by_name() 225 prog_type, test->sec_name); in test_prog_type_by_name() 229 expected_attach_type, test->sec_name); in test_prog_type_by_name() 239 CHECK(rc != test->expected_attach.rc, "check_ret", in test_attach_type_by_name() 245 CHECK(attach_type != test->expected_attach.attach_type, in test_attach_type_by_name() 247 attach_type, test->sec_name); in test_attach_type_by_name() 255 struct sec_name_test *test = &tests[i]; in test_section_names() local 257 test_prog_type_by_name(test); in test_section_names() [all …]
|
| A D | helper_restricted.c | 11 struct test_helper_restricted *test; in test_helper_restricted() local 14 test = test_helper_restricted__open(); in test_helper_restricted() 15 if (!ASSERT_OK_PTR(test, "open")) in test_helper_restricted() 18 prog_cnt = test->skeleton->prog_cnt; in test_helper_restricted() 21 struct bpf_program *prog = *test->skeleton->progs[j].prog; in test_helper_restricted() 26 err = test_helper_restricted__load(test); in test_helper_restricted() 29 test_helper_restricted__destroy(test); in test_helper_restricted()
|
| /tools/testing/selftests/ftrace/ |
| A D | README | 5 shell scripts for testing. Feel free to add new test cases. 17 # ./ftracetest test.d/basic3.tc 21 # ./ftracetest test.d/kprobe/ 27 extension) and rewrite the test description line. 52 * PASS: The test succeeded as expected. The test which exits with 0 is 53 counted as passed test. 55 * FAIL: The test failed, but was expected to succeed. The test which exits 56 with !0 is counted as failed test. 59 for example, the test was interrupted 60 or the test depends on a previous test, which failed. [all …]
|
| /tools/perf/arch/x86/include/ |
| A D | arch-tests.h | 10 int test__rdpmc(struct test_suite *test, int subtest); 12 int test__insn_x86(struct test_suite *test, int subtest); 14 int test__intel_pt_pkt_decoder(struct test_suite *test, int subtest); 15 int test__intel_pt_hybrid_compat(struct test_suite *test, int subtest); 16 int test__bp_modify(struct test_suite *test, int subtest); 17 int test__amd_ibs_via_core_pmu(struct test_suite *test, int subtest); 18 int test__amd_ibs_period(struct test_suite *test, int subtest); 19 int test__hybrid(struct test_suite *test, int subtest);
|
| /tools/testing/selftests/lkdtm/ |
| A D | run.sh | 29 line=$(grep -E '^#?'"$test"'\b' tests.txt) 31 echo "Skipped: missing test '$test' in tests.txt" 35 if ! grep -E -q '^'"$test"'$' "$TRIGGER" ; then 36 echo "Skipped: test '$test' missing in $TRIGGER!" 49 if echo "$test" | grep -q '^#' ; then 50 test=$(echo "$test" | cut -c2-) 54 echo "Skipping $test: $expect" 93 echo "$test" | cat >"$TRIGGER" || true 102 echo "$test: saw '$expect': ok" 106 echo "$test: saw 'XFAIL': [SKIP]" [all …]
|
| /tools/testing/selftests/kvm/ |
| A D | system_counter_offset_test.c | 39 &test->tsc_offset); in setup_system_counter() 42 static uint64_t guest_read_system_counter(struct test_case *test) in guest_read_system_counter() argument 47 static uint64_t host_read_guest_system_counter(struct test_case *test) in host_read_guest_system_counter() argument 49 return rdtsc() + test->tsc_offset; in host_read_guest_system_counter() 54 #error test not implemented for this architecture! 66 struct test_case *test = &test_cases[i]; in guest_main() local 68 GUEST_SYNC_CLOCK(i, guest_read_system_counter(test)); in guest_main() 96 struct test_case *test = &test_cases[i]; in enter_guest() local 98 setup_system_counter(vcpu, test); in enter_guest() 99 start = host_read_guest_system_counter(test); in enter_guest() [all …]
|
| /tools/perf/Documentation/ |
| A D | perf-test.txt | 1 perf-test(1) 6 perf-test - Runs sanity tests. 11 'perf test [<options>] [{list <test-name-fragment>|[<test-name-fragments>|<test-numbers>]}]' 18 To get a list of available tests use 'perf test list', specifying a test name 22 from 'perf test list'. 34 With a single '-v', verbose level 1, only failing test output 35 is displayed. With '-vv' and higher all test output is shown. 44 --runs-per-test:: 45 Run each test the given number of times, by default once. This 46 option can be useful to determine if a test is flaky. [all …]
|
| /tools/testing/selftests/resctrl/ |
| A D | resctrl_tests.c | 88 res = signal_handler_register(test); in test_prepare() 105 if (test->cleanup) in test_cleanup() 106 test->cleanup(); in test_cleanup() 113 if (!test->vendor_specific) in test_vendor_specific_check() 123 if (test->disabled) in run_single_test() 126 if (!test_vendor_specific_check(test)) { in run_single_test() 140 if (test_prepare(test)) { in run_single_test() 145 if (!test->feature_check(test)) { in run_single_test() 147 test->name, test->name); in run_single_test() 151 ret = test->run_test(test, uparams); in run_single_test() [all …]
|
| /tools/testing/selftests/arm64/fp/ |
| A D | .gitignore | 4 fpsimd-test 5 kernel-test 10 sve-test 11 ssve-test 16 za-test 18 zt-test
|
| /tools/perf/tests/workloads/ |
| A D | Build | 3 perf-test-y += noploop.o 4 perf-test-y += thloop.o 5 perf-test-y += leafloop.o 6 perf-test-y += sqrtloop.o 7 perf-test-y += brstack.o 8 perf-test-y += datasym.o 9 perf-test-y += landlock.o
|
| /tools/testing/selftests/timens/ |
| A D | gettime_perf.c | 47 static void test(clock_t clockid, char *clockstr, bool in_ns) in test() function 76 test(CLOCK_MONOTONIC, "monotonic", false); in main() 77 test(CLOCK_MONOTONIC_COARSE, "monotonic-coarse", false); in main() 78 test(CLOCK_MONOTONIC_RAW, "monotonic-raw", false); in main() 79 test(CLOCK_BOOTTIME, "boottime", false); in main() 98 test(CLOCK_MONOTONIC, "monotonic", true); in main() 99 test(CLOCK_MONOTONIC_COARSE, "monotonic-coarse", true); in main() 100 test(CLOCK_MONOTONIC_RAW, "monotonic-raw", true); in main() 101 test(CLOCK_BOOTTIME, "boottime", true); in main()
|
| /tools/testing/selftests/openat2/ |
| A D | openat2_test.c | 122 if (test->err >= 0) in test_openat2_struct() 125 failed = (fd != test->err); in test_openat2_struct() 142 if (test->err >= 0) in test_openat2_struct() 144 test->name, misalign); in test_openat2_struct() 147 test->name, misalign, test->err, in test_openat2_struct() 148 strerror(-test->err)); in test_openat2_struct() 273 if (test->err >= 0) in test_openat2_flags() 276 failed = (fd != test->err); in test_openat2_flags() 306 test->how.flags); in test_openat2_flags() 312 if (test->err >= 0) in test_openat2_flags() [all …]
|
| /tools/perf/tests/shell/ |
| A D | record_lbr.sh | 35 echo "$test" 38 echo "$test [Failed support missing]" 49 echo "$test [Failed in perf report]" 54 echo "$test [Success]" 67 echo "$test" 70 echo "$test [Failed support missing]" 83 echo "$test [Failed no samples captured]" 87 echo "$test: $sam_nr samples" 92 echo "$test [Failed samples missing branch stacks]" 100 echo "$test [Failed empty br stack ratio exceed $threshold%: $r%]" [all …]
|