| /linux/tools/testing/selftests/bpf/prog_tests/ |
| A D | arg_parsing.c | 9 set->tests = NULL; in init_test_filter_set() 17 for (j = 0; j < set->tests[i].subtest_cnt; j++) in free_test_filter_set() 18 free((void *)set->tests[i].subtests[j]); in free_test_filter_set() 19 free(set->tests[i].subtests); in free_test_filter_set() 20 free(set->tests[i].name); in free_test_filter_set() 23 free(set->tests); in free_test_filter_set() 36 if (!ASSERT_OK_PTR(set.tests, "test filters initialized")) in test_parse_test_list() 46 if (!ASSERT_OK_PTR(set.tests, "test filters initialized")) in test_parse_test_list() 60 if (!ASSERT_OK_PTR(set.tests, "test filters initialized")) in test_parse_test_list() 66 ASSERT_OK(strcmp("arg_parsing", set.tests[0].subtests[0]), in test_parse_test_list() [all …]
|
| A D | obj_name.c | 10 } tests[] = { in test_obj_name() local 23 for (i = 0; i < ARRAY_SIZE(tests); i++) { in test_obj_name() 24 size_t name_len = strlen(tests[i].name) + 1; in test_obj_name() 40 CHECK((tests[i].success && fd < 0) || in test_obj_name() 41 (!tests[i].success && fd >= 0) || in test_obj_name() 42 (!tests[i].success && errno != tests[i].expected_errno), in test_obj_name() 45 fd, tests[i].success, errno, tests[i].expected_errno); in test_obj_name() 61 CHECK((tests[i].success && fd < 0) || in test_obj_name() 62 (!tests[i].success && fd >= 0) || in test_obj_name() 63 (!tests[i].success && errno != tests[i].expected_errno), in test_obj_name() [all …]
|
| A D | global_data.c | 18 } tests[] = { in test_global_data_number() local 32 for (i = 0; i < ARRAY_SIZE(tests); i++) { in test_global_data_number() 34 CHECK(err || num != tests[i].num, tests[i].name, in test_global_data_number() 36 err, num, tests[i].num); in test_global_data_number() 53 } tests[] = { in test_global_data_string() local 61 for (i = 0; i < ARRAY_SIZE(tests); i++) { in test_global_data_string() 63 CHECK(err || memcmp(str, tests[i].str, sizeof(str)), in test_global_data_string() 65 err, str, tests[i].str); in test_global_data_string() 88 } tests[] = { in test_global_data_struct() local 95 for (i = 0; i < ARRAY_SIZE(tests); i++) { in test_global_data_struct() [all …]
|
| A D | empty_skb.c | 29 } tests[] = { in test_empty_skb() local 115 for (i = 0; i < ARRAY_SIZE(tests); i++) { in test_empty_skb() 122 expected_ret = at_egress && !at_tc ? tests[i].lwt_egress_ret : tests[i].ret; in test_empty_skb() 124 tattr.data_in = tests[i].data_in; in test_empty_skb() 125 tattr.data_size_in = tests[i].data_size_in; in test_empty_skb() 128 bpf_obj->bss->ifindex = *tests[i].ifindex; in test_empty_skb() 131 sprintf(buf, "err: %s [%s]", tests[i].msg, bpf_program__name(prog)); in test_empty_skb() 133 if (at_tc && tests[i].success_on_tc) in test_empty_skb() 136 ASSERT_EQ(err, tests[i].err, buf); in test_empty_skb() 137 sprintf(buf, "ret: %s [%s]", tests[i].msg, bpf_program__name(prog)); in test_empty_skb() [all …]
|
| /linux/fs/btrfs/ |
| A D | Makefile | 43 btrfs-$(CONFIG_BTRFS_FS_RUN_SANITY_TESTS) += tests/free-space-tests.o \ 44 tests/extent-buffer-tests.o tests/btrfs-tests.o \ 45 tests/extent-io-tests.o tests/inode-tests.o tests/qgroup-tests.o \ 46 tests/free-space-tree-tests.o tests/extent-map-tests.o
|
| /linux/tools/perf/Documentation/ |
| A D | perf-test.txt | 6 perf-test - Runs sanity tests. 15 This command does assorted sanity tests, initially through linked routines but 16 also will look for a directory with more tests in the form of scripts. 18 To get a list of available tests use 'perf test list', specifying a test name 19 fragment will show all tests that have it. 21 To run just specific tests, inform test name fragments or the numbers obtained 36 Run tests one after the other, this is the default mode. 40 Run tests in parallel, speeds up the whole process but is not safe with 41 the current infrastructure, where some tests that compete for some resources, 42 for instance, 'perf probe' tests that add/remove probes or clean all probes, etc. [all …]
|
| /linux/drivers/of/unittest-data/ |
| A D | testcases_common.dtsi | 15 #include "tests-phandle.dtsi" 16 #include "tests-interrupts.dtsi" 17 #include "tests-match.dtsi" 18 #include "tests-address.dtsi" 19 #include "tests-platform.dtsi" 20 #include "tests-overlay.dtsi" 21 #include "tests-lifecycle.dtsi"
|
| /linux/tools/testing/selftests/clone3/ |
| A D | clone3.c | 173 static const struct test tests[] = { variable 319 ksft_set_plan(ARRAY_SIZE(tests)); in main() 322 for (i = 0; i < ARRAY_SIZE(tests); i++) { in main() 323 if (tests[i].filter && tests[i].filter()) { in main() 324 ksft_test_result_skip("%s\n", tests[i].name); in main() 328 if (tests[i].size_function) in main() 329 size = tests[i].size_function(); in main() 331 size = tests[i].size; in main() 336 tests[i].expected, in main() 337 tests[i].test_mode), in main() [all …]
|
| /linux/Documentation/rust/ |
| A D | testing.rst | 9 There are three sorts of tests: 11 - The KUnit tests. 12 - The ``#[test]`` tests. 15 The KUnit tests 18 These are the tests that come from the examples in the Rust documentation. They 19 get transformed into KUnit tests. 43 KUnit tests are documentation tests argument 126 The ``#[test]`` tests 135 (currently) and thus is fairly limited in what these tests can test. 142 The kernel config options required for the tests are listed in the [all …]
|
| /linux/tools/testing/selftests/riscv/hwprobe/ |
| A D | cbo.c | 169 } tests[] = { variable 190 tests[TEST_NO_ZICBOZ].enabled = true; in main() 191 tests[TEST_NO_ZICBOM].enabled = true; in main() 206 tests[TEST_ZICBOZ].enabled = true; in main() 207 tests[TEST_NO_ZICBOZ].enabled = false; in main() 212 for (i = 0; i < ARRAY_SIZE(tests); ++i) in main() 213 plan += tests[i].enabled ? tests[i].nr_tests : 0; in main() 220 for (i = 0; i < ARRAY_SIZE(tests); ++i) { in main() 221 if (tests[i].enabled) in main() 222 tests[i].test_fn(&cpus); in main()
|
| /linux/net/mptcp/ |
| A D | crypto_test.c | 15 static struct test_case tests[] = { variable 41 for (i = 0; i < ARRAY_SIZE(tests); ++i) { in mptcp_crypto_test_basic() 43 key1 = be64_to_cpu(*((__be64 *)&tests[i].key[0])); in mptcp_crypto_test_basic() 44 key2 = be64_to_cpu(*((__be64 *)&tests[i].key[8])); in mptcp_crypto_test_basic() 45 nonce1 = be32_to_cpu(*((__be32 *)&tests[i].msg[0])); in mptcp_crypto_test_basic() 46 nonce2 = be32_to_cpu(*((__be32 *)&tests[i].msg[4])); in mptcp_crypto_test_basic() 56 KUNIT_EXPECT_STREQ(test, &hmac_hex[0], tests[i].result); in mptcp_crypto_test_basic()
|
| /linux/lib/kunit/ |
| A D | Kconfig | 6 tristate "KUnit - Enable support for unit tests" 9 Enables support for kernel unit tests (KUnit), a lightweight unit 10 testing and mocking framework for the Linux kernel. These tests are 28 bool "Enable KUnit tests which print BUG stacktraces" 33 Enables fault handling tests for the KUnit framework. These tests may 42 Enables the unit tests for the KUnit test framework. These tests test 43 the KUnit test framework itself; the tests are both written using 60 tristate "All KUnit tests with satisfied dependencies" 62 Enables all KUnit tests, if they can be enabled. 63 KUnit tests run during boot and output the results to the debug log [all …]
|
| /linux/Documentation/dev-tools/kunit/ |
| A D | style.rst | 7 To make finding, writing, and using KUnit tests as simple as possible, it is 36 unsure, follow the conventions set by tests in similar areas. 89 subsystem (for example, both unit tests and integration tests), they should be 124 Individual tests consist of a single function which tests a constrained 180 This builds unit tests for foo. 182 For more information on KUnit and unit tests in general, 203 between KUnit and non-KUnit tests clearer. 206 ``tests/<suite>_kunit.c``. The ``tests`` directory should be placed at 207 the same level as the code under test. For example, tests for 208 ``lib/string.c`` live in ``lib/tests/string_kunit.c``. [all …]
|
| A D | run_wrapper.rst | 4 Running tests with kunit_tool 7 We can either run KUnit tests using kunit_tool or can run tests 8 manually, and then use kunit_tool to parse the results. To run tests 13 tests, and formats the test results. 35 - ``--timeout`` sets a maximum amount of time for tests to run. 101 Configuring, building, and running tests 155 Filtering tests 160 example: if we only want to run KUnit resource tests, use: 170 Running tests on QEMU 173 kunit_tool supports running tests on qemu as well as [all …]
|
| A D | run_manual.rst | 8 with other systems, or run tests on real hardware), we can 12 possible that tests may reduce the stability or security of 18 KUnit tests can run without kunit_tool. This can be useful, if: 26 tests can also be built by enabling their config options in our 27 ``.config``. KUnit tests usually (but don't always) have config options 28 ending in ``_KUNIT_TEST``. Most tests can either be built as a module, 34 automatically enable all tests with satisfied dependencies. This is 39 the tests. If the tests are built-in, they will run automatically on the 43 If the tests are built as modules, they will run when the module is 78 You can use the debugfs filesystem to trigger built-in tests to run after [all …]
|
| /linux/tools/lib/perf/tests/ |
| A D | Build | 1 tests-y += main.o 2 tests-y += test-evsel.o 3 tests-y += test-evlist.o 4 tests-y += test-cpumap.o 5 tests-y += test-threadmap.o
|
| /linux/tools/testing/selftests/mm/ |
| A D | run_vmtests.sh | 22 -d: run destructive tests 25 will run all tests. 31 tests for mmap(2) 33 tests for gup 35 tests for userfaultfd(2) 39 tests for mlock(2) 41 tests for mremap(2) 45 vmalloc smoke tests 47 hmm smoke tests 57 ksm tests that require >=2 NUMA nodes [all …]
|
| /linux/drivers/net/ethernet/sfc/falcon/ |
| A D | selftest.c | 107 tests->phy_alive = rc ? -1 : 1; in ef4_test_phy_alive() 122 tests->nvram = rc ? -1 : 1; in ef4_test_nvram() 136 struct ef4_self_tests *tests) in ef4_test_interrupts() argument 143 tests->interrupt = -1; in ef4_test_interrupts() 149 tests->interrupt = 0; in ef4_test_interrupts() 172 tests->interrupt = 1; in ef4_test_interrupts() 178 struct ef4_self_tests *tests) in ef4_test_eventq_irq() argument 676 &tests->loopback[mode]); in ef4_test_loopbacks() 713 rc = ef4_test_phy_alive(efx, tests); in ef4_selftest() 717 rc = ef4_test_nvram(efx, tests); in ef4_selftest() [all …]
|
| /linux/drivers/net/ethernet/sfc/ |
| A D | selftest.c | 108 tests->phy_alive = rc ? -1 : 1; in efx_test_phy_alive() 122 tests->nvram = rc ? -1 : 1; in efx_test_nvram() 136 struct efx_self_tests *tests) in efx_test_interrupts() argument 143 tests->interrupt = -1; in efx_test_interrupts() 149 tests->interrupt = 0; in efx_test_interrupts() 172 tests->interrupt = 1; in efx_test_interrupts() 178 struct efx_self_tests *tests) in efx_test_eventq_irq() argument 669 &tests->loopback[mode]); in efx_test_loopbacks() 706 rc = efx_test_phy_alive(efx, tests); in efx_selftest() 710 rc = efx_test_nvram(efx, tests); in efx_selftest() [all …]
|
| /linux/drivers/net/ethernet/sfc/siena/ |
| A D | selftest.c | 109 tests->phy_alive = rc ? -1 : 1; in efx_test_phy_alive() 123 tests->nvram = rc ? -1 : 1; in efx_test_nvram() 137 struct efx_self_tests *tests) in efx_test_interrupts() argument 144 tests->interrupt = -1; in efx_test_interrupts() 150 tests->interrupt = 0; in efx_test_interrupts() 173 tests->interrupt = 1; in efx_test_interrupts() 179 struct efx_self_tests *tests) in efx_test_eventq_irq() argument 674 &tests->loopback[mode]); in efx_test_loopbacks() 711 rc = efx_test_phy_alive(efx, tests); in efx_siena_selftest() 715 rc = efx_test_nvram(efx, tests); in efx_siena_selftest() [all …]
|
| /linux/drivers/firewire/ |
| A D | Kconfig | 26 This builds the KUnit tests whether structures exposed to user 29 KUnit tests run during boot and output the results to the debug 34 For more information on KUnit and unit tests in general, refer 38 tristate "KUnit tests for device attributes" if !KUNIT_ALL_TESTS 42 This builds the KUnit tests for device attribute for node and 45 KUnit tests run during boot and output the results to the debug 50 For more information on KUnit and unit tests in general, refer 58 This builds the KUnit tests for packet serialization and 66 For more information on KUnit and unit tests in general, refer 74 This builds the KUnit tests for helpers of self ID sequence. [all …]
|
| /linux/drivers/gpu/drm/ci/xfails/ |
| A D | panfrost-g12b-skips.txt | 1 # Skip driver specific tests 9 # Skip intel specific tests 14 # Panfrost is not a KMS driver, so skip the KMS tests 17 # Skip display functionality tests for GPU-only drivers 21 # Currently fails and causes coverage loss for other tests
|
| A D | panfrost-mt8183-skips.txt | 1 # Skip driver specific tests 9 # Skip intel specific tests 14 # Panfrost is not a KMS driver, so skip the KMS tests 17 # Skip display functionality tests for GPU-only drivers 21 # Currently fails and causes coverage loss for other tests
|
| /linux/Documentation/dev-tools/ |
| A D | kselftest.rst | 25 to run the full range of hot-plug tests. In default mode, hot-plug tests run 37 To build the tests:: 42 To run the tests:: 83 single test to run, or a list of tests to run. 104 To run all tests but a single subsystem:: 108 You can specify multiple tests to skip:: 123 To build the hotplug tests:: 127 To run the hotplug tests:: 156 note some tests will require root privileges:: 217 Contributing new tests [all …]
|
| /linux/tools/testing/selftests/kvm/s390x/ |
| A D | tprot.c | 77 } tests[] = { variable 141 enum stage stage = tests[*i].stage; in perform_next_stage() 145 for (; tests[*i].stage == stage; (*i)++) { in perform_next_stage() 153 skip = tests[*i].addr < (void *)PAGE_SIZE && in perform_next_stage() 154 tests[*i].expected != TRANSL_UNAVAIL && in perform_next_stage() 157 result = test_protection(tests[*i].addr, tests[*i].key); in perform_next_stage() 158 __GUEST_ASSERT(result == tests[*i].expected, in perform_next_stage() 160 tests[*i].expected, result, *i); in perform_next_stage()
|