/tools/memory-model/scripts/ |
D | checklitmushist.sh | 34 mkdir $T/results 35 find litmus -type d -print | ( cd $T/results; sed -e 's/^/mkdir -p /' | sh ) 46 LKMM_DESTDIR=$T/results; export LKMM_DESTDIR 55 ( cd $T/results; find litmus -type f -name '*.litmus.out' -print |
|
D | README | 11 the results against the expected results recorded in the 18 documenting expected results, comparing the actual results to 41 of processes given a specified timeout, recording the results 55 and record the results in .litmus.out files.
|
/tools/perf/tests/shell/ |
D | stat+shadow_stat.sh | 57 results="$results $cpu:$num" 61 cyc=${results##* $cpu:}
|
/tools/testing/selftests/bpf/ |
D | test_tcp_check_syncookie_user.c | 208 int results = -1; in main() local 217 results = get_map_fd_by_prog_id(atoi(argv[1]), &xdp); in main() 218 if (results < 0) { in main() 242 if (run_test(server, results, xdp)) in main() 245 if (run_test(server_v6, results, xdp)) in main() 255 close(results); in main()
|
D | bench.c | 300 struct bench_res *results; member 373 state.results = calloc(env.duration_sec + env.warmup_sec + 2, in setup_benchmark() 374 sizeof(*state.results)); in setup_benchmark() 375 if (!state.producers || !state.consumers || !state.results) in setup_benchmark() 423 struct bench_res *res = &state.results[iter]; in collect_measurements() 461 bench->report_final(state.results + env.warmup_sec, in main()
|
/tools/testing/selftests/ |
D | kselftest_harness.h | 130 snprintf(_metadata->results->reason, \ 131 sizeof(_metadata->results->reason), fmt, ##__VA_ARGS__); \ 134 _metadata->results->reason); \ 833 struct __test_results *results; member 976 memset(t->results->reason, 0, sizeof(t->results->reason)); in __run_test() 1006 ksft_test_result_skip("%s\n", t->results->reason[0] ? in __run_test() 1007 t->results->reason : "unknown"); in __run_test() 1019 struct __test_results *results; in test_harness_run() local 1034 results = mmap(NULL, sizeof(*results), PROT_READ | PROT_WRITE, in test_harness_run() 1045 t->results = results; in test_harness_run() [all …]
|
/tools/testing/selftests/bpf/progs/ |
D | test_core_retro.c | 23 } results SEC(".maps"); 38 bpf_map_update_elem(&results, &zero, &tgid, 0); in handle_sys_enter()
|
D | test_tcp_check_syncookie_kern.c | 19 struct bpf_map_def SEC("maps") results = { 137 bpf_map_update_elem(&results, &key_gen, &cookie, 0); in check_syncookie() 138 bpf_map_update_elem(&results, &key_mss, &mss, 0); in check_syncookie() 144 bpf_map_update_elem(&results, &key, &cookie, 0); in check_syncookie()
|
/tools/testing/selftests/arm64/mte/ |
D | check_gcr_el1_cswitch.c | 102 int results[NUM_ITERATIONS]; in mte_gcr_fork_test() local 125 results[i] = WEXITSTATUS(res); in mte_gcr_fork_test() 131 if (results[i] == KSFT_FAIL) in mte_gcr_fork_test()
|
/tools/testing/selftests/rcutorture/bin/ |
D | kvm-recheck-refscale.sh | 17 echo Unreadable results directory: $i
|
D | kvm-recheck-rcuscale.sh | 17 echo Unreadable results directory: $i
|
D | kvm-recheck-scf.sh | 17 echo Unreadable results directory: $i
|
D | kvm-recheck-lock.sh | 17 echo Unreadable results directory: $i
|
D | kvm-recheck-rcu.sh | 17 echo Unreadable results directory: $i
|
D | kvm-check-branches.sh | 84 echo Run results: $resdir/$ds/$idir/$rrd
|
D | kvm-again.sh | 136 echo ---- Re-run results directory: $rundir
|
/tools/testing/selftests/bpf/prog_tests/ |
D | core_retro.c | 29 err = bpf_map_lookup_elem(bpf_map__fd(skel->maps.results), &zero, &res); in test_core_retro()
|
D | select_reuseport.c | 345 __u32 results[NR_RESULTS]; in check_results() local 350 err = bpf_map_lookup_elem(result_map, &i, &results[i]); in check_results() 356 if (results[i] != expected_results[i]) { in check_results() 367 printf("%u", results[0]); in check_results() 369 printf(", %u", results[i]); in check_results()
|
/tools/testing/selftests/tc-testing/plugin-lib/ |
D | README-PLUGINS | 21 …Run the command under test with valgrind, and produce an extra set of TAP results for the memory t…
|
/tools/perf/Documentation/ |
D | perf-lock.txt | 21 results of lock events.
|
D | perf-script-perl.txt | 18 displays the results of the trace analysis implemented in the given 32 do basic things like aggregate event data, print results, etc. Also, 33 the check-perf-script.pl script, while not interesting for its results, 143 as display results:
|
/tools/testing/selftests/tc-testing/creating-testcases/ |
D | AddingTestCases.txt | 38 this test case will still appear in the results output but 80 the tests are run will be destroyed. This is to prevent inaccurate results 81 in the test cases. tdc will output a series of TAP results for the skipped
|
/tools/testing/selftests/tc-testing/ |
D | TODO.txt | 13 - Allow tdc to write its results to file.
|
/tools/power/cpupower/bench/ |
D | README-BENCH | 64 Example of expected results with ondemand governor: 66 This shows expected results of the first two test run rounds from
|
/tools/kvm/kvm_stat/ |
D | kvm_stat | 804 results = {} 808 return results 834 if field in results: 835 results[field] += increment 837 results[field] = increment 839 return results
|