/external/vboot_reference/tests/ |
D | test_common.c | 21 int TEST_EQ(int result, int expected_result, const char* testname) { in TEST_EQ() argument 23 fprintf(stderr, "%s Test " COL_GREEN "PASSED\n" COL_STOP, testname); in TEST_EQ() 26 fprintf(stderr, "%s Test " COL_RED "FAILED\n" COL_STOP, testname); in TEST_EQ() 34 int TEST_NEQ(int result, int not_expected_result, const char* testname) { in TEST_NEQ() argument 36 fprintf(stderr, "%s Test " COL_GREEN "PASSED\n" COL_STOP, testname); in TEST_NEQ() 39 fprintf(stderr, "%s Test " COL_RED "FAILED\n" COL_STOP, testname); in TEST_NEQ() 48 const char* testname) { in TEST_PTR_EQ() argument 50 fprintf(stderr, "%s Test " COL_GREEN "PASSED\n" COL_STOP, testname); in TEST_PTR_EQ() 53 fprintf(stderr, "%s Test " COL_RED "FAILED\n" COL_STOP, testname); in TEST_PTR_EQ() 62 const char* testname) { in TEST_PTR_NEQ() argument [all …]
|
D | test_common.h | 14 int TEST_EQ(int result, int expected_result, const char* testname); 18 int TEST_NEQ(int result, int not_expected_result, const char* testname); 24 const char* testname); 30 const char* testname); 35 const char* testname); 39 int TEST_TRUE(int result, const char* testname); 43 int TEST_FALSE(int result, const char* testname); 47 int TEST_SUCC(int result, const char* testname);
|
/external/ltp/testcases/kernel/syscalls/membarrier/ |
D | membarrier01.c | 29 char testname[80]; member 50 .testname = "cmd_fail", 62 .testname = "cmd_flags_fail", 75 .testname = "cmd_global_success", 90 .testname = "cmd_private_expedited_fail", 102 .testname = "cmd_private_expedited_register_success", 118 .testname = "cmd_private_expedited_success", 137 .testname = "cmd_private_expedited_sync_core_fail", 149 .testname = "cmd_private_expedited_sync_core_register_success", 162 .testname = "cmd_private_expedited_sync_core_success", [all …]
|
/external/curl/tests/ |
D | appveyor.pm | 37 my ($curl, $testnum, $testname)=@_; 38 $testname =~ s/\\/\\\\/g; 39 $testname =~ s/\'/\\\'/g; 40 $testname =~ s/\"/\\\"/g; 46 'testName': '$testname', 54 $APPVEYOR_TEST_NAMES{$testnum}=$testname; 59 my $testname=$APPVEYOR_TEST_NAMES{$testnum}; 60 if(!defined $testname) { 90 'testName': '$testname', 105 'message': '$appveyor_outcome: $testname',
|
D | azure.pm | 60 my ($curl, $azure_run_id, $testnum, $testname)=@_; 61 $testname =~ s/\\/\\\\/g; 62 $testname =~ s/\'/\\\'/g; 63 $testname =~ s/\"/\\\"/g; 74 'testCaseTitle': '$title_testnum: $testname',
|
/external/skqp/modules/pathkit/tests/ |
D | testReporter.js | 6 function reportCanvas(canvas, testname, outputType='canvas') { argument 8 return _report(b64, outputType, testname); 11 function reportSVG(svg, testname) { argument 27 _report(b64, 'svg', testname).then(() => { 38 function reportSVGString(svgstr, testname, fillRule='nofill') { argument 54 return reportSVG(newSVG, testname); 58 function reportPath(path, testname, done) { argument 67 return reportCanvas(canvas, testname).then(() => { 68 reportSVGString(svgStr, testname).then(() => { 76 function _report(data, outputType, testname) { argument [all …]
|
/external/skia/modules/canvaskit/tests/ |
D | testReporter.js | 6 function reportCanvas(canvas, testname, outputType='canvas') { argument 8 return _report(b64, outputType, testname); 11 function reportSVG(svg, testname) { argument 27 _report(b64, 'svg', testname).then(() => { 38 function reportSVGString(svgstr, testname, fillRule='nofill') { argument 54 return reportSVG(newSVG, testname); 58 function reportPath(path, testname, done) { argument 67 return reportCanvas(canvas, testname).then(() => { 68 reportSVGString(svgStr, testname).then(() => { 76 function _report(data, outputType, testname) { argument [all …]
|
/external/skia/modules/pathkit/tests/ |
D | testReporter.js | 6 function reportCanvas(canvas, testname, outputType='canvas') { argument 8 return _report(b64, outputType, testname); 11 function reportSVG(svg, testname) { argument 27 _report(b64, 'svg', testname).then(() => { 38 function reportSVGString(svgstr, testname, fillRule='nofill') { argument 54 return reportSVG(newSVG, testname); 58 function reportPath(path, testname, done) { argument 67 return reportCanvas(canvas, testname).then(() => { 68 reportSVGString(svgStr, testname).then(() => { 76 function _report(data, outputType, testname) { argument [all …]
|
/external/autotest/tko/parsers/ |
D | version_1.py | 186 def __init__(self, indent, status, subdir, testname, reason, argument 192 self.status = self.subdir = self.testname = self.reason = None 197 testname, reason, 212 if self.testname != 'reboot': 254 def make_dummy_abort(indent, subdir, testname, timestamp, reason): argument 270 if not testname: 271 testname = '----' 279 return msg % (subdir, testname, timestamp_field, reason) 284 line_buffer, line, indent, subdir, testname, timestamp, reason): argument 300 indent, subdir, testname, timestamp, reason) [all …]
|
D | version_0.py | 205 def __init__(self, subdir, testname, status, reason, test_kernel, argument 217 super(test, self).__init__(subdir, testname, status, reason, 245 def __init__(self, indent, status, subdir, testname, reason, argument 263 self.testname = self.parse_name(testname) 297 status, subdir, testname = parts[0:part_index] 313 return cls(indent, status, subdir, testname, reason, 351 not line.testname): 360 not line.testname): 383 line.testname, line.reason]) 392 if (line.testname == "Autotest.install" and [all …]
|
/external/autotest/client/site_tests/graphics_GLBench/ |
D | graphics_GLBench.py | 66 def is_no_checksum_test(self, testname): argument 72 if testname.startswith(prefix): 179 testname = key.strip() 188 (unit, testname)) 191 perf_value_name = '%s_%s' % (unit, testname) 204 keyvals[testname] = -3.0 205 failed_tests[testname] = 'GLError' 209 keyvals[testname] = 0.0 213 keyvals[testname] = -2.0 215 failed_tests[testname] = imagefile [all …]
|
/external/autotest/client/bin/ |
D | job_unittest.py | 386 testname = "error_test" 387 outputdir = os.path.join(self.job.resultdir, testname) 389 testname, 'test').and_return(("", testname)) 391 self.job.record.expect_call("START", testname, testname, 393 self.job._runtest.expect_call(testname, "", None, (), {}).and_raises( 395 self.job.record.expect_call("ERROR", testname, testname, 397 self.job.record.expect_call("END ERROR", testname, testname) 402 self.job.run_test(testname) 421 testname = "error_test" 422 outputdir = os.path.join(self.job.resultdir, testname) [all …]
|
/external/python/cpython3/Lib/test/ |
D | test_regrtest.py | 951 testname = self.create_test(code=code) 954 all_methods = ['%s.Tests.test_method1' % testname, 955 '%s.Tests.test_method2' % testname] 956 output = self.run_tests('--list-cases', testname) 960 all_methods = ['%s.Tests.test_method1' % testname] 963 testname) 997 testname = self.create_test(code=code) 1007 '%s.Tests.test_method3' % testname] 1012 output = self.run_tests("-v", "--ignorefile", filename, testname) 1033 testname = self.create_test(code=code) [all …]
|
/external/autotest/server/cros/dynamic_suite/ |
D | tools.py | 270 def _testname_to_keyval_key(testname): argument 282 return testname.replace('/', '_') 289 def create_bug_keyvals(job_id, testname, bug_info): argument 299 testname = _testname_to_keyval_key(testname) 300 keyval_base = '%s_%s' % (job_id, testname) if job_id else testname 307 def get_test_failure_bug_info(keyvals, job_id, testname): argument 330 testname = _testname_to_keyval_key(testname) 331 keyval_base = '%s_%s' % (job_id, testname) if job_id else testname
|
/external/icu/icu4j/perf-tests/ |
D | perftests.pl | 83 my $testname = $shortNames[$index]; 95 my $testName = "DateFmt-$testname-pat$patternCounter-JDK"; 105 my $testName = "DateFmt-$testname-pat$patternCounter"; 128 my $testname = $shortNames[$index]; 139 my $testName = "NumFmt-$testname-$patternName-JDK"; 149 my $testName = "NumFmt-$testname-$patternName"; 230 my $testname = "Coll-$locale-data$counter-StrCol"; 232 $ICU->setAttribute("test"=> $testname); 238 my $testname = "Coll-$locale-data$counter-keyGen"; 239 $Key->setAttribute("test"=> $testname); [all …]
|
/external/rust/crates/grpcio-sys/grpc/third_party/cares/cares/test/ |
D | ares-test.h | 375 #define ICLASS_NAME(casename, testname) Contained##casename##_##testname argument 376 #define CONTAINED_TEST_F(casename, testname, hostname, domainname, files) \ argument 377 class ICLASS_NAME(casename, testname) : public casename { \ 379 ICLASS_NAME(casename, testname)() {} \ 382 TEST_F(ICLASS_NAME(casename, testname), _) { \ 384 VoidToIntFn fn(ICLASS_NAME(casename, testname)::InnerTestBody); \ 387 int ICLASS_NAME(casename, testname)::InnerTestBody() 412 #define VCLASS_NAME(casename, testname) Virt##casename##_##testname argument 413 #define VIRT_NONVIRT_TEST_F(casename, testname) \ argument 414 class VCLASS_NAME(casename, testname) : public casename { \ [all …]
|
/external/eigen/blas/testing/ |
D | CMakeLists.txt | 2 macro(ei_add_blas_test testname) 4 set(targetname ${testname}) 6 set(filename ${testname}.f) 17 …add_test(${testname} "${Eigen_SOURCE_DIR}/blas/testing/runblastest.sh" "${testname}" "${Eigen_SOUR…
|
/external/eigen/cmake/ |
D | EigenTesting.cmake | 12 macro(ei_add_test_internal testname testname_with_suffix) 16 set(filename ${testname}.${EIGEN_ADD_TEST_FILENAME_EXTENSION}) 18 set(filename ${testname}.cpp) 62 ei_add_target_property(${targetname} COMPILE_FLAGS "-DEIGEN_TEST_FUNC=${testname}") 113 macro(ei_add_test_internal_sycl testname testname_with_suffix) 118 set(filename ${testname}.${EIGEN_ADD_TEST_FILENAME_EXTENSION}) 120 set(filename ${testname}.cpp) 135 add_custom_target(${testname}_integration_header_sycl DEPENDS ${include_file}) 138 add_dependencies(${targetname} ${testname}_integration_header_sycl) 157 ei_add_target_property(${targetname} COMPILE_FLAGS "-DEIGEN_TEST_FUNC=${testname}") [all …]
|
/external/elfutils/tests/ |
D | backtrace-subr.sh | 76 testname=$3 77 check_main $bt $testname 78 check_gsignal $bt $testname 79 check_err $err $testname 85 testname=$2 87 echo >&2 $testname: arch not supported 96 testname=$2 97 check_unsupported $err $testname 105 echo >&2 $testname: arm needs debuginfo installed for all libraries
|
/external/autotest/server/site_tests/stress_ClientTestReboot/ |
D | control | 23 --args="testname=NAME loops=N" 29 if 'testname' in args_dict and 'loops' in args_dict: 30 testname = args_dict.get('testname') 33 testname=testname, loops=loops) 37 '"testname=<test> loops=<number>"')
|
/external/ltp/testcases/open_posix_testsuite/bin/ |
D | run-tests.sh | 45 testname="$TEST_PATH/${1%.*}" 47 complog=$(basename $testname).log.$$ 54 echo "$testname: execution: PASS" >> "${LOGFILE}" 79 echo "$testname: execution: $msg: Output: " >> "${LOGFILE}" 81 echo "$testname: execution: $msg " 83 echo "$testname: execution: SKIPPED (test not present)"
|
/external/autotest/server/cros/tradefed/ |
D | tradefed_utils.py | 142 for testname, fail_count in waived_count.items(): 143 waived += [testname] * fail_count 145 testname, fail_count) 192 testname = match.group(1) 193 if waivers and testname in waivers: 194 waived_count[testname] = waived_count.get(testname, 0) + 1 196 failed_tests.add(testname) 211 for testname, fail_count in waived_count.items(): 218 'abis: %s', fail_count, testname, len(abis), abis) 220 waived += [testname] * fail_count [all …]
|
/external/igt-gpu-tools/tests/ |
D | igt_command_line.sh | 55 testname="$test" 93 …if [ "$testname" != "i915_selftest" -a "$testname" != "drm_mm" -a "$testname" != "kms_selftest" -a…
|
/external/mesa3d/src/freedreno/decode/scripts/ |
D | analyze.lua | 34 local testname = nil 48 testname = posix.basename(name) 57 gpu["tests"][testname] = test 95 table.insert(drawlist, testname .. "." .. didx) 108 testname = nil 113 for testname,test in pairs(gpu["tests"]) do 114 io.write(" " .. testname .. ", draws=" .. #test["draws"] .. "\n")
|
/external/llvm-project/lldb/packages/Python/lldbsuite/test/builders/ |
D | builder.py | 136 testname=None): argument 140 self.getMake(testdir, testname) + [ 162 testname=None): argument 166 self.getMake(testdir, testname) + [ 187 testname=None): argument 191 self.getMake(testdir, testname) + [ 212 testname=None): argument 216 self.getMake(testdir, testname) + [ 237 testname=None): argument
|