/external/vboot_reference/tests/ |
D | test_common.c | 21 int TEST_EQ(int result, int expected_result, const char* testname) { in TEST_EQ() argument 23 fprintf(stderr, "%s Test " COL_GREEN "PASSED\n" COL_STOP, testname); in TEST_EQ() 26 fprintf(stderr, "%s Test " COL_RED "FAILED\n" COL_STOP, testname); in TEST_EQ() 34 int TEST_NEQ(int result, int not_expected_result, const char* testname) { in TEST_NEQ() argument 36 fprintf(stderr, "%s Test " COL_GREEN "PASSED\n" COL_STOP, testname); in TEST_NEQ() 39 fprintf(stderr, "%s Test " COL_RED "FAILED\n" COL_STOP, testname); in TEST_NEQ() 48 const char* testname) { in TEST_PTR_EQ() argument 50 fprintf(stderr, "%s Test " COL_GREEN "PASSED\n" COL_STOP, testname); in TEST_PTR_EQ() 53 fprintf(stderr, "%s Test " COL_RED "FAILED\n" COL_STOP, testname); in TEST_PTR_EQ() 62 const char* testname) { in TEST_PTR_NEQ() argument [all …]
|
D | test_common.h | 14 int TEST_EQ(int result, int expected_result, const char* testname); 18 int TEST_NEQ(int result, int not_expected_result, const char* testname); 24 const char* testname); 30 const char* testname); 35 const char* testname); 39 int TEST_TRUE(int result, const char* testname); 43 int TEST_FALSE(int result, const char* testname); 47 int TEST_SUCC(int result, const char* testname);
|
/external/autotest/server/site_tests/platform_GCC/ |
D | platform_GCC.py | 25 result, testname = line.split(': ', 1) 26 testname = testname.strip() 27 if testname in results: 28 counts[testname] += 1 29 unique_testname = '%s (%d)' % (testname, counts[testname]) 31 counts[testname] = 1 32 unique_testname = testname 47 for testname in new_results.keys(): 48 if testname not in baseline_results: 49 differences.append((testname, 'NOTEXECUTED', [all …]
|
/external/autotest/tko/parsers/ |
D | version_1.py | 183 def __init__(self, indent, status, subdir, testname, reason, argument 189 self.status = self.subdir = self.testname = self.reason = None 194 testname, reason, 209 if self.testname != 'reboot': 251 def make_dummy_abort(indent, subdir, testname, timestamp, reason): argument 267 if not testname: 268 testname = '----' 276 return msg % (subdir, testname, timestamp_field, reason) 281 line_buffer, line, indent, subdir, testname, timestamp, reason): argument 297 indent, subdir, testname, timestamp, reason) [all …]
|
D | version_0.py | 203 def __init__(self, subdir, testname, status, reason, test_kernel, argument 215 super(test, self).__init__(subdir, testname, status, reason, 243 def __init__(self, indent, status, subdir, testname, reason, argument 261 self.testname = self.parse_name(testname) 295 status, subdir, testname = parts[0:part_index] 311 return cls(indent, status, subdir, testname, reason, 349 not line.testname): 358 not line.testname): 381 line.testname, line.reason]) 390 if (line.testname == "Autotest.install" and [all …]
|
D | version_1_unittest.py | 69 self.assertEquals(line.testname, 'test_name') 87 self.assertEquals(line.testname, None) 99 self.assertEquals(line.testname, None) 115 self.assertEquals(line.testname, None) 132 self.assertEquals(line.testname, 'fields') 148 self.assertEquals(line.testname, 'second field') 329 self.testname = 'testname' 337 self.indent, self.subdir, self.testname, self.timestamp, 341 '\t' * self.indent, self.subdir, self.testname, self.timestamp, 347 self.indent, None, self.testname, self.timestamp, self.reason) [all …]
|
/external/autotest/client/tests/ltp/ |
D | ltp-diff.py | 49 testname = s.group(1) 51 runs[i][testname] = status 52 testnames[testname] = 1 75 for testname in all_testnames: 80 if not runs[i].has_key(testname): 81 runs[i][testname] = "null" 82 if not runs[i-1].has_key(testname): 83 runs[i-1][testname] = "null" 85 if runs[i][testname] != runs[i-1][testname]: 96 testname_cleaned = re.sub('\s+', ',', testname) [all …]
|
/external/autotest/client/site_tests/kernel_LTP/ |
D | ltp-diff.py | 52 testname = s.group(1) 54 runs[i][testname] = status 55 testnames[testname] = 1 80 for testname in all_testnames: 85 if not runs[i].has_key(testname): 86 runs[i][testname] = "null" 87 if not runs[i-1].has_key(testname): 88 runs[i-1][testname] = "null" 90 if runs[i][testname] != runs[i-1][testname]: 101 testname_cleaned = re.sub('\s+', ',', testname) [all …]
|
/external/autotest/client/bin/ |
D | job_unittest.py | 383 testname = "error_test" 384 outputdir = os.path.join(self.job.resultdir, testname) 386 testname, 'test').and_return(("", testname)) 388 self.job.record.expect_call("START", testname, testname, 390 self.job._runtest.expect_call(testname, "", None, (), {}).and_raises( 392 self.job.record.expect_call("ERROR", testname, testname, 394 self.job.record.expect_call("END ERROR", testname, testname) 399 self.job.run_test(testname) 418 testname = "error_test" 419 outputdir = os.path.join(self.job.resultdir, testname) [all …]
|
D | job.py | 525 _group, testname = self.pkgmgr.get_package_name(url, 'test') 526 testname, subdir, tag = self._build_tagged_test_name(testname, dargs) 534 self.record("WARN", subdir, testname, reason) 541 self.record(detail.exit_status, subdir, testname, str(detail)) 544 self.record('GOOD', subdir, testname, 'completed successfully') 546 return (subdir, testname, group_func, timeout) 562 (subdir, testname, group_func, timeout) = self._run_test_base(url, 566 self._rungroup(subdir, testname, group_func, timeout) 591 (subdir, testname, group_func, timeout) = self._run_test_base(url, 595 self._rungroup(subdir, testname, group_func, timeout) [all …]
|
/external/vulkan-validation-layers/scripts/ |
D | vk_validation_stats.py | 123 testname = db_line[2] 132 self.db_dict[error_enum]['testname'] = testname 145 if testname.lower() not in ['unknown', 'none', 'nottestable']: 146 self.db_enum_to_tests[error_enum] = testname.split(',') 257 testname = '' 266 testname = line.split(',')[-1] 267 testname = testname.strip().strip(' {)') 269 if ('' == testname): 272 self.test_to_errors[testname] = [] 275 testname = testname.strip().strip(' {)') [all …]
|
/external/autotest/client/site_tests/graphics_GLBench/ |
D | graphics_GLBench.py | 92 def is_no_checksum_test(self, testname): argument 98 if testname.startswith(prefix): 206 testname = key.strip() 214 (unit, testname)) 219 perf_value_name = '%s_%s' % (unit, testname) 241 keyvals[testname] = -3.0 242 failed_tests[testname] = 'GLError' 246 keyvals[testname] = 0.0 250 keyvals[testname] = -2.0 252 failed_tests[testname] = imagefile [all …]
|
/external/autotest/server/cros/dynamic_suite/ |
D | tools.py | 266 def _testname_to_keyval_key(testname): argument 278 return testname.replace('/', '_') 285 def create_bug_keyvals(job_id, testname, bug_info): argument 295 testname = _testname_to_keyval_key(testname) 296 keyval_base = '%s_%s' % (job_id, testname) if job_id else testname 303 def get_test_failure_bug_info(keyvals, job_id, testname): argument 326 testname = _testname_to_keyval_key(testname) 327 keyval_base = '%s_%s' % (job_id, testname) if job_id else testname
|
/external/icu/icu4j/perf-tests/ |
D | perftests.pl | 83 my $testname = $shortNames[$index]; 95 my $testName = "DateFmt-$testname-pat$patternCounter-JDK"; 105 my $testName = "DateFmt-$testname-pat$patternCounter"; 128 my $testname = $shortNames[$index]; 139 my $testName = "NumFmt-$testname-$patternName-JDK"; 149 my $testName = "NumFmt-$testname-$patternName"; 230 my $testname = "Coll-$locale-data$counter-StrCol"; 232 $ICU->setAttribute("test"=> $testname); 238 my $testname = "Coll-$locale-data$counter-keyGen"; 239 $Key->setAttribute("test"=> $testname); [all …]
|
/external/elfutils/tests/ |
D | backtrace-subr.sh | 63 testname=$3 64 check_main $bt $testname 65 check_gsignal $bt $testname 66 check_err $err $testname 72 testname=$2 74 echo >&2 $testname: arch not supported 82 testname=$2 83 check_unsupported $err $testname 91 echo >&2 $testname: arm needs debuginfo installed for all libraries
|
/external/eigen/cmake/ |
D | EigenTesting.cmake | 12 macro(ei_add_test_internal testname testname_with_suffix) 16 set(filename ${testname}.${EIGEN_ADD_TEST_FILENAME_EXTENSION}) 18 set(filename ${testname}.cpp) 62 ei_add_target_property(${targetname} COMPILE_FLAGS "-DEIGEN_TEST_FUNC=${testname}") 113 macro(ei_add_test_internal_sycl testname testname_with_suffix) 118 set(filename ${testname}.${EIGEN_ADD_TEST_FILENAME_EXTENSION}) 120 set(filename ${testname}.cpp) 135 add_custom_target(${testname}_integration_header_sycl DEPENDS ${include_file}) 138 add_dependencies(${targetname} ${testname}_integration_header_sycl) 157 ei_add_target_property(${targetname} COMPILE_FLAGS "-DEIGEN_TEST_FUNC=${testname}") [all …]
|
/external/webrtc/webrtc/tools/rtcbot/ |
D | main.js | 74 function runTest(testname) { argument 75 if (testname in testSuites) { 76 console.log("Running test: " + testname); 78 testSuites[testname](test); 80 console.log("Unknown test: " + testname); 89 for (var testname in testSuites) 90 console.log(' ' + testname);
|
/external/eigen/blas/testing/ |
D | CMakeLists.txt | 2 macro(ei_add_blas_test testname) 4 set(targetname ${testname}) 6 set(filename ${testname}.f) 17 …add_test(${testname} "${Eigen_SOURCE_DIR}/blas/testing/runblastest.sh" "${testname}" "${Eigen_SOUR…
|
/external/autotest/server/site_tests/stress_ClientTestReboot/ |
D | control | 23 --args="testname=NAME loops=N" 29 if 'testname' in args_dict and 'loops' in args_dict: 30 testname = args_dict.get('testname') 33 testname=testname, loops=loops) 37 '"testname=<test> loops=<number>"')
|
/external/ltp/testcases/open_posix_testsuite/bin/ |
D | run-tests.sh | 45 testname="$TEST_PATH/${1%.*}" 47 complog=$(basename $testname).log.$$ 54 echo "$testname: execution: PASS" >> "${LOGFILE}" 79 echo "$testname: execution: $msg: Output: " >> "${LOGFILE}" 81 echo "$testname: execution: $msg " 83 echo "$testname: execution: SKIPPED (test not present)"
|
/external/autotest/client/cros/ |
D | chrome_binary_test.py | 76 testname = '%s.%s' % (suite.get('name'), case.get('name')) 77 reasons[testname] = failure.attrib['message'] 85 testname, reason = m.group(1, 2) 87 if testname not in reasons: 88 reasons[testname] = reason 92 for testname, reason in sorted(reasons.items()): 93 message += '; <%s>: %s' % (testname, reason.replace('\n', '; '))
|
/external/autotest/tko/ |
D | models.py | 97 def __init__(self, subdir, testname, status, reason, test_kernel, argument 101 self.testname = testname 139 def parse_test(cls, job, subdir, testname, status, reason, test_kernel, argument 162 tko_utils.dprint("parsing test %s %s" % (subdir, testname)) 201 return constructor(subdir, testname, status, reason, test_kernel, 207 def parse_partial_test(cls, job, subdir, testname, reason, test_kernel, argument 227 tko_utils.dprint('parsing partial test %s %s' % (subdir, testname)) 229 return cls(subdir, testname, 'RUNNING', reason, test_kernel,
|
/external/autotest/client/site_tests/network_RackWiFiConnect/ |
D | network_RackWiFiConnect.py | 44 if test == service_test.testname: 62 service_test.testname) 158 if service_test.testname == test: 159 self._connect_and_assert(service_test.testname, 166 service_test.testname) 168 self._connect_and_assert(service_test.testname,
|
/external/mesa3d/src/gallium/drivers/llvmpipe/ |
D | SConscript | 36 testname = 'lp_test_' + test 38 target = testname, 39 source = [testname + '.c', 'lp_test_main.c'], 41 env.UnitTest(testname, target)
|
/external/protobuf/jenkins/ |
D | make_test_output.py | 49 for testname in os.listdir(basedir + "/logs/1"): 50 test = tests[testname] 52 with open(basedir + "/logs/1/" + testname + "/stdout") as f: 55 with open(basedir + "/logs/1/" + testname + "/stderr") as f:
|