/external/grpc-grpc/tools/run_tests/python_utils/ |
D | upload_test_results.py | 70 def _get_build_metadata(test_results): argument 80 test_results['build_id'] = build_id 82 test_results['build_url'] = build_url 84 test_results['job_name'] = job_name 129 test_results = {} 130 _get_build_metadata(test_results) 131 test_results['cpu_estimated'] = result.cpu_estimated 132 test_results['cpu_measured'] = result.cpu_measured 133 test_results['elapsed_time'] = '%.2f' % result.elapsed_time 134 test_results['result'] = result.state [all …]
|
/external/autotest/venv/skylab_suite/ |
D | suite_tracking.py | 59 test_results = _parse_test_results(suite_handler) 60 suite_state, return_code = _get_suite_state(test_results, suite_handler) 61 if not test_results: 68 _log_test_results(test_results) 73 _log_test_result_links(test_results) 75 _log_buildbot_links(suite_handler, suite_name, test_results) 168 def _log_buildbot_links(suite_handler, suite_name, test_results): argument 182 failed_results = [t for t in test_results if _is_failed_result(t)] 189 def _log_buildbot_links_for_provision_tasks(test_results): argument 190 for result in test_results: [all …]
|
/external/autotest/client/site_tests/graphics_dEQP/ |
D | graphics_dEQP.py | 97 test_results=None, 113 if test_results is None: 114 test_results = {} 118 return test_results 149 test_results[result] = test_results.get(result, 0) + 1 158 return test_results 327 test_results = {} 414 test_results[result] = test_results.get(result, 0) + 1 416 return test_results 513 test_results = self._run_tests_hasty(test_cases, failing_test) [all …]
|
/external/autotest/server/ |
D | results_mocker.py | 64 self.results_dir, self.test_results, self.host_keyval_dir) 67 if not os.path.exists(self.test_results): 68 os.makedirs(self.test_results) 71 assert(os.path.exists(self.test_results) and 85 self.test_results = os.path.join(results_dir, test_name) 95 self.test_status = os.path.join(self.test_results, 'status') 99 self.test_keyvals = os.path.join(self.test_results, 'keyval')
|
/external/vulkan-validation-layers/scripts/ |
D | parse_test_results.py | 48 self.test_results = defaultdict(defaultdict) 72 for test_name, results in self.test_results.items(): 98 num_tests = len(self.test_results) 132 self.test_results[self.current_test][self.current_profile] = "skip" 137 if self.test_results.get(self.current_test, {}).get(self.current_profile, "") != "skip": 138 self.test_results[self.current_test][self.current_profile] = "pass" 143 self.test_results[self.current_test][self.current_profile] = "fail"
|
/external/autotest/client/site_tests/platform_GesturesRegressionTest/ |
D | platform_GesturesRegressionTest.py | 52 self.test_results = {} 58 self.test_results[key.replace('/', '-')] = score 62 if self.test_results: 63 self.write_perf_keyval(self.test_results)
|
/external/tensorflow/tensorflow/tools/test/ |
D | run_and_gather_logs.py | 72 test_results, _ = run_and_gather_logs_lib.run_and_gather_logs( 77 test_results.build_configuration.CopyFrom(gather_build_configuration()) 80 print(text_format.MessageToString(test_results)) 94 json_test_results = json_format.MessageToJson(test_results)
|
/external/webrtc/webrtc/modules/audio_coding/neteq/ |
D | neteq_unittest.cc | 120 template<class T> void ProcessReference(const T& test_results); 122 const T (&test_results)[n], 125 const T (&test_results)[n], 128 const T (&test_results)[n], 162 void RefFiles::ProcessReference(const T& test_results) { in ProcessReference() argument 163 WriteToFile(test_results); in ProcessReference() 164 ReadFromFileAndCompare(test_results); in ProcessReference() 168 void RefFiles::ProcessReference(const T (&test_results)[n], size_t length) { in ProcessReference() 169 WriteToFile(test_results, length); in ProcessReference() 170 ReadFromFileAndCompare(test_results, length); in ProcessReference() [all …]
|
/external/toolchain-utils/crosperf/ |
D | results_report.py | 360 def _GetHTMLCharts(label_names, test_results): argument 362 for item, runs in test_results.iteritems(): 412 test_results = self.benchmark_results.run_keyvals 413 charts = _GetHTMLCharts(label_names, test_results) 635 for test, test_results in benchmark_results.run_keyvals.iteritems(): 636 for label_name, label_results in zip(label_names, test_results):
|
/external/autotest/server/cros/dynamic_suite/ |
D | suite_unittest.py | 769 test_results = self._createSuiteMockResults() 772 [test_results[0] + test_results[1]], 804 test_results = self._createSuiteMockResults(result_status='WARN') 807 [test_results[0] + test_results[1]], 831 test_results = self._createSuiteMockResults() 834 [test_results[0] + test_results[1]],
|
/external/fonttools/Lib/fonttools.egg-info/ |
D | SOURCES.txt | 1724 Tests/varLib/data/test_results/Build.ttx 1725 Tests/varLib/data/test_results/BuildAvarEmptyAxis.ttx 1726 Tests/varLib/data/test_results/BuildAvarIdentityMaps.ttx 1727 Tests/varLib/data/test_results/BuildAvarSingleAxis.ttx 1728 Tests/varLib/data/test_results/BuildGvarCompositeExplicitDelta.ttx 1729 Tests/varLib/data/test_results/BuildMain.ttx 1730 Tests/varLib/data/test_results/BuildTestCFF2.ttx 1731 Tests/varLib/data/test_results/FeatureVars.ttx 1732 Tests/varLib/data/test_results/InterpolateLayout.ttx 1733 Tests/varLib/data/test_results/InterpolateLayout2.ttx [all …]
|
/external/autotest/site_utils/ |
D | run_suite.py | 1504 test_results = [] 1509 test_results.append(test_result) 1510 return test_results 1516 test_results = self._make_test_results() 1517 if len(test_results) == 0: 1520 max_name_length = max(len(t.name) for t in test_results) 1521 for test_result in test_results:
|
/external/deqp/external/openglcts/modules/glesext/tessellation_shader/ |
D | esextcTessellationShaderIsolines.cpp | 502 _test_results& test_results = m_test_results[vertex_spacing_mode]; in findTestResult() local 506 for (_test_results_iterator test_results_iterator = test_results.begin(); in findTestResult() 507 test_results_iterator != test_results.end(); test_results_iterator++) in findTestResult()
|
/external/nos/host/generic/nugget/proto/nugget/app/protoapi/ |
D | testing_api.proto | 98 oneof test_results {
|