/external/python/google-api-python-client/samples/analytics/ |
D | core_reporting_v3_reference.py | 83 # Try to make a request to the API. Print the results or handle errors. 85 results = get_api_query(service, flags.table_id).execute() 86 print_results(results) 123 def print_results(results): argument 124 """Prints all the results in the Core Reporting API Response. 127 results: The response returned from the Core Reporting API. 130 print_report_info(results) 131 print_pagination_info(results) 132 print_profile_info(results) 133 print_query(results) [all …]
|
/external/deqp/scripts/ |
D | testset.py | 29 def all (results, predicate): argument 30 for result in results: 35 def any (results, predicate): argument 36 for result in results: 48 def __init__ (self, name, results): argument 50 self.results = results 68 results = [] 70 results.append(TestCaseResult(case, [])) 71 return results 73 def addResultsToCaseList (caselist, results): argument [all …]
|
/external/v8/tools/testrunner/testproc/ |
D | result.py | 30 """Result consisting of multiple results. It can be used by processors that 31 create multiple subtests for each test and want to pass all results back. 35 def create(results): argument 36 """Create grouped result from the list of results. It filters out skipped 37 results. If all results are skipped results it returns skipped result. 40 results: list of pairs (test, result) 42 results = [(t, r) for (t, r) in results if not r.is_skipped] 43 if not results: 45 return GroupedResult(results) 47 def __init__(self, results): argument [all …]
|
/external/clang/lib/Sema/ |
D | SemaCodeComplete.cpp | 41 /// \brief A container of code-completion results. 53 /// \brief The actual results we have found. 54 std::vector<Result> Results; member in __anon28dbc1b40111::ResultBuilder 90 // 1 -> 2 elements: create the vector of results and push in the in Add() 118 /// results. 121 /// \brief The semantic analysis object for which results are being 131 /// results that are not desirable. 159 /// \brief The completion context in which we are gathering results. 204 /// results. 210 /// \brief Set the filter used for code-completion results. [all …]
|
/external/skia/tests/ |
D | StringTest.cpp | 210 SkTArray<SkString> results; in DEF_TEST() local 212 SkStrSplit("a-_b_c-dee--f-_-_-g-", "-_", &results); in DEF_TEST() 213 REPORTER_ASSERT(r, results.count() == 6); in DEF_TEST() 214 REPORTER_ASSERT(r, results[0].equals("a")); in DEF_TEST() 215 REPORTER_ASSERT(r, results[1].equals("b")); in DEF_TEST() 216 REPORTER_ASSERT(r, results[2].equals("c")); in DEF_TEST() 217 REPORTER_ASSERT(r, results[3].equals("dee")); in DEF_TEST() 218 REPORTER_ASSERT(r, results[4].equals("f")); in DEF_TEST() 219 REPORTER_ASSERT(r, results[5].equals("g")); in DEF_TEST() 221 results.reset(); in DEF_TEST() [all …]
|
/external/skqp/tests/ |
D | StringTest.cpp | 210 SkTArray<SkString> results; in DEF_TEST() local 212 SkStrSplit("a-_b_c-dee--f-_-_-g-", "-_", &results); in DEF_TEST() 213 REPORTER_ASSERT(r, results.count() == 6); in DEF_TEST() 214 REPORTER_ASSERT(r, results[0].equals("a")); in DEF_TEST() 215 REPORTER_ASSERT(r, results[1].equals("b")); in DEF_TEST() 216 REPORTER_ASSERT(r, results[2].equals("c")); in DEF_TEST() 217 REPORTER_ASSERT(r, results[3].equals("dee")); in DEF_TEST() 218 REPORTER_ASSERT(r, results[4].equals("f")); in DEF_TEST() 219 REPORTER_ASSERT(r, results[5].equals("g")); in DEF_TEST() 221 results.reset(); in DEF_TEST() [all …]
|
/external/python/cpython3/Lib/lib2to3/fixes/ |
D | fix_operator.py | 43 def transform(self, node, results): argument 44 method = self._check_method(node, results) 46 return method(node, results) 49 def _sequenceIncludes(self, node, results): argument 50 return self._handle_rename(node, results, "contains") 53 def _isCallable(self, node, results): argument 54 obj = results["obj"] 58 def _repeat(self, node, results): argument 59 return self._handle_rename(node, results, "mul") 62 def _irepeat(self, node, results): argument [all …]
|
/external/python/cpython2/Lib/lib2to3/fixes/ |
D | fix_operator.py | 41 def transform(self, node, results): argument 42 method = self._check_method(node, results) 44 return method(node, results) 47 def _sequenceIncludes(self, node, results): argument 48 return self._handle_rename(node, results, u"contains") 51 def _isCallable(self, node, results): argument 52 obj = results["obj"] 57 def _repeat(self, node, results): argument 58 return self._handle_rename(node, results, u"mul") 61 def _irepeat(self, node, results): argument [all …]
|
/external/dagger2/producers/src/test/java/dagger/producers/internal/ |
D | SetOfProducedProducerTest.java | 68 Results<Integer> results = Results.create(producer.get().get()); in delegateSetNpe() local 69 assertThat(results.successes).isEmpty(); in delegateSetNpe() 70 assertThat(results.failures).hasSize(1); in delegateSetNpe() 71 assertThat(Iterables.getOnlyElement(results.failures).getCause()) in delegateSetNpe() 81 Results<Integer> results = Results.create(producer.get().get()); in oneOfDelegateSetNpe() local 82 assertThat(results.successes).containsExactly(3, 7); in oneOfDelegateSetNpe() 83 assertThat(results.failures).hasSize(1); in oneOfDelegateSetNpe() 84 assertThat(Iterables.getOnlyElement(results.failures).getCause()) in oneOfDelegateSetNpe() 93 Results<Integer> results = Results.create(producer.get().get()); in delegateElementNpe() local 94 assertThat(results.successes).isEmpty(); in delegateElementNpe() [all …]
|
/external/autotest/server/ |
D | autoserv.py | 124 @param resultsdir: Folder to store results. This could be different from 125 parser.options.results: parser.options.results can be set to None 126 for results to be stored in a temp folder. resultsdir can be None 154 def _run_with_ssp(job, container_id, job_id, results, parser, ssp_url, argument 161 @param results: Folder to store results. This could be different from 162 parser.options.results: 163 parser.options.results can be set to None for results to be 165 results can be None for autoserv run requires no logging. 181 results, control=control, 212 if parser.options.results: [all …]
|
D | autoserv | 124 @param resultsdir: Folder to store results. This could be different from 125 parser.options.results: parser.options.results can be set to None 126 for results to be stored in a temp folder. resultsdir can be None 154 def _run_with_ssp(job, container_id, job_id, results, parser, ssp_url, argument 161 @param results: Folder to store results. This could be different from 162 parser.options.results: 163 parser.options.results can be set to None for results to be 165 results can be None for autoserv run requires no logging. 181 results, control=control, 212 if parser.options.results: [all …]
|
/external/skia/ |
D | PRESUBMIT.py | 79 results = [] 81 results.append( 85 return results 119 results = [] 121 results.append( 126 return results 130 results = [] 142 results.append(output_api.PresubmitError( 144 return results 149 results = [] [all …]
|
/external/skqp/ |
D | PRESUBMIT.py | 79 results = [] 81 results.append( 85 return results 119 results = [] 121 results.append( 126 return results 130 results = [] 142 results.append(output_api.PresubmitError( 144 return results 149 results = [] [all …]
|
/external/eigen/bench/ |
D | dense_solvers.cpp | 10 std::map<std::string,Array<float,1,8,DontAlign|RowMajor> > results; variable 74 results["LLT"][id] = t_llt.best(); in bench() 75 results["LDLT"][id] = t_ldlt.best(); in bench() 76 results["PartialPivLU"][id] = t_lu.best(); in bench() 77 results["FullPivLU"][id] = t_fplu.best(); in bench() 78 results["HouseholderQR"][id] = t_qr.best(); in bench() 79 results["ColPivHouseholderQR"][id] = t_cpqr.best(); in bench() 80 results["CompleteOrthogonalDecomposition"][id] = t_cod.best(); in bench() 81 results["FullPivHouseholderQR"][id] = t_fpqr.best(); in bench() 82 results["JacobiSVD"][id] = t_jsvd.best(); in bench() [all …]
|
/external/tensorflow/tensorflow/python/training/ |
D | input_test.py | 495 results = self.evaluate(batched_fetch) 496 self.assertAllEqual(results[0], 499 results[1].indices, 506 self.assertAllEqual(results[1].values, expected) 507 self.assertAllEqual(results[1].dense_shape, [batch_size, 2]) 508 self.assertAllEqual(results[2], [b"string"] * batch_size) 566 results = self.evaluate(batched) 569 self.assertAllEqual(results[0], expected_results) 572 self.assertAllEqual(results[1], expected_strings) 599 results = self.evaluate(batched) [all …]
|
/external/autotest/client/site_tests/platform_BootPerf/ |
D | platform_BootPerf.py | 117 """Copy raw data files to the test results.""" 202 def _gather_vboot_times(self, results): argument 211 seconds and record the following keyvals in `results`: 221 @param results Keyvals dictionary. 231 results['mhz_primary_cpu'] = khz / 1000.0 245 results['seconds_power_on_to_lf_start'] = times[0] 246 results['seconds_power_on_to_lf_end'] = times[1] 247 results['seconds_power_on_to_lk_start'] = times[2] 248 results['seconds_power_on_to_lk_end'] = times[3] 251 def _gather_firmware_boot_time(self, results): argument [all …]
|
/external/autotest/client/tests/ltp/ |
D | ltp-diff.py | 5 # Input: Two or more files containing results from different executions of 7 # of the ltp.results file. 11 # where the results of those runs differ 13 # 0 if all runs had identical results 14 # Non-zero if results differ, or bad input 23 Note: location[1,2,N] may be local files or URLs of LTP results\n" 28 Download the results if needed. 29 Return results of each run in a numerically-indexed dictionary 41 results = fh.readlines() 44 print "ERROR: reading results resource [%s]" % (file) [all …]
|
/external/autotest/client/site_tests/kernel_LTP/ |
D | ltp-diff.py | 5 # Input: Two or more files containing results from different executions of 7 # of the ltp.results file. 11 # where the results of those runs differ 13 # 0 if all runs had identical results 14 # Non-zero if results differ, or bad input 24 Note: location[1,2,N] may be local files or URLs of LTP results\n" 29 Download the results if needed. 30 Return results of each run in a numerically-indexed dictionary 44 results = fh.readlines() 47 print "ERROR: reading results resource [%s]" % (file) [all …]
|
/external/v8/tools/unittests/ |
D | run_perf_test.py | 149 self._test_output = path.join(TEST_WORKSPACE, "results.json") 151 "--json-test-results", 166 "results": trace["results"], 197 {"name": "Richards", "results": ["1.234"], "stddev": ""}, 198 {"name": "DeltaBlue", "results": ["10657567.0"], "stddev": ""}, 210 {"name": "Richards", "results": ["1.234"], "stddev": ""}, 211 {"name": "DeltaBlue", "results": ["10657567.0"], "stddev": ""}, 228 {"name": "Richards", "results": ["50.0", "100.0"], "stddev": ""}, 229 {"name": "DeltaBlue", "results": ["300.0", "200.0"], "stddev": ""}, 246 {"name": "Richards", "results": ["50.0", "100.0"], "stddev": ""}, [all …]
|
/external/libchrome/base/strings/ |
D | string_split_unittest.cc | 153 std::vector<std::string> results = SplitStringUsingSubstr( in TEST() local 155 ASSERT_EQ(1u, results.size()); in TEST() 156 EXPECT_THAT(results, ElementsAre("")); in TEST() 234 std::vector<std::string> results = SplitStringUsingSubstr( in TEST() local 237 ASSERT_EQ(1u, results.size()); in TEST() 238 EXPECT_THAT(results, ElementsAre("alongwordwithnodelimiter")); in TEST() 242 std::vector<std::string> results = SplitStringUsingSubstr( in TEST() local 245 ASSERT_EQ(6u, results.size()); in TEST() 246 EXPECT_THAT(results, ElementsAre("", "", "", "one", "two", "three")); in TEST() 250 std::vector<std::string> results = SplitStringUsingSubstr( in TEST() local [all …]
|
/external/autotest/client/common_lib/ |
D | gtest_parser.py | 76 status: test results status to search for. 173 results = self._master_name_re.search(line) 174 if results: 175 self.master_name = results.group(1) 178 results = self._disabled.search(line) 179 if results: 181 disabled = int(results.group(1)) 194 results = self._flaky.search(line) 195 if results: 197 flaky = int(results.group(1)) [all …]
|
/external/tensorflow/tensorflow/core/util/ |
D | tensor_slice_set_test.cc | 90 float results[10]; in TEST() local 91 EXPECT_TRUE(tss.Query(s, results)); in TEST() 93 EXPECT_EQ(expected[i], results[i]); in TEST() 105 float results[5]; in TEST() local 106 EXPECT_TRUE(tss.Query(s, results)); in TEST() 108 EXPECT_EQ(expected[i], results[i]); in TEST() 121 float results[6]; in TEST() local 122 EXPECT_TRUE(tss.Query(s, results)); in TEST() 124 EXPECT_EQ(expected[i], results[i]); in TEST() 135 float results[6]; in TEST() local [all …]
|
/external/ltp/utils/ffsb-6.0-rc2/ |
D | ffsb_op.c | 58 void init_ffsb_op_results(ffsb_op_results_t * results) in init_ffsb_op_results() argument 60 memset(results, 0, sizeof(ffsb_op_results_t)); in init_ffsb_op_results() 63 static int exclusive_op(ffsb_op_results_t * results, unsigned int op_num) in exclusive_op() argument 70 ret += results->ops[i]; in exclusive_op() 85 static void print_op_results(unsigned int op_num, ffsb_op_results_t * results, in print_op_results() argument 91 double op_pcnt = 100 * (double)results->ops[op_num] / (double)total_ops; in print_op_results() 92 double weight_pcnt = 100 * (double)results->op_weight[op_num] / in print_op_results() 95 ffsb_printsize(buf, results->bytes[op_num] / runtime, 256); in print_op_results() 99 generic_op_print(ffsb_op_list[op_num].op_name, results->ops[op_num], in print_op_results() 105 ffsb_op_results_t * results, double runtime) [all …]
|
/external/google-benchmark/src/ |
D | benchmark_runner.cc | 66 const internal::ThreadManager::Result& results, size_t memory_iterations, in CreateRunReport() argument 72 report.error_occurred = results.has_error_; in CreateRunReport() 73 report.error_message = results.error_message_; in CreateRunReport() 74 report.report_label = results.report_label_; in CreateRunReport() 76 report.iterations = results.iterations; in CreateRunReport() 81 report.real_accumulated_time = results.manual_time_used; in CreateRunReport() 83 report.real_accumulated_time = results.real_time_used; in CreateRunReport() 85 report.cpu_accumulated_time = results.cpu_time_used; in CreateRunReport() 86 report.complexity_n = results.complexity_n; in CreateRunReport() 90 report.counters = results.counters; in CreateRunReport() [all …]
|
/external/libcxx/utils/google-benchmark/src/ |
D | benchmark_runner.cc | 66 const internal::ThreadManager::Result& results, size_t memory_iterations, in CreateRunReport() argument 72 report.error_occurred = results.has_error_; in CreateRunReport() 73 report.error_message = results.error_message_; in CreateRunReport() 74 report.report_label = results.report_label_; in CreateRunReport() 76 report.iterations = results.iterations; in CreateRunReport() 81 report.real_accumulated_time = results.manual_time_used; in CreateRunReport() 83 report.real_accumulated_time = results.real_time_used; in CreateRunReport() 85 report.cpu_accumulated_time = results.cpu_time_used; in CreateRunReport() 86 report.complexity_n = results.complexity_n; in CreateRunReport() 90 report.counters = results.counters; in CreateRunReport() [all …]
|