/external/autotest/server/cros/network/ |
D | perf_test_manager.py | 46 def get_config(self, test_type, test_time=DEFAULT_TEST_TIME): argument 57 if self._use_iperf and test_type not in [ 62 self._iperf_type_from_perf_type(test_type)) 64 self._netperf_type_from_perf_type(test_type), 68 test_type, argument 87 if self._use_iperf and test_type not in [ 91 if test_type in [ 101 if test_type in [ 113 'Test type %s is not supported by this test.' % test_type) 138 def _netperf_type_from_perf_type(self, test_type): argument [all …]
|
D | netperf_runner.py | 24 def from_netperf_results(test_type, results, duration_seconds): argument 41 if test_type in NetperfConfig.TCP_STREAM_TESTS: 57 result = NetperfResult(test_type, duration_seconds, 59 elif test_type in NetperfConfig.UDP_STREAM_TESTS: 76 result = NetperfResult(test_type, duration_seconds, 79 elif test_type in NetperfConfig.REQUEST_RESPONSE_TESTS: 96 result = NetperfResult(test_type, duration_seconds, 98 elif test_type in NetperfConfig.BIDIRECTIONAL_TESTS: 107 result = NetperfResult(test_type, 112 raise error.TestFail('Invalid netperf test type: %r.' % test_type) [all …]
|
D | expected_performance_results.py | 152 def get_expected_throughput_wifi(test_type, mode, channel_width): argument 164 if test_type in expected_throughput_wifi: 165 if mode in expected_throughput_wifi[test_type]: 166 if channel_width in expected_throughput_wifi[test_type][mode]: 167 return expected_throughput_wifi[test_type][mode][channel_width] 173 % (test_type, mode, ret_mode)) 219 def get_board_max_expectation(test_type, board_name): argument 233 return board_maximums.get(test_type)
|
D | iperf_runner.py | 220 test_type, argument 233 if test_type == IperfConfig.IPERF_TEST_TYPE_TCP_TX: 236 elif test_type == IperfConfig.IPERF_TEST_TYPE_TCP_RX: 239 elif test_type == IperfConfig.IPERF_TEST_TYPE_TCP_BIDIRECTIONAL: 242 elif test_type == IperfConfig.IPERF_TEST_TYPE_UDP_TX: 245 elif test_type == IperfConfig.IPERF_TEST_TYPE_UDP_RX: 248 elif test_type == IperfConfig.IPERF_TEST_TYPE_UDP_BIDIRECTIONAL: 254 test_type) 258 self.test_type = test_type
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | split_test.cc | 80 void Check(TestType test_type, int axis, int num_splits, in Check() argument 92 if (test_type == TestType::kDynamic) { in Check() 130 for (TestType test_type : SplitOpTest<TestType>::range_) { in TYPED_TEST() 131 Check<TypeParam>(/*axis_as_tensor*/ test_type, in TYPED_TEST() 139 Check<TypeParam>(/*axis_as_tensor*/ test_type, in TYPED_TEST() 147 Check<TypeParam>(/*axis_as_tensor*/ test_type, in TYPED_TEST() 155 Check<TypeParam>(/*axis_as_tensor*/ test_type, in TYPED_TEST() 167 for (TestType test_type : SplitOpTest<TestType>::range_) { in TYPED_TEST() 168 Check<TypeParam>(/*axis_as_tensor*/ test_type, in TYPED_TEST() 176 Check<TypeParam>(/*axis_as_tensor*/ test_type, in TYPED_TEST() [all …]
|
D | split_v_test.cc | 91 void Check(TestType test_type, int axis, std::initializer_list<int> input_shape, in Check() argument 99 switch (test_type) { in Check() 149 void Check##TestSuiteName##CaseName(TestType test_type); \ 162 void Check##TestSuiteName##CaseName(TestType test_type) 172 Check<TypeParam>(test_type, in TYPED_SPLIT_V_TEST() 179 Check<TypeParam>(test_type, in TYPED_SPLIT_V_TEST() 187 Check<TypeParam>(test_type, in TYPED_SPLIT_V_TEST() 195 Check<TypeParam>(test_type, in TYPED_SPLIT_V_TEST() 203 Check<TypeParam>(test_type, in TYPED_SPLIT_V_TEST() 214 Check<TypeParam>(test_type, in TYPED_SPLIT_V_TEST() [all …]
|
D | expand_dims_test.cc | 87 for (TestType test_type : ExpandDimsOpTest<TestType>::range_) { in TYPED_TEST() 90 ExpandDimsOpModel<TypeParam> axis_0(0, {2, 2}, values, test_type); in TYPED_TEST() 95 ExpandDimsOpModel<TypeParam> axis_1(1, {2, 2}, values, test_type); in TYPED_TEST() 100 ExpandDimsOpModel<TypeParam> axis_2(2, {2, 2}, values, test_type); in TYPED_TEST() 108 for (TestType test_type : ExpandDimsOpTest<TestType>::range_) { in TYPED_TEST() 111 ExpandDimsOpModel<TypeParam> m(-1, {2, 2}, values, test_type); in TYPED_TEST()
|
/external/cronet/buildtools/third_party/libc++/trunk/test/std/containers/sequences/array/ |
D | size_and_alignment.pass.cpp | 44 void test_type() { in test_type() function 65 test_type<char>(); in main() 66 test_type<int>(); in main() 67 test_type<double>(); in main() 68 test_type<long double>(); in main() 71 test_type<std::max_align_t>(); in main() 72 test_type<TestType1>(); in main() 73 test_type<TestType2>(); in main() 74 test_type<TestType3>(); in main()
|
/external/autotest/client/bin/ |
D | fio_util.py | 81 test_type = '' variable in fio_graph_generator 149 def _gen_data_row(cls, test_type, pass_list, percentile): argument 168 def _write_data(cls, f, test_type, pass_list, percentile): argument 179 row = cls._gen_data_row(test_type, pass_list, percentile) 187 def _write_option(cls, f, test_name, test_type, percentile): argument 199 (cls.graph_title[test_type], test_name)) 202 (cls.graph_title[test_type], test_name)) 205 option['vAxis'] = {'title': cls.v_title[test_type]} 212 def _write_graph(cls, test_name, test_type, pass_list, percentile=False): argument 222 test_name, test_type, str(pass_list)) [all …]
|
/external/libcxx/test/std/containers/sequences/array/ |
D | size_and_alignment.pass.cpp | 47 void test_type() { in test_type() function 64 test_type<char>(); in main() 65 test_type<int>(); in main() 66 test_type<double>(); in main() 67 test_type<long double>(); in main() 68 test_type<std::max_align_t>(); in main() 69 test_type<TestType1>(); in main() 70 test_type<TestType2>(); in main()
|
/external/cronet/build/android/pylib/results/ |
D | report_results.py | 17 def _LogToFile(results, test_type, suite_name): argument 23 log_file_path, re.sub(r'\W', '_', test_type).lower() + '.log') 28 (test_type, os.environ.get('BUILDBOT_BUILDERNAME'), 41 def _LogToFlakinessDashboard(results, test_type, test_package, argument 45 test_type, test_package, flakiness_server) 50 if test_type in ('instrumentation', 'Instrumentation'): 63 elif test_type == 'gtest': 77 def LogFull(results, test_type, test_package, annotation=None, argument 131 _LogToFile(results, test_type, suite_name) 134 _LogToFlakinessDashboard(results, test_type, test_package,
|
/external/angle/build/android/pylib/results/ |
D | report_results.py | 17 def _LogToFile(results, test_type, suite_name): argument 23 log_file_path, re.sub(r'\W', '_', test_type).lower() + '.log') 28 (test_type, os.environ.get('BUILDBOT_BUILDERNAME'), 41 def _LogToFlakinessDashboard(results, test_type, test_package, argument 45 test_type, test_package, flakiness_server) 50 if test_type in ('instrumentation', 'Instrumentation'): 63 elif test_type == 'gtest': 77 def LogFull(results, test_type, test_package, annotation=None, argument 131 _LogToFile(results, test_type, suite_name) 134 _LogToFlakinessDashboard(results, test_type, test_package,
|
/external/angle/build/android/pylib/base/ |
D | base_test_result.py | 43 def __init__(self, name, test_type, duration=0, log='', failure_reason=None): argument 53 assert test_type in ResultType.GetTypes() 55 self._test_type = test_type 99 def SetType(self, test_type): argument 101 assert test_type in ResultType.GetTypes() 102 self._test_type = test_type 171 for test_type in ResultType.GetTypes(): 172 if test_type != ResultType.PASS: 173 for t in sorted(self._GetType(test_type)): 176 s.append('[%s] %s:' % (test_type, t)) [all …]
|
/external/cronet/build/android/pylib/base/ |
D | base_test_result.py | 43 def __init__(self, name, test_type, duration=0, log='', failure_reason=None): argument 53 assert test_type in ResultType.GetTypes() 55 self._test_type = test_type 99 def SetType(self, test_type): argument 101 assert test_type in ResultType.GetTypes() 102 self._test_type = test_type 171 for test_type in ResultType.GetTypes(): 172 if test_type != ResultType.PASS: 173 for t in sorted(self._GetType(test_type)): 176 s.append('[%s] %s:' % (test_type, t)) [all …]
|
/external/clang/test/CXX/dcl.dcl/dcl.attr/dcl.attr.noreturn/ |
D | p1.cpp | 30 template<typename T> void test_type(T) { T::error; } // expected-error {{has no members}} in test_type() function 31 template<> void test_type(int (*)()) {} in test_type() function 40 test_type(e); in check() 41 test_type(f); in check() 42 test_type(g); in check() 43 test_type(h); // expected-note {{instantiation}} in check()
|
/external/autotest/server/site_tests/network_WiFi_Perf/ |
D | network_WiFi_Perf.py | 78 should_expected_throughput, test_type, failed_test_types, argument 97 test_type, self.context.client.board) 104 test_type, must_expected_throughput, result.throughput) 110 test_type, should_expected_throughput, 116 test_type, should_expected_throughput, 120 '[test_type=%s' % test_type, 185 for test_type in self.PERF_TEST_TYPES: 186 config = manager.get_config(test_type) 189 session = manager.get_session(test_type, 199 test_type, ap_config.mode, ch_width) [all …]
|
/external/vixl/tools/test_generator/ |
D | generator.py | 312 def __init__(self, test_name, test_isa, test_type, mnemonics, operands, argument 316 self.test_type = test_type 322 if self.test_type == "simulator": 331 if self.test_type in ["simulator", "macro-assembler"]: 475 if self.test_type == "simulator": 477 elif self.test_type == "assembler": 479 elif self.test_type == "macro-assembler": 481 elif self.test_type == "assembler-negative": 484 raise Exception("Unrecognized test type \"{}\".".format(self.test_type)) 529 if self.test_type in ["simulator", "macro-assembler"]: [all …]
|
/external/rust/crates/ring/src/rsa/ |
D | convert_nist_rsa_test_vectors.py | 146 def main(fn, test_type, padding_alg): argument 158 if test_type == "verify": 160 elif test_type == "sign": 163 print("Invalid test_type: %s" % test_type) 174 if test_type == "verify": 205 if test_type == 'sign': 239 test_type = 'sign' variable 241 test_type = 'verify' variable 246 main(sys.argv[1], test_type, pad_alg)
|
/external/webrtc/tools_webrtc/mb/ |
D | mb.py | 52 test_type = isolate_map[target]['type'] 60 if test_type == 'nontest': 63 if test_type not in ('console_test_launcher', 'windowed_test_launcher', 68 (target, test_type), 91 elif is_ios or is_fuchsia or test_type == 'raw': 109 xvfb = use_x11 and test_type == 'windowed_test_launcher' 133 if test_type == 'non_parallel_console_test_launcher':
|
/external/OpenCL-CTS/test_conformance/basic/ |
D | test_vector_swizzle.cpp | 678 static int test_type(const char* type_name, cl_device_id device, in test_type() function 694 result |= test_type<cl_char>("char", device, context, queue); in test_vector_swizzle() 695 result |= test_type<cl_uchar>("uchar", device, context, queue); in test_vector_swizzle() 696 result |= test_type<cl_short>("short", device, context, queue); in test_vector_swizzle() 697 result |= test_type<cl_ushort>("ushort", device, context, queue); in test_vector_swizzle() 698 result |= test_type<cl_int>("int", device, context, queue); in test_vector_swizzle() 699 result |= test_type<cl_uint>("uint", device, context, queue); in test_vector_swizzle() 702 result |= test_type<cl_long>("long", device, context, queue); in test_vector_swizzle() 703 result |= test_type<cl_ulong>("ulong", device, context, queue); in test_vector_swizzle() 705 result |= test_type<cl_float>("float", device, context, queue); in test_vector_swizzle() [all …]
|
/external/igt-gpu-tools/tests/ |
D | kms_atomic_interruptible.c | 42 static int block_plane(igt_display_t *display, igt_output_t *output, enum plane_test_type test_type… in block_plane() argument 50 if (test_type == test_legacy_modeset || test_type == test_atomic_modeset) { in block_plane() 79 enum plane_test_type test_type, unsigned plane_type) in run_plane_test() argument 113 if (test_type != test_legacy_modeset && test_type != test_atomic_modeset) { in run_plane_test() 120 block = block_plane(display, output, test_type, plane); in run_plane_test() 138 switch (test_type) { in run_plane_test() 259 if (test_type == test_legacy_modeset || test_type == test_atomic_modeset) in run_plane_test()
|
/external/ltp/testcases/network/stress/route/ |
D | route6-rmmod | 197 test_type=$1 199 TCID=route6-rmmod0${test_type} 200 TST_COUNT=$test_type 202 case $test_type in 226 case $test_type in
|
/external/angle/build/android/pylib/results/flakiness_dashboard/ |
D | results_uploader.py | 27 test_results_map, test_results_server, test_type, master_name): argument 37 test_type=test_type, 151 test_type=self._tests_type, 164 def Upload(results, flakiness_dashboard_server, test_type): argument 172 uploader = ResultsUploader(test_type)
|
/external/cronet/build/android/pylib/results/flakiness_dashboard/ |
D | results_uploader.py | 27 test_results_map, test_results_server, test_type, master_name): argument 37 test_type=test_type, 151 test_type=self._tests_type, 164 def Upload(results, flakiness_dashboard_server, test_type): argument 172 uploader = ResultsUploader(test_type)
|
/external/igt-gpu-tools/lib/tests/ |
D | igt_exit_handler.c | 55 enum test_type { enum 63 static int testfunc(enum test_type test_type) in testfunc() argument 88 switch (test_type) { in testfunc()
|