/external/grpc-grpc-java/benchmarks/src/generated/main/grpc/io/grpc/benchmarks/proto/ |
D | BenchmarkServiceGrpc.java | 1 package io.grpc.benchmarks.proto; 30 private static volatile io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Messages.SimpleRequest, 31 io.grpc.benchmarks.proto.Messages.SimpleResponse> getUnaryCallMethod; 35 requestType = io.grpc.benchmarks.proto.Messages.SimpleRequest.class, 36 responseType = io.grpc.benchmarks.proto.Messages.SimpleResponse.class, 38 public static io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Messages.SimpleRequest, 39 io.grpc.benchmarks.proto.Messages.SimpleResponse> getUnaryCallMethod() { in getUnaryCallMethod() 40 …io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Messages.SimpleRequest, io.grpc.benchmarks.proto… in getUnaryCallMethod() 45 …io.grpc.MethodDescriptor.<io.grpc.benchmarks.proto.Messages.SimpleRequest, io.grpc.benchmarks.prot… in getUnaryCallMethod() 51 io.grpc.benchmarks.proto.Messages.SimpleRequest.getDefaultInstance())) in getUnaryCallMethod() [all …]
|
D | WorkerServiceGrpc.java | 1 package io.grpc.benchmarks.proto; 30 private static volatile io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ServerArgs, 31 io.grpc.benchmarks.proto.Control.ServerStatus> getRunServerMethod; 35 requestType = io.grpc.benchmarks.proto.Control.ServerArgs.class, 36 responseType = io.grpc.benchmarks.proto.Control.ServerStatus.class, 38 public static io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ServerArgs, 39 io.grpc.benchmarks.proto.Control.ServerStatus> getRunServerMethod() { in getRunServerMethod() 40 …io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ServerArgs, io.grpc.benchmarks.proto.Con… in getRunServerMethod() 45 …io.grpc.MethodDescriptor.<io.grpc.benchmarks.proto.Control.ServerArgs, io.grpc.benchmarks.proto.Co… in getRunServerMethod() 51 io.grpc.benchmarks.proto.Control.ServerArgs.getDefaultInstance())) in getRunServerMethod() [all …]
|
D | ReportQpsScenarioServiceGrpc.java | 1 package io.grpc.benchmarks.proto; 30 private static volatile io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ScenarioResult, 31 io.grpc.benchmarks.proto.Control.Void> getReportScenarioMethod; 35 requestType = io.grpc.benchmarks.proto.Control.ScenarioResult.class, 36 responseType = io.grpc.benchmarks.proto.Control.Void.class, 38 public static io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ScenarioResult, 39 io.grpc.benchmarks.proto.Control.Void> getReportScenarioMethod() { in getReportScenarioMethod() 40 …io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ScenarioResult, io.grpc.benchmarks.proto… in getReportScenarioMethod() 45 …io.grpc.MethodDescriptor.<io.grpc.benchmarks.proto.Control.ScenarioResult, io.grpc.benchmarks.prot… in getReportScenarioMethod() 51 io.grpc.benchmarks.proto.Control.ScenarioResult.getDefaultInstance())) in getReportScenarioMethod() [all …]
|
/external/grpc-grpc-java/benchmarks/src/generated/main/java/io/grpc/benchmarks/proto/ |
D | Payloads.java | 4 package io.grpc.benchmarks.proto; 103 … return io.grpc.benchmarks.proto.Payloads.internal_static_grpc_testing_ByteBufferParams_descriptor; in getDescriptor() 108 …return io.grpc.benchmarks.proto.Payloads.internal_static_grpc_testing_ByteBufferParams_fieldAccess… in internalGetFieldAccessorTable() 110 …io.grpc.benchmarks.proto.Payloads.ByteBufferParams.class, io.grpc.benchmarks.proto.Payloads.ByteBu… in internalGetFieldAccessorTable() 175 if (!(obj instanceof io.grpc.benchmarks.proto.Payloads.ByteBufferParams)) { in equals() 178 …io.grpc.benchmarks.proto.Payloads.ByteBufferParams other = (io.grpc.benchmarks.proto.Payloads.Byte… in equals() 205 public static io.grpc.benchmarks.proto.Payloads.ByteBufferParams parseFrom( in parseFrom() 210 public static io.grpc.benchmarks.proto.Payloads.ByteBufferParams parseFrom( in parseFrom() 216 public static io.grpc.benchmarks.proto.Payloads.ByteBufferParams parseFrom( in parseFrom() 221 public static io.grpc.benchmarks.proto.Payloads.ByteBufferParams parseFrom( in parseFrom() [all …]
|
D | Control.java | 4 package io.grpc.benchmarks.proto; 116 return io.grpc.benchmarks.proto.Control.getDescriptor().getEnumTypes().get(0); in getDescriptor() 240 return io.grpc.benchmarks.proto.Control.getDescriptor().getEnumTypes().get(1); in getDescriptor() 365 return io.grpc.benchmarks.proto.Control.getDescriptor().getEnumTypes().get(2); in getDescriptor() 475 return io.grpc.benchmarks.proto.Control.internal_static_grpc_testing_PoissonParams_descriptor; in getDescriptor() 480 …return io.grpc.benchmarks.proto.Control.internal_static_grpc_testing_PoissonParams_fieldAccessorTa… in internalGetFieldAccessorTable() 482 …io.grpc.benchmarks.proto.Control.PoissonParams.class, io.grpc.benchmarks.proto.Control.PoissonPara… in internalGetFieldAccessorTable() 535 if (!(obj instanceof io.grpc.benchmarks.proto.Control.PoissonParams)) { in equals() 538 …io.grpc.benchmarks.proto.Control.PoissonParams other = (io.grpc.benchmarks.proto.Control.PoissonPa… in equals() 564 public static io.grpc.benchmarks.proto.Control.PoissonParams parseFrom( in parseFrom() [all …]
|
D | Messages.java | 4 package io.grpc.benchmarks.proto; 93 return io.grpc.benchmarks.proto.Messages.getDescriptor().getEnumTypes().get(0); in getDescriptor() 204 return io.grpc.benchmarks.proto.Messages.internal_static_grpc_testing_BoolValue_descriptor; in getDescriptor() 209 … return io.grpc.benchmarks.proto.Messages.internal_static_grpc_testing_BoolValue_fieldAccessorTable in internalGetFieldAccessorTable() 211 …io.grpc.benchmarks.proto.Messages.BoolValue.class, io.grpc.benchmarks.proto.Messages.BoolValue.Bui… in internalGetFieldAccessorTable() 264 if (!(obj instanceof io.grpc.benchmarks.proto.Messages.BoolValue)) { in equals() 267 …io.grpc.benchmarks.proto.Messages.BoolValue other = (io.grpc.benchmarks.proto.Messages.BoolValue) … in equals() 291 public static io.grpc.benchmarks.proto.Messages.BoolValue parseFrom( in parseFrom() 296 public static io.grpc.benchmarks.proto.Messages.BoolValue parseFrom( in parseFrom() 302 public static io.grpc.benchmarks.proto.Messages.BoolValue parseFrom( in parseFrom() [all …]
|
D | Stats.java | 4 package io.grpc.benchmarks.proto; 172 return io.grpc.benchmarks.proto.Stats.internal_static_grpc_testing_ServerStats_descriptor; in getDescriptor() 177 … return io.grpc.benchmarks.proto.Stats.internal_static_grpc_testing_ServerStats_fieldAccessorTable in internalGetFieldAccessorTable() 179 …io.grpc.benchmarks.proto.Stats.ServerStats.class, io.grpc.benchmarks.proto.Stats.ServerStats.Build… in internalGetFieldAccessorTable() 333 if (!(obj instanceof io.grpc.benchmarks.proto.Stats.ServerStats)) { in equals() 336 …io.grpc.benchmarks.proto.Stats.ServerStats other = (io.grpc.benchmarks.proto.Stats.ServerStats) ob… in equals() 391 public static io.grpc.benchmarks.proto.Stats.ServerStats parseFrom( in parseFrom() 396 public static io.grpc.benchmarks.proto.Stats.ServerStats parseFrom( in parseFrom() 402 public static io.grpc.benchmarks.proto.Stats.ServerStats parseFrom( in parseFrom() 407 public static io.grpc.benchmarks.proto.Stats.ServerStats parseFrom( in parseFrom() [all …]
|
D | Services.java | 4 package io.grpc.benchmarks.proto; 62 io.grpc.benchmarks.proto.Messages.getDescriptor(), in internalBuildGeneratedFileFrom() 63 io.grpc.benchmarks.proto.Control.getDescriptor(), in internalBuildGeneratedFileFrom() 65 io.grpc.benchmarks.proto.Messages.getDescriptor(); in io.grpc.benchmarks.proto.Messages.getDescriptor() 66 io.grpc.benchmarks.proto.Control.getDescriptor(); in io.grpc.benchmarks.proto.Control.getDescriptor()
|
/external/eigen/bench/ |
D | benchmark-blocking-sizes.cpp | 345 void serialize_benchmarks(const char* filename, const vector<benchmark_t>& benchmarks, size_t first… in serialize_benchmarks() argument 353 size_t benchmarks_vector_size = benchmarks.size(); in serialize_benchmarks() 357 fwrite(benchmarks.data(), sizeof(benchmark_t), benchmarks.size(), file); in serialize_benchmarks() 361 bool deserialize_benchmarks(const char* filename, vector<benchmark_t>& benchmarks, size_t& first_be… in deserialize_benchmarks() argument 377 benchmarks.resize(benchmarks_vector_size); in deserialize_benchmarks() 378 if (benchmarks.size() != fread(benchmarks.data(), sizeof(benchmark_t), benchmarks.size(), file)) { in deserialize_benchmarks() 386 vector<benchmark_t>& benchmarks, in try_run_some_benchmarks() argument 390 if (first_benchmark_to_run == benchmarks.size()) { in try_run_some_benchmarks() 401 float ratio_done = float(benchmark_index) / benchmarks.size(); in try_run_some_benchmarks() 405 if (benchmark_index == benchmarks.size() || in try_run_some_benchmarks() [all …]
|
/external/okhttp/okio/benchmarks/ |
D | README.md | 4 … used to measure various aspects of performance for Okio buffers. Okio benchmarks are written usin… 9 To run benchmarks locally, first build and package the project modules: 15 This should create a `benchmarks.jar` file in the `target` directory, which is a typical JMH benchm… 18 $ java -jar benchmarks/target/benchmarks.jar -l 20 com.squareup.okio.benchmarks.BufferPerformanceBench.cold 21 com.squareup.okio.benchmarks.BufferPerformanceBench.threads16hot 22 com.squareup.okio.benchmarks.BufferPerformanceBench.threads1hot 23 com.squareup.okio.benchmarks.BufferPerformanceBench.threads2hot 24 com.squareup.okio.benchmarks.BufferPerformanceBench.threads32hot 25 com.squareup.okio.benchmarks.BufferPerformanceBench.threads4hot [all …]
|
/external/toolchain-utils/crosperf/ |
D | results_organizer_unittest.py | 138 benchmarks = [mock_instance.benchmark1, mock_instance.benchmark2] 140 benchmark_runs[0] = BenchmarkRun('b1', benchmarks[0], labels[0], 1, '', '', 142 benchmark_runs[1] = BenchmarkRun('b2', benchmarks[0], labels[0], 2, '', '', 144 benchmark_runs[2] = BenchmarkRun('b3', benchmarks[0], labels[1], 1, '', '', 146 benchmark_runs[3] = BenchmarkRun('b4', benchmarks[0], labels[1], 2, '', '', 148 benchmark_runs[4] = BenchmarkRun('b5', benchmarks[1], labels[0], 1, '', '', 150 benchmark_runs[5] = BenchmarkRun('b6', benchmarks[1], labels[0], 2, '', '', 152 benchmark_runs[6] = BenchmarkRun('b7', benchmarks[1], labels[1], 1, '', '', 154 benchmark_runs[7] = BenchmarkRun('b8', benchmarks[1], labels[1], 2, '', '', 163 organized = OrganizeResults(benchmark_runs, labels, benchmarks)
|
D | experiment_factory.py | 94 def AppendBenchmarkSet(self, benchmarks, benchmark_list, test_args, argument 102 benchmarks.append(telemetry_benchmark) 157 benchmarks = [] 172 self.AppendBenchmarkSet(benchmarks, telemetry_perfv2_tests, test_args, 176 self.AppendBenchmarkSet(benchmarks, telemetry_pagecycler_tests, 181 self.AppendBenchmarkSet(benchmarks, telemetry_toolchain_perf_tests, 186 benchmarks.append( 199 self.AppendBenchmarkSet(benchmarks, 207 benchmarks.append(benchmark) 211 benchmarks, [all …]
|
D | experiment_factory_unittest.py | 56 self.assertEqual(len(exp.benchmarks), 1) 57 self.assertEqual(exp.benchmarks[0].name, 'PageCycler') 58 self.assertEqual(exp.benchmarks[0].test_name, 'PageCycler') 59 self.assertEqual(exp.benchmarks[0].iterations, 3) 160 self.assertEqual(len(exp.benchmarks), 1) 161 self.assertEqual(exp.benchmarks[0].name, 'kraken') 162 self.assertEqual(exp.benchmarks[0].test_name, 'kraken') 163 self.assertEqual(exp.benchmarks[0].iterations, 1) 164 self.assertEqual(exp.benchmarks[0].suite, 'telemetry_Crosperf') 165 self.assertFalse(exp.benchmarks[0].show_all_results)
|
D | results_organizer.py | 20 def _AdjustIteration(benchmarks, max_dup, bench): argument 22 for benchmark in benchmarks: 93 def _DuplicatePass(result, benchmarks): argument 102 _AdjustIteration(benchmarks, max_dup, bench) 154 def OrganizeResults(benchmark_runs, labels, benchmarks=None, json_report=False): argument 172 if benchmarks is None: 173 benchmarks = [] 209 _DuplicatePass(result, benchmarks)
|
/external/grpc-grpc-java/benchmarks/src/main/java/io/grpc/benchmarks/qps/ |
D | OpenLoopClient.java | 17 package io.grpc.benchmarks.qps; 20 import static io.grpc.benchmarks.Utils.HISTOGRAM_MAX_VALUE; 21 import static io.grpc.benchmarks.Utils.HISTOGRAM_PRECISION; 22 import static io.grpc.benchmarks.Utils.saveHistogram; 23 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.ADDRESS; 24 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.CLIENT_PAYLOAD; 25 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.DURATION; 26 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.FLOW_CONTROL_WINDOW; 27 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.SAVE_HISTOGRAM; 28 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.SERVER_PAYLOAD; [all …]
|
D | AsyncClient.java | 17 package io.grpc.benchmarks.qps; 19 import static io.grpc.benchmarks.Utils.HISTOGRAM_MAX_VALUE; 20 import static io.grpc.benchmarks.Utils.HISTOGRAM_PRECISION; 21 import static io.grpc.benchmarks.Utils.saveHistogram; 22 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.ADDRESS; 23 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.CHANNELS; 24 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.CLIENT_PAYLOAD; 25 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.DIRECTEXECUTOR; 26 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.DURATION; 27 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.FLOW_CONTROL_WINDOW; [all …]
|
/external/grpc-grpc/tools/run_tests/ |
D | run_microbenchmark.py | 80 benchmarks = [] 92 benchmarks.append( 117 if len(benchmarks) >= min(16, multiprocessing.cpu_count()): 121 benchmarks, maxjobs=max(1, 125 benchmarks = [] 129 if len(benchmarks): 130 jobset.run(benchmarks, maxjobs=max(1, multiprocessing.cpu_count() / 2)) 142 benchmarks = [] 148 benchmarks.append( 172 if len(benchmarks) >= 20: [all …]
|
/external/grpc-grpc/tools/profiling/microbenchmarks/bm_diff/ |
D | bm_build.py | 57 def _make_cmd(cfg, benchmarks, jobs): argument 58 return ['make'] + benchmarks + ['CONFIG=%s' % cfg, '-j', '%d' % jobs] 61 def build(name, benchmarks, jobs, counters): argument 65 subprocess.check_call(_make_cmd('opt', benchmarks, jobs)) 67 subprocess.check_call(_make_cmd('counters', benchmarks, jobs)) 70 subprocess.check_call(_make_cmd('opt', benchmarks, jobs)) 72 subprocess.check_call(_make_cmd('counters', benchmarks, jobs)) 81 build(args.name, args.benchmarks, args.jobs, args.counters)
|
/external/tensorflow/tensorflow/contrib/eager/python/examples/resnet50/ |
D | README.md | 7 - `resnet50_test.py`: Sanity unittests and benchmarks for using the model with 9 - `resnet50_graph_test.py`: Sanity unittests and benchmarks when using the same 18 python resnet50_test.py --benchmarks=. 21 python resnet50_graph_test.py --benchmarks=. 25 package. To build (and run benchmarks) from source: 29 bazel run -c opt --config=cuda :resnet50_test -- --benchmarks=. 32 bazel run -c opt --config=cuda :resnet50_graph_test -- --benchmarks=. 37 On October 31, 2017, the benchmarks demonstrated comparable performance
|
/external/v8/benchmarks/ |
D | base.js | 66 function BenchmarkSuite(name, reference, benchmarks) { argument 69 this.benchmarks = benchmarks; 141 result += suites[i].benchmarks.length; 238 var length = this.benchmarks.length; 250 suite.benchmarks[index].Setup(); 263 data = suite.RunSingleBenchmark(suite.benchmarks[index], data); 274 suite.benchmarks[index++].TearDown();
|
/external/tensorflow/tensorflow/tools/test/ |
D | performance.bzl | 9 benchmarks = "..", 40 "--test_args=--benchmarks=%s" % benchmarks, 56 benchmarks = "..", 65 benchmarks = benchmarks,
|
/external/tensorflow/tensorflow/contrib/eager/python/examples/rnn_ptb/ |
D | README.md | 23 python rnn_ptb_test.py --benchmarks=. 26 python rnn_ptb_graph_test.py --benchmarks=. 30 package. To build (and run benchmarks) from source: 35 bazel run -c opt --config=cuda :rnn_ptb_test -- --benchmarks=. 38 bazel run -c opt --config=cuda :rnn_ptb_graph_test -- --benchmarks=. 43 On October 31, 2017, the benchmarks demonstrated slightly better performance
|
/external/protobuf/benchmarks/ |
D | readme.txt | 10 If we end up with a lot of different benchmarks it may be worth 29 $ javac -d . -cp ../protobuf.jar benchmarks/*.java 34 benchmarks.GoogleSize$SizeMessage1 ../google_message1.dat 35 benchmarks.GoogleSpeed$SpeedMessage1 ../google_message1.dat 36 benchmarks.GoogleSize$SizeMessage2 ../google_message2.dat 37 benchmarks.GoogleSpeed$SpeedMessage2 ../google_message2.dat
|
/external/v8/tools/ |
D | try_perf.py | 66 benchmarks = ['"%s"' % benchmark for benchmark in options.benchmarks] 67 cmd += ['-p \'testfilter=[%s]\'' % ','.join(benchmarks)] 90 if not options.benchmarks: 94 for benchmark in options.benchmarks:
|
/external/libcxx/docs/ |
D | TestingLibcxx.rst | 205 The benchmarks are written using the `Google Benchmark`_ library, a copy of which 216 The benchmark tests are not built by default. The benchmarks can be built using 217 the ``cxx-benchmarks`` target. 225 $ make cxx-benchmarks 227 This will build all of the benchmarks under ``<libcxx-src>/benchmarks`` to be 229 ``build/benchmarks``. 231 The benchmarks can also be built against the platforms native standard library 234 The compiled benchmarks are named ``<test>.libcxx.out`` if they test libc++ and 245 The benchmarks must be run manually by the user. Currently there is no way 252 $ cd build/benchmarks [all …]
|