Home
last modified time | relevance | path

Searched refs:benchmarks (Results 1 – 25 of 260) sorted by relevance

1234567891011

/external/grpc-grpc-java/benchmarks/src/generated/main/grpc/io/grpc/benchmarks/proto/
DBenchmarkServiceGrpc.java1 package io.grpc.benchmarks.proto;
30 private static volatile io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Messages.SimpleRequest,
31 io.grpc.benchmarks.proto.Messages.SimpleResponse> getUnaryCallMethod;
35 requestType = io.grpc.benchmarks.proto.Messages.SimpleRequest.class,
36 responseType = io.grpc.benchmarks.proto.Messages.SimpleResponse.class,
38 public static io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Messages.SimpleRequest,
39 io.grpc.benchmarks.proto.Messages.SimpleResponse> getUnaryCallMethod() { in getUnaryCallMethod()
40 …io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Messages.SimpleRequest, io.grpc.benchmarks.proto… in getUnaryCallMethod()
45 …io.grpc.MethodDescriptor.<io.grpc.benchmarks.proto.Messages.SimpleRequest, io.grpc.benchmarks.prot… in getUnaryCallMethod()
51 io.grpc.benchmarks.proto.Messages.SimpleRequest.getDefaultInstance())) in getUnaryCallMethod()
[all …]
DWorkerServiceGrpc.java1 package io.grpc.benchmarks.proto;
30 private static volatile io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ServerArgs,
31 io.grpc.benchmarks.proto.Control.ServerStatus> getRunServerMethod;
35 requestType = io.grpc.benchmarks.proto.Control.ServerArgs.class,
36 responseType = io.grpc.benchmarks.proto.Control.ServerStatus.class,
38 public static io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ServerArgs,
39 io.grpc.benchmarks.proto.Control.ServerStatus> getRunServerMethod() { in getRunServerMethod()
40 …io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ServerArgs, io.grpc.benchmarks.proto.Con… in getRunServerMethod()
45 …io.grpc.MethodDescriptor.<io.grpc.benchmarks.proto.Control.ServerArgs, io.grpc.benchmarks.proto.Co… in getRunServerMethod()
51 io.grpc.benchmarks.proto.Control.ServerArgs.getDefaultInstance())) in getRunServerMethod()
[all …]
DReportQpsScenarioServiceGrpc.java1 package io.grpc.benchmarks.proto;
30 private static volatile io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ScenarioResult,
31 io.grpc.benchmarks.proto.Control.Void> getReportScenarioMethod;
35 requestType = io.grpc.benchmarks.proto.Control.ScenarioResult.class,
36 responseType = io.grpc.benchmarks.proto.Control.Void.class,
38 public static io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ScenarioResult,
39 io.grpc.benchmarks.proto.Control.Void> getReportScenarioMethod() { in getReportScenarioMethod()
40 …io.grpc.MethodDescriptor<io.grpc.benchmarks.proto.Control.ScenarioResult, io.grpc.benchmarks.proto… in getReportScenarioMethod()
45 …io.grpc.MethodDescriptor.<io.grpc.benchmarks.proto.Control.ScenarioResult, io.grpc.benchmarks.prot… in getReportScenarioMethod()
51 io.grpc.benchmarks.proto.Control.ScenarioResult.getDefaultInstance())) in getReportScenarioMethod()
[all …]
/external/grpc-grpc-java/benchmarks/src/generated/main/java/io/grpc/benchmarks/proto/
DPayloads.java4 package io.grpc.benchmarks.proto;
103 … return io.grpc.benchmarks.proto.Payloads.internal_static_grpc_testing_ByteBufferParams_descriptor; in getDescriptor()
108 …return io.grpc.benchmarks.proto.Payloads.internal_static_grpc_testing_ByteBufferParams_fieldAccess… in internalGetFieldAccessorTable()
110 …io.grpc.benchmarks.proto.Payloads.ByteBufferParams.class, io.grpc.benchmarks.proto.Payloads.ByteBu… in internalGetFieldAccessorTable()
175 if (!(obj instanceof io.grpc.benchmarks.proto.Payloads.ByteBufferParams)) { in equals()
178 …io.grpc.benchmarks.proto.Payloads.ByteBufferParams other = (io.grpc.benchmarks.proto.Payloads.Byte… in equals()
205 public static io.grpc.benchmarks.proto.Payloads.ByteBufferParams parseFrom( in parseFrom()
210 public static io.grpc.benchmarks.proto.Payloads.ByteBufferParams parseFrom( in parseFrom()
216 public static io.grpc.benchmarks.proto.Payloads.ByteBufferParams parseFrom( in parseFrom()
221 public static io.grpc.benchmarks.proto.Payloads.ByteBufferParams parseFrom( in parseFrom()
[all …]
DControl.java4 package io.grpc.benchmarks.proto;
116 return io.grpc.benchmarks.proto.Control.getDescriptor().getEnumTypes().get(0); in getDescriptor()
240 return io.grpc.benchmarks.proto.Control.getDescriptor().getEnumTypes().get(1); in getDescriptor()
365 return io.grpc.benchmarks.proto.Control.getDescriptor().getEnumTypes().get(2); in getDescriptor()
475 return io.grpc.benchmarks.proto.Control.internal_static_grpc_testing_PoissonParams_descriptor; in getDescriptor()
480 …return io.grpc.benchmarks.proto.Control.internal_static_grpc_testing_PoissonParams_fieldAccessorTa… in internalGetFieldAccessorTable()
482 …io.grpc.benchmarks.proto.Control.PoissonParams.class, io.grpc.benchmarks.proto.Control.PoissonPara… in internalGetFieldAccessorTable()
535 if (!(obj instanceof io.grpc.benchmarks.proto.Control.PoissonParams)) { in equals()
538 …io.grpc.benchmarks.proto.Control.PoissonParams other = (io.grpc.benchmarks.proto.Control.PoissonPa… in equals()
564 public static io.grpc.benchmarks.proto.Control.PoissonParams parseFrom( in parseFrom()
[all …]
DMessages.java4 package io.grpc.benchmarks.proto;
93 return io.grpc.benchmarks.proto.Messages.getDescriptor().getEnumTypes().get(0); in getDescriptor()
204 return io.grpc.benchmarks.proto.Messages.internal_static_grpc_testing_BoolValue_descriptor; in getDescriptor()
209 … return io.grpc.benchmarks.proto.Messages.internal_static_grpc_testing_BoolValue_fieldAccessorTable in internalGetFieldAccessorTable()
211 …io.grpc.benchmarks.proto.Messages.BoolValue.class, io.grpc.benchmarks.proto.Messages.BoolValue.Bui… in internalGetFieldAccessorTable()
264 if (!(obj instanceof io.grpc.benchmarks.proto.Messages.BoolValue)) { in equals()
267 …io.grpc.benchmarks.proto.Messages.BoolValue other = (io.grpc.benchmarks.proto.Messages.BoolValue) … in equals()
291 public static io.grpc.benchmarks.proto.Messages.BoolValue parseFrom( in parseFrom()
296 public static io.grpc.benchmarks.proto.Messages.BoolValue parseFrom( in parseFrom()
302 public static io.grpc.benchmarks.proto.Messages.BoolValue parseFrom( in parseFrom()
[all …]
DStats.java4 package io.grpc.benchmarks.proto;
172 return io.grpc.benchmarks.proto.Stats.internal_static_grpc_testing_ServerStats_descriptor; in getDescriptor()
177 … return io.grpc.benchmarks.proto.Stats.internal_static_grpc_testing_ServerStats_fieldAccessorTable in internalGetFieldAccessorTable()
179 …io.grpc.benchmarks.proto.Stats.ServerStats.class, io.grpc.benchmarks.proto.Stats.ServerStats.Build… in internalGetFieldAccessorTable()
333 if (!(obj instanceof io.grpc.benchmarks.proto.Stats.ServerStats)) { in equals()
336 …io.grpc.benchmarks.proto.Stats.ServerStats other = (io.grpc.benchmarks.proto.Stats.ServerStats) ob… in equals()
391 public static io.grpc.benchmarks.proto.Stats.ServerStats parseFrom( in parseFrom()
396 public static io.grpc.benchmarks.proto.Stats.ServerStats parseFrom( in parseFrom()
402 public static io.grpc.benchmarks.proto.Stats.ServerStats parseFrom( in parseFrom()
407 public static io.grpc.benchmarks.proto.Stats.ServerStats parseFrom( in parseFrom()
[all …]
DServices.java4 package io.grpc.benchmarks.proto;
62 io.grpc.benchmarks.proto.Messages.getDescriptor(), in internalBuildGeneratedFileFrom()
63 io.grpc.benchmarks.proto.Control.getDescriptor(), in internalBuildGeneratedFileFrom()
65 io.grpc.benchmarks.proto.Messages.getDescriptor(); in io.grpc.benchmarks.proto.Messages.getDescriptor()
66 io.grpc.benchmarks.proto.Control.getDescriptor(); in io.grpc.benchmarks.proto.Control.getDescriptor()
/external/eigen/bench/
Dbenchmark-blocking-sizes.cpp345 void serialize_benchmarks(const char* filename, const vector<benchmark_t>& benchmarks, size_t first… in serialize_benchmarks() argument
353 size_t benchmarks_vector_size = benchmarks.size(); in serialize_benchmarks()
357 fwrite(benchmarks.data(), sizeof(benchmark_t), benchmarks.size(), file); in serialize_benchmarks()
361 bool deserialize_benchmarks(const char* filename, vector<benchmark_t>& benchmarks, size_t& first_be… in deserialize_benchmarks() argument
377 benchmarks.resize(benchmarks_vector_size); in deserialize_benchmarks()
378 if (benchmarks.size() != fread(benchmarks.data(), sizeof(benchmark_t), benchmarks.size(), file)) { in deserialize_benchmarks()
386 vector<benchmark_t>& benchmarks, in try_run_some_benchmarks() argument
390 if (first_benchmark_to_run == benchmarks.size()) { in try_run_some_benchmarks()
401 float ratio_done = float(benchmark_index) / benchmarks.size(); in try_run_some_benchmarks()
405 if (benchmark_index == benchmarks.size() || in try_run_some_benchmarks()
[all …]
/external/okhttp/okio/benchmarks/
DREADME.md4 … used to measure various aspects of performance for Okio buffers. Okio benchmarks are written usin…
9 To run benchmarks locally, first build and package the project modules:
15 This should create a `benchmarks.jar` file in the `target` directory, which is a typical JMH benchm…
18 $ java -jar benchmarks/target/benchmarks.jar -l
20 com.squareup.okio.benchmarks.BufferPerformanceBench.cold
21 com.squareup.okio.benchmarks.BufferPerformanceBench.threads16hot
22 com.squareup.okio.benchmarks.BufferPerformanceBench.threads1hot
23 com.squareup.okio.benchmarks.BufferPerformanceBench.threads2hot
24 com.squareup.okio.benchmarks.BufferPerformanceBench.threads32hot
25 com.squareup.okio.benchmarks.BufferPerformanceBench.threads4hot
[all …]
/external/toolchain-utils/crosperf/
Dresults_organizer_unittest.py138 benchmarks = [mock_instance.benchmark1, mock_instance.benchmark2]
140 benchmark_runs[0] = BenchmarkRun('b1', benchmarks[0], labels[0], 1, '', '',
142 benchmark_runs[1] = BenchmarkRun('b2', benchmarks[0], labels[0], 2, '', '',
144 benchmark_runs[2] = BenchmarkRun('b3', benchmarks[0], labels[1], 1, '', '',
146 benchmark_runs[3] = BenchmarkRun('b4', benchmarks[0], labels[1], 2, '', '',
148 benchmark_runs[4] = BenchmarkRun('b5', benchmarks[1], labels[0], 1, '', '',
150 benchmark_runs[5] = BenchmarkRun('b6', benchmarks[1], labels[0], 2, '', '',
152 benchmark_runs[6] = BenchmarkRun('b7', benchmarks[1], labels[1], 1, '', '',
154 benchmark_runs[7] = BenchmarkRun('b8', benchmarks[1], labels[1], 2, '', '',
163 organized = OrganizeResults(benchmark_runs, labels, benchmarks)
Dexperiment_factory.py94 def AppendBenchmarkSet(self, benchmarks, benchmark_list, test_args, argument
102 benchmarks.append(telemetry_benchmark)
157 benchmarks = []
172 self.AppendBenchmarkSet(benchmarks, telemetry_perfv2_tests, test_args,
176 self.AppendBenchmarkSet(benchmarks, telemetry_pagecycler_tests,
181 self.AppendBenchmarkSet(benchmarks, telemetry_toolchain_perf_tests,
186 benchmarks.append(
199 self.AppendBenchmarkSet(benchmarks,
207 benchmarks.append(benchmark)
211 benchmarks,
[all …]
Dexperiment_factory_unittest.py56 self.assertEqual(len(exp.benchmarks), 1)
57 self.assertEqual(exp.benchmarks[0].name, 'PageCycler')
58 self.assertEqual(exp.benchmarks[0].test_name, 'PageCycler')
59 self.assertEqual(exp.benchmarks[0].iterations, 3)
160 self.assertEqual(len(exp.benchmarks), 1)
161 self.assertEqual(exp.benchmarks[0].name, 'kraken')
162 self.assertEqual(exp.benchmarks[0].test_name, 'kraken')
163 self.assertEqual(exp.benchmarks[0].iterations, 1)
164 self.assertEqual(exp.benchmarks[0].suite, 'telemetry_Crosperf')
165 self.assertFalse(exp.benchmarks[0].show_all_results)
Dresults_organizer.py20 def _AdjustIteration(benchmarks, max_dup, bench): argument
22 for benchmark in benchmarks:
93 def _DuplicatePass(result, benchmarks): argument
102 _AdjustIteration(benchmarks, max_dup, bench)
154 def OrganizeResults(benchmark_runs, labels, benchmarks=None, json_report=False): argument
172 if benchmarks is None:
173 benchmarks = []
209 _DuplicatePass(result, benchmarks)
/external/grpc-grpc-java/benchmarks/src/main/java/io/grpc/benchmarks/qps/
DOpenLoopClient.java17 package io.grpc.benchmarks.qps;
20 import static io.grpc.benchmarks.Utils.HISTOGRAM_MAX_VALUE;
21 import static io.grpc.benchmarks.Utils.HISTOGRAM_PRECISION;
22 import static io.grpc.benchmarks.Utils.saveHistogram;
23 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.ADDRESS;
24 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.CLIENT_PAYLOAD;
25 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.DURATION;
26 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.FLOW_CONTROL_WINDOW;
27 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.SAVE_HISTOGRAM;
28 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.SERVER_PAYLOAD;
[all …]
DAsyncClient.java17 package io.grpc.benchmarks.qps;
19 import static io.grpc.benchmarks.Utils.HISTOGRAM_MAX_VALUE;
20 import static io.grpc.benchmarks.Utils.HISTOGRAM_PRECISION;
21 import static io.grpc.benchmarks.Utils.saveHistogram;
22 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.ADDRESS;
23 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.CHANNELS;
24 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.CLIENT_PAYLOAD;
25 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.DIRECTEXECUTOR;
26 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.DURATION;
27 import static io.grpc.benchmarks.qps.ClientConfiguration.ClientParam.FLOW_CONTROL_WINDOW;
[all …]
/external/grpc-grpc/tools/run_tests/
Drun_microbenchmark.py80 benchmarks = []
92 benchmarks.append(
117 if len(benchmarks) >= min(16, multiprocessing.cpu_count()):
121 benchmarks, maxjobs=max(1,
125 benchmarks = []
129 if len(benchmarks):
130 jobset.run(benchmarks, maxjobs=max(1, multiprocessing.cpu_count() / 2))
142 benchmarks = []
148 benchmarks.append(
172 if len(benchmarks) >= 20:
[all …]
/external/grpc-grpc/tools/profiling/microbenchmarks/bm_diff/
Dbm_build.py57 def _make_cmd(cfg, benchmarks, jobs): argument
58 return ['make'] + benchmarks + ['CONFIG=%s' % cfg, '-j', '%d' % jobs]
61 def build(name, benchmarks, jobs, counters): argument
65 subprocess.check_call(_make_cmd('opt', benchmarks, jobs))
67 subprocess.check_call(_make_cmd('counters', benchmarks, jobs))
70 subprocess.check_call(_make_cmd('opt', benchmarks, jobs))
72 subprocess.check_call(_make_cmd('counters', benchmarks, jobs))
81 build(args.name, args.benchmarks, args.jobs, args.counters)
/external/tensorflow/tensorflow/contrib/eager/python/examples/resnet50/
DREADME.md7 - `resnet50_test.py`: Sanity unittests and benchmarks for using the model with
9 - `resnet50_graph_test.py`: Sanity unittests and benchmarks when using the same
18 python resnet50_test.py --benchmarks=.
21 python resnet50_graph_test.py --benchmarks=.
25 package. To build (and run benchmarks) from source:
29 bazel run -c opt --config=cuda :resnet50_test -- --benchmarks=.
32 bazel run -c opt --config=cuda :resnet50_graph_test -- --benchmarks=.
37 On October 31, 2017, the benchmarks demonstrated comparable performance
/external/v8/benchmarks/
Dbase.js66 function BenchmarkSuite(name, reference, benchmarks) { argument
69 this.benchmarks = benchmarks;
141 result += suites[i].benchmarks.length;
238 var length = this.benchmarks.length;
250 suite.benchmarks[index].Setup();
263 data = suite.RunSingleBenchmark(suite.benchmarks[index], data);
274 suite.benchmarks[index++].TearDown();
/external/tensorflow/tensorflow/tools/test/
Dperformance.bzl9 benchmarks = "..",
40 "--test_args=--benchmarks=%s" % benchmarks,
56 benchmarks = "..",
65 benchmarks = benchmarks,
/external/tensorflow/tensorflow/contrib/eager/python/examples/rnn_ptb/
DREADME.md23 python rnn_ptb_test.py --benchmarks=.
26 python rnn_ptb_graph_test.py --benchmarks=.
30 package. To build (and run benchmarks) from source:
35 bazel run -c opt --config=cuda :rnn_ptb_test -- --benchmarks=.
38 bazel run -c opt --config=cuda :rnn_ptb_graph_test -- --benchmarks=.
43 On October 31, 2017, the benchmarks demonstrated slightly better performance
/external/protobuf/benchmarks/
Dreadme.txt10 If we end up with a lot of different benchmarks it may be worth
29 $ javac -d . -cp ../protobuf.jar benchmarks/*.java
34 benchmarks.GoogleSize$SizeMessage1 ../google_message1.dat
35 benchmarks.GoogleSpeed$SpeedMessage1 ../google_message1.dat
36 benchmarks.GoogleSize$SizeMessage2 ../google_message2.dat
37 benchmarks.GoogleSpeed$SpeedMessage2 ../google_message2.dat
/external/v8/tools/
Dtry_perf.py66 benchmarks = ['"%s"' % benchmark for benchmark in options.benchmarks]
67 cmd += ['-p \'testfilter=[%s]\'' % ','.join(benchmarks)]
90 if not options.benchmarks:
94 for benchmark in options.benchmarks:
/external/libcxx/docs/
DTestingLibcxx.rst205 The benchmarks are written using the `Google Benchmark`_ library, a copy of which
216 The benchmark tests are not built by default. The benchmarks can be built using
217 the ``cxx-benchmarks`` target.
225 $ make cxx-benchmarks
227 This will build all of the benchmarks under ``<libcxx-src>/benchmarks`` to be
229 ``build/benchmarks``.
231 The benchmarks can also be built against the platforms native standard library
234 The compiled benchmarks are named ``<test>.libcxx.out`` if they test libc++ and
245 The benchmarks must be run manually by the user. Currently there is no way
252 $ cd build/benchmarks
[all …]

1234567891011