Home
last modified time | relevance | path

Searched refs:benchmark_name (Results 1 – 25 of 25) sorted by relevance

/external/libcxx/utils/google-benchmark/src/
Dcomplexity.cc182 std::string benchmark_name = in ComputeBigO() local
183 reports[0].benchmark_name.substr(0, reports[0].benchmark_name.find('/')); in ComputeBigO()
187 big_o.benchmark_name = benchmark_name + "_BigO"; in ComputeBigO()
204 rms.benchmark_name = benchmark_name + "_RMS"; in ComputeBigO()
Dstatistics.cc124 CHECK_EQ(reports[0].benchmark_name, run.benchmark_name); in ComputeStats()
151 data.benchmark_name = reports[0].benchmark_name + "_" + Stat.name_; in ComputeStats()
Dcsv_reporter.cc91 std::string name = run.benchmark_name; in PrintRunData()
Dconsole_reporter.cc109 result.benchmark_name.c_str()); in PrintRunData()
Djson_reporter.cc153 out << indent << FormatKV("name", run.benchmark_name) << ",\n"; in PrintRunData()
Dbenchmark.cc227 report.benchmark_name = b.name; in CreateRunReport()
/external/google-benchmark/src/
Dcomplexity.cc182 std::string benchmark_name = in ComputeBigO() local
183 reports[0].benchmark_name.substr(0, reports[0].benchmark_name.find('/')); in ComputeBigO()
187 big_o.benchmark_name = benchmark_name + "_BigO"; in ComputeBigO()
204 rms.benchmark_name = benchmark_name + "_RMS"; in ComputeBigO()
Dstatistics.cc124 CHECK_EQ(reports[0].benchmark_name, run.benchmark_name); in ComputeStats()
151 data.benchmark_name = reports[0].benchmark_name + "_" + Stat.name_; in ComputeStats()
Dcsv_reporter.cc91 std::string name = run.benchmark_name; in PrintRunData()
Dconsole_reporter.cc109 result.benchmark_name.c_str()); in PrintRunData()
Djson_reporter.cc153 out << indent << FormatKV("name", run.benchmark_name) << ",\n"; in PrintRunData()
Dbenchmark.cc227 report.benchmark_name = b.name; in CreateRunReport()
/external/tensorflow/tensorflow/tools/benchmark/
Dbenchmark_model.cc234 const string& benchmark_name, const string& postfix, in RecordBenchmarkEntry() argument
238 stream << benchmark_name; in RecordBenchmarkEntry()
358 string benchmark_name = ""; in Main() local
387 Flag("benchmark_name", &benchmark_name, "benchmark name"), in Main()
453 LOG(INFO) << "Benchmark name: [" << benchmark_name << "]"; in Main()
603 if (!benchmark_name.empty() && !output_prefix.empty()) { in Main()
613 RecordBenchmarkEntry(output_prefix, benchmark_name, "", no_stat_num_runs, in Main()
617 RecordBenchmarkEntry(output_prefix, benchmark_name, "meta-init", 1, in Main()
622 RecordBenchmarkEntry(output_prefix, benchmark_name, "meta-first-inference", in Main()
628 output_prefix, benchmark_name, "meta-init-plus-first-inference", 1, in Main()
[all …]
/external/python/cpython3/Tools/importbench/
Dimportbench.py212 benchmark_name = benchmark.__doc__
213 old_result = max(prev_results[benchmark_name])
214 new_result = max(new_results[benchmark_name])
218 print(benchmark_name, ':', result)
/external/toolchain-utils/crosperf/
Dresults_report.py136 def _ProcessPerfReport(self, perf_report, label, benchmark_name, iteration): argument
141 if benchmark_name not in self.perf_data:
142 self.perf_data[benchmark_name] = {event: [] for event in perf_of_run}
143 ben_data = self.perf_data[benchmark_name]
510 def _ReadExperimentPerfReport(results_directory, label_name, benchmark_name, argument
518 raw_dir_name = label_name + benchmark_name + str(benchmark_iteration + 1)
Dexperiment_factory.py164 benchmark_name = benchmark_settings.name
167 test_name = benchmark_name
244 benchmark_name,
Dexperiment_status.py137 benchmark_name = benchmark_run.benchmark.name
138 benchmark_iterations[benchmark_name].append(benchmark_run.iteration)
/external/libcxx/utils/google-benchmark/test/
Dregister_benchmark_test.cc32 CHECK(name == run.benchmark_name) << "expected " << name << " got " in CheckRun()
33 << run.benchmark_name; in CheckRun()
Dskip_with_error_test.cc36 CHECK(name == run.benchmark_name) << "expected " << name << " got " in CheckRun()
37 << run.benchmark_name; in CheckRun()
/external/google-benchmark/test/
Dregister_benchmark_test.cc32 CHECK(name == run.benchmark_name) << "expected " << name << " got " in CheckRun()
33 << run.benchmark_name; in CheckRun()
Dskip_with_error_test.cc36 CHECK(name == run.benchmark_name) << "expected " << name << " got " in CheckRun()
37 << run.benchmark_name; in CheckRun()
/external/tensorflow/tensorflow/python/platform/
Dbenchmark.py298 benchmark_name = "%s.%s" % (benchmark.__module__, benchmark.__name__)
309 full_benchmark_name = "%s.%s" % (benchmark_name, attr)
/external/tensorflow/tensorflow/compiler/aot/
Dtfcompile.bzl287 benchmark_name = name + "_benchmark"
288 benchmark_file = benchmark_name + ".cc"
294 name=("gen_" + benchmark_name),
316 name=benchmark_name,
/external/google-benchmark/include/benchmark/
Dbenchmark.h1210 std::string benchmark_name; member
/external/libcxx/utils/google-benchmark/include/benchmark/
Dbenchmark.h1210 std::string benchmark_name; member