Lines Matching refs:benchmark_results
173 def _GetTables(benchmark_results, columns, table_type): argument
174 iter_counts = benchmark_results.iter_counts
175 result = benchmark_results.run_keyvals
188 table = TableGenerator(runs, benchmark_results.label_names).GetTable()
197 def _GetPerfTables(benchmark_results, columns, table_type): argument
198 p_table = _PerfTable(benchmark_results.benchmark_names_and_iterations,
199 benchmark_results.label_names,
200 benchmark_results.read_perf_report)
204 iterations = benchmark_results.iter_counts[benchmark]
212 benchmark_results.label_names,
232 self.benchmark_results = results
236 return get_tables(self.benchmark_results, columns, table_type)
401 def __init__(self, benchmark_results, experiment=None): argument
402 super(HTMLResultsReport, self).__init__(benchmark_results)
411 label_names = self.benchmark_results.label_names
412 test_results = self.benchmark_results.run_keyvals
589 benchmark_results, argument
598 super(JSONResultsReport, self).__init__(benchmark_results)
621 benchmark_results = BenchmarkResults.FromExperiment(
623 return JSONResultsReport(benchmark_results, date, time, experiment,
631 benchmark_results = self.benchmark_results
632 label_names = benchmark_results.label_names
635 for test, test_results in benchmark_results.run_keyvals.iteritems():