• Home
  • Raw
  • Download

Lines Matching refs:benchmark_name

137     for benchmark_name, directories in benchmark_directory_map.items():
140 is_ref = '.reference' in benchmark_name
164 logging.error('Failed to obtain test results for %s: %s', benchmark_name, e)
169 logging.info('Benchmark %s ran no tests on at least one shard' % benchmark_name)
171 benchmark_enabled_map[benchmark_name] = True
187 for benchmark_name, directories in benchmark_directory_map.items():
193 name=_generate_unique_logdog_filename(benchmark_name), data=f.read())
194 benchmark_logs_links[benchmark_name].append(uploaded_link)
241 benchmark_name = _get_benchmark_name(directory)
242 logging.debug('Found benchmark %s directory %s' % (benchmark_name, directory))
243 if benchmark_name in benchmark_directory_map.keys():
244 benchmark_directory_map[benchmark_name].append(directory)
246 benchmark_directory_map[benchmark_name] = [directory]
254 for benchmark_name, directories in benchmark_directory_map.items():
255 if not benchmark_enabled_map.get(benchmark_name, False):
390 def _add_build_info(results, benchmark_name, build_properties): argument
401 benchmark_name,
414 def _merge_perf_results(benchmark_name, results_filename, directories, build_properties): argument
427 logging.error('Failed to obtain any perf results from %s.', benchmark_name)
437 merged_results = _add_build_info(merged_results, benchmark_name, build_properties)
443 print_duration(('%s results merging' % (benchmark_name)), begin_time, end_time)
446 def _upload_individual(benchmark_name, directories, configuration_name, build_properties, argument
455 merge_perf_dir = os.path.join(os.path.abspath(tmpfile_dir), benchmark_name)
459 _merge_perf_results(benchmark_name, results_filename, directories, build_properties)
466 (benchmark_name, results_size_in_mib))
468 upload_return_code = _upload_perf_results(results_filename, benchmark_name,
471 print_duration(('%s upload time' % (benchmark_name)), upload_begin_time,
473 return (benchmark_name, upload_return_code == 0)
482 benchmark_name = params[0]
484 logging.exception('Error uploading perf result of %s' % benchmark_name)
485 return benchmark_name, upload_succeed
586 for benchmark_name, directories in benchmark_directory_map.items():
587 if not benchmark_enabled_map.get(benchmark_name, False):
591 output_json_file = os.path.join(output_results_dir, (str(uuid.uuid4()) + benchmark_name))
592 results_dict[benchmark_name] = output_json_file
596 (benchmark_name, directories, configuration_name, build_properties, output_json_file))
613 for benchmark_name in benchmark_directory_map:
614 results.append((benchmark_name, False))
627 for benchmark_name, output_file in results_dict.items():
628 upload_succeed = benchmark_upload_result_map[benchmark_name]
631 is_reference = '.reference' in benchmark_name
633 benchmark_name,
656 def _write_perf_data_to_logfile(benchmark_name, output_file, configuration_name, build_properties, argument
666 logging.error('Error parsing perf results JSON for benchmark %s' % benchmark_name)
669 json_fname = _generate_unique_logdog_filename(benchmark_name)
678 logging.warning("Perf results JSON file doesn't exist for benchmark %s" % benchmark_name)
680 base_benchmark_name = benchmark_name.replace('.reference', '')