• Home
  • Raw
  • Download

Lines Matching +full:baseline +full:- +full:file

5 compare.py - versatile benchmark output compare tool
23 output_file = find_benchmark_flag('--benchmark_out=', flags)
24 output_type = find_benchmark_flag('--benchmark_out_format=', flags)
26 print(("WARNING: '--benchmark_out=%s' will be passed to both "
32 print(("ERROR: passing '--benchmark_out_format=%s' to 'compare.py`"
42 '-a',
43 '--display_aggregates_only',
46 help="If there are repetitions, by default, we display everything - the"
53 '--no-color',
61 '-d',
62 '--dump_to_json',
64 help="Additionally, dump benchmark comparison output to this file in JSON format.")
68 '--no-utest',
72 …help="The tool can do a two-tailed Mann-Whitney U test with the null hypothesis that it is equally…
75 "--alpha",
79 …help=("significance level alpha. if the calculated p-value is below this value, then the result is…
88 help='The most simple use-case, compare all the output of these two benchmarks')
89 baseline = parser_a.add_argument_group(
90 'baseline', 'The benchmark baseline')
91 baseline.add_argument(
96 help='A benchmark executable or JSON output file')
98 'contender', 'The benchmark that will be compared against the baseline')
104 help='A benchmark executable or JSON output file')
113 baseline = parser_b.add_argument_group(
114 'baseline', 'The benchmark baseline')
115 baseline.add_argument(
120 help='A benchmark executable or JSON output file')
121 baseline.add_argument(
126 help='The first filter, that will be used as baseline')
128 'contender', 'The benchmark that will be compared against the baseline')
134 help='The second filter, that will be compared against the baseline')
144 baseline = parser_c.add_argument_group(
145 'baseline', 'The benchmark baseline')
146 baseline.add_argument(
151 help='A benchmark executable or JSON output file')
152 baseline.add_argument(
157 help='The first filter, that will be used as baseline')
159 'contender', 'The benchmark that will be compared against the baseline')
165 …elp='The second benchmark executable or JSON output file, that will be compared against the baseli…
171 help='The second filter, that will be compared against the baseline')
231 benchmark_options += ['--benchmark_display_aggregates_only=true']
237 options_baseline = ['--benchmark_filter=%s' % filter_baseline]
238 options_contender = ['--benchmark_filter=%s' % filter_contender]
292 ['--no-utest', 'benchmarks', self.testInput0, self.testInput1])
303 ['-a', 'benchmarks', self.testInput0, self.testInput1])
313 ['--alpha=0.314', 'benchmarks', self.testInput0, self.testInput1])
324 ['--no-utest', '--alpha=0.314', 'benchmarks', self.testInput0, self.testInput1])
345 ['benchmarks', self.testInput0, self.testInput1, '--', 'e'])
377 ['filters', self.testInput0, 'c', 'd', '--', 'f'])
412 ['benchmarksfiltered', self.testInput0, 'c', self.testInput1, 'e', '--', 'g'])
428 # kate: tab-width: 4; replace-tabs on; indent-width 4; tab-indents: off;
429 # kate: indent-mode python; remove-trailing-spaces modified;