1"""report.py - Utilities for reporting statistics about benchmark results 2""" 3import os 4import re 5import copy 6 7class BenchmarkColor(object): 8 def __init__(self, name, code): 9 self.name = name 10 self.code = code 11 12 def __repr__(self): 13 return '%s%r' % (self.__class__.__name__, 14 (self.name, self.code)) 15 16 def __format__(self, format): 17 return self.code 18 19# Benchmark Colors Enumeration 20BC_NONE = BenchmarkColor('NONE', '') 21BC_MAGENTA = BenchmarkColor('MAGENTA', '\033[95m') 22BC_CYAN = BenchmarkColor('CYAN', '\033[96m') 23BC_OKBLUE = BenchmarkColor('OKBLUE', '\033[94m') 24BC_HEADER = BenchmarkColor('HEADER', '\033[92m') 25BC_WARNING = BenchmarkColor('WARNING', '\033[93m') 26BC_WHITE = BenchmarkColor('WHITE', '\033[97m') 27BC_FAIL = BenchmarkColor('FAIL', '\033[91m') 28BC_ENDC = BenchmarkColor('ENDC', '\033[0m') 29BC_BOLD = BenchmarkColor('BOLD', '\033[1m') 30BC_UNDERLINE = BenchmarkColor('UNDERLINE', '\033[4m') 31 32def color_format(use_color, fmt_str, *args, **kwargs): 33 """ 34 Return the result of 'fmt_str.format(*args, **kwargs)' after transforming 35 'args' and 'kwargs' according to the value of 'use_color'. If 'use_color' 36 is False then all color codes in 'args' and 'kwargs' are replaced with 37 the empty string. 38 """ 39 assert use_color is True or use_color is False 40 if not use_color: 41 args = [arg if not isinstance(arg, BenchmarkColor) else BC_NONE 42 for arg in args] 43 kwargs = {key: arg if not isinstance(arg, BenchmarkColor) else BC_NONE 44 for key, arg in kwargs.items()} 45 return fmt_str.format(*args, **kwargs) 46 47 48def find_longest_name(benchmark_list): 49 """ 50 Return the length of the longest benchmark name in a given list of 51 benchmark JSON objects 52 """ 53 longest_name = 1 54 for bc in benchmark_list: 55 if len(bc['name']) > longest_name: 56 longest_name = len(bc['name']) 57 return longest_name 58 59 60def calculate_change(old_val, new_val): 61 """ 62 Return a float representing the decimal change between old_val and new_val. 63 """ 64 if old_val == 0 and new_val == 0: 65 return 0.0 66 if old_val == 0: 67 return float(new_val - old_val) / (float(old_val + new_val) / 2) 68 return float(new_val - old_val) / abs(old_val) 69 70 71def filter_benchmark(json_orig, family, replacement=""): 72 """ 73 Apply a filter to the json, and only leave the 'family' of benchmarks. 74 """ 75 regex = re.compile(family) 76 filtered = {} 77 filtered['benchmarks'] = [] 78 for be in json_orig['benchmarks']: 79 if not regex.search(be['name']): 80 continue 81 filteredbench = copy.deepcopy(be) # Do NOT modify the old name! 82 filteredbench['name'] = regex.sub(replacement, filteredbench['name']) 83 filtered['benchmarks'].append(filteredbench) 84 return filtered 85 86 87def generate_difference_report(json1, json2, use_color=True): 88 """ 89 Calculate and report the difference between each test of two benchmarks 90 runs specified as 'json1' and 'json2'. 91 """ 92 first_col_width = find_longest_name(json1['benchmarks']) 93 def find_test(name): 94 for b in json2['benchmarks']: 95 if b['name'] == name: 96 return b 97 return None 98 first_col_width = max(first_col_width, len('Benchmark')) 99 first_line = "{:<{}s}Time CPU Time Old Time New CPU Old CPU New".format( 100 'Benchmark', 12 + first_col_width) 101 output_strs = [first_line, '-' * len(first_line)] 102 103 gen = (bn for bn in json1['benchmarks'] if 'real_time' in bn and 'cpu_time' in bn) 104 for bn in gen: 105 other_bench = find_test(bn['name']) 106 if not other_bench: 107 continue 108 109 if bn['time_unit'] != other_bench['time_unit']: 110 continue 111 112 def get_color(res): 113 if res > 0.05: 114 return BC_FAIL 115 elif res > -0.07: 116 return BC_WHITE 117 else: 118 return BC_CYAN 119 fmt_str = "{}{:<{}s}{endc}{}{:+16.4f}{endc}{}{:+16.4f}{endc}{:14.0f}{:14.0f}{endc}{:14.0f}{:14.0f}" 120 tres = calculate_change(bn['real_time'], other_bench['real_time']) 121 cpures = calculate_change(bn['cpu_time'], other_bench['cpu_time']) 122 output_strs += [color_format(use_color, fmt_str, 123 BC_HEADER, bn['name'], first_col_width, 124 get_color(tres), tres, get_color(cpures), cpures, 125 bn['real_time'], other_bench['real_time'], 126 bn['cpu_time'], other_bench['cpu_time'], 127 endc=BC_ENDC)] 128 return output_strs 129 130############################################################################### 131# Unit tests 132 133import unittest 134 135class TestReportDifference(unittest.TestCase): 136 def load_results(self): 137 import json 138 testInputs = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'Inputs') 139 testOutput1 = os.path.join(testInputs, 'test1_run1.json') 140 testOutput2 = os.path.join(testInputs, 'test1_run2.json') 141 with open(testOutput1, 'r') as f: 142 json1 = json.load(f) 143 with open(testOutput2, 'r') as f: 144 json2 = json.load(f) 145 return json1, json2 146 147 def test_basic(self): 148 expect_lines = [ 149 ['BM_SameTimes', '+0.0000', '+0.0000', '10', '10', '10', '10'], 150 ['BM_2xFaster', '-0.5000', '-0.5000', '50', '25', '50', '25'], 151 ['BM_2xSlower', '+1.0000', '+1.0000', '50', '100', '50', '100'], 152 ['BM_1PercentFaster', '-0.0100', '-0.0100', '100', '99', '100', '99'], 153 ['BM_1PercentSlower', '+0.0100', '+0.0100', '100', '101', '100', '101'], 154 ['BM_10PercentFaster', '-0.1000', '-0.1000', '100', '90', '100', '90'], 155 ['BM_10PercentSlower', '+0.1000', '+0.1000', '100', '110', '100', '110'], 156 ['BM_100xSlower', '+99.0000', '+99.0000', '100', '10000', '100', '10000'], 157 ['BM_100xFaster', '-0.9900', '-0.9900', '10000', '100', '10000', '100'], 158 ['BM_10PercentCPUToTime', '+0.1000', '-0.1000', '100', '110', '100', '90'], 159 ['BM_ThirdFaster', '-0.3333', '-0.3334', '100', '67', '100', '67'], 160 ['BM_BadTimeUnit', '-0.9000', '+0.2000', '0', '0', '0', '1'], 161 ] 162 json1, json2 = self.load_results() 163 output_lines_with_header = generate_difference_report(json1, json2, use_color=False) 164 output_lines = output_lines_with_header[2:] 165 print("\n".join(output_lines_with_header)) 166 self.assertEqual(len(output_lines), len(expect_lines)) 167 for i in range(0, len(output_lines)): 168 parts = [x for x in output_lines[i].split(' ') if x] 169 self.assertEqual(len(parts), 7) 170 self.assertEqual(parts, expect_lines[i]) 171 172 173class TestReportDifferenceBetweenFamilies(unittest.TestCase): 174 def load_result(self): 175 import json 176 testInputs = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'Inputs') 177 testOutput = os.path.join(testInputs, 'test2_run.json') 178 with open(testOutput, 'r') as f: 179 json = json.load(f) 180 return json 181 182 def test_basic(self): 183 expect_lines = [ 184 ['.', '-0.5000', '-0.5000', '10', '5', '10', '5'], 185 ['./4', '-0.5000', '-0.5000', '40', '20', '40', '20'], 186 ['Prefix/.', '-0.5000', '-0.5000', '20', '10', '20', '10'], 187 ['Prefix/./3', '-0.5000', '-0.5000', '30', '15', '30', '15'], 188 ] 189 json = self.load_result() 190 json1 = filter_benchmark(json, "BM_Z.ro", ".") 191 json2 = filter_benchmark(json, "BM_O.e", ".") 192 output_lines_with_header = generate_difference_report(json1, json2, use_color=False) 193 output_lines = output_lines_with_header[2:] 194 print("\n") 195 print("\n".join(output_lines_with_header)) 196 self.assertEqual(len(output_lines), len(expect_lines)) 197 for i in range(0, len(output_lines)): 198 parts = [x for x in output_lines[i].split(' ') if x] 199 self.assertEqual(len(parts), 7) 200 self.assertEqual(parts, expect_lines[i]) 201 202 203if __name__ == '__main__': 204 unittest.main() 205 206# vim: tabstop=4 expandtab shiftwidth=4 softtabstop=4 207# kate: tab-width: 4; replace-tabs on; indent-width 4; tab-indents: off; 208# kate: indent-mode python; remove-trailing-spaces modified; 209