# Copyright (c) 2012 The Chromium OS Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. # # GLMark outputs a final performance score, and it checks the performance score # against minimum requirement if min_score is set. import logging import os import re import string from autotest_lib.client.bin import test, utils from autotest_lib.client.common_lib import error from autotest_lib.client.cros import service_stopper from autotest_lib.client.cros.graphics import graphics_utils GLMARK2_TEST_RE = ( r'^\[(?P.*)\] (?P.*): FPS: (?P\d+) FrameTime: ' r'(?P\d+.\d+) ms$') GLMARK2_SCORE_RE = r'glmark2 Score: (\d+)' # perf value description strings may only contain letters, numbers, periods, # dashes and underscores. # But glmark2 test names are usually in the form: # scene-name:opt=val:opt=v1,v2;v3,v4 or scene: # which we convert to: # scene-name.opt_val.opt_v1-v2_v3-v4 or scene.default description_table = string.maketrans(':,=;', '.-__') description_delete = '<>' class graphics_GLMark2(graphics_utils.GraphicsTest): """Runs glmark2, which benchmarks only calls compatible with OpenGL ES 2.0""" version = 1 preserve_srcdir = True _services = None def setup(self): self.job.setup_dep(['glmark2']) def initialize(self): super(graphics_GLMark2, self).initialize() # If UI is running, we must stop it and restore later. self._services = service_stopper.ServiceStopper(['ui']) self._services.stop_services() def cleanup(self): if self._services: self._services.restore_services() super(graphics_GLMark2, self).cleanup() @graphics_utils.GraphicsTest.failure_report_decorator('graphics_GLMark2') def run_once(self, size='800x600', hasty=False, min_score=None): dep = 'glmark2' dep_dir = os.path.join(self.autodir, 'deps', dep) self.job.install_pkg(dep, 'dep', dep_dir) glmark2 = os.path.join(self.autodir, 'deps/glmark2/glmark2') if not os.path.exists(glmark2): raise error.TestFail('Failed: Could not find test binary.') glmark2_data = os.path.join(self.autodir, 'deps/glmark2/data') options = [] options.append('--data-path %s' % glmark2_data) options.append('--size %s' % size) options.append('--annotate') if hasty: options.append('-b :duration=0.2') else: options.append('-b :duration=2') cmd = glmark2 + ' ' + ' '.join(options) if os.environ.get('CROS_FACTORY'): from autotest_lib.client.cros import factory_setup_modules from cros.factory.test import ui ui.start_reposition_thread('^glmark') # TODO(ihf): Switch this test to use perf.PerfControl like # graphics_GLBench once it is stable. crbug.com/344766. if not hasty: if not utils.wait_for_idle_cpu(60.0, 0.1): if not utils.wait_for_idle_cpu(20.0, 0.2): raise error.TestFail('Failed: Could not get idle CPU.') if not utils.wait_for_cool_machine(): raise error.TestFail('Failed: Could not get cold machine.') # In this test we are manually handling stderr, so expected=True. # Strangely autotest takes CmdError/CmdTimeoutError as warning only. try: result = utils.run(cmd, stderr_is_expected=True, stdout_tee=utils.TEE_TO_LOGS, stderr_tee=utils.TEE_TO_LOGS) except error.CmdError: raise error.TestFail('Failed: CmdError running %s' % cmd) except error.CmdTimeoutError: raise error.TestFail('Failed: CmdTimeout running %s' % cmd) logging.info(result) for line in result.stderr.splitlines(): if line.startswith('Error:'): # Line already starts with 'Error: ", not need to prepend. raise error.TestFail(line) # Numbers in hasty mode are not as reliable, so don't send them to # the dashboard etc. if not hasty: keyvals = {} score = None # glmark2 output the final performance score as: # glmark2 Score: 530 for line in result.stdout.splitlines(): match = re.findall(GLMARK2_SCORE_RE, line) if match: score = int(match[0]) if not score: raise error.TestFail('Failed: Unable to read benchmark score') # Output numbers for plotting by harness. logging.info('GLMark2 score: %d', score) if os.environ.get('CROS_FACTORY'): from autotest_lib.client.cros import factory_setup_modules from cros.factory.event_log import EventLog EventLog('graphics_GLMark2').Log('glmark2_score', score=score) keyvals['glmark2_score'] = score self.write_perf_keyval(keyvals) self.output_perf_value( description='Score', value=score, units='score', higher_is_better=True) if min_score is not None and score < min_score: raise error.TestFail( 'Failed: Benchmark score %d < %d (minimum score ' 'requirement)' % (score, min_score))