Lines Matching +full:- +full:- +full:fail +full:- +full:env +full:- +full:changed
1 #!/usr/bin/env python
92 # ---------------------------------------------
93 # --- P r o g r e s s I n d i c a t o r s ---
94 # ---------------------------------------------
122 output += ["--- stderr ---" ]
125 output += ["--- stdout ---"]
129 output += ["--- %s ---" % PrintCrashed(failure.output.exit_code)]
131 output += ["--- TIMEOUT ---"]
152 # Spawn N-1 threads and then use this thread as the last one.
153 # That way -j1 avoids threading altogether which is a nice fallback
155 for i in range(tasks - 1):
163 # Use a timeout so that signals (ctrl-c) will be processed.
198 # See https://smartos.org/bugview/OS-2767
205 case.duration = (datetime.now() - start)
220 self.remaining -= 1
271 outcome = 'FAIL'
303 find_full_path = re.search(r' +at .*\(.*%s:([0-9]+):([0-9]+)' % test.file, traceback)
321 logger.info(' stack: |-')
340 # Print test name as (for example) "parallel/test-assert". Tests that are
344 command = output.command[-1]
349 self.severity = 'fail'
363 self.severity = 'fail'
392 logger.info(' ---')
413 # Print test name as (for example) "parallel/test-assert". Tests that are
417 command = output.command[-1]
466 print("--- %s ---" % PrintCrashed(output.output.exit_code))
468 print("--- TIMEOUT ---")
471 if length and (len(str) > (length - 3)):
472 return str[:(length-3)] + "..."
478 elapsed = time.time() - self.start_time
481 'remaining': (((self.total - self.remaining) * 100) // self.total),
497 …s)02i|\033[34m%%%(remaining) 4d\033[0m|\033[32m+%(passed) 4d\033[0m|\033[31m-%(failed) 4d\033[0m]:…
511 … 'status_line': "[%(mins)02i:%(secs)02i|%%%(remaining) 4d|+%(passed) 4d|-%(failed) 4d]: %(test)s",
534 # -------------------------
535 # --- F r a m e w o r k ---
536 # -------------------------
576 def RunCommand(self, command, env): argument
581 env,
596 # Tests can leave the tty in non-blocking mode. If the test runner
623 outcome = FAIL
632 # Timed out tests will have exit_code -signal.SIGTERM.
659 SEM_INVALID_VALUE = -1
676 # looked at post-mortem, which helps for investigating failures that are
701 # Compute the end time - if the process crosses this limit we
745 def Execute(args, context, timeout=None, env=None, disable_core_files=False, stdin=None): argument
749 if env is None:
750 env = {}
762 for key, value in env.items():
780 env = env_copy,
953 # -------------------------------------------
954 # --- T e s t C o n f i g u r a t i o n ---
955 # -------------------------------------------
960 FAIL = 'fail' variable
978 def Evaluate(self, env, defs): argument
987 def GetOutcomes(self, env, defs): argument
988 if self.name in env: return set([env[self.name]])
997 def GetOutcomes(self, env, defs): argument
999 return defs[self.name].GetOutcomes(env, defs)
1011 def Evaluate(self, env, defs): argument
1013 return self.left.Evaluate(env, defs) or self.right.Evaluate(env, defs)
1017 inter = self.left.GetOutcomes(env, defs) & self.right.GetOutcomes(env, defs)
1021 return self.left.Evaluate(env, defs) and self.right.Evaluate(env, defs)
1023 def GetOutcomes(self, env, defs): argument
1025 return self.left.GetOutcomes(env, defs) | self.right.GetOutcomes(env, defs)
1027 if self.right.Evaluate(env, defs):
1028 return self.left.GetOutcomes(env, defs)
1033 return self.left.GetOutcomes(env, defs) & self.right.GetOutcomes(env, defs)
1058 return self.index < self.length + (length - 1)
1219 def ClassifyTests(self, cases, env): argument
1220 sections = [ s for s in self.sections if s.condition.Evaluate(env, self.defs) ]
1226 outcomes_list = [ r.GetOutcomes(env, self.defs) for r in matches ]
1258 def GetOutcomes(self, env, defs): argument
1259 return self.value.GetOutcomes(env, defs)
1273 PREFIX_PATTERN = re.compile(r'^\s*prefix\s+([\w_.\-/]+)$')
1313 # ---------------
1314 # --- M a i n ---
1315 # ---------------
1323 result.add_option("-m", "--mode", help="The test modes in which to run (comma-separated)",
1325 result.add_option("-v", "--verbose", help="Verbose output",
1327 result.add_option('--logfile', dest='logfile',
1329 result.add_option("-p", "--progress",
1332 result.add_option("--report", help="Print a summary of the tests to be run",
1334 result.add_option("-s", "--suite", help="A test suite",
1336 result.add_option("-t", "--timeout", help="Timeout in seconds",
1338 result.add_option("--arch", help='The architecture to run tests for',
1340 result.add_option("--snapshot", help="Run the tests with snapshot turned on",
1342 result.add_option("--special-command", default=None)
1343 result.add_option("--node-args", dest="node_args", help="Args to pass through to Node",
1345 result.add_option("--expect-fail", dest="expect_fail",
1346 help="Expect test cases to fail", default=False, action="store_true")
1347 result.add_option("--valgrind", help="Run tests through valgrind",
1349 result.add_option("--worker", help="Run parallel tests inside a worker context",
1351 result.add_option("--check-deopts", help="Check tests for permanent deoptimizations",
1353 result.add_option("--cat", help="Print the source of the tests",
1355 result.add_option("--flaky-tests",
1358 result.add_option("--skip-tests",
1359 help="Tests that should not be executed (comma-separated)",
1361 result.add_option("--warn-unused", help="Report unused rules",
1363 result.add_option("-j", help="The number of parallel tasks to run",
1365 result.add_option("-J", help="Run tasks in parallel on all cores",
1367 result.add_option("--time", help="Print timing information after running",
1369 result.add_option("--suppress-dialogs", help="Suppress Windows dialogs for crashing tests",
1371 result.add_option("--no-suppress-dialogs", help="Display Windows dialogs for crashing tests",
1373 result.add_option("--shell", help="Path to node executable", default=None)
1374 result.add_option("--store-unexpected-output",
1377 result.add_option("--no-store-unexpected-output",
1380 result.add_option("-r", "--run",
1381 …help="Divide the tests in m groups (interleaved) and run tests from group n (--run=n,m with n < m)…
1383 result.add_option('--temp-dir',
1385 result.add_option('--test-root',
1387 result.add_option('--repeat',
1390 result.add_option('--abort-on-timeout',
1393 result.add_option("--type",
1410 print("The run argument must be two comma-separated integers.")
1430 print("Unknown flaky-tests mode %s" % options.flaky_tests)
1439 * %(fail_ok)4d tests are expected to fail that we won't fix
1440 * %(fail)4d tests are expected to fail that we should fix\
1471 path = path[:-3]
1473 path = path[:-4]
1477 if (not value) or (value.find('@') == -1):
1512 return "CRASHED (Signal: %d)" % -code
1516 # default JavaScript test-run, e.g., internet/ requires a network connection,
1524 'js-native-api',
1525 'node-api',
1527 'tick-processor',
1528 'v8-updates'
1536 subsystem_regex = re.compile(r'^[a-zA-Z-]*$')
1538 mapped_args = ["*/test*-%s-*" % arg if check(arg) else arg for arg in args]
1549 ssl_ver = Execute([vm, '-p', 'process.versions.openssl'], context).stdout
1566 fh = logging.FileHandler(options.logfile, encoding='utf-8', mode='w')
1580 # Check for --valgrind option. If enabled, we overwrite the special
1581 # command flag with a command that uses the run-valgrind.py script.
1583 run_valgrind = join(workspace, "tools", "run-valgrind.py")
1584 options.special_command = "python -u " + run_valgrind + " @"
1587 options.node_args.append("--trace-opt")
1588 options.node_args.append("--trace-file-names")
1589 # --always-opt is needed because many tests do not run long enough for the
1591 options.node_args.append("--always-opt")
1595 run_worker = join(workspace, "tools", "run-worker.js")
1630 archEngineContext = Execute([vm, "-p", "process.arch"], context)
1636 env = {
1644 cases, unused_rules = config.ClassifyTests(test_list, env)
1655 '-p', 'process.features.inspector'], context)
1660 '-p', 'process.versions.openssl'], context)
1671 print("--- begin source: %s ---" % test.GetLabel())
1674 print("--- end source: %s ---" % test.GetLabel())
1708 'skipped': len(all_cases) - len(cases_to_run),
1710 'fail_ok': len([t for t in cases_to_run if t.outcomes == set([FAIL, OKAY])]),
1711 'fail': len([t for t in cases_to_run if t.outcomes == set([FAIL])])
1716 # silent errors if this file is changed to list the tests in a way that
1733 duration = time.time() - start
1742 sys.stderr.write("--- Total time: %s ---\n" % FormatTime(duration))