diff --git a/test/debugger/testcfg.py b/test/debugger/testcfg.py index 1f3e49d24f..fb7169bdfa 100644 --- a/test/debugger/testcfg.py +++ b/test/debugger/testcfg.py @@ -55,7 +55,7 @@ class TestCase(testcase.D8TestCase): def _get_files_params(self): files = self._source_files - if self._test_config.isolates: + if self.test_config.isolates: files = files + ['--isolate'] + files return files diff --git a/test/debugging/testcfg.py b/test/debugging/testcfg.py index f5d15a9f93..4976768b2f 100644 --- a/test/debugging/testcfg.py +++ b/test/debugging/testcfg.py @@ -23,7 +23,11 @@ class PYTestCase(testcase.TestCase): return super(PYTestCase, self).get_command() def _get_cmd_params(self): - return self._get_files_params() + ['--', os.path.join(self._test_config.shell_dir, 'd8')] + self._get_source_flags() + return ( + self._get_files_params() + + ['--', os.path.join(self.test_config.shell_dir, 'd8')] + + self._get_source_flags() + ) def _get_shell_flags(self): return [] diff --git a/test/inspector/testcfg.py b/test/inspector/testcfg.py index 9ad99c47c6..9d703aa97d 100644 --- a/test/inspector/testcfg.py +++ b/test/inspector/testcfg.py @@ -67,4 +67,4 @@ class TestCase(testcase.TestCase): return outproc.ExpectedOutProc( self.expected_outcomes, os.path.join(self.suite.root, self.path) + EXPECTED_SUFFIX, - self.suite.test_config.regenerate_expected_files) + self.test_config.regenerate_expected_files) diff --git a/test/intl/testcfg.py b/test/intl/testcfg.py index 1b0c36e0c0..d8acad5033 100644 --- a/test/intl/testcfg.py +++ b/test/intl/testcfg.py @@ -80,7 +80,7 @@ class TestCase(testcase.D8TestCase): ] ] - if self._test_config.isolates: + if self.test_config.isolates: files += ['--isolate'] + files return files diff --git a/test/message/testcfg.py b/test/message/testcfg.py index dda414c8a4..00d60a6bd9 100644 --- a/test/message/testcfg.py +++ b/test/message/testcfg.py @@ -98,4 +98,4 @@ class TestCase(testcase.D8TestCase): self._base_path, self._expected_fail(), self._base_path + '.out', - self.suite.test_config.regenerate_expected_files) + self.test_config.regenerate_expected_files) diff --git a/test/mjsunit/testcfg.py b/test/mjsunit/testcfg.py index 3afd9e2bce..1a703d330d 100644 --- a/test/mjsunit/testcfg.py +++ b/test/mjsunit/testcfg.py @@ -105,7 +105,7 @@ class TestCase(testcase.D8TestCase): else: mjsunit_files = [os.path.join(self.suite.root, "mjsunit.js")] - if self.suite.framework_name == 'num_fuzzer': + if self.framework_name == 'num_fuzzer': mjsunit_files.append(os.path.join(self.suite.root, "mjsunit_numfuzz.js")) self._source_files = files @@ -128,10 +128,10 @@ class TestCase(testcase.D8TestCase): def _get_files_params(self): files = list(self._source_files) - if not self._test_config.no_harness: + if not self.test_config.no_harness: files += self._mjsunit_files files += self._files_suffix - if self._test_config.isolates: + if self.test_config.isolates: files += ['--isolate'] + files return files @@ -190,8 +190,7 @@ class CombinedTest(testcase.D8TestCase): passed as arguments. """ def __init__(self, name, tests): - super(CombinedTest, self).__init__(tests[0].suite, '', name, - tests[0]._test_config) + super(CombinedTest, self).__init__(tests[0].suite, '', name) self._tests = tests def _prepare_outcomes(self, force_update=True): diff --git a/test/webkit/testcfg.py b/test/webkit/testcfg.py index 4958d705ed..ef666531d5 100644 --- a/test/webkit/testcfg.py +++ b/test/webkit/testcfg.py @@ -81,7 +81,7 @@ class TestCase(testcase.D8TestCase): def _get_files_params(self): files = self._source_files - if self._test_config.isolates: + if self.test_config.isolates: files = files + ['--isolate'] + files return files diff --git a/tools/testrunner/base_runner.py b/tools/testrunner/base_runner.py index 698be9926b..423ca009ec 100644 --- a/tools/testrunner/base_runner.py +++ b/tools/testrunner/base_runner.py @@ -528,8 +528,7 @@ class BaseTestRunner(object): if self.options.verbose: print('>>> Loading test suite: %s' % name) suite = testsuite.TestSuite.Load( - ctx, os.path.join(self.options.test_root, name), test_config, - self.framework_name) + ctx, os.path.join(self.options.test_root, name), test_config) if self._is_testsuite_supported(suite): tests = suite.load_tests_from_disk(variables) @@ -659,17 +658,21 @@ class BaseTestRunner(object): return [] # pragma: no cover def _create_test_config(self): + shard_id, shard_count = self.options.shard_info timeout = self.build_config.timeout_scalefactor( self.options.timeout * self.mode_options.timeout_scalefactor) return TestConfig( command_prefix=self.options.command_prefix, extra_flags=self.options.extra_flags, + framework_name=self.framework_name, isolates=self.options.isolates, mode_flags=self.mode_options.flags + self._runner_flags(), no_harness=self.options.no_harness, noi18n=self.build_config.no_i18n, random_seed=self.options.random_seed, run_skipped=self.options.run_skipped, + shard_count=shard_count, + shard_id=shard_id, shell_dir=self.outdir, timeout=timeout, verbose=self.options.verbose, diff --git a/tools/testrunner/local/testsuite.py b/tools/testrunner/local/testsuite.py index 06439b507c..60d52e8d30 100644 --- a/tools/testrunner/local/testsuite.py +++ b/tools/testrunner/local/testsuite.py @@ -115,9 +115,7 @@ class TestLoader(object): def _create_test(self, path, suite, **kwargs): """Converts paths into test objects using the given options""" - return self.test_class(suite, path, self._path_to_name(path), - self.test_config, self.suite.framework_name, - **kwargs) + return self.test_class(suite, path, self._path_to_name(path), **kwargs) def list_tests(self): """Loads and returns the test objects for a TestSuite""" @@ -248,22 +246,25 @@ def _load_testsuite_module(name, root): class TestSuite(object): @staticmethod - def Load(ctx, root, test_config, framework_name): + def Load(ctx, root, test_config): name = root.split(os.path.sep)[-1] with _load_testsuite_module(name, root) as module: - return module.TestSuite(ctx, name, root, test_config, framework_name) + return module.TestSuite(ctx, name, root, test_config) - def __init__(self, ctx, name, root, test_config, framework_name): + def __init__(self, ctx, name, root, test_config): self.name = name # string self.root = root # string containing path self.test_config = test_config - self.framework_name = framework_name # name of the test runner impl self.tests = None # list of TestCase objects self.statusfile = None self._test_loader = self._test_loader_class()(ctx, self, self._test_class(), self.test_config, self.root) + @property + def framework_name(self): + return self.test_config.framework_name + def status_file(self): return "%s/%s.status" % (self.root, self.name) diff --git a/tools/testrunner/local/testsuite_test.py b/tools/testrunner/local/testsuite_test.py index fa7374218b..e2c34a55ee 100755 --- a/tools/testrunner/local/testsuite_test.py +++ b/tools/testrunner/local/testsuite_test.py @@ -26,20 +26,22 @@ class TestSuiteTest(unittest.TestCase): self.test_config = TestConfig( command_prefix=[], extra_flags=[], + framework_name='standard_runner', isolates=False, mode_flags=[], no_harness=False, noi18n=False, random_seed=0, run_skipped=False, + shard_count=1, + shard_id=0, shell_dir='fake_testsuite/fake_d8', timeout=10, verbose=False, ) self.suite = TestSuite.Load( - DefaultOSContext(PosixCommand), self.test_root, self.test_config, - "standard_runner") + DefaultOSContext(PosixCommand), self.test_root, self.test_config) def testLoadingTestSuites(self): self.assertEqual(self.suite.name, "fake_testsuite") diff --git a/tools/testrunner/objects/testcase.py b/tools/testrunner/objects/testcase.py index dfd25132c8..7b87033010 100644 --- a/tools/testrunner/objects/testcase.py +++ b/tools/testrunner/objects/testcase.py @@ -79,7 +79,7 @@ def read_file(file): class TestCase(object): - def __init__(self, suite, path, name, test_config, framework_name): + def __init__(self, suite, path, name): self.suite = suite # TestSuite object self.path = path # string, e.g. 'div-mod', 'test-api/foo' @@ -95,9 +95,6 @@ class TestCase(object): self.processor = DuckProcessor() self.procid = '%s/%s' % (self.suite.name, self.name) # unique id self.keep_output = False # Can output of this test be dropped - - # Test config contains information needed to build the command. - self._test_config = test_config self._random_seed = None # Overrides test config value if not None # Outcomes @@ -107,8 +104,6 @@ class TestCase(object): self._statusfile_flags = None self.expected_failure_reason = None - self.framework_name = framework_name - self._prepare_outcomes() def create_subtest(self, processor, subtest_id, variant=None, flags=None, @@ -271,10 +266,26 @@ class TestCase(object): "INCOMPATIBLE_FLAGS_PER_EXTRA_FLAG[\"" + extra_flag + "\"]") return self._expected_outcomes + @property + def test_config(self): + return self.suite.test_config + + @property + def framework_name(self): + return self.test_config.framework_name + + @property + def shard_id(self): + return self.test_config.shard_id + + @property + def shard_count(self): + return self.test_config.shard_count + @property def do_skip(self): return (statusfile.SKIP in self._statusfile_outcomes and - not self.suite.test_config.run_skipped) + not self.test_config.run_skipped) @property def is_heavy(self): @@ -357,10 +368,10 @@ class TestCase(object): @property def random_seed(self): - return self._random_seed or self._test_config.random_seed + return self._random_seed or self.test_config.random_seed def _get_extra_flags(self): - return self._test_config.extra_flags + return self.test_config.extra_flags def _get_variant_flags(self): return self.variant_flags @@ -373,7 +384,7 @@ class TestCase(object): return self._statusfile_flags def _get_mode_flags(self): - return self._test_config.mode_flags + return self.test_config.mode_flags def _get_source_flags(self): return [] @@ -385,7 +396,7 @@ class TestCase(object): return [] def _get_timeout(self, params): - timeout = self._test_config.timeout + timeout = self.test_config.timeout if "--jitless" in params: timeout *= 2 if "--no-turbofan" in params: @@ -406,12 +417,12 @@ class TestCase(object): def _create_cmd(self, ctx, shell, params, env, timeout): return ctx.command( - cmd_prefix=self._test_config.command_prefix, - shell=os.path.abspath(os.path.join(self._test_config.shell_dir, shell)), + cmd_prefix=self.test_config.command_prefix, + shell=os.path.abspath(os.path.join(self.test_config.shell_dir, shell)), args=params, env=env, timeout=timeout, - verbose=self._test_config.verbose, + verbose=self.test_config.verbose, resources_func=self._get_resources, handle_sigterm=True, ) diff --git a/tools/testrunner/objects/testcase_test.py b/tools/testrunner/objects/testcase_test.py index 98eeae8883..b1031c6ea4 100644 --- a/tools/testrunner/objects/testcase_test.py +++ b/tools/testrunner/objects/testcase_test.py @@ -21,9 +21,7 @@ class TestCaseTest(unittest.TestCase): test = TestCase( suite=FakeSuite(), path='far/away', - name='parent', - test_config=None, - framework_name='none') + name='parent') self.assertEqual(test.rdb_test_id, 'fakeSuite/parent') # provide by DuckProcessor self.assertEqual(test.processor.name, None) diff --git a/tools/testrunner/test_config.py b/tools/testrunner/test_config.py index 7cf65b9726..00a68e2acd 100644 --- a/tools/testrunner/test_config.py +++ b/tools/testrunner/test_config.py @@ -9,18 +9,22 @@ class TestConfig(object): def __init__(self, command_prefix, extra_flags, + framework_name, isolates, mode_flags, no_harness, noi18n, random_seed, run_skipped, + shard_count, + shard_id, shell_dir, timeout, verbose, regenerate_expected_files=False): self.command_prefix = command_prefix self.extra_flags = extra_flags + self.framework_name = framework_name self.isolates = isolates self.mode_flags = mode_flags self.no_harness = no_harness @@ -28,6 +32,8 @@ class TestConfig(object): # random_seed is always not None. self.random_seed = random_seed or random_utils.random_seed() self.run_skipped = run_skipped + self.shard_count = shard_count + self.shard_id = shard_id self.shell_dir = shell_dir self.timeout = timeout self.verbose = verbose diff --git a/tools/testrunner/testdata/expected_test_results1.json b/tools/testrunner/testdata/expected_test_results1.json index e0641739bc..9ce5797a60 100644 --- a/tools/testrunner/testdata/expected_test_results1.json +++ b/tools/testrunner/testdata/expected_test_results1.json @@ -1,164 +1,176 @@ { - "duration_mean": 1, + "duration_mean": 1, "results": [ { - "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", + "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", "crash_state": "", "crash_type": "", - "duration": 1, + "duration": 1, "error_details": "+Mock diff", - "exit_code": 1, + "exit_code": 1, "expected": [ "PASS" - ], + ], "flags": [ - "--test", - "strawberries", - "--random-seed=123", - "--nohard-abort", + "--test", + "strawberries", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "name": "sweet/strawberries", - "random_seed": 123, - "result": "FAIL", - "run": 1, - "stderr": "", - "stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n", - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "name": "sweet/strawberries", + "random_seed": 123, + "result": "FAIL", + "run": 1, + "shard_count": 1, + "shard_id": 0, + "stderr": "", + "stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n", + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] - }, + }, { - "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", + "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", "crash_state": "", "crash_type": "", - "duration": 1, + "duration": 1, "error_details": "+Mock diff", - "exit_code": 1, + "exit_code": 1, "expected": [ "PASS" - ], + ], "flags": [ - "--test", - "strawberries", - "--random-seed=123", - "--nohard-abort", + "--test", + "strawberries", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "name": "sweet/strawberries", - "random_seed": 123, - "result": "FAIL", - "run": 2, - "stderr": "", - "stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n", - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "name": "sweet/strawberries", + "random_seed": 123, + "result": "FAIL", + "run": 2, + "shard_count": 1, + "shard_id": 0, + "stderr": "", + "stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n", + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] - }, + }, { - "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", + "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", "crash_state": "", "crash_type": "", - "duration": 1, + "duration": 1, "error_details": "+Mock diff", - "exit_code": 1, + "exit_code": 1, "expected": [ "PASS" - ], + ], "flags": [ - "--test", - "strawberries", - "--random-seed=123", - "--nohard-abort", + "--test", + "strawberries", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "name": "sweet/strawberries", - "random_seed": 123, - "result": "FAIL", - "run": 3, - "stderr": "", - "stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n", - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "name": "sweet/strawberries", + "random_seed": 123, + "result": "FAIL", + "run": 3, + "shard_count": 1, + "shard_id": 0, + "stderr": "", + "stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n", + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] } - ], - "slowest_tests": [ + ], + "slowest_tests": [ { - "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", - "duration": 1, - "exit_code": 1, + "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", + "duration": 1, + "exit_code": 1, "expected": [ "PASS" - ], + ], "flags": [ - "--test", - "strawberries", - "--random-seed=123", - "--nohard-abort", + "--test", + "strawberries", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "marked_slow": true, - "name": "sweet/strawberries", - "random_seed": 123, - "result": "FAIL", - "run": 1, - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "marked_slow": true, + "name": "sweet/strawberries", + "random_seed": 123, + "result": "FAIL", + "run": 1, + "shard_count": 1, + "shard_id": 0, + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] - }, + }, { - "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", - "duration": 1, - "exit_code": 1, + "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", + "duration": 1, + "exit_code": 1, "expected": [ "PASS" - ], + ], "flags": [ - "--test", - "strawberries", - "--random-seed=123", - "--nohard-abort", + "--test", + "strawberries", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "marked_slow": true, - "name": "sweet/strawberries", - "random_seed": 123, - "result": "FAIL", - "run": 2, - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "marked_slow": true, + "name": "sweet/strawberries", + "random_seed": 123, + "result": "FAIL", + "run": 2, + "shard_count": 1, + "shard_id": 0, + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] - }, + }, { - "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", - "duration": 1, - "exit_code": 1, + "command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner", + "duration": 1, + "exit_code": 1, "expected": [ "PASS" - ], + ], "flags": [ - "--test", - "strawberries", - "--random-seed=123", - "--nohard-abort", + "--test", + "strawberries", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "marked_slow": true, - "name": "sweet/strawberries", - "random_seed": 123, - "result": "FAIL", - "run": 3, - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "marked_slow": true, + "name": "sweet/strawberries", + "random_seed": 123, + "result": "FAIL", + "run": 3, + "shard_count": 1, + "shard_id": 0, + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] } - ], + ], "test_total": 3 -} \ No newline at end of file +} diff --git a/tools/testrunner/testdata/expected_test_results2.json b/tools/testrunner/testdata/expected_test_results2.json index a023d825ab..8eaf6a577a 100644 --- a/tools/testrunner/testdata/expected_test_results2.json +++ b/tools/testrunner/testdata/expected_test_results2.json @@ -1,108 +1,116 @@ { - "duration_mean": 1, + "duration_mean": 1, "results": [ { - "command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner", + "command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner", "crash_state": "", "crash_type": "", - "duration": 1, + "duration": 1, "error_details": null, - "exit_code": 1, + "exit_code": 1, "expected": [ "PASS" - ], + ], "flags": [ - "bananaflakes", - "--random-seed=123", - "--nohard-abort", + "bananaflakes", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "name": "sweet/bananaflakes", - "random_seed": 123, - "result": "FAIL", - "run": 1, - "stderr": "", - "stdout": "bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner\n", - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "name": "sweet/bananaflakes", + "random_seed": 123, + "result": "FAIL", + "run": 1, + "shard_count": 1, + "shard_id": 0, + "stderr": "", + "stdout": "bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner\n", + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] - }, + }, { - "command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner", + "command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner", "crash_state": "", "crash_type": "", - "duration": 1, + "duration": 1, "error_details": null, - "exit_code": 0, + "exit_code": 0, "expected": [ "PASS" - ], + ], "flags": [ - "bananaflakes", - "--random-seed=123", - "--nohard-abort", + "bananaflakes", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "name": "sweet/bananaflakes", - "random_seed": 123, - "result": "PASS", - "run": 2, - "stderr": "", - "stdout": "bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner\n", - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "name": "sweet/bananaflakes", + "random_seed": 123, + "result": "PASS", + "run": 2, + "shard_count": 1, + "shard_id": 0, + "stderr": "", + "stdout": "bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner\n", + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] } - ], + ], "slowest_tests": [ { - "command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner", - "duration": 1, - "exit_code": 0, + "command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner", + "duration": 1, + "exit_code": 0, "expected": [ "PASS" - ], + ], "flags": [ - "bananaflakes", - "--random-seed=123", - "--nohard-abort", + "bananaflakes", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "marked_slow": false, - "name": "sweet/bananaflakes", - "random_seed": 123, - "result": "", - "run": 2, - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "marked_slow": false, + "name": "sweet/bananaflakes", + "random_seed": 123, + "result": "", + "run": 2, + "shard_count": 1, + "shard_id": 0, + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] - }, + }, { - "command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner", - "duration": 1, - "exit_code": 1, + "command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner", + "duration": 1, + "exit_code": 1, "expected": [ "PASS" - ], + ], "flags": [ - "bananaflakes", - "--random-seed=123", - "--nohard-abort", + "bananaflakes", + "--random-seed=123", + "--nohard-abort", "--testing-d8-test-runner" - ], - "framework_name": "standard_runner", - "marked_slow": false, - "name": "sweet/bananaflakes", - "random_seed": 123, - "result": "FAIL", - "run": 1, - "target_name": "d8_mocked.py", - "variant": "default", + ], + "framework_name": "standard_runner", + "marked_slow": false, + "name": "sweet/bananaflakes", + "random_seed": 123, + "result": "FAIL", + "run": 1, + "shard_count": 1, + "shard_id": 0, + "target_name": "d8_mocked.py", + "variant": "default", "variant_flags": [] } - ], + ], "test_total": 2 -} +} \ No newline at end of file diff --git a/tools/testrunner/testproc/shard.py b/tools/testrunner/testproc/shard.py index 3b67ec133f..e31a2ea2fb 100644 --- a/tools/testrunner/testproc/shard.py +++ b/tools/testrunner/testproc/shard.py @@ -20,7 +20,7 @@ def radix_hash(capacity, key): class ShardProc(base.TestProcFilter): @staticmethod def create(options): - myid, count = options.shard_info() + myid, count = options.shard_info if count == 1: return None return ShardProc(myid, count) diff --git a/tools/testrunner/testproc/util.py b/tools/testrunner/testproc/util.py index 1bbd707570..e969a0347d 100644 --- a/tools/testrunner/testproc/util.py +++ b/tools/testrunner/testproc/util.py @@ -56,15 +56,17 @@ def kill_processes_linux(): def base_test_record(test, result, run): record = { - 'name': test.full_name, - 'flags': result.cmd.args, - 'run': run + 1, 'expected': test.expected_outcomes, + 'flags': result.cmd.args, + 'framework_name': test.framework_name, + 'name': test.full_name, 'random_seed': test.random_seed, + 'run': run + 1, + 'shard_id': test.shard_id, + 'shard_count': test.shard_count, 'target_name': test.get_shell(), 'variant': test.variant, 'variant_flags': test.variant_flags, - 'framework_name': test.framework_name, } if result.output: record.update( diff --git a/tools/testrunner/utils/augmented_options.py b/tools/testrunner/utils/augmented_options.py index 0af04031c5..69fd5305e5 100644 --- a/tools/testrunner/utils/augmented_options.py +++ b/tools/testrunner/utils/augmented_options.py @@ -5,8 +5,12 @@ import optparse import os import random + +from functools import cached_property + from testrunner.testproc import fuzzer + class AugmentedOptions(optparse.Values): """This class will augment exiting options object with a couple of convenient methods and properties. @@ -21,6 +25,7 @@ class AugmentedOptions(optparse.Values): self._fuzzer_rng = random.Random(self.fuzzer_random_seed) return self._fuzzer_rng + @cached_property def shard_info(self): """ Returns pair: