[test] Include shard info in test records

This will allow on the infra side to easier link to the respective
shard on a test failure. Without that it's a hassle to find out on
which shard the failing test ran.

This also simplifies how the global test_config stores information.
Some information was duplicated, but is now rather shared through
properties if the owning object is already present.

Bug: v8:13681
Change-Id: I52f01a4fac74627575d80f25923faba99eb6a1fb
Reviewed-on: https://chromium-review.googlesource.com/c/v8/v8/+/4181030
Reviewed-by: Liviu Rau <liviurau@google.com>
Commit-Queue: Michael Achenbach <machenbach@chromium.org>
Cr-Commit-Position: refs/heads/main@{#85429}
This commit is contained in:
Michael Achenbach 2023-01-20 17:22:22 +01:00 committed by V8 LUCI CQ
parent 7a3a6e88bd
commit 29b3bd3826
18 changed files with 281 additions and 230 deletions

View File

@ -55,7 +55,7 @@ class TestCase(testcase.D8TestCase):
def _get_files_params(self):
files = self._source_files
if self._test_config.isolates:
if self.test_config.isolates:
files = files + ['--isolate'] + files
return files

View File

@ -23,7 +23,11 @@ class PYTestCase(testcase.TestCase):
return super(PYTestCase, self).get_command()
def _get_cmd_params(self):
return self._get_files_params() + ['--', os.path.join(self._test_config.shell_dir, 'd8')] + self._get_source_flags()
return (
self._get_files_params() +
['--', os.path.join(self.test_config.shell_dir, 'd8')] +
self._get_source_flags()
)
def _get_shell_flags(self):
return []

View File

@ -67,4 +67,4 @@ class TestCase(testcase.TestCase):
return outproc.ExpectedOutProc(
self.expected_outcomes,
os.path.join(self.suite.root, self.path) + EXPECTED_SUFFIX,
self.suite.test_config.regenerate_expected_files)
self.test_config.regenerate_expected_files)

View File

@ -80,7 +80,7 @@ class TestCase(testcase.D8TestCase):
]
]
if self._test_config.isolates:
if self.test_config.isolates:
files += ['--isolate'] + files
return files

View File

@ -98,4 +98,4 @@ class TestCase(testcase.D8TestCase):
self._base_path,
self._expected_fail(),
self._base_path + '.out',
self.suite.test_config.regenerate_expected_files)
self.test_config.regenerate_expected_files)

View File

@ -105,7 +105,7 @@ class TestCase(testcase.D8TestCase):
else:
mjsunit_files = [os.path.join(self.suite.root, "mjsunit.js")]
if self.suite.framework_name == 'num_fuzzer':
if self.framework_name == 'num_fuzzer':
mjsunit_files.append(os.path.join(self.suite.root, "mjsunit_numfuzz.js"))
self._source_files = files
@ -128,10 +128,10 @@ class TestCase(testcase.D8TestCase):
def _get_files_params(self):
files = list(self._source_files)
if not self._test_config.no_harness:
if not self.test_config.no_harness:
files += self._mjsunit_files
files += self._files_suffix
if self._test_config.isolates:
if self.test_config.isolates:
files += ['--isolate'] + files
return files
@ -190,8 +190,7 @@ class CombinedTest(testcase.D8TestCase):
passed as arguments.
"""
def __init__(self, name, tests):
super(CombinedTest, self).__init__(tests[0].suite, '', name,
tests[0]._test_config)
super(CombinedTest, self).__init__(tests[0].suite, '', name)
self._tests = tests
def _prepare_outcomes(self, force_update=True):

View File

@ -81,7 +81,7 @@ class TestCase(testcase.D8TestCase):
def _get_files_params(self):
files = self._source_files
if self._test_config.isolates:
if self.test_config.isolates:
files = files + ['--isolate'] + files
return files

View File

@ -528,8 +528,7 @@ class BaseTestRunner(object):
if self.options.verbose:
print('>>> Loading test suite: %s' % name)
suite = testsuite.TestSuite.Load(
ctx, os.path.join(self.options.test_root, name), test_config,
self.framework_name)
ctx, os.path.join(self.options.test_root, name), test_config)
if self._is_testsuite_supported(suite):
tests = suite.load_tests_from_disk(variables)
@ -659,17 +658,21 @@ class BaseTestRunner(object):
return [] # pragma: no cover
def _create_test_config(self):
shard_id, shard_count = self.options.shard_info
timeout = self.build_config.timeout_scalefactor(
self.options.timeout * self.mode_options.timeout_scalefactor)
return TestConfig(
command_prefix=self.options.command_prefix,
extra_flags=self.options.extra_flags,
framework_name=self.framework_name,
isolates=self.options.isolates,
mode_flags=self.mode_options.flags + self._runner_flags(),
no_harness=self.options.no_harness,
noi18n=self.build_config.no_i18n,
random_seed=self.options.random_seed,
run_skipped=self.options.run_skipped,
shard_count=shard_count,
shard_id=shard_id,
shell_dir=self.outdir,
timeout=timeout,
verbose=self.options.verbose,

View File

@ -115,9 +115,7 @@ class TestLoader(object):
def _create_test(self, path, suite, **kwargs):
"""Converts paths into test objects using the given options"""
return self.test_class(suite, path, self._path_to_name(path),
self.test_config, self.suite.framework_name,
**kwargs)
return self.test_class(suite, path, self._path_to_name(path), **kwargs)
def list_tests(self):
"""Loads and returns the test objects for a TestSuite"""
@ -248,22 +246,25 @@ def _load_testsuite_module(name, root):
class TestSuite(object):
@staticmethod
def Load(ctx, root, test_config, framework_name):
def Load(ctx, root, test_config):
name = root.split(os.path.sep)[-1]
with _load_testsuite_module(name, root) as module:
return module.TestSuite(ctx, name, root, test_config, framework_name)
return module.TestSuite(ctx, name, root, test_config)
def __init__(self, ctx, name, root, test_config, framework_name):
def __init__(self, ctx, name, root, test_config):
self.name = name # string
self.root = root # string containing path
self.test_config = test_config
self.framework_name = framework_name # name of the test runner impl
self.tests = None # list of TestCase objects
self.statusfile = None
self._test_loader = self._test_loader_class()(ctx, self, self._test_class(),
self.test_config, self.root)
@property
def framework_name(self):
return self.test_config.framework_name
def status_file(self):
return "%s/%s.status" % (self.root, self.name)

View File

@ -26,20 +26,22 @@ class TestSuiteTest(unittest.TestCase):
self.test_config = TestConfig(
command_prefix=[],
extra_flags=[],
framework_name='standard_runner',
isolates=False,
mode_flags=[],
no_harness=False,
noi18n=False,
random_seed=0,
run_skipped=False,
shard_count=1,
shard_id=0,
shell_dir='fake_testsuite/fake_d8',
timeout=10,
verbose=False,
)
self.suite = TestSuite.Load(
DefaultOSContext(PosixCommand), self.test_root, self.test_config,
"standard_runner")
DefaultOSContext(PosixCommand), self.test_root, self.test_config)
def testLoadingTestSuites(self):
self.assertEqual(self.suite.name, "fake_testsuite")

View File

@ -79,7 +79,7 @@ def read_file(file):
class TestCase(object):
def __init__(self, suite, path, name, test_config, framework_name):
def __init__(self, suite, path, name):
self.suite = suite # TestSuite object
self.path = path # string, e.g. 'div-mod', 'test-api/foo'
@ -95,9 +95,6 @@ class TestCase(object):
self.processor = DuckProcessor()
self.procid = '%s/%s' % (self.suite.name, self.name) # unique id
self.keep_output = False # Can output of this test be dropped
# Test config contains information needed to build the command.
self._test_config = test_config
self._random_seed = None # Overrides test config value if not None
# Outcomes
@ -107,8 +104,6 @@ class TestCase(object):
self._statusfile_flags = None
self.expected_failure_reason = None
self.framework_name = framework_name
self._prepare_outcomes()
def create_subtest(self, processor, subtest_id, variant=None, flags=None,
@ -271,10 +266,26 @@ class TestCase(object):
"INCOMPATIBLE_FLAGS_PER_EXTRA_FLAG[\"" + extra_flag + "\"]")
return self._expected_outcomes
@property
def test_config(self):
return self.suite.test_config
@property
def framework_name(self):
return self.test_config.framework_name
@property
def shard_id(self):
return self.test_config.shard_id
@property
def shard_count(self):
return self.test_config.shard_count
@property
def do_skip(self):
return (statusfile.SKIP in self._statusfile_outcomes and
not self.suite.test_config.run_skipped)
not self.test_config.run_skipped)
@property
def is_heavy(self):
@ -357,10 +368,10 @@ class TestCase(object):
@property
def random_seed(self):
return self._random_seed or self._test_config.random_seed
return self._random_seed or self.test_config.random_seed
def _get_extra_flags(self):
return self._test_config.extra_flags
return self.test_config.extra_flags
def _get_variant_flags(self):
return self.variant_flags
@ -373,7 +384,7 @@ class TestCase(object):
return self._statusfile_flags
def _get_mode_flags(self):
return self._test_config.mode_flags
return self.test_config.mode_flags
def _get_source_flags(self):
return []
@ -385,7 +396,7 @@ class TestCase(object):
return []
def _get_timeout(self, params):
timeout = self._test_config.timeout
timeout = self.test_config.timeout
if "--jitless" in params:
timeout *= 2
if "--no-turbofan" in params:
@ -406,12 +417,12 @@ class TestCase(object):
def _create_cmd(self, ctx, shell, params, env, timeout):
return ctx.command(
cmd_prefix=self._test_config.command_prefix,
shell=os.path.abspath(os.path.join(self._test_config.shell_dir, shell)),
cmd_prefix=self.test_config.command_prefix,
shell=os.path.abspath(os.path.join(self.test_config.shell_dir, shell)),
args=params,
env=env,
timeout=timeout,
verbose=self._test_config.verbose,
verbose=self.test_config.verbose,
resources_func=self._get_resources,
handle_sigterm=True,
)

View File

@ -21,9 +21,7 @@ class TestCaseTest(unittest.TestCase):
test = TestCase(
suite=FakeSuite(),
path='far/away',
name='parent',
test_config=None,
framework_name='none')
name='parent')
self.assertEqual(test.rdb_test_id, 'fakeSuite/parent')
# provide by DuckProcessor
self.assertEqual(test.processor.name, None)

View File

@ -9,18 +9,22 @@ class TestConfig(object):
def __init__(self,
command_prefix,
extra_flags,
framework_name,
isolates,
mode_flags,
no_harness,
noi18n,
random_seed,
run_skipped,
shard_count,
shard_id,
shell_dir,
timeout,
verbose,
regenerate_expected_files=False):
self.command_prefix = command_prefix
self.extra_flags = extra_flags
self.framework_name = framework_name
self.isolates = isolates
self.mode_flags = mode_flags
self.no_harness = no_harness
@ -28,6 +32,8 @@ class TestConfig(object):
# random_seed is always not None.
self.random_seed = random_seed or random_utils.random_seed()
self.run_skipped = run_skipped
self.shard_count = shard_count
self.shard_id = shard_id
self.shell_dir = shell_dir
self.timeout = timeout
self.verbose = verbose

View File

@ -1,164 +1,176 @@
{
"duration_mean": 1,
"duration_mean": 1,
"results": [
{
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"crash_state": "",
"crash_type": "",
"duration": 1,
"duration": 1,
"error_details": "+Mock diff",
"exit_code": 1,
"exit_code": 1,
"expected": [
"PASS"
],
],
"flags": [
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 1,
"stderr": "",
"stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 1,
"shard_count": 1,
"shard_id": 0,
"stderr": "",
"stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
},
},
{
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"crash_state": "",
"crash_type": "",
"duration": 1,
"duration": 1,
"error_details": "+Mock diff",
"exit_code": 1,
"exit_code": 1,
"expected": [
"PASS"
],
],
"flags": [
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 2,
"stderr": "",
"stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 2,
"shard_count": 1,
"shard_id": 0,
"stderr": "",
"stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
},
},
{
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"crash_state": "",
"crash_type": "",
"duration": 1,
"duration": 1,
"error_details": "+Mock diff",
"exit_code": 1,
"exit_code": 1,
"expected": [
"PASS"
],
],
"flags": [
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 3,
"stderr": "",
"stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 3,
"shard_count": 1,
"shard_id": 0,
"stderr": "",
"stdout": "--test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
}
],
"slowest_tests": [
],
"slowest_tests": [
{
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 1,
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 1,
"expected": [
"PASS"
],
],
"flags": [
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"marked_slow": true,
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 1,
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"marked_slow": true,
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 1,
"shard_count": 1,
"shard_id": 0,
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
},
},
{
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 1,
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 1,
"expected": [
"PASS"
],
],
"flags": [
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"marked_slow": true,
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 2,
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"marked_slow": true,
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 2,
"shard_count": 1,
"shard_id": 0,
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
},
},
{
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 1,
"command": "/usr/bin/python out/build/d8_mocked.py --test strawberries --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 1,
"expected": [
"PASS"
],
],
"flags": [
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--test",
"strawberries",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"marked_slow": true,
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 3,
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"marked_slow": true,
"name": "sweet/strawberries",
"random_seed": 123,
"result": "FAIL",
"run": 3,
"shard_count": 1,
"shard_id": 0,
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
}
],
],
"test_total": 3
}
}

View File

@ -1,108 +1,116 @@
{
"duration_mean": 1,
"duration_mean": 1,
"results": [
{
"command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner",
"command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner",
"crash_state": "",
"crash_type": "",
"duration": 1,
"duration": 1,
"error_details": null,
"exit_code": 1,
"exit_code": 1,
"expected": [
"PASS"
],
],
"flags": [
"bananaflakes",
"--random-seed=123",
"--nohard-abort",
"bananaflakes",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"name": "sweet/bananaflakes",
"random_seed": 123,
"result": "FAIL",
"run": 1,
"stderr": "",
"stdout": "bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"name": "sweet/bananaflakes",
"random_seed": 123,
"result": "FAIL",
"run": 1,
"shard_count": 1,
"shard_id": 0,
"stderr": "",
"stdout": "bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
},
},
{
"command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner",
"command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner",
"crash_state": "",
"crash_type": "",
"duration": 1,
"duration": 1,
"error_details": null,
"exit_code": 0,
"exit_code": 0,
"expected": [
"PASS"
],
],
"flags": [
"bananaflakes",
"--random-seed=123",
"--nohard-abort",
"bananaflakes",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"name": "sweet/bananaflakes",
"random_seed": 123,
"result": "PASS",
"run": 2,
"stderr": "",
"stdout": "bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"name": "sweet/bananaflakes",
"random_seed": 123,
"result": "PASS",
"run": 2,
"shard_count": 1,
"shard_id": 0,
"stderr": "",
"stdout": "bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner\n",
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
}
],
],
"slowest_tests": [
{
"command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 0,
"command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 0,
"expected": [
"PASS"
],
],
"flags": [
"bananaflakes",
"--random-seed=123",
"--nohard-abort",
"bananaflakes",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"marked_slow": false,
"name": "sweet/bananaflakes",
"random_seed": 123,
"result": "",
"run": 2,
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"marked_slow": false,
"name": "sweet/bananaflakes",
"random_seed": 123,
"result": "",
"run": 2,
"shard_count": 1,
"shard_id": 0,
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
},
},
{
"command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 1,
"command": "/usr/bin/python out/build/d8_mocked.py bananaflakes --random-seed=123 --nohard-abort --testing-d8-test-runner",
"duration": 1,
"exit_code": 1,
"expected": [
"PASS"
],
],
"flags": [
"bananaflakes",
"--random-seed=123",
"--nohard-abort",
"bananaflakes",
"--random-seed=123",
"--nohard-abort",
"--testing-d8-test-runner"
],
"framework_name": "standard_runner",
"marked_slow": false,
"name": "sweet/bananaflakes",
"random_seed": 123,
"result": "FAIL",
"run": 1,
"target_name": "d8_mocked.py",
"variant": "default",
],
"framework_name": "standard_runner",
"marked_slow": false,
"name": "sweet/bananaflakes",
"random_seed": 123,
"result": "FAIL",
"run": 1,
"shard_count": 1,
"shard_id": 0,
"target_name": "d8_mocked.py",
"variant": "default",
"variant_flags": []
}
],
],
"test_total": 2
}
}

View File

@ -20,7 +20,7 @@ def radix_hash(capacity, key):
class ShardProc(base.TestProcFilter):
@staticmethod
def create(options):
myid, count = options.shard_info()
myid, count = options.shard_info
if count == 1:
return None
return ShardProc(myid, count)

View File

@ -56,15 +56,17 @@ def kill_processes_linux():
def base_test_record(test, result, run):
record = {
'name': test.full_name,
'flags': result.cmd.args,
'run': run + 1,
'expected': test.expected_outcomes,
'flags': result.cmd.args,
'framework_name': test.framework_name,
'name': test.full_name,
'random_seed': test.random_seed,
'run': run + 1,
'shard_id': test.shard_id,
'shard_count': test.shard_count,
'target_name': test.get_shell(),
'variant': test.variant,
'variant_flags': test.variant_flags,
'framework_name': test.framework_name,
}
if result.output:
record.update(

View File

@ -5,8 +5,12 @@
import optparse
import os
import random
from functools import cached_property
from testrunner.testproc import fuzzer
class AugmentedOptions(optparse.Values):
"""This class will augment exiting options object with
a couple of convenient methods and properties.
@ -21,6 +25,7 @@ class AugmentedOptions(optparse.Values):
self._fuzzer_rng = random.Random(self.fuzzer_random_seed)
return self._fuzzer_rng
@cached_property
def shard_info(self):
"""
Returns pair: