a2b6dfb2d1
Without this change, wildcards always overwrite the outcomes of more specific rules. Now we always merge. Review URL: https://codereview.chromium.org/1153073002 Cr-Commit-Position: refs/heads/master@{#28629}
280 lines
9.4 KiB
Python
280 lines
9.4 KiB
Python
# Copyright 2012 the V8 project authors. All rights reserved.
|
|
# Redistribution and use in source and binary forms, with or without
|
|
# modification, are permitted provided that the following conditions are
|
|
# met:
|
|
#
|
|
# * Redistributions of source code must retain the above copyright
|
|
# notice, this list of conditions and the following disclaimer.
|
|
# * Redistributions in binary form must reproduce the above
|
|
# copyright notice, this list of conditions and the following
|
|
# disclaimer in the documentation and/or other materials provided
|
|
# with the distribution.
|
|
# * Neither the name of Google Inc. nor the names of its
|
|
# contributors may be used to endorse or promote products derived
|
|
# from this software without specific prior written permission.
|
|
#
|
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
|
|
import imp
|
|
import os
|
|
|
|
from . import commands
|
|
from . import statusfile
|
|
from . import utils
|
|
from ..objects import testcase
|
|
|
|
# Use this to run several variants of the tests.
|
|
VARIANT_FLAGS = {
|
|
"default": [],
|
|
"stress": ["--stress-opt", "--always-opt"],
|
|
"turbofan": ["--turbo", "--always-opt"],
|
|
"nocrankshaft": ["--nocrankshaft"]}
|
|
|
|
FAST_VARIANT_FLAGS = [
|
|
f for v, f in VARIANT_FLAGS.iteritems() if v in ["default", "turbofan"]
|
|
]
|
|
|
|
class TestSuite(object):
|
|
|
|
@staticmethod
|
|
def LoadTestSuite(root):
|
|
name = root.split(os.path.sep)[-1]
|
|
f = None
|
|
try:
|
|
(f, pathname, description) = imp.find_module("testcfg", [root])
|
|
module = imp.load_module("testcfg", f, pathname, description)
|
|
return module.GetSuite(name, root)
|
|
except:
|
|
# Use default if no testcfg is present.
|
|
return GoogleTestSuite(name, root)
|
|
finally:
|
|
if f:
|
|
f.close()
|
|
|
|
def __init__(self, name, root):
|
|
self.name = name # string
|
|
self.root = root # string containing path
|
|
self.tests = None # list of TestCase objects
|
|
self.rules = None # dictionary mapping test path to list of outcomes
|
|
self.wildcards = None # dictionary mapping test paths to list of outcomes
|
|
self.total_duration = None # float, assigned on demand
|
|
|
|
def shell(self):
|
|
return "d8"
|
|
|
|
def suffix(self):
|
|
return ".js"
|
|
|
|
def status_file(self):
|
|
return "%s/%s.status" % (self.root, self.name)
|
|
|
|
# Used in the status file and for stdout printing.
|
|
def CommonTestName(self, testcase):
|
|
if utils.IsWindows():
|
|
return testcase.path.replace("\\", "/")
|
|
else:
|
|
return testcase.path
|
|
|
|
def ListTests(self, context):
|
|
raise NotImplementedError
|
|
|
|
def VariantFlags(self, testcase, default_flags):
|
|
if testcase.outcomes and statusfile.OnlyStandardVariant(testcase.outcomes):
|
|
return [[]]
|
|
if testcase.outcomes and statusfile.OnlyFastVariants(testcase.outcomes):
|
|
# FAST_VARIANTS implies no --always-opt.
|
|
return [ filter(lambda flag: flag != "--always-opt", f)
|
|
for f in filter(lambda flags: flags in FAST_VARIANT_FLAGS,
|
|
default_flags) ]
|
|
return default_flags
|
|
|
|
def DownloadData(self):
|
|
pass
|
|
|
|
def ReadStatusFile(self, variables):
|
|
(self.rules, self.wildcards) = \
|
|
statusfile.ReadStatusFile(self.status_file(), variables)
|
|
|
|
def ReadTestCases(self, context):
|
|
self.tests = self.ListTests(context)
|
|
|
|
@staticmethod
|
|
def _FilterFlaky(flaky, mode):
|
|
return (mode == "run" and not flaky) or (mode == "skip" and flaky)
|
|
|
|
@staticmethod
|
|
def _FilterSlow(slow, mode):
|
|
return (mode == "run" and not slow) or (mode == "skip" and slow)
|
|
|
|
@staticmethod
|
|
def _FilterPassFail(pass_fail, mode):
|
|
return (mode == "run" and not pass_fail) or (mode == "skip" and pass_fail)
|
|
|
|
def FilterTestCasesByStatus(self, warn_unused_rules,
|
|
flaky_tests="dontcare",
|
|
slow_tests="dontcare",
|
|
pass_fail_tests="dontcare"):
|
|
filtered = []
|
|
used_rules = set()
|
|
for t in self.tests:
|
|
flaky = False
|
|
slow = False
|
|
pass_fail = False
|
|
testname = self.CommonTestName(t)
|
|
if testname in self.rules:
|
|
used_rules.add(testname)
|
|
# Even for skipped tests, as the TestCase object stays around and
|
|
# PrintReport() uses it.
|
|
t.outcomes = self.rules[testname]
|
|
if statusfile.DoSkip(t.outcomes):
|
|
continue # Don't add skipped tests to |filtered|.
|
|
for outcome in t.outcomes:
|
|
if outcome.startswith('Flags: '):
|
|
t.flags += outcome[7:].split()
|
|
flaky = statusfile.IsFlaky(t.outcomes)
|
|
slow = statusfile.IsSlow(t.outcomes)
|
|
pass_fail = statusfile.IsPassOrFail(t.outcomes)
|
|
skip = False
|
|
for rule in self.wildcards:
|
|
assert rule[-1] == '*'
|
|
if testname.startswith(rule[:-1]):
|
|
used_rules.add(rule)
|
|
t.outcomes |= self.wildcards[rule]
|
|
if statusfile.DoSkip(t.outcomes):
|
|
skip = True
|
|
break # "for rule in self.wildcards"
|
|
flaky = flaky or statusfile.IsFlaky(t.outcomes)
|
|
slow = slow or statusfile.IsSlow(t.outcomes)
|
|
pass_fail = pass_fail or statusfile.IsPassOrFail(t.outcomes)
|
|
if (skip or self._FilterFlaky(flaky, flaky_tests)
|
|
or self._FilterSlow(slow, slow_tests)
|
|
or self._FilterPassFail(pass_fail, pass_fail_tests)):
|
|
continue # "for t in self.tests"
|
|
filtered.append(t)
|
|
self.tests = filtered
|
|
|
|
if not warn_unused_rules:
|
|
return
|
|
|
|
for rule in self.rules:
|
|
if rule not in used_rules:
|
|
print("Unused rule: %s -> %s" % (rule, self.rules[rule]))
|
|
for rule in self.wildcards:
|
|
if rule not in used_rules:
|
|
print("Unused rule: %s -> %s" % (rule, self.wildcards[rule]))
|
|
|
|
def FilterTestCasesByArgs(self, args):
|
|
filtered = []
|
|
filtered_args = []
|
|
for a in args:
|
|
argpath = a.split(os.path.sep)
|
|
if argpath[0] != self.name:
|
|
continue
|
|
if len(argpath) == 1 or (len(argpath) == 2 and argpath[1] == '*'):
|
|
return # Don't filter, run all tests in this suite.
|
|
path = os.path.sep.join(argpath[1:])
|
|
if path[-1] == '*':
|
|
path = path[:-1]
|
|
filtered_args.append(path)
|
|
for t in self.tests:
|
|
for a in filtered_args:
|
|
if t.path.startswith(a):
|
|
filtered.append(t)
|
|
break
|
|
self.tests = filtered
|
|
|
|
def GetFlagsForTestCase(self, testcase, context):
|
|
raise NotImplementedError
|
|
|
|
def GetSourceForTest(self, testcase):
|
|
return "(no source available)"
|
|
|
|
def IsFailureOutput(self, output, testpath):
|
|
return output.exit_code != 0
|
|
|
|
def IsNegativeTest(self, testcase):
|
|
return False
|
|
|
|
def HasFailed(self, testcase):
|
|
execution_failed = self.IsFailureOutput(testcase.output, testcase.path)
|
|
if self.IsNegativeTest(testcase):
|
|
return not execution_failed
|
|
else:
|
|
return execution_failed
|
|
|
|
def GetOutcome(self, testcase):
|
|
if testcase.output.HasCrashed():
|
|
return statusfile.CRASH
|
|
elif testcase.output.HasTimedOut():
|
|
return statusfile.TIMEOUT
|
|
elif self.HasFailed(testcase):
|
|
return statusfile.FAIL
|
|
else:
|
|
return statusfile.PASS
|
|
|
|
def HasUnexpectedOutput(self, testcase):
|
|
outcome = self.GetOutcome(testcase)
|
|
return not outcome in (testcase.outcomes or [statusfile.PASS])
|
|
|
|
def StripOutputForTransmit(self, testcase):
|
|
if not self.HasUnexpectedOutput(testcase):
|
|
testcase.output.stdout = ""
|
|
testcase.output.stderr = ""
|
|
|
|
def CalculateTotalDuration(self):
|
|
self.total_duration = 0.0
|
|
for t in self.tests:
|
|
self.total_duration += t.duration
|
|
return self.total_duration
|
|
|
|
|
|
class GoogleTestSuite(TestSuite):
|
|
def __init__(self, name, root):
|
|
super(GoogleTestSuite, self).__init__(name, root)
|
|
|
|
def ListTests(self, context):
|
|
shell = os.path.abspath(os.path.join(context.shell_dir, self.shell()))
|
|
if utils.IsWindows():
|
|
shell += ".exe"
|
|
output = commands.Execute(context.command_prefix +
|
|
[shell, "--gtest_list_tests"] +
|
|
context.extra_flags)
|
|
if output.exit_code != 0:
|
|
print output.stdout
|
|
print output.stderr
|
|
raise Exception("Test executable failed to list the tests.")
|
|
tests = []
|
|
test_case = ''
|
|
for line in output.stdout.splitlines():
|
|
test_desc = line.strip().split()[0]
|
|
if test_desc.endswith('.'):
|
|
test_case = test_desc
|
|
elif test_case and test_desc:
|
|
test = testcase.TestCase(self, test_case + test_desc, dependency=None)
|
|
tests.append(test)
|
|
tests.sort()
|
|
return tests
|
|
|
|
def GetFlagsForTestCase(self, testcase, context):
|
|
return (testcase.flags + ["--gtest_filter=" + testcase.path] +
|
|
["--gtest_random_seed=%s" % context.random_seed] +
|
|
["--gtest_print_time=0"] +
|
|
context.mode_flags)
|
|
|
|
def VariantFlags(self, testcase, default_flags):
|
|
return [[]]
|
|
|
|
def shell(self):
|
|
return self.name
|