af120db4af
We define a TestFailedError exception and raise it when we can reliably detect that a test has crashed. All other exceptions are treated as infra failures and are captured by the try-catch clause in MainWrapper function. This also fixes all tests in run_perf_test.py, run_tests_test.py and makes sure that both are run on any changes in tools directory. R=machenbach@chromium.org Bug: chromium:899028 Change-Id: I283bc87b31c814be476bebe9fdda414975494183 Reviewed-on: https://chromium-review.googlesource.com/c/1303293 Commit-Queue: Sergiy Byelozyorov <sergiyb@chromium.org> Reviewed-by: Michael Achenbach <machenbach@chromium.org> Cr-Commit-Position: refs/heads/master@{#57134}
1119 lines
38 KiB
Python
Executable File
1119 lines
38 KiB
Python
Executable File
#!/usr/bin/env python
|
|
# Copyright 2014 the V8 project authors. All rights reserved.
|
|
# Use of this source code is governed by a BSD-style license that can be
|
|
# found in the LICENSE file.
|
|
|
|
"""
|
|
Performance runner for d8.
|
|
|
|
Call e.g. with tools/run-perf.py --arch ia32 some_suite.json
|
|
|
|
The suite json format is expected to be:
|
|
{
|
|
"path": <relative path chunks to perf resources and main file>,
|
|
"owners": [<list of email addresses of benchmark owners (required)>],
|
|
"name": <optional suite name, file name is default>,
|
|
"archs": [<architecture name for which this suite is run>, ...],
|
|
"binary": <name of binary to run, default "d8">,
|
|
"flags": [<flag to d8>, ...],
|
|
"test_flags": [<flag to the test file>, ...],
|
|
"run_count": <how often will this suite run (optional)>,
|
|
"run_count_XXX": <how often will this suite run for arch XXX (optional)>,
|
|
"resources": [<js file to be moved to android device>, ...]
|
|
"main": <main js perf runner file>,
|
|
"results_regexp": <optional regexp>,
|
|
"results_processor": <optional python results processor script>,
|
|
"units": <the unit specification for the performance dashboard>,
|
|
"process_size": <flag - collect maximum memory used by the process>,
|
|
"tests": [
|
|
{
|
|
"name": <name of the trace>,
|
|
"results_regexp": <optional more specific regexp>,
|
|
"results_processor": <optional python results processor script>,
|
|
"units": <the unit specification for the performance dashboard>,
|
|
"process_size": <flag - collect maximum memory used by the process>,
|
|
}, ...
|
|
]
|
|
}
|
|
|
|
The tests field can also nest other suites in arbitrary depth. A suite
|
|
with a "main" file is a leaf suite that can contain one more level of
|
|
tests.
|
|
|
|
A suite's results_regexp is expected to have one string place holder
|
|
"%s" for the trace name. A trace's results_regexp overwrites suite
|
|
defaults.
|
|
|
|
A suite's results_processor may point to an optional python script. If
|
|
specified, it is called after running the tests (with a path relative to the
|
|
suite level's path). It is expected to read the measurement's output text
|
|
on stdin and print the processed output to stdout.
|
|
|
|
The results_regexp will be applied to the processed output.
|
|
|
|
A suite without "tests" is considered a performance test itself.
|
|
|
|
Full example (suite with one runner):
|
|
{
|
|
"path": ["."],
|
|
"owners": ["username@chromium.org"],
|
|
"flags": ["--expose-gc"],
|
|
"test_flags": ["5"],
|
|
"archs": ["ia32", "x64"],
|
|
"run_count": 5,
|
|
"run_count_ia32": 3,
|
|
"main": "run.js",
|
|
"results_regexp": "^%s: (.+)$",
|
|
"units": "score",
|
|
"tests": [
|
|
{"name": "Richards"},
|
|
{"name": "DeltaBlue"},
|
|
{"name": "NavierStokes",
|
|
"results_regexp": "^NavierStokes: (.+)$"}
|
|
]
|
|
}
|
|
|
|
Full example (suite with several runners):
|
|
{
|
|
"path": ["."],
|
|
"owners": ["username@chromium.org", "otherowner@google.com"],
|
|
"flags": ["--expose-gc"],
|
|
"archs": ["ia32", "x64"],
|
|
"run_count": 5,
|
|
"units": "score",
|
|
"tests": [
|
|
{"name": "Richards",
|
|
"path": ["richards"],
|
|
"main": "run.js",
|
|
"run_count": 3,
|
|
"results_regexp": "^Richards: (.+)$"},
|
|
{"name": "NavierStokes",
|
|
"path": ["navier_stokes"],
|
|
"main": "run.js",
|
|
"results_regexp": "^NavierStokes: (.+)$"}
|
|
]
|
|
}
|
|
|
|
Path pieces are concatenated. D8 is always run with the suite's path as cwd.
|
|
|
|
The test flags are passed to the js test file after '--'.
|
|
"""
|
|
|
|
from collections import OrderedDict
|
|
import json
|
|
import logging
|
|
import math
|
|
import optparse
|
|
import os
|
|
import re
|
|
import subprocess
|
|
import sys
|
|
import traceback
|
|
|
|
from testrunner.local import android
|
|
from testrunner.local import command
|
|
from testrunner.local import utils
|
|
|
|
ARCH_GUESS = utils.DefaultArch()
|
|
SUPPORTED_ARCHS = ["arm",
|
|
"ia32",
|
|
"mips",
|
|
"mipsel",
|
|
"x64",
|
|
"arm64"]
|
|
|
|
GENERIC_RESULTS_RE = re.compile(r"^RESULT ([^:]+): ([^=]+)= ([^ ]+) ([^ ]*)$")
|
|
RESULT_STDDEV_RE = re.compile(r"^\{([^\}]+)\}$")
|
|
RESULT_LIST_RE = re.compile(r"^\[([^\]]+)\]$")
|
|
TOOLS_BASE = os.path.abspath(os.path.dirname(__file__))
|
|
INFRA_FAILURE_RETCODE = 87
|
|
|
|
|
|
def GeometricMean(values):
|
|
"""Returns the geometric mean of a list of values.
|
|
|
|
The mean is calculated using log to avoid overflow.
|
|
"""
|
|
values = map(float, values)
|
|
return str(math.exp(sum(map(math.log, values)) / len(values)))
|
|
|
|
|
|
class TestFailedError(Exception):
|
|
"""Error raised when a test has failed due to a non-infra issue."""
|
|
pass
|
|
|
|
|
|
class Results(object):
|
|
"""Place holder for result traces."""
|
|
def __init__(self, traces=None, errors=None):
|
|
self.traces = traces or []
|
|
self.errors = errors or []
|
|
|
|
def ToDict(self):
|
|
return {"traces": self.traces, "errors": self.errors}
|
|
|
|
def WriteToFile(self, file_name):
|
|
with open(file_name, "w") as f:
|
|
f.write(json.dumps(self.ToDict()))
|
|
|
|
def __add__(self, other):
|
|
self.traces += other.traces
|
|
self.errors += other.errors
|
|
return self
|
|
|
|
def __str__(self): # pragma: no cover
|
|
return str(self.ToDict())
|
|
|
|
|
|
class Measurement(object):
|
|
"""Represents a series of results of one trace.
|
|
|
|
The results are from repetitive runs of the same executable. They are
|
|
gathered by repeated calls to ConsumeOutput.
|
|
"""
|
|
def __init__(self, graphs, units, results_regexp, stddev_regexp):
|
|
self.name = '/'.join(graphs)
|
|
self.graphs = graphs
|
|
self.units = units
|
|
self.results_regexp = results_regexp
|
|
self.stddev_regexp = stddev_regexp
|
|
self.results = []
|
|
self.errors = []
|
|
self.stddev = ""
|
|
self.process_size = False
|
|
|
|
def ConsumeOutput(self, stdout):
|
|
try:
|
|
result = re.search(self.results_regexp, stdout, re.M).group(1)
|
|
self.results.append(str(float(result)))
|
|
except ValueError:
|
|
self.errors.append("Regexp \"%s\" returned a non-numeric for test %s."
|
|
% (self.results_regexp, self.name))
|
|
except:
|
|
self.errors.append("Regexp \"%s\" didn't match for test %s."
|
|
% (self.results_regexp, self.name))
|
|
|
|
try:
|
|
if self.stddev_regexp and self.stddev:
|
|
self.errors.append("Test %s should only run once since a stddev "
|
|
"is provided by the test." % self.name)
|
|
if self.stddev_regexp:
|
|
self.stddev = re.search(self.stddev_regexp, stdout, re.M).group(1)
|
|
except:
|
|
self.errors.append("Regexp \"%s\" didn't match for test %s."
|
|
% (self.stddev_regexp, self.name))
|
|
|
|
def GetResults(self):
|
|
return Results([{
|
|
"graphs": self.graphs,
|
|
"units": self.units,
|
|
"results": self.results,
|
|
"stddev": self.stddev,
|
|
}], self.errors)
|
|
|
|
|
|
class NullMeasurement(object):
|
|
"""Null object to avoid having extra logic for configurations that don't
|
|
require secondary run, e.g. CI bots.
|
|
"""
|
|
def ConsumeOutput(self, stdout):
|
|
pass
|
|
|
|
def GetResults(self):
|
|
return Results()
|
|
|
|
|
|
def Unzip(iterable):
|
|
left = []
|
|
right = []
|
|
for l, r in iterable:
|
|
left.append(l)
|
|
right.append(r)
|
|
return lambda: iter(left), lambda: iter(right)
|
|
|
|
|
|
def RunResultsProcessor(results_processor, stdout, count):
|
|
# Dummy pass through for null-runs.
|
|
if stdout is None:
|
|
return None
|
|
|
|
# We assume the results processor is relative to the suite.
|
|
assert os.path.exists(results_processor)
|
|
p = subprocess.Popen(
|
|
[sys.executable, results_processor],
|
|
stdin=subprocess.PIPE,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.PIPE,
|
|
)
|
|
result, _ = p.communicate(input=stdout)
|
|
logging.info(">>> Processed stdout (#%d):\n%s", count, result)
|
|
return result
|
|
|
|
|
|
def AccumulateResults(
|
|
graph_names, trace_configs, iter_output, perform_measurement, calc_total):
|
|
"""Iterates over the output of multiple benchmark reruns and accumulates
|
|
results for a configured list of traces.
|
|
|
|
Args:
|
|
graph_names: List of names that configure the base path of the traces. E.g.
|
|
['v8', 'Octane'].
|
|
trace_configs: List of "TraceConfig" instances. Each trace config defines
|
|
how to perform a measurement.
|
|
iter_output: Iterator over the standard output of each test run.
|
|
perform_measurement: Whether to actually run tests and perform measurements.
|
|
This is needed so that we reuse this script for both CI
|
|
and trybot, but want to ignore second run on CI without
|
|
having to spread this logic throughout the script.
|
|
calc_total: Boolean flag to speficy the calculation of a summary trace.
|
|
Returns: A "Results" object.
|
|
"""
|
|
measurements = [
|
|
trace.CreateMeasurement(perform_measurement) for trace in trace_configs]
|
|
for stdout in iter_output():
|
|
for measurement in measurements:
|
|
measurement.ConsumeOutput(stdout)
|
|
|
|
res = reduce(lambda r, m: r + m.GetResults(), measurements, Results())
|
|
|
|
if not res.traces or not calc_total:
|
|
return res
|
|
|
|
# Assume all traces have the same structure.
|
|
if len(set(map(lambda t: len(t["results"]), res.traces))) != 1:
|
|
res.errors.append("Not all traces have the same number of results.")
|
|
return res
|
|
|
|
# Calculate the geometric means for all traces. Above we made sure that
|
|
# there is at least one trace and that the number of results is the same
|
|
# for each trace.
|
|
n_results = len(res.traces[0]["results"])
|
|
total_results = [GeometricMean(t["results"][i] for t in res.traces)
|
|
for i in range(0, n_results)]
|
|
res.traces.append({
|
|
"graphs": graph_names + ["Total"],
|
|
"units": res.traces[0]["units"],
|
|
"results": total_results,
|
|
"stddev": "",
|
|
})
|
|
return res
|
|
|
|
|
|
def AccumulateGenericResults(graph_names, suite_units, iter_output):
|
|
"""Iterates over the output of multiple benchmark reruns and accumulates
|
|
generic results.
|
|
|
|
Args:
|
|
graph_names: List of names that configure the base path of the traces. E.g.
|
|
['v8', 'Octane'].
|
|
suite_units: Measurement default units as defined by the benchmark suite.
|
|
iter_output: Iterator over the standard output of each test run.
|
|
Returns: A "Results" object.
|
|
"""
|
|
traces = OrderedDict()
|
|
for stdout in iter_output():
|
|
if stdout is None:
|
|
# The None value is used as a null object to simplify logic.
|
|
continue
|
|
for line in stdout.strip().splitlines():
|
|
match = GENERIC_RESULTS_RE.match(line)
|
|
if match:
|
|
stddev = ""
|
|
graph = match.group(1)
|
|
trace = match.group(2)
|
|
body = match.group(3)
|
|
units = match.group(4)
|
|
match_stddev = RESULT_STDDEV_RE.match(body)
|
|
match_list = RESULT_LIST_RE.match(body)
|
|
errors = []
|
|
if match_stddev:
|
|
result, stddev = map(str.strip, match_stddev.group(1).split(","))
|
|
results = [result]
|
|
elif match_list:
|
|
results = map(str.strip, match_list.group(1).split(","))
|
|
else:
|
|
results = [body.strip()]
|
|
|
|
try:
|
|
results = map(lambda r: str(float(r)), results)
|
|
except ValueError:
|
|
results = []
|
|
errors = ["Found non-numeric in %s" %
|
|
"/".join(graph_names + [graph, trace])]
|
|
|
|
trace_result = traces.setdefault(trace, Results([{
|
|
"graphs": graph_names + [graph, trace],
|
|
"units": (units or suite_units).strip(),
|
|
"results": [],
|
|
"stddev": "",
|
|
}], errors))
|
|
trace_result.traces[0]["results"].extend(results)
|
|
trace_result.traces[0]["stddev"] = stddev
|
|
|
|
return reduce(lambda r, t: r + t, traces.itervalues(), Results())
|
|
|
|
|
|
class Node(object):
|
|
"""Represents a node in the suite tree structure."""
|
|
def __init__(self, *args):
|
|
self._children = []
|
|
|
|
def AppendChild(self, child):
|
|
self._children.append(child)
|
|
|
|
|
|
class DefaultSentinel(Node):
|
|
"""Fake parent node with all default values."""
|
|
def __init__(self, binary = "d8"):
|
|
super(DefaultSentinel, self).__init__()
|
|
self.binary = binary
|
|
self.run_count = 10
|
|
self.timeout = 60
|
|
self.path = []
|
|
self.graphs = []
|
|
self.flags = []
|
|
self.test_flags = []
|
|
self.process_size = False
|
|
self.resources = []
|
|
self.results_processor = None
|
|
self.results_regexp = None
|
|
self.stddev_regexp = None
|
|
self.units = "score"
|
|
self.total = False
|
|
self.owners = []
|
|
|
|
|
|
class GraphConfig(Node):
|
|
"""Represents a suite definition.
|
|
|
|
Can either be a leaf or an inner node that provides default values.
|
|
"""
|
|
def __init__(self, suite, parent, arch):
|
|
super(GraphConfig, self).__init__()
|
|
self._suite = suite
|
|
|
|
assert isinstance(suite.get("path", []), list)
|
|
assert isinstance(suite.get("owners", []), list)
|
|
assert isinstance(suite["name"], basestring)
|
|
assert isinstance(suite.get("flags", []), list)
|
|
assert isinstance(suite.get("test_flags", []), list)
|
|
assert isinstance(suite.get("resources", []), list)
|
|
|
|
# Accumulated values.
|
|
self.path = parent.path[:] + suite.get("path", [])
|
|
self.graphs = parent.graphs[:] + [suite["name"]]
|
|
self.flags = parent.flags[:] + suite.get("flags", [])
|
|
self.test_flags = parent.test_flags[:] + suite.get("test_flags", [])
|
|
self.owners = parent.owners[:] + suite.get("owners", [])
|
|
|
|
# Values independent of parent node.
|
|
self.resources = suite.get("resources", [])
|
|
|
|
# Descrete values (with parent defaults).
|
|
self.binary = suite.get("binary", parent.binary)
|
|
self.run_count = suite.get("run_count", parent.run_count)
|
|
self.run_count = suite.get("run_count_%s" % arch, self.run_count)
|
|
self.timeout = suite.get("timeout", parent.timeout)
|
|
self.timeout = suite.get("timeout_%s" % arch, self.timeout)
|
|
self.units = suite.get("units", parent.units)
|
|
self.total = suite.get("total", parent.total)
|
|
self.results_processor = suite.get(
|
|
"results_processor", parent.results_processor)
|
|
self.process_size = suite.get("process_size", parent.process_size)
|
|
|
|
# A regular expression for results. If the parent graph provides a
|
|
# regexp and the current suite has none, a string place holder for the
|
|
# suite name is expected.
|
|
# TODO(machenbach): Currently that makes only sense for the leaf level.
|
|
# Multiple place holders for multiple levels are not supported.
|
|
if parent.results_regexp:
|
|
regexp_default = parent.results_regexp % re.escape(suite["name"])
|
|
else:
|
|
regexp_default = None
|
|
self.results_regexp = suite.get("results_regexp", regexp_default)
|
|
|
|
# A similar regular expression for the standard deviation (optional).
|
|
if parent.stddev_regexp:
|
|
stddev_default = parent.stddev_regexp % re.escape(suite["name"])
|
|
else:
|
|
stddev_default = None
|
|
self.stddev_regexp = suite.get("stddev_regexp", stddev_default)
|
|
|
|
|
|
class TraceConfig(GraphConfig):
|
|
"""Represents a leaf in the suite tree structure."""
|
|
def __init__(self, suite, parent, arch):
|
|
super(TraceConfig, self).__init__(suite, parent, arch)
|
|
assert self.results_regexp
|
|
assert self.owners
|
|
|
|
def CreateMeasurement(self, perform_measurement):
|
|
if not perform_measurement:
|
|
return NullMeasurement()
|
|
|
|
return Measurement(
|
|
self.graphs,
|
|
self.units,
|
|
self.results_regexp,
|
|
self.stddev_regexp,
|
|
)
|
|
|
|
|
|
class RunnableConfig(GraphConfig):
|
|
"""Represents a runnable suite definition (i.e. has a main file).
|
|
"""
|
|
@property
|
|
def main(self):
|
|
return self._suite.get("main", "")
|
|
|
|
def PostProcess(self, stdouts_iter):
|
|
if self.results_processor:
|
|
def it():
|
|
for i, stdout in enumerate(stdouts_iter()):
|
|
yield RunResultsProcessor(self.results_processor, stdout, i + 1)
|
|
return it
|
|
else:
|
|
return stdouts_iter
|
|
|
|
def ChangeCWD(self, suite_path):
|
|
"""Changes the cwd to to path defined in the current graph.
|
|
|
|
The tests are supposed to be relative to the suite configuration.
|
|
"""
|
|
suite_dir = os.path.abspath(os.path.dirname(suite_path))
|
|
bench_dir = os.path.normpath(os.path.join(*self.path))
|
|
os.chdir(os.path.join(suite_dir, bench_dir))
|
|
|
|
def GetCommandFlags(self, extra_flags=None):
|
|
suffix = ["--"] + self.test_flags if self.test_flags else []
|
|
return self.flags + (extra_flags or []) + [self.main] + suffix
|
|
|
|
def GetCommand(self, cmd_prefix, shell_dir, extra_flags=None):
|
|
# TODO(machenbach): This requires +.exe if run on windows.
|
|
extra_flags = extra_flags or []
|
|
if self.binary != 'd8' and '--prof' in extra_flags:
|
|
logging.info("Profiler supported only on a benchmark run with d8")
|
|
|
|
if self.process_size:
|
|
cmd_prefix = ["/usr/bin/time", "--format=MaxMemory: %MKB"] + cmd_prefix
|
|
if self.binary.endswith('.py'):
|
|
# Copy cmd_prefix instead of update (+=).
|
|
cmd_prefix = cmd_prefix + [sys.executable]
|
|
|
|
return command.Command(
|
|
cmd_prefix=cmd_prefix,
|
|
shell=os.path.join(shell_dir, self.binary),
|
|
args=self.GetCommandFlags(extra_flags=extra_flags),
|
|
timeout=self.timeout or 60)
|
|
|
|
def Run(self, runner, trybot):
|
|
"""Iterates over several runs and handles the output for all traces."""
|
|
stdout, stdout_secondary = Unzip(runner())
|
|
return (
|
|
AccumulateResults(
|
|
self.graphs,
|
|
self._children,
|
|
iter_output=self.PostProcess(stdout),
|
|
perform_measurement=True,
|
|
calc_total=self.total,
|
|
),
|
|
AccumulateResults(
|
|
self.graphs,
|
|
self._children,
|
|
iter_output=self.PostProcess(stdout_secondary),
|
|
perform_measurement=trybot, # only run second time on trybots
|
|
calc_total=self.total,
|
|
),
|
|
)
|
|
|
|
|
|
class RunnableTraceConfig(TraceConfig, RunnableConfig):
|
|
"""Represents a runnable suite definition that is a leaf."""
|
|
def __init__(self, suite, parent, arch):
|
|
super(RunnableTraceConfig, self).__init__(suite, parent, arch)
|
|
|
|
def Run(self, runner, trybot):
|
|
"""Iterates over several runs and handles the output."""
|
|
measurement = self.CreateMeasurement(perform_measurement=True)
|
|
measurement_secondary = self.CreateMeasurement(perform_measurement=trybot)
|
|
for stdout, stdout_secondary in runner():
|
|
measurement.ConsumeOutput(stdout)
|
|
measurement_secondary.ConsumeOutput(stdout_secondary)
|
|
return (
|
|
measurement.GetResults(),
|
|
measurement_secondary.GetResults(),
|
|
)
|
|
|
|
|
|
class RunnableGenericConfig(RunnableConfig):
|
|
"""Represents a runnable suite definition with generic traces."""
|
|
def __init__(self, suite, parent, arch):
|
|
super(RunnableGenericConfig, self).__init__(suite, parent, arch)
|
|
|
|
def Run(self, runner, trybot):
|
|
stdout, stdout_secondary = Unzip(runner())
|
|
return (
|
|
AccumulateGenericResults(self.graphs, self.units, stdout),
|
|
AccumulateGenericResults(self.graphs, self.units, stdout_secondary),
|
|
)
|
|
|
|
|
|
def MakeGraphConfig(suite, arch, parent):
|
|
"""Factory method for making graph configuration objects."""
|
|
if isinstance(parent, RunnableConfig):
|
|
# Below a runnable can only be traces.
|
|
return TraceConfig(suite, parent, arch)
|
|
elif suite.get("main") is not None:
|
|
# A main file makes this graph runnable. Empty strings are accepted.
|
|
if suite.get("tests"):
|
|
# This graph has subgraphs (traces).
|
|
return RunnableConfig(suite, parent, arch)
|
|
else:
|
|
# This graph has no subgraphs, it's a leaf.
|
|
return RunnableTraceConfig(suite, parent, arch)
|
|
elif suite.get("generic"):
|
|
# This is a generic suite definition. It is either a runnable executable
|
|
# or has a main js file.
|
|
return RunnableGenericConfig(suite, parent, arch)
|
|
elif suite.get("tests"):
|
|
# This is neither a leaf nor a runnable.
|
|
return GraphConfig(suite, parent, arch)
|
|
else: # pragma: no cover
|
|
raise Exception("Invalid suite configuration.")
|
|
|
|
|
|
def BuildGraphConfigs(suite, arch, parent):
|
|
"""Builds a tree structure of graph objects that corresponds to the suite
|
|
configuration.
|
|
"""
|
|
|
|
# TODO(machenbach): Implement notion of cpu type?
|
|
if arch not in suite.get("archs", SUPPORTED_ARCHS):
|
|
return None
|
|
|
|
graph = MakeGraphConfig(suite, arch, parent)
|
|
for subsuite in suite.get("tests", []):
|
|
BuildGraphConfigs(subsuite, arch, graph)
|
|
parent.AppendChild(graph)
|
|
return graph
|
|
|
|
|
|
def FlattenRunnables(node, node_cb):
|
|
"""Generator that traverses the tree structure and iterates over all
|
|
runnables.
|
|
"""
|
|
node_cb(node)
|
|
if isinstance(node, RunnableConfig):
|
|
yield node
|
|
elif isinstance(node, Node):
|
|
for child in node._children:
|
|
for result in FlattenRunnables(child, node_cb):
|
|
yield result
|
|
else: # pragma: no cover
|
|
raise Exception("Invalid suite configuration.")
|
|
|
|
|
|
class Platform(object):
|
|
def __init__(self, options):
|
|
self.shell_dir = options.shell_dir
|
|
self.shell_dir_secondary = options.shell_dir_secondary
|
|
self.extra_flags = options.extra_flags.split()
|
|
self.options = options
|
|
|
|
@staticmethod
|
|
def ReadBuildConfig(options):
|
|
config_path = os.path.join(options.shell_dir, 'v8_build_config.json')
|
|
if not os.path.isfile(config_path):
|
|
return {}
|
|
with open(config_path) as f:
|
|
return json.load(f)
|
|
|
|
@staticmethod
|
|
def GetPlatform(options):
|
|
if Platform.ReadBuildConfig(options).get('is_android', False):
|
|
return AndroidPlatform(options)
|
|
else:
|
|
return DesktopPlatform(options)
|
|
|
|
def _Run(self, runnable, count, secondary=False):
|
|
raise NotImplementedError() # pragma: no cover
|
|
|
|
def Run(self, runnable, count):
|
|
"""Execute the benchmark's main file.
|
|
|
|
If options.shell_dir_secondary is specified, the benchmark is run twice,
|
|
e.g. with and without patch.
|
|
Args:
|
|
runnable: A Runnable benchmark instance.
|
|
count: The number of this (repeated) run.
|
|
Returns: A tuple with the two benchmark outputs. The latter will be None if
|
|
options.shell_dir_secondary was not specified.
|
|
"""
|
|
stdout = self._Run(runnable, count, secondary=False)
|
|
if self.shell_dir_secondary:
|
|
return stdout, self._Run(runnable, count, secondary=True)
|
|
else:
|
|
return stdout, None
|
|
|
|
|
|
class DesktopPlatform(Platform):
|
|
def __init__(self, options):
|
|
super(DesktopPlatform, self).__init__(options)
|
|
self.command_prefix = []
|
|
|
|
# Setup command class to OS specific version.
|
|
command.setup(utils.GuessOS())
|
|
|
|
if options.prioritize or options.affinitize != None:
|
|
self.command_prefix = ["schedtool"]
|
|
if options.prioritize:
|
|
self.command_prefix += ["-n", "-20"]
|
|
if options.affinitize != None:
|
|
# schedtool expects a bit pattern when setting affinity, where each
|
|
# bit set to '1' corresponds to a core where the process may run on.
|
|
# First bit corresponds to CPU 0. Since the 'affinitize' parameter is
|
|
# a core number, we need to map to said bit pattern.
|
|
cpu = int(options.affinitize)
|
|
core = 1 << cpu
|
|
self.command_prefix += ["-a", ("0x%x" % core)]
|
|
self.command_prefix += ["-e"]
|
|
|
|
def PreExecution(self):
|
|
pass
|
|
|
|
def PostExecution(self):
|
|
pass
|
|
|
|
def PreTests(self, node, path):
|
|
if isinstance(node, RunnableConfig):
|
|
node.ChangeCWD(path)
|
|
|
|
def _Run(self, runnable, count, secondary=False):
|
|
suffix = ' - secondary' if secondary else ''
|
|
shell_dir = self.shell_dir_secondary if secondary else self.shell_dir
|
|
title = ">>> %%s (#%d)%s:" % ((count + 1), suffix)
|
|
cmd = runnable.GetCommand(self.command_prefix, shell_dir, self.extra_flags)
|
|
try:
|
|
output = cmd.execute()
|
|
except OSError: # pragma: no cover
|
|
logging.exception(title % "OSError")
|
|
raise
|
|
|
|
logging.info(title % "Stdout" + "\n%s", output.stdout)
|
|
if output.stderr: # pragma: no cover
|
|
# Print stderr for debugging.
|
|
logging.info(title % "Stderr" + "\n%s", output.stderr)
|
|
if output.timed_out:
|
|
logging.warning(">>> Test timed out after %ss.", runnable.timeout)
|
|
raise TestFailedError()
|
|
if output.exit_code != 0:
|
|
logging.warning(">>> Test crashed.")
|
|
raise TestFailedError()
|
|
if '--prof' in self.extra_flags:
|
|
os_prefix = {"linux": "linux", "macos": "mac"}.get(utils.GuessOS())
|
|
if os_prefix:
|
|
tick_tools = os.path.join(TOOLS_BASE, "%s-tick-processor" % os_prefix)
|
|
subprocess.check_call(tick_tools + " --only-summary", shell=True)
|
|
else: # pragma: no cover
|
|
logging.warning(
|
|
"Profiler option currently supported on Linux and Mac OS.")
|
|
|
|
# time outputs to stderr
|
|
if runnable.process_size:
|
|
return output.stdout + output.stderr
|
|
return output.stdout
|
|
|
|
|
|
class AndroidPlatform(Platform): # pragma: no cover
|
|
|
|
def __init__(self, options):
|
|
super(AndroidPlatform, self).__init__(options)
|
|
self.driver = android.android_driver(options.device)
|
|
|
|
def PreExecution(self):
|
|
self.driver.set_high_perf_mode()
|
|
|
|
def PostExecution(self):
|
|
self.driver.set_default_perf_mode()
|
|
self.driver.tear_down()
|
|
|
|
def PreTests(self, node, path):
|
|
if isinstance(node, RunnableConfig):
|
|
node.ChangeCWD(path)
|
|
suite_dir = os.path.abspath(os.path.dirname(path))
|
|
if node.path:
|
|
bench_rel = os.path.normpath(os.path.join(*node.path))
|
|
bench_abs = os.path.join(suite_dir, bench_rel)
|
|
else:
|
|
bench_rel = "."
|
|
bench_abs = suite_dir
|
|
|
|
self.driver.push_executable(self.shell_dir, "bin", node.binary)
|
|
if self.shell_dir_secondary:
|
|
self.driver.push_executable(
|
|
self.shell_dir_secondary, "bin_secondary", node.binary)
|
|
|
|
if isinstance(node, RunnableConfig):
|
|
self.driver.push_file(bench_abs, node.main, bench_rel)
|
|
for resource in node.resources:
|
|
self.driver.push_file(bench_abs, resource, bench_rel)
|
|
|
|
def _Run(self, runnable, count, secondary=False):
|
|
suffix = ' - secondary' if secondary else ''
|
|
target_dir = "bin_secondary" if secondary else "bin"
|
|
title = ">>> %%s (#%d)%s:" % ((count + 1), suffix)
|
|
self.driver.drop_ram_caches()
|
|
|
|
# Relative path to benchmark directory.
|
|
if runnable.path:
|
|
bench_rel = os.path.normpath(os.path.join(*runnable.path))
|
|
else:
|
|
bench_rel = "."
|
|
|
|
logcat_file = None
|
|
if self.options.dump_logcats_to:
|
|
runnable_name = '-'.join(runnable.graphs)
|
|
logcat_file = os.path.join(
|
|
self.options.dump_logcats_to, 'logcat-%s-#%d%s.log' % (
|
|
runnable_name, count + 1, '-secondary' if secondary else ''))
|
|
logging.debug('Dumping logcat into %s', logcat_file)
|
|
|
|
try:
|
|
stdout = self.driver.run(
|
|
target_dir=target_dir,
|
|
binary=runnable.binary,
|
|
args=runnable.GetCommandFlags(self.extra_flags),
|
|
rel_path=bench_rel,
|
|
timeout=runnable.timeout,
|
|
logcat_file=logcat_file,
|
|
)
|
|
logging.info(title % "Stdout" + "\n%s", stdout)
|
|
except android.CommandFailedException as e:
|
|
logging.info(title % "Stdout" + "\n%s", e.output)
|
|
logging.warning('>>> Test crashed.')
|
|
raise TestFailedError()
|
|
except android.TimeoutException as e:
|
|
if e.output is not None:
|
|
logging.info(title % "Stdout" + "\n%s", e.output)
|
|
logging.warning(">>> Test timed out after %ss.", runnable.timeout)
|
|
raise TestFailedError()
|
|
if runnable.process_size:
|
|
return stdout + "MaxMemory: Unsupported"
|
|
return stdout
|
|
|
|
class CustomMachineConfiguration:
|
|
def __init__(self, disable_aslr = False, governor = None):
|
|
self.aslr_backup = None
|
|
self.governor_backup = None
|
|
self.disable_aslr = disable_aslr
|
|
self.governor = governor
|
|
|
|
def __enter__(self):
|
|
if self.disable_aslr:
|
|
self.aslr_backup = CustomMachineConfiguration.GetASLR()
|
|
CustomMachineConfiguration.SetASLR(0)
|
|
if self.governor != None:
|
|
self.governor_backup = CustomMachineConfiguration.GetCPUGovernor()
|
|
CustomMachineConfiguration.SetCPUGovernor(self.governor)
|
|
return self
|
|
|
|
def __exit__(self, type, value, traceback):
|
|
if self.aslr_backup != None:
|
|
CustomMachineConfiguration.SetASLR(self.aslr_backup)
|
|
if self.governor_backup != None:
|
|
CustomMachineConfiguration.SetCPUGovernor(self.governor_backup)
|
|
|
|
@staticmethod
|
|
def GetASLR():
|
|
try:
|
|
with open("/proc/sys/kernel/randomize_va_space", "r") as f:
|
|
return int(f.readline().strip())
|
|
except Exception:
|
|
logging.exception("Failed to get current ASLR settings.")
|
|
raise
|
|
|
|
@staticmethod
|
|
def SetASLR(value):
|
|
try:
|
|
with open("/proc/sys/kernel/randomize_va_space", "w") as f:
|
|
f.write(str(value))
|
|
except Exception:
|
|
logging.exception(
|
|
"Failed to update ASLR to %s. Are we running under sudo?", value)
|
|
raise
|
|
|
|
new_value = CustomMachineConfiguration.GetASLR()
|
|
if value != new_value:
|
|
raise Exception("Present value is %s" % new_value)
|
|
|
|
@staticmethod
|
|
def GetCPUCoresRange():
|
|
try:
|
|
with open("/sys/devices/system/cpu/present", "r") as f:
|
|
indexes = f.readline()
|
|
r = map(int, indexes.split("-"))
|
|
if len(r) == 1:
|
|
return range(r[0], r[0] + 1)
|
|
return range(r[0], r[1] + 1)
|
|
except Exception:
|
|
logging.exception("Failed to retrieve number of CPUs.")
|
|
raise
|
|
|
|
@staticmethod
|
|
def GetCPUPathForId(cpu_index):
|
|
ret = "/sys/devices/system/cpu/cpu"
|
|
ret += str(cpu_index)
|
|
ret += "/cpufreq/scaling_governor"
|
|
return ret
|
|
|
|
@staticmethod
|
|
def GetCPUGovernor():
|
|
try:
|
|
cpu_indices = CustomMachineConfiguration.GetCPUCoresRange()
|
|
ret = None
|
|
for cpu_index in cpu_indices:
|
|
cpu_device = CustomMachineConfiguration.GetCPUPathForId(cpu_index)
|
|
with open(cpu_device, "r") as f:
|
|
# We assume the governors of all CPUs are set to the same value
|
|
val = f.readline().strip()
|
|
if ret == None:
|
|
ret = val
|
|
elif ret != val:
|
|
raise Exception("CPU cores have differing governor settings")
|
|
return ret
|
|
except Exception:
|
|
logging.exception("Failed to get the current CPU governor. Is the CPU "
|
|
"governor disabled? Check BIOS.")
|
|
raise
|
|
|
|
@staticmethod
|
|
def SetCPUGovernor(value):
|
|
try:
|
|
cpu_indices = CustomMachineConfiguration.GetCPUCoresRange()
|
|
for cpu_index in cpu_indices:
|
|
cpu_device = CustomMachineConfiguration.GetCPUPathForId(cpu_index)
|
|
with open(cpu_device, "w") as f:
|
|
f.write(value)
|
|
|
|
except Exception:
|
|
logging.exception("Failed to change CPU governor to %s. Are we "
|
|
"running under sudo?", value)
|
|
raise
|
|
|
|
cur_value = CustomMachineConfiguration.GetCPUGovernor()
|
|
if cur_value != value:
|
|
raise Exception("Could not set CPU governor. Present value is %s"
|
|
% cur_value )
|
|
|
|
def Main(args):
|
|
parser = optparse.OptionParser()
|
|
parser.add_option("--android-build-tools", help="Deprecated.")
|
|
parser.add_option("--arch",
|
|
help=("The architecture to run tests for, "
|
|
"'auto' or 'native' for auto-detect"),
|
|
default="x64")
|
|
parser.add_option("--buildbot",
|
|
help="Adapt to path structure used on buildbots and adds "
|
|
"timestamps/level to all logged status messages",
|
|
default=False, action="store_true")
|
|
parser.add_option("--device",
|
|
help="The device ID to run Android tests on. If not given "
|
|
"it will be autodetected.")
|
|
parser.add_option("--extra-flags",
|
|
help="Additional flags to pass to the test executable",
|
|
default="")
|
|
parser.add_option("--json-test-results",
|
|
help="Path to a file for storing json results.")
|
|
parser.add_option("--json-test-results-secondary",
|
|
"--json-test-results-no-patch", # TODO(sergiyb): Deprecate.
|
|
help="Path to a file for storing json results from run "
|
|
"without patch or for reference build run.")
|
|
parser.add_option("--outdir", help="Base directory with compile output",
|
|
default="out")
|
|
parser.add_option("--outdir-secondary",
|
|
"--outdir-no-patch", # TODO(sergiyb): Deprecate.
|
|
help="Base directory with compile output without patch or "
|
|
"for reference build")
|
|
parser.add_option("--binary-override-path",
|
|
help="JavaScript engine binary. By default, d8 under "
|
|
"architecture-specific build dir. "
|
|
"Not supported in conjunction with outdir-secondary.")
|
|
parser.add_option("--prioritize",
|
|
help="Raise the priority to nice -20 for the benchmarking "
|
|
"process.Requires Linux, schedtool, and sudo privileges.",
|
|
default=False, action="store_true")
|
|
parser.add_option("--affinitize",
|
|
help="Run benchmarking process on the specified core. "
|
|
"For example: "
|
|
"--affinitize=0 will run the benchmark process on core 0. "
|
|
"--affinitize=3 will run the benchmark process on core 3. "
|
|
"Requires Linux, schedtool, and sudo privileges.",
|
|
default=None)
|
|
parser.add_option("--noaslr",
|
|
help="Disable ASLR for the duration of the benchmarked "
|
|
"process. Requires Linux and sudo privileges.",
|
|
default=False, action="store_true")
|
|
parser.add_option("--cpu-governor",
|
|
help="Set cpu governor to specified policy for the "
|
|
"duration of the benchmarked process. Typical options: "
|
|
"'powersave' for more stable results, or 'performance' "
|
|
"for shorter completion time of suite, with potentially "
|
|
"more noise in results.")
|
|
parser.add_option("--filter",
|
|
help="Only run the benchmarks beginning with this string. "
|
|
"For example: "
|
|
"--filter=JSTests/TypedArrays/ will run only TypedArray "
|
|
"benchmarks from the JSTests suite.",
|
|
default="")
|
|
parser.add_option("--run-count-multiplier", default=1, type="int",
|
|
help="Multipled used to increase number of times each test "
|
|
"is retried.")
|
|
parser.add_option("--dump-logcats-to",
|
|
help="Writes logcat output from each test into specified "
|
|
"directory. Only supported for android targets.")
|
|
|
|
(options, args) = parser.parse_args(args)
|
|
|
|
logging.basicConfig(level=logging.INFO, format="%(levelname)-8s %(message)s")
|
|
|
|
if len(args) == 0: # pragma: no cover
|
|
parser.print_help()
|
|
return INFRA_FAILURE_RETCODE
|
|
|
|
if options.arch in ["auto", "native"]: # pragma: no cover
|
|
options.arch = ARCH_GUESS
|
|
|
|
if not options.arch in SUPPORTED_ARCHS: # pragma: no cover
|
|
logging.error("Unknown architecture %s", options.arch)
|
|
return INFRA_FAILURE_RETCODE
|
|
|
|
if (options.json_test_results_secondary and
|
|
not options.outdir_secondary): # pragma: no cover
|
|
logging.error("For writing secondary json test results, a secondary outdir "
|
|
"patch must be specified.")
|
|
return INFRA_FAILURE_RETCODE
|
|
|
|
workspace = os.path.abspath(os.path.join(os.path.dirname(__file__), ".."))
|
|
|
|
if options.buildbot:
|
|
build_config = "Release"
|
|
else:
|
|
build_config = "%s.release" % options.arch
|
|
|
|
if options.binary_override_path == None:
|
|
options.shell_dir = os.path.join(workspace, options.outdir, build_config)
|
|
default_binary_name = "d8"
|
|
else:
|
|
if not os.path.isfile(options.binary_override_path):
|
|
logging.error("binary-override-path must be a file name")
|
|
return INFRA_FAILURE_RETCODE
|
|
if options.outdir_secondary:
|
|
logging.error("specify either binary-override-path or outdir-secondary")
|
|
return INFRA_FAILURE_RETCODE
|
|
options.shell_dir = os.path.abspath(
|
|
os.path.dirname(options.binary_override_path))
|
|
default_binary_name = os.path.basename(options.binary_override_path)
|
|
|
|
if options.outdir_secondary:
|
|
options.shell_dir_secondary = os.path.join(
|
|
workspace, options.outdir_secondary, build_config)
|
|
else:
|
|
options.shell_dir_secondary = None
|
|
|
|
if options.json_test_results:
|
|
options.json_test_results = os.path.abspath(options.json_test_results)
|
|
|
|
if options.json_test_results_secondary:
|
|
options.json_test_results_secondary = os.path.abspath(
|
|
options.json_test_results_secondary)
|
|
|
|
# Ensure all arguments have absolute path before we start changing current
|
|
# directory.
|
|
args = map(os.path.abspath, args)
|
|
|
|
prev_aslr = None
|
|
prev_cpu_gov = None
|
|
platform = Platform.GetPlatform(options)
|
|
|
|
results = Results()
|
|
results_secondary = Results()
|
|
# We use list here to allow modification in nested function below.
|
|
have_failed_tests = [False]
|
|
with CustomMachineConfiguration(governor = options.cpu_governor,
|
|
disable_aslr = options.noaslr) as conf:
|
|
for path in args:
|
|
if not os.path.exists(path): # pragma: no cover
|
|
results.errors.append("Configuration file %s does not exist." % path)
|
|
continue
|
|
|
|
with open(path) as f:
|
|
suite = json.loads(f.read())
|
|
|
|
# If no name is given, default to the file name without .json.
|
|
suite.setdefault("name", os.path.splitext(os.path.basename(path))[0])
|
|
|
|
# Setup things common to one test suite.
|
|
platform.PreExecution()
|
|
|
|
# Build the graph/trace tree structure.
|
|
default_parent = DefaultSentinel(default_binary_name)
|
|
root = BuildGraphConfigs(suite, options.arch, default_parent)
|
|
|
|
# Callback to be called on each node on traversal.
|
|
def NodeCB(node):
|
|
platform.PreTests(node, path)
|
|
|
|
# Traverse graph/trace tree and iterate over all runnables.
|
|
for runnable in FlattenRunnables(root, NodeCB):
|
|
runnable_name = "/".join(runnable.graphs)
|
|
if (not runnable_name.startswith(options.filter) and
|
|
runnable_name + "/" != options.filter):
|
|
continue
|
|
logging.info(">>> Running suite: %s", runnable_name)
|
|
|
|
def Runner():
|
|
"""Output generator that reruns several times."""
|
|
total_runs = runnable.run_count * options.run_count_multiplier
|
|
for i in xrange(0, max(1, total_runs)):
|
|
# TODO(machenbach): Allow timeout per arch like with run_count per
|
|
# arch.
|
|
try:
|
|
yield platform.Run(runnable, i)
|
|
except TestFailedError:
|
|
have_failed_tests[0] = True
|
|
|
|
# Let runnable iterate over all runs and handle output.
|
|
result, result_secondary = runnable.Run(
|
|
Runner, trybot=options.shell_dir_secondary)
|
|
results += result
|
|
results_secondary += result_secondary
|
|
platform.PostExecution()
|
|
|
|
if options.json_test_results:
|
|
results.WriteToFile(options.json_test_results)
|
|
else: # pragma: no cover
|
|
print results
|
|
|
|
if options.json_test_results_secondary:
|
|
results_secondary.WriteToFile(options.json_test_results_secondary)
|
|
else: # pragma: no cover
|
|
print results_secondary
|
|
|
|
if results.errors or have_failed_tests[0]:
|
|
return 1
|
|
|
|
return 0
|
|
|
|
|
|
def MainWrapper():
|
|
try:
|
|
return Main(sys.argv[1:])
|
|
except:
|
|
# Log uncaptured exceptions and report infra failure to the caller.
|
|
traceback.print_exc()
|
|
return INFRA_FAILURE_RETCODE
|
|
|
|
|
|
if __name__ == "__main__": # pragma: no cover
|
|
sys.exit(MainWrapper())
|