2012-10-08 13:49:57 +00:00
|
|
|
# Copyright 2012 the V8 project authors. All rights reserved.
|
|
|
|
# Redistribution and use in source and binary forms, with or without
|
|
|
|
# modification, are permitted provided that the following conditions are
|
|
|
|
# met:
|
|
|
|
#
|
|
|
|
# * Redistributions of source code must retain the above copyright
|
|
|
|
# notice, this list of conditions and the following disclaimer.
|
|
|
|
# * Redistributions in binary form must reproduce the above
|
|
|
|
# copyright notice, this list of conditions and the following
|
|
|
|
# disclaimer in the documentation and/or other materials provided
|
|
|
|
# with the distribution.
|
|
|
|
# * Neither the name of Google Inc. nor the names of its
|
|
|
|
# contributors may be used to endorse or promote products derived
|
|
|
|
# from this software without specific prior written permission.
|
|
|
|
#
|
|
|
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
|
|
|
|
"""Top-level presubmit script for V8.
|
|
|
|
|
|
|
|
See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts
|
|
|
|
for more details about the presubmit API built into gcl.
|
|
|
|
"""
|
|
|
|
|
2017-05-22 13:58:02 +00:00
|
|
|
import json
|
2019-09-20 12:21:46 +00:00
|
|
|
import os
|
2017-03-14 15:41:00 +00:00
|
|
|
import re
|
2014-06-03 13:39:42 +00:00
|
|
|
import sys
|
|
|
|
|
|
|
|
|
2014-09-23 08:14:06 +00:00
|
|
|
_EXCLUDED_PATHS = (
|
|
|
|
r"^test[\\\/].*",
|
|
|
|
r"^testing[\\\/].*",
|
|
|
|
r"^third_party[\\\/].*",
|
|
|
|
r"^tools[\\\/].*",
|
|
|
|
)
|
|
|
|
|
2019-01-09 14:38:22 +00:00
|
|
|
_LICENSE_FILE = (
|
|
|
|
r"LICENSE"
|
|
|
|
)
|
2014-09-23 08:14:06 +00:00
|
|
|
|
2017-05-03 09:58:03 +00:00
|
|
|
# Regular expression that matches code which should not be run through cpplint.
|
|
|
|
_NO_LINT_PATHS = (
|
|
|
|
r'src[\\\/]base[\\\/]export-template\.h',
|
|
|
|
)
|
|
|
|
|
|
|
|
|
2014-09-23 08:14:06 +00:00
|
|
|
# Regular expression that matches code only used for test binaries
|
|
|
|
# (best effort).
|
|
|
|
_TEST_CODE_EXCLUDED_PATHS = (
|
|
|
|
r'.+-unittest\.cc',
|
|
|
|
# Has a method VisitForTest().
|
|
|
|
r'src[\\\/]compiler[\\\/]ast-graph-builder\.cc',
|
|
|
|
# Test extension.
|
|
|
|
r'src[\\\/]extensions[\\\/]gc-extension\.cc',
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
_TEST_ONLY_WARNING = (
|
|
|
|
'You might be calling functions intended only for testing from\n'
|
|
|
|
'production code. It is OK to ignore this warning if you know what\n'
|
|
|
|
'you are doing, as the heuristics used to detect the situation are\n'
|
|
|
|
'not perfect. The commit queue will not block on this warning.')
|
|
|
|
|
|
|
|
|
2012-10-08 13:49:57 +00:00
|
|
|
def _V8PresubmitChecks(input_api, output_api):
|
|
|
|
"""Runs the V8 presubmit checks."""
|
|
|
|
import sys
|
|
|
|
sys.path.append(input_api.os_path.join(
|
|
|
|
input_api.PresubmitLocalPath(), 'tools'))
|
2018-10-30 14:23:38 +00:00
|
|
|
from v8_presubmit import CppLintProcessor
|
2019-01-02 10:15:28 +00:00
|
|
|
from v8_presubmit import TorqueLintProcessor
|
2018-10-30 14:23:38 +00:00
|
|
|
from v8_presubmit import SourceProcessor
|
|
|
|
from v8_presubmit import StatusFilesProcessor
|
2012-10-08 13:49:57 +00:00
|
|
|
|
2017-05-03 09:58:03 +00:00
|
|
|
def FilterFile(affected_file):
|
|
|
|
return input_api.FilterSourceFile(
|
|
|
|
affected_file,
|
|
|
|
white_list=None,
|
|
|
|
black_list=_NO_LINT_PATHS)
|
|
|
|
|
2018-09-25 10:59:03 +00:00
|
|
|
def FilterTorqueFile(affected_file):
|
|
|
|
return input_api.FilterSourceFile(
|
|
|
|
affected_file,
|
|
|
|
white_list=(r'.+\.tq'))
|
|
|
|
|
2012-10-08 13:49:57 +00:00
|
|
|
results = []
|
2016-11-28 08:30:09 +00:00
|
|
|
if not CppLintProcessor().RunOnFiles(
|
2017-05-03 09:58:03 +00:00
|
|
|
input_api.AffectedFiles(file_filter=FilterFile, include_deletes=False)):
|
2012-10-08 13:49:57 +00:00
|
|
|
results.append(output_api.PresubmitError("C++ lint check failed"))
|
2019-02-20 14:06:04 +00:00
|
|
|
if not TorqueLintProcessor().RunOnFiles(
|
|
|
|
input_api.AffectedFiles(file_filter=FilterTorqueFile,
|
|
|
|
include_deletes=False)):
|
2019-02-20 14:06:46 +00:00
|
|
|
results.append(output_api.PresubmitError("Torque format check failed"))
|
2016-11-28 08:30:09 +00:00
|
|
|
if not SourceProcessor().RunOnFiles(
|
|
|
|
input_api.AffectedFiles(include_deletes=False)):
|
2012-10-08 13:49:57 +00:00
|
|
|
results.append(output_api.PresubmitError(
|
2013-07-05 09:52:11 +00:00
|
|
|
"Copyright header, trailing whitespaces and two empty lines " \
|
|
|
|
"between declarations check failed"))
|
2016-12-06 11:41:35 +00:00
|
|
|
if not StatusFilesProcessor().RunOnFiles(
|
2017-02-16 13:11:14 +00:00
|
|
|
input_api.AffectedFiles(include_deletes=True)):
|
2015-11-25 13:18:49 +00:00
|
|
|
results.append(output_api.PresubmitError("Status file check failed"))
|
2016-12-06 11:31:09 +00:00
|
|
|
results.extend(input_api.canned_checks.CheckAuthorizedAuthor(
|
2018-05-26 00:37:50 +00:00
|
|
|
input_api, output_api, bot_whitelist=[
|
|
|
|
'v8-ci-autoroll-builder@chops-service-accounts.iam.gserviceaccount.com'
|
|
|
|
]))
|
2012-10-08 13:49:57 +00:00
|
|
|
return results
|
|
|
|
|
|
|
|
|
2014-06-03 13:39:42 +00:00
|
|
|
def _CheckUnwantedDependencies(input_api, output_api):
|
|
|
|
"""Runs checkdeps on #include statements added in this
|
|
|
|
change. Breaking - rules is an error, breaking ! rules is a
|
|
|
|
warning.
|
|
|
|
"""
|
|
|
|
# We need to wait until we have an input_api object and use this
|
|
|
|
# roundabout construct to import checkdeps because this file is
|
|
|
|
# eval-ed and thus doesn't have __file__.
|
|
|
|
original_sys_path = sys.path
|
|
|
|
try:
|
|
|
|
sys.path = sys.path + [input_api.os_path.join(
|
|
|
|
input_api.PresubmitLocalPath(), 'buildtools', 'checkdeps')]
|
|
|
|
import checkdeps
|
|
|
|
from cpp_checker import CppChecker
|
|
|
|
from rules import Rule
|
|
|
|
finally:
|
|
|
|
# Restore sys.path to what it was before.
|
|
|
|
sys.path = original_sys_path
|
|
|
|
|
2019-09-20 12:21:46 +00:00
|
|
|
def _FilesImpactedByDepsChange(files):
|
|
|
|
all_files = [f.AbsoluteLocalPath() for f in files]
|
|
|
|
deps_files = [p for p in all_files if IsDepsFile(p)]
|
|
|
|
impacted_files = union([_CollectImpactedFiles(path) for path in deps_files])
|
|
|
|
impacted_file_objs = [ImpactedFile(path) for path in impacted_files]
|
|
|
|
return impacted_file_objs
|
|
|
|
|
|
|
|
def IsDepsFile(p):
|
|
|
|
return os.path.isfile(p) and os.path.basename(p) == 'DEPS'
|
|
|
|
|
|
|
|
def union(list_of_lists):
|
|
|
|
"""Ensure no duplicates"""
|
|
|
|
return set(sum(list_of_lists, []))
|
|
|
|
|
|
|
|
def _CollectImpactedFiles(deps_file):
|
|
|
|
# TODO(liviurau): Do not walk paths twice. Then we have no duplicates.
|
|
|
|
# Higher level DEPS changes may dominate lower level DEPS changes.
|
|
|
|
# TODO(liviurau): Check if DEPS changed in the right way.
|
|
|
|
# 'include_rules' impact c++ files but 'vars' or 'deps' do not.
|
|
|
|
# Maybe we just eval both old and new DEPS content and check
|
|
|
|
# if the list are the same.
|
|
|
|
result = []
|
|
|
|
parent_dir = os.path.dirname(deps_file)
|
|
|
|
for relative_f in input_api.change.AllFiles(parent_dir):
|
|
|
|
abs_f = os.path.join(parent_dir, relative_f)
|
|
|
|
if CppChecker.IsCppFile(abs_f):
|
|
|
|
result.append(abs_f)
|
|
|
|
return result
|
|
|
|
|
|
|
|
class ImpactedFile(object):
|
|
|
|
"""Duck type version of AffectedFile needed to check files under directories
|
|
|
|
where a DEPS file changed. Extend the interface along the line of
|
|
|
|
AffectedFile if you need it for other checks."""
|
|
|
|
|
|
|
|
def __init__(self, path):
|
|
|
|
self._path = path
|
|
|
|
|
|
|
|
def LocalPath(self):
|
|
|
|
path = self._path.replace(os.sep, '/')
|
|
|
|
return os.path.normpath(path)
|
|
|
|
|
|
|
|
def ChangedContents(self):
|
|
|
|
with open(self._path) as f:
|
|
|
|
# TODO(liviurau): read only '#include' lines
|
|
|
|
lines = f.readlines()
|
|
|
|
return enumerate(lines, start=1)
|
|
|
|
|
|
|
|
def _FilterDuplicates(impacted_files, affected_files):
|
|
|
|
""""We include all impacted files but exclude affected files that are also
|
|
|
|
impacted. Files impacted by DEPS changes take precedence before files
|
|
|
|
affected by direct changes."""
|
|
|
|
result = impacted_files[:]
|
|
|
|
only_paths = set([imf.LocalPath() for imf in impacted_files])
|
|
|
|
for af in affected_files:
|
|
|
|
if not af.LocalPath() in only_paths:
|
|
|
|
result.append(af)
|
|
|
|
return result
|
|
|
|
|
2014-06-03 13:39:42 +00:00
|
|
|
added_includes = []
|
2019-09-20 12:21:46 +00:00
|
|
|
affected_files = input_api.AffectedFiles()
|
|
|
|
impacted_by_deps = _FilesImpactedByDepsChange(affected_files)
|
|
|
|
for f in _FilterDuplicates(impacted_by_deps, affected_files):
|
2014-06-03 13:39:42 +00:00
|
|
|
if not CppChecker.IsCppFile(f.LocalPath()):
|
|
|
|
continue
|
|
|
|
|
|
|
|
changed_lines = [line for line_num, line in f.ChangedContents()]
|
|
|
|
added_includes.append([f.LocalPath(), changed_lines])
|
|
|
|
|
|
|
|
deps_checker = checkdeps.DepsChecker(input_api.PresubmitLocalPath())
|
|
|
|
|
|
|
|
error_descriptions = []
|
|
|
|
warning_descriptions = []
|
|
|
|
for path, rule_type, rule_description in deps_checker.CheckAddedCppIncludes(
|
|
|
|
added_includes):
|
|
|
|
description_with_path = '%s\n %s' % (path, rule_description)
|
|
|
|
if rule_type == Rule.DISALLOW:
|
|
|
|
error_descriptions.append(description_with_path)
|
|
|
|
else:
|
|
|
|
warning_descriptions.append(description_with_path)
|
|
|
|
|
|
|
|
results = []
|
|
|
|
if error_descriptions:
|
|
|
|
results.append(output_api.PresubmitError(
|
|
|
|
'You added one or more #includes that violate checkdeps rules.',
|
|
|
|
error_descriptions))
|
|
|
|
if warning_descriptions:
|
|
|
|
results.append(output_api.PresubmitPromptOrNotify(
|
|
|
|
'You added one or more #includes of files that are temporarily\n'
|
|
|
|
'allowed but being removed. Can you avoid introducing the\n'
|
|
|
|
'#include? See relevant DEPS file(s) for details and contacts.',
|
|
|
|
warning_descriptions))
|
|
|
|
return results
|
|
|
|
|
|
|
|
|
2018-02-02 12:05:19 +00:00
|
|
|
def _CheckHeadersHaveIncludeGuards(input_api, output_api):
|
|
|
|
"""Ensures that all header files have include guards."""
|
|
|
|
file_inclusion_pattern = r'src/.+\.h'
|
|
|
|
|
|
|
|
def FilterFile(affected_file):
|
|
|
|
black_list = (_EXCLUDED_PATHS +
|
|
|
|
input_api.DEFAULT_BLACK_LIST)
|
|
|
|
return input_api.FilterSourceFile(
|
|
|
|
affected_file,
|
|
|
|
white_list=(file_inclusion_pattern, ),
|
|
|
|
black_list=black_list)
|
|
|
|
|
|
|
|
leading_src_pattern = input_api.re.compile(r'^src/')
|
|
|
|
dash_dot_slash_pattern = input_api.re.compile(r'[-./]')
|
|
|
|
def PathToGuardMacro(path):
|
|
|
|
"""Guards should be of the form V8_PATH_TO_FILE_WITHOUT_SRC_H_."""
|
|
|
|
x = input_api.re.sub(leading_src_pattern, 'v8_', path)
|
|
|
|
x = input_api.re.sub(dash_dot_slash_pattern, '_', x)
|
|
|
|
x = x.upper() + "_"
|
|
|
|
return x
|
|
|
|
|
|
|
|
problems = []
|
|
|
|
for f in input_api.AffectedSourceFiles(FilterFile):
|
|
|
|
local_path = f.LocalPath()
|
|
|
|
guard_macro = PathToGuardMacro(local_path)
|
|
|
|
guard_patterns = [
|
|
|
|
input_api.re.compile(r'^#ifndef ' + guard_macro + '$'),
|
|
|
|
input_api.re.compile(r'^#define ' + guard_macro + '$'),
|
|
|
|
input_api.re.compile(r'^#endif // ' + guard_macro + '$')]
|
|
|
|
skip_check_pattern = input_api.re.compile(
|
|
|
|
r'^// PRESUBMIT_INTENTIONALLY_MISSING_INCLUDE_GUARD')
|
|
|
|
found_patterns = [ False, False, False ]
|
|
|
|
file_omitted = False
|
|
|
|
|
|
|
|
for line in f.NewContents():
|
|
|
|
for i in range(len(guard_patterns)):
|
|
|
|
if guard_patterns[i].match(line):
|
|
|
|
found_patterns[i] = True
|
|
|
|
if skip_check_pattern.match(line):
|
|
|
|
file_omitted = True
|
|
|
|
break
|
|
|
|
|
|
|
|
if not file_omitted and not all(found_patterns):
|
|
|
|
problems.append(
|
|
|
|
'%s: Missing include guard \'%s\'' % (local_path, guard_macro))
|
|
|
|
|
|
|
|
if problems:
|
|
|
|
return [output_api.PresubmitError(
|
|
|
|
'You added one or more header files without an appropriate\n'
|
|
|
|
'include guard. Add the include guard {#ifndef,#define,#endif}\n'
|
|
|
|
'triplet or omit the check entirely through the magic comment:\n'
|
|
|
|
'"// PRESUBMIT_INTENTIONALLY_MISSING_INCLUDE_GUARD".', problems)]
|
|
|
|
else:
|
|
|
|
return []
|
|
|
|
|
|
|
|
|
2017-09-14 09:42:32 +00:00
|
|
|
# TODO(mstarzinger): Similar checking should be made available as part of
|
|
|
|
# tools/presubmit.py (note that tools/check-inline-includes.sh exists).
|
2015-08-18 16:41:02 +00:00
|
|
|
def _CheckNoInlineHeaderIncludesInNormalHeaders(input_api, output_api):
|
|
|
|
"""Attempts to prevent inclusion of inline headers into normal header
|
|
|
|
files. This tries to establish a layering where inline headers can be
|
|
|
|
included by other inline headers or compilation units only."""
|
|
|
|
file_inclusion_pattern = r'(?!.+-inl\.h).+\.h'
|
|
|
|
include_directive_pattern = input_api.re.compile(r'#include ".+-inl.h"')
|
2017-09-14 09:42:32 +00:00
|
|
|
include_error = (
|
|
|
|
'You are including an inline header (e.g. foo-inl.h) within a normal\n'
|
|
|
|
'header (e.g. bar.h) file. This violates layering of dependencies.')
|
2015-08-18 16:41:02 +00:00
|
|
|
|
|
|
|
def FilterFile(affected_file):
|
|
|
|
black_list = (_EXCLUDED_PATHS +
|
|
|
|
input_api.DEFAULT_BLACK_LIST)
|
|
|
|
return input_api.FilterSourceFile(
|
|
|
|
affected_file,
|
|
|
|
white_list=(file_inclusion_pattern, ),
|
|
|
|
black_list=black_list)
|
|
|
|
|
|
|
|
problems = []
|
|
|
|
for f in input_api.AffectedSourceFiles(FilterFile):
|
|
|
|
local_path = f.LocalPath()
|
|
|
|
for line_number, line in f.ChangedContents():
|
|
|
|
if (include_directive_pattern.search(line)):
|
|
|
|
problems.append(
|
|
|
|
'%s:%d\n %s' % (local_path, line_number, line.strip()))
|
|
|
|
|
|
|
|
if problems:
|
2017-09-14 09:42:32 +00:00
|
|
|
return [output_api.PresubmitError(include_error, problems)]
|
2015-08-18 16:41:02 +00:00
|
|
|
else:
|
|
|
|
return []
|
|
|
|
|
|
|
|
|
2014-09-23 08:14:06 +00:00
|
|
|
def _CheckNoProductionCodeUsingTestOnlyFunctions(input_api, output_api):
|
|
|
|
"""Attempts to prevent use of functions intended only for testing in
|
|
|
|
non-testing code. For now this is just a best-effort implementation
|
|
|
|
that ignores header files and may have some false positives. A
|
|
|
|
better implementation would probably need a proper C++ parser.
|
|
|
|
"""
|
|
|
|
# We only scan .cc files, as the declaration of for-testing functions in
|
|
|
|
# header files are hard to distinguish from calls to such functions without a
|
|
|
|
# proper C++ parser.
|
|
|
|
file_inclusion_pattern = r'.+\.cc'
|
|
|
|
|
|
|
|
base_function_pattern = r'[ :]test::[^\s]+|ForTest(ing)?|for_test(ing)?'
|
|
|
|
inclusion_pattern = input_api.re.compile(r'(%s)\s*\(' % base_function_pattern)
|
|
|
|
comment_pattern = input_api.re.compile(r'//.*(%s)' % base_function_pattern)
|
|
|
|
exclusion_pattern = input_api.re.compile(
|
|
|
|
r'::[A-Za-z0-9_]+(%s)|(%s)[^;]+\{' % (
|
|
|
|
base_function_pattern, base_function_pattern))
|
|
|
|
|
|
|
|
def FilterFile(affected_file):
|
|
|
|
black_list = (_EXCLUDED_PATHS +
|
|
|
|
_TEST_CODE_EXCLUDED_PATHS +
|
|
|
|
input_api.DEFAULT_BLACK_LIST)
|
|
|
|
return input_api.FilterSourceFile(
|
|
|
|
affected_file,
|
|
|
|
white_list=(file_inclusion_pattern, ),
|
|
|
|
black_list=black_list)
|
|
|
|
|
|
|
|
problems = []
|
|
|
|
for f in input_api.AffectedSourceFiles(FilterFile):
|
|
|
|
local_path = f.LocalPath()
|
|
|
|
for line_number, line in f.ChangedContents():
|
|
|
|
if (inclusion_pattern.search(line) and
|
|
|
|
not comment_pattern.search(line) and
|
|
|
|
not exclusion_pattern.search(line)):
|
|
|
|
problems.append(
|
|
|
|
'%s:%d\n %s' % (local_path, line_number, line.strip()))
|
|
|
|
|
|
|
|
if problems:
|
|
|
|
return [output_api.PresubmitPromptOrNotify(_TEST_ONLY_WARNING, problems)]
|
|
|
|
else:
|
|
|
|
return []
|
|
|
|
|
|
|
|
|
2019-09-13 12:32:03 +00:00
|
|
|
def _CheckGenderNeutralInLicenses(input_api, output_api):
|
|
|
|
# License files are taken as is, even if they include gendered pronouns.
|
|
|
|
def LicenseFilter(path):
|
|
|
|
input_api.FilterSourceFile(path, black_list=_LICENSE_FILE)
|
|
|
|
|
|
|
|
return input_api.canned_checks.CheckGenderNeutral(
|
|
|
|
input_api, output_api, source_file_filter=LicenseFilter)
|
|
|
|
|
|
|
|
|
|
|
|
def _RunTestsWithVPythonSpec(input_api, output_api):
|
|
|
|
return input_api.RunTests(
|
|
|
|
input_api.canned_checks.CheckVPythonSpec(input_api, output_api))
|
|
|
|
|
|
|
|
|
2012-10-08 13:49:57 +00:00
|
|
|
def _CommonChecks(input_api, output_api):
|
|
|
|
"""Checks common to both upload and commit."""
|
2018-04-26 08:34:07 +00:00
|
|
|
# TODO(machenbach): Replace some of those checks, e.g. owners and copyright,
|
|
|
|
# with the canned PanProjectChecks. Need to make sure that the checks all
|
|
|
|
# pass on all existing files.
|
2019-09-13 12:32:03 +00:00
|
|
|
checks = [
|
|
|
|
input_api.canned_checks.CheckOwnersFormat,
|
|
|
|
input_api.canned_checks.CheckOwners,
|
|
|
|
_CheckCommitMessageBugEntry,
|
|
|
|
input_api.canned_checks.CheckPatchFormatted,
|
|
|
|
_CheckGenderNeutralInLicenses,
|
|
|
|
_V8PresubmitChecks,
|
|
|
|
_CheckUnwantedDependencies,
|
|
|
|
_CheckNoProductionCodeUsingTestOnlyFunctions,
|
|
|
|
_CheckHeadersHaveIncludeGuards,
|
|
|
|
_CheckNoInlineHeaderIncludesInNormalHeaders,
|
|
|
|
_CheckJSONFiles,
|
|
|
|
_CheckMacroUndefs,
|
|
|
|
_CheckNoexceptAnnotations,
|
|
|
|
_RunTestsWithVPythonSpec,
|
|
|
|
]
|
|
|
|
|
|
|
|
return sum([check(input_api, output_api) for check in checks], [])
|
2012-10-08 13:49:57 +00:00
|
|
|
|
|
|
|
|
2013-09-24 13:35:45 +00:00
|
|
|
def _SkipTreeCheck(input_api, output_api):
|
|
|
|
"""Check the env var whether we want to skip tree check.
|
2015-02-26 08:59:17 +00:00
|
|
|
Only skip if include/v8-version.h has been updated."""
|
|
|
|
src_version = 'include/v8-version.h'
|
2014-01-10 15:17:49 +00:00
|
|
|
if not input_api.AffectedSourceFiles(
|
|
|
|
lambda file: file.LocalPath() == src_version):
|
2013-09-24 13:35:45 +00:00
|
|
|
return False
|
|
|
|
return input_api.environ.get('PRESUBMIT_TREE_CHECK') == 'skip'
|
|
|
|
|
|
|
|
|
2017-03-14 14:46:13 +00:00
|
|
|
def _CheckCommitMessageBugEntry(input_api, output_api):
|
|
|
|
"""Check that bug entries are well-formed in commit message."""
|
2017-03-14 15:41:00 +00:00
|
|
|
bogus_bug_msg = (
|
|
|
|
'Bogus BUG entry: %s. Please specify the issue tracker prefix and the '
|
|
|
|
'issue number, separated by a colon, e.g. v8:123 or chromium:12345.')
|
2017-03-14 14:46:13 +00:00
|
|
|
results = []
|
|
|
|
for bug in (input_api.change.BUG or '').split(','):
|
|
|
|
bug = bug.strip()
|
|
|
|
if 'none'.startswith(bug.lower()):
|
|
|
|
continue
|
|
|
|
if ':' not in bug:
|
|
|
|
try:
|
|
|
|
if int(bug) > 100000:
|
|
|
|
# Rough indicator for current chromium bugs.
|
2017-03-14 15:41:00 +00:00
|
|
|
prefix_guess = 'chromium'
|
|
|
|
else:
|
|
|
|
prefix_guess = 'v8'
|
|
|
|
results.append('BUG entry requires issue tracker prefix, e.g. %s:%s' %
|
|
|
|
(prefix_guess, bug))
|
2017-03-14 14:46:13 +00:00
|
|
|
except ValueError:
|
2017-03-14 15:41:00 +00:00
|
|
|
results.append(bogus_bug_msg % bug)
|
|
|
|
elif not re.match(r'\w+:\d+', bug):
|
|
|
|
results.append(bogus_bug_msg % bug)
|
2017-03-14 14:46:13 +00:00
|
|
|
return [output_api.PresubmitError(r) for r in results]
|
|
|
|
|
|
|
|
|
2017-05-22 13:58:02 +00:00
|
|
|
def _CheckJSONFiles(input_api, output_api):
|
|
|
|
def FilterFile(affected_file):
|
|
|
|
return input_api.FilterSourceFile(
|
|
|
|
affected_file,
|
|
|
|
white_list=(r'.+\.json',))
|
|
|
|
|
|
|
|
results = []
|
|
|
|
for f in input_api.AffectedFiles(
|
|
|
|
file_filter=FilterFile, include_deletes=False):
|
|
|
|
with open(f.LocalPath()) as j:
|
|
|
|
try:
|
|
|
|
json.load(j)
|
|
|
|
except Exception as e:
|
|
|
|
results.append(
|
|
|
|
'JSON validation failed for %s. Error:\n%s' % (f.LocalPath(), e))
|
|
|
|
|
|
|
|
return [output_api.PresubmitError(r) for r in results]
|
|
|
|
|
|
|
|
|
2017-09-11 12:14:14 +00:00
|
|
|
def _CheckMacroUndefs(input_api, output_api):
|
|
|
|
"""
|
|
|
|
Checks that each #define in a .cc file is eventually followed by an #undef.
|
|
|
|
|
2019-09-26 08:16:50 +00:00
|
|
|
TODO(clemensb): This check should eventually be enabled for all cc files via
|
2017-09-11 12:14:14 +00:00
|
|
|
tools/presubmit.py (https://crbug.com/v8/6811).
|
|
|
|
"""
|
|
|
|
def FilterFile(affected_file):
|
|
|
|
# Skip header files, as they often define type lists which are used in
|
|
|
|
# other files.
|
|
|
|
white_list = (r'.+\.cc',r'.+\.cpp',r'.+\.c')
|
|
|
|
return input_api.FilterSourceFile(affected_file, white_list=white_list)
|
|
|
|
|
2019-09-13 12:32:03 +00:00
|
|
|
def Touches(line):
|
|
|
|
return line.startswith('+') or line.startswith('-')
|
|
|
|
|
|
|
|
def InvolvesMacros(text):
|
|
|
|
return define_pattern.match(text) or undef_pattern.match(text)
|
|
|
|
|
2017-09-11 12:14:14 +00:00
|
|
|
def TouchesMacros(f):
|
2019-09-13 12:32:03 +00:00
|
|
|
return any(Touches(line) and InvolvesMacros(line[1:])
|
|
|
|
for line in f.GenerateScmDiff().splitlines())
|
|
|
|
|
|
|
|
def CollectUndefsWithNoDef(defined_macros, errors, f, line, line_nr):
|
|
|
|
define_match = define_pattern.match(line)
|
|
|
|
if define_match:
|
|
|
|
name = define_match.group(1)
|
|
|
|
defined_macros[name] = line_nr
|
|
|
|
undef_match = undef_pattern.match(line)
|
|
|
|
if undef_match and not "// NOLINT" in line:
|
|
|
|
name = undef_match.group(1)
|
|
|
|
if name in defined_macros:
|
|
|
|
del defined_macros[name]
|
|
|
|
else:
|
|
|
|
errors.append('{}:{}: Macro named \'{}\' was not defined before.'
|
|
|
|
.format(f.LocalPath(), line_nr, name))
|
2017-09-11 12:14:14 +00:00
|
|
|
|
|
|
|
define_pattern = input_api.re.compile(r'#define (\w+)')
|
|
|
|
undef_pattern = input_api.re.compile(r'#undef (\w+)')
|
|
|
|
errors = []
|
|
|
|
for f in input_api.AffectedFiles(
|
|
|
|
file_filter=FilterFile, include_deletes=False):
|
|
|
|
if not TouchesMacros(f):
|
|
|
|
continue
|
|
|
|
|
|
|
|
defined_macros = dict()
|
|
|
|
with open(f.LocalPath()) as fh:
|
2019-09-13 12:32:03 +00:00
|
|
|
for line_nr, line in enumerate(fh, start=1):
|
|
|
|
CollectUndefsWithNoDef(defined_macros, errors, f, line, line_nr)
|
|
|
|
|
2017-09-11 12:14:14 +00:00
|
|
|
for name, line_nr in sorted(defined_macros.items(), key=lambda e: e[1]):
|
|
|
|
errors.append('{}:{}: Macro missing #undef: {}'
|
|
|
|
.format(f.LocalPath(), line_nr, name))
|
|
|
|
|
|
|
|
if errors:
|
|
|
|
return [output_api.PresubmitPromptOrNotify(
|
|
|
|
'Detected mismatches in #define / #undef in the file(s) where you '
|
|
|
|
'modified preprocessor macros.',
|
|
|
|
errors)]
|
|
|
|
return []
|
|
|
|
|
|
|
|
|
2018-12-20 13:20:59 +00:00
|
|
|
def _CheckNoexceptAnnotations(input_api, output_api):
|
|
|
|
"""
|
|
|
|
Checks that all user-defined constructors and assignment operators are marked
|
|
|
|
V8_NOEXCEPT.
|
|
|
|
|
|
|
|
This is required for standard containers to pick the right constructors. Our
|
|
|
|
macros (like MOVE_ONLY_WITH_DEFAULT_CONSTRUCTORS) add this automatically.
|
|
|
|
Omitting it at some places can result in weird compiler errors if this is
|
|
|
|
mixed with other classes that have the annotation.
|
|
|
|
|
2019-09-26 08:16:50 +00:00
|
|
|
TODO(clemensb): This check should eventually be enabled for all files via
|
2018-12-20 13:20:59 +00:00
|
|
|
tools/presubmit.py (https://crbug.com/v8/8616).
|
|
|
|
"""
|
|
|
|
|
2019-01-03 13:11:38 +00:00
|
|
|
def FilterFile(affected_file):
|
|
|
|
return input_api.FilterSourceFile(
|
|
|
|
affected_file,
|
|
|
|
white_list=(r'src/.*', r'test/.*'))
|
|
|
|
|
|
|
|
|
2018-12-20 13:20:59 +00:00
|
|
|
# matches any class name.
|
2019-01-07 12:35:49 +00:00
|
|
|
class_name = r'\b([A-Z][A-Za-z0-9_:]*)(?:::\1)?'
|
2018-12-20 13:20:59 +00:00
|
|
|
# initial class name is potentially followed by this to declare an assignment
|
|
|
|
# operator.
|
2019-01-07 12:35:49 +00:00
|
|
|
potential_assignment = r'(?:&\s+(?:\1::)?operator=)?\s*'
|
2018-12-20 13:20:59 +00:00
|
|
|
# matches an argument list that contains only a reference to a class named
|
|
|
|
# like the first capture group, potentially const.
|
2019-01-07 12:35:49 +00:00
|
|
|
single_class_ref_arg = r'\(\s*(?:const\s+)?\1(?:::\1)?&&?[^,;)]*\)'
|
2018-12-20 14:35:56 +00:00
|
|
|
# matches anything but a sequence of whitespaces followed by either
|
|
|
|
# V8_NOEXCEPT or "= delete".
|
|
|
|
not_followed_by_noexcept = r'(?!\s+(?:V8_NOEXCEPT|=\s+delete)\b)'
|
2018-12-20 13:20:59 +00:00
|
|
|
full_pattern = r'^.*?' + class_name + potential_assignment + \
|
|
|
|
single_class_ref_arg + not_followed_by_noexcept + '.*?$'
|
|
|
|
regexp = input_api.re.compile(full_pattern, re.MULTILINE)
|
|
|
|
|
|
|
|
errors = []
|
2019-01-03 13:11:38 +00:00
|
|
|
for f in input_api.AffectedFiles(file_filter=FilterFile,
|
|
|
|
include_deletes=False):
|
2018-12-20 13:20:59 +00:00
|
|
|
with open(f.LocalPath()) as fh:
|
|
|
|
for match in re.finditer(regexp, fh.read()):
|
2018-12-20 14:35:56 +00:00
|
|
|
errors.append('in {}: {}'.format(f.LocalPath(),
|
|
|
|
match.group().strip()))
|
2018-12-20 13:20:59 +00:00
|
|
|
|
|
|
|
if errors:
|
|
|
|
return [output_api.PresubmitPromptOrNotify(
|
|
|
|
'Copy constructors, move constructors, copy assignment operators and '
|
|
|
|
'move assignment operators should be marked V8_NOEXCEPT.\n'
|
|
|
|
'Please report false positives on https://crbug.com/v8/8616.',
|
|
|
|
errors)]
|
|
|
|
return []
|
|
|
|
|
|
|
|
|
2012-10-08 13:49:57 +00:00
|
|
|
def CheckChangeOnUpload(input_api, output_api):
|
|
|
|
results = []
|
|
|
|
results.extend(_CommonChecks(input_api, output_api))
|
|
|
|
return results
|
|
|
|
|
|
|
|
|
|
|
|
def CheckChangeOnCommit(input_api, output_api):
|
|
|
|
results = []
|
|
|
|
results.extend(_CommonChecks(input_api, output_api))
|
|
|
|
results.extend(input_api.canned_checks.CheckChangeHasDescription(
|
|
|
|
input_api, output_api))
|
2013-09-24 13:35:45 +00:00
|
|
|
if not _SkipTreeCheck(input_api, output_api):
|
|
|
|
results.extend(input_api.canned_checks.CheckTreeIsOpen(
|
|
|
|
input_api, output_api,
|
|
|
|
json_url='http://v8-status.appspot.com/current?format=json'))
|
2012-10-08 13:49:57 +00:00
|
|
|
return results
|