9766f00166
BUG=skia:5719 GOLD_TRYBOT_URL= https://gold.skia.org/search?issue=2360203004 Review-Url: https://codereview.chromium.org/2360203004
750 lines
24 KiB
Python
750 lines
24 KiB
Python
# Copyright 2016 The Chromium Authors. All rights reserved.
|
|
# Use of this source code is governed by a BSD-style license that can be
|
|
# found in the LICENSE file.
|
|
|
|
|
|
# Recipe module for Skia Swarming trigger.
|
|
|
|
|
|
import os
|
|
import json
|
|
|
|
|
|
DEPS = [
|
|
'build/file',
|
|
'build/gsutil',
|
|
'builder_name_schema',
|
|
'core',
|
|
'depot_tools/depot_tools',
|
|
'depot_tools/git',
|
|
'depot_tools/tryserver',
|
|
'recipe_engine/json',
|
|
'recipe_engine/path',
|
|
'recipe_engine/properties',
|
|
'recipe_engine/python',
|
|
'recipe_engine/raw_io',
|
|
'recipe_engine/step',
|
|
'recipe_engine/time',
|
|
'run',
|
|
'swarming',
|
|
'vars',
|
|
]
|
|
|
|
|
|
TEST_BUILDERS = {
|
|
'client.skia': {
|
|
'skiabot-linux-swarm-000': [
|
|
'Build-Mac-Clang-Arm64-Release-Android',
|
|
'Build-Mac-Clang-x86_64-Release',
|
|
'Build-Ubuntu-GCC-Arm64-Debug-Android_Vulkan',
|
|
'Build-Ubuntu-GCC-x86_64-Debug',
|
|
'Build-Ubuntu-GCC-x86_64-Debug-GN',
|
|
'Build-Ubuntu-GCC-x86_64-Release-RemoteRun',
|
|
'Build-Ubuntu-GCC-x86_64-Release-Trybot',
|
|
'Build-Win-MSVC-x86_64-Release',
|
|
'Build-Win-MSVC-x86_64-Release-Vulkan',
|
|
'Housekeeper-Nightly-RecreateSKPs_Canary',
|
|
'Housekeeper-PerCommit',
|
|
'Infra-PerCommit',
|
|
'Perf-Ubuntu-GCC-GCE-CPU-AVX2-x86_64-Release-Trybot',
|
|
'Perf-Ubuntu-GCC-Golo-GPU-GT610-x86_64-Release-CT_BENCH_1k_SKPs',
|
|
'Test-Android-Clang-AndroidOne-CPU-MT6582-arm-Debug-GN_Android',
|
|
'Test-Android-GCC-NVIDIA_Shield-GPU-TegraX1-Arm64-Debug-Vulkan',
|
|
'Test-Android-GCC-Nexus7v2-GPU-Tegra3-Arm7-Release',
|
|
'Test-Mac-Clang-MacMini6.2-CPU-AVX-x86_64-Release',
|
|
'Test-Ubuntu-Clang-GCE-CPU-AVX2-x86_64-Coverage-Trybot',
|
|
'Test-Ubuntu-Clang-GCE-CPU-AVX2-x86_64-Debug-MSAN',
|
|
'Test-Ubuntu-GCC-GCE-CPU-AVX2-x86_64-Debug',
|
|
'Test-Ubuntu-GCC-GCE-CPU-AVX2-x86_64-Release-Shared',
|
|
'Test-Ubuntu-GCC-ShuttleA-GPU-GTX550Ti-x86_64-Release-Valgrind',
|
|
'Test-Win8-MSVC-ShuttleA-GPU-HD7770-x86_64-Release',
|
|
'Test-Win8-MSVC-ShuttleB-CPU-AVX2-x86_64-Release',
|
|
'Test-iOS-Clang-iPad4-GPU-SGX554-Arm7-Release',
|
|
],
|
|
},
|
|
}
|
|
|
|
|
|
UPLOAD_DIMENSIONS = {
|
|
'pool': 'Skia',
|
|
'os': 'Linux',
|
|
'cpu': 'x86-64-avx2',
|
|
'gpu': 'none',
|
|
}
|
|
|
|
|
|
def derive_compile_bot_name(api):
|
|
builder_name = api.properties['buildername']
|
|
builder_cfg = api.builder_name_schema.DictForBuilderName(builder_name)
|
|
if builder_cfg['role'] == 'Housekeeper':
|
|
return 'Build-Ubuntu-GCC-x86_64-Release-Shared'
|
|
if builder_cfg['role'] in ('Test', 'Perf'):
|
|
task_os = builder_cfg['os']
|
|
extra_config = builder_cfg.get('extra_config', '')
|
|
if task_os == 'Android':
|
|
if extra_config == 'Vulkan':
|
|
extra_config = '%s_%s' % (task_os, 'Vulkan')
|
|
elif 'GN_Android' in extra_config:
|
|
pass # i.e. extra_config stays GN_Android or GN_Android_Vulkan
|
|
else:
|
|
extra_config = task_os
|
|
task_os = 'Ubuntu'
|
|
elif task_os == 'iOS':
|
|
extra_config = task_os
|
|
task_os = 'Mac'
|
|
elif 'Win' in task_os:
|
|
task_os = 'Win'
|
|
return api.builder_name_schema.MakeBuilderName(
|
|
role=api.builder_name_schema.BUILDER_ROLE_BUILD,
|
|
os=task_os,
|
|
compiler=builder_cfg['compiler'],
|
|
target_arch=builder_cfg['arch'],
|
|
configuration=builder_cfg['configuration'],
|
|
extra_config=extra_config,
|
|
is_trybot=api.builder_name_schema.IsTrybot(builder_name))
|
|
return builder_name
|
|
|
|
|
|
def swarm_dimensions(builder_cfg):
|
|
"""Return a dict of keys and values to be used as Swarming bot dimensions."""
|
|
dimensions = {
|
|
'pool': 'Skia',
|
|
}
|
|
dimensions['os'] = builder_cfg.get('os', 'Ubuntu')
|
|
if builder_cfg.get('extra_config', '').startswith('CT_'):
|
|
dimensions['pool'] = 'SkiaCT'
|
|
return dimensions # Do not need any more dimensions for CT builders.
|
|
if 'Win' in builder_cfg.get('os', ''):
|
|
dimensions['os'] = 'Windows'
|
|
if builder_cfg['role'] in ('Test', 'Perf'):
|
|
if 'Android' in builder_cfg['os']:
|
|
# For Android, the device type is a better dimension than CPU or GPU.
|
|
dimensions['device_type'] = {
|
|
'AndroidOne': 'sprout',
|
|
'GalaxyS3': 'm0', #'smdk4x12', Detected incorrectly by swarming?
|
|
'GalaxyS4': None, # TODO(borenet,kjlubick)
|
|
'GalaxyS7': 'heroqlteatt',
|
|
'NVIDIA_Shield': 'foster',
|
|
'Nexus10': 'manta',
|
|
'Nexus5': 'hammerhead',
|
|
'Nexus6': 'shamu',
|
|
'Nexus6p': 'angler',
|
|
'Nexus7': 'grouper',
|
|
'Nexus7v2': 'flo',
|
|
'Nexus9': 'flounder',
|
|
'NexusPlayer': 'fugu',
|
|
}[builder_cfg['model']]
|
|
elif 'iOS' in builder_cfg['os']:
|
|
# For iOS, the device type is a better dimension than CPU or GPU.
|
|
dimensions['device'] = {
|
|
'iPad4': 'iPad4,1',
|
|
}[builder_cfg['model']]
|
|
# TODO(borenet): Replace this hack with something better.
|
|
dimensions['os'] = 'iOS-9.2'
|
|
elif builder_cfg['cpu_or_gpu'] == 'CPU':
|
|
dimensions['gpu'] = 'none'
|
|
dimensions['cpu'] = {
|
|
'AVX': 'x86-64',
|
|
'AVX2': 'x86-64-avx2',
|
|
'SSE4': 'x86-64',
|
|
}[builder_cfg['cpu_or_gpu_value']]
|
|
if ('Win' in builder_cfg['os'] and
|
|
builder_cfg['cpu_or_gpu_value'] == 'AVX2'):
|
|
# AVX2 is not correctly detected on Windows. Fall back on other
|
|
# dimensions to ensure that we correctly target machines which we know
|
|
# have AVX2 support.
|
|
dimensions['cpu'] = 'x86-64'
|
|
dimensions['os'] = 'Windows-2008ServerR2-SP1'
|
|
else:
|
|
dimensions['gpu'] = {
|
|
'GeForce320M': '10de:08a4',
|
|
'GT610': '10de:104a',
|
|
'GTX550Ti': '10de:1244',
|
|
'GTX660': '10de:11c0',
|
|
'GTX960': '10de:1401',
|
|
'HD4000': '8086:0a2e',
|
|
'HD4600': '8086:0412',
|
|
'HD7770': '1002:683d',
|
|
'iHD530': '8086:1912',
|
|
}[builder_cfg['cpu_or_gpu_value']]
|
|
else:
|
|
dimensions['gpu'] = 'none'
|
|
return dimensions
|
|
|
|
|
|
def fix_filemodes(api, path):
|
|
"""Set all filemodes to 644 or 755 in the given directory path."""
|
|
api.python.inline(
|
|
name='fix filemodes',
|
|
program='''import os
|
|
for r, _, files in os.walk(os.getcwd()):
|
|
for fname in files:
|
|
f = os.path.join(r, fname)
|
|
if os.path.isfile(f):
|
|
if os.access(f, os.X_OK):
|
|
os.chmod(f, 0755)
|
|
else:
|
|
os.chmod(f, 0644)
|
|
''',
|
|
cwd=path)
|
|
|
|
|
|
def trigger_task(api, task_name, builder, master, slave, buildnumber,
|
|
builder_cfg, got_revision, infrabots_dir, idempotent=False,
|
|
store_output=True, extra_isolate_hashes=None, expiration=None,
|
|
hard_timeout=None, io_timeout=None, cipd_packages=None,
|
|
recipe_name=None, isolate_file=None, dimensions=None):
|
|
"""Trigger the given bot to run as a Swarming task."""
|
|
# TODO(borenet): We're using Swarming directly to run the recipe through
|
|
# recipes.py. Once it's possible to track the state of a Buildbucket build,
|
|
# we should switch to use the trigger recipe module instead.
|
|
|
|
properties = {
|
|
'buildername': builder,
|
|
'mastername': master,
|
|
'buildnumber': buildnumber,
|
|
'reason': 'Triggered by Skia swarm_trigger Recipe',
|
|
'revision': got_revision,
|
|
'slavename': slave,
|
|
'swarm_out_dir': '${ISOLATED_OUTDIR}',
|
|
}
|
|
if builder_cfg['is_trybot']:
|
|
if api.properties.get('patch_storage') == 'gerrit':
|
|
properties['patch_storage'] = api.properties['patch_storage']
|
|
properties['repository'] = api.properties['repository']
|
|
properties['event.patchSet.ref'] = api.properties['event.patchSet.ref']
|
|
properties['event.change.number'] = api.properties['event.change.number']
|
|
else:
|
|
properties['issue'] = str(api.properties['issue'])
|
|
properties['patchset'] = str(api.properties['patchset'])
|
|
properties['rietveld'] = api.properties['rietveld']
|
|
|
|
extra_args = [
|
|
'--workdir', '../../..',
|
|
recipe_name or 'swarm_%s' % task_name,
|
|
]
|
|
for k, v in properties.iteritems():
|
|
extra_args.append('%s=%s' % (k, v))
|
|
|
|
isolate_base_dir = api.path['slave_build']
|
|
dimensions = dimensions or swarm_dimensions(builder_cfg)
|
|
isolate_blacklist = ['.git', 'out', '*.pyc', '.recipe_deps']
|
|
isolate_vars = {
|
|
'WORKDIR': api.path['slave_build'],
|
|
}
|
|
|
|
isolate_file = isolate_file or '%s_skia.isolate' % task_name
|
|
if 'Coverage' == builder_cfg.get('configuration'):
|
|
isolate_file = 'coverage_skia.isolate'
|
|
if 'RecreateSKPs' in builder:
|
|
isolate_file = 'compile_skia.isolate'
|
|
return api.swarming.isolate_and_trigger_task(
|
|
infrabots_dir.join(isolate_file),
|
|
isolate_base_dir,
|
|
'%s_skia' % task_name,
|
|
isolate_vars,
|
|
dimensions,
|
|
isolate_blacklist=isolate_blacklist,
|
|
extra_isolate_hashes=extra_isolate_hashes,
|
|
idempotent=idempotent,
|
|
store_output=store_output,
|
|
extra_args=extra_args,
|
|
expiration=expiration,
|
|
hard_timeout=hard_timeout,
|
|
io_timeout=io_timeout,
|
|
cipd_packages=cipd_packages)
|
|
|
|
|
|
def checkout_steps(api):
|
|
"""Run the steps to obtain a checkout of Skia."""
|
|
# In this case, we're already running inside a checkout of Skia, so just
|
|
# report the currently-checked-out commit.
|
|
checkout_path = api.path['slave_build'].join('skia')
|
|
got_revision = api.git(
|
|
'rev-parse', 'HEAD', cwd=checkout_path,
|
|
stdout=api.raw_io.output(),
|
|
step_test_data=lambda: api.raw_io.test_api.stream_output('abc123\n'),
|
|
).stdout.rstrip()
|
|
cmd = ['python', '-c', '"print \'%s\'"' % got_revision]
|
|
res = api.step('got_revision', cmd=cmd)
|
|
res.presentation.properties['got_revision'] = got_revision
|
|
api.path['checkout'] = checkout_path
|
|
|
|
# Write a fake .gclient file if none exists. This is required by .isolates.
|
|
dot_gclient = api.path['slave_build'].join('.gclient')
|
|
if not api.path.exists(dot_gclient):
|
|
api.run.writefile(dot_gclient, '')
|
|
|
|
fix_filemodes(api, api.path['checkout'])
|
|
return got_revision
|
|
|
|
|
|
def housekeeper_swarm(api, builder_cfg, got_revision, infrabots_dir,
|
|
extra_isolate_hashes):
|
|
task = trigger_task(
|
|
api,
|
|
'housekeeper',
|
|
api.properties['buildername'],
|
|
api.properties['mastername'],
|
|
api.properties['slavename'],
|
|
api.properties['buildnumber'],
|
|
builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
idempotent=False,
|
|
store_output=False,
|
|
extra_isolate_hashes=extra_isolate_hashes,
|
|
cipd_packages=[cipd_pkg(api, infrabots_dir, 'go')],
|
|
)
|
|
return api.swarming.collect_swarming_task(task)
|
|
|
|
|
|
def recreate_skps_swarm(api, builder_cfg, got_revision, infrabots_dir,
|
|
extra_isolate_hashes):
|
|
task = trigger_task(
|
|
api,
|
|
'RecreateSKPs',
|
|
api.properties['buildername'],
|
|
api.properties['mastername'],
|
|
api.properties['slavename'],
|
|
api.properties['buildnumber'],
|
|
builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
idempotent=False,
|
|
store_output=False,
|
|
extra_isolate_hashes=extra_isolate_hashes)
|
|
return api.swarming.collect_swarming_task(task)
|
|
|
|
|
|
def ct_skps_swarm(api, builder_cfg, got_revision, infrabots_dir,
|
|
extra_isolate_hashes):
|
|
expiration, hard_timeout, io_timeout = get_timeouts(builder_cfg)
|
|
task = trigger_task(
|
|
api,
|
|
'ct_skps',
|
|
api.properties['buildername'],
|
|
api.properties['mastername'],
|
|
api.properties['slavename'],
|
|
api.properties['buildnumber'],
|
|
builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
idempotent=False,
|
|
store_output=False,
|
|
extra_isolate_hashes=extra_isolate_hashes,
|
|
expiration=expiration,
|
|
hard_timeout=hard_timeout,
|
|
io_timeout=io_timeout)
|
|
return api.swarming.collect_swarming_task(task)
|
|
|
|
|
|
def infra_swarm(api, got_revision, infrabots_dir, extra_isolate_hashes):
|
|
# Fake the builder cfg.
|
|
builder_cfg = {
|
|
'role': 'Infra',
|
|
'is_trybot': api.builder_name_schema.IsTrybot(
|
|
api.properties['buildername'])
|
|
}
|
|
task = trigger_task(
|
|
api,
|
|
'infra',
|
|
api.properties['buildername'],
|
|
api.properties['mastername'],
|
|
api.properties['slavename'],
|
|
api.properties['buildnumber'],
|
|
builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
idempotent=False,
|
|
store_output=False,
|
|
extra_isolate_hashes=extra_isolate_hashes)
|
|
return api.swarming.collect_swarming_task(task)
|
|
|
|
|
|
def compile_steps_swarm(api, builder_cfg, got_revision, infrabots_dir):
|
|
builder_name = derive_compile_bot_name(api)
|
|
compile_builder_cfg = api.builder_name_schema.DictForBuilderName(builder_name)
|
|
|
|
cipd_packages = []
|
|
|
|
# Android bots require a toolchain.
|
|
if 'Android' in builder_name:
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'android_sdk'))
|
|
if 'Mac' in builder_name:
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'android_ndk_darwin'))
|
|
else:
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'android_ndk_linux'))
|
|
|
|
if 'Ubuntu' in builder_name and 'Clang' in builder_name:
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'clang_linux'))
|
|
|
|
# Windows bots require a toolchain.
|
|
if 'Win' in builder_name:
|
|
version_file = infrabots_dir.join('assets', 'win_toolchain', 'VERSION')
|
|
version = api.run.readfile(version_file,
|
|
name='read win_toolchain VERSION',
|
|
test_data='0').rstrip()
|
|
version = 'version:%s' % version
|
|
pkg = ('t', 'skia/bots/win_toolchain', version)
|
|
cipd_packages.append(pkg)
|
|
|
|
if 'Vulkan' in builder_name:
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'win_vulkan_sdk'))
|
|
|
|
# Fake these properties for compile tasks so that they can be de-duped.
|
|
master = 'client.skia.compile'
|
|
slave = 'skiabot-dummy-compile-slave'
|
|
buildnumber = 1
|
|
|
|
task = trigger_task(
|
|
api,
|
|
'compile',
|
|
builder_name,
|
|
master,
|
|
slave,
|
|
buildnumber,
|
|
compile_builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
idempotent=True,
|
|
store_output=False,
|
|
cipd_packages=cipd_packages)
|
|
|
|
# Wait for compile to finish, record the results hash.
|
|
return api.swarming.collect_swarming_task_isolate_hash(task)
|
|
|
|
|
|
def get_timeouts(builder_cfg):
|
|
"""Some builders require longer than the default timeouts.
|
|
|
|
Returns tuple of (expiration, hard_timeout, io_timeout). If those values are
|
|
none then default timeouts should be used.
|
|
"""
|
|
expiration = None
|
|
hard_timeout = None
|
|
io_timeout = None
|
|
if 'Valgrind' in builder_cfg.get('extra_config', ''):
|
|
expiration = 2*24*60*60
|
|
hard_timeout = 9*60*60
|
|
io_timeout = 60*60
|
|
if builder_cfg.get('extra_config', '').startswith('CT_'):
|
|
hard_timeout = 24*60*60
|
|
io_timeout = 60*60
|
|
return expiration, hard_timeout, io_timeout
|
|
|
|
|
|
def perf_steps_trigger(api, builder_cfg, got_revision, infrabots_dir,
|
|
extra_hashes, cipd_packages):
|
|
"""Trigger perf tests via Swarming."""
|
|
|
|
expiration, hard_timeout, io_timeout = get_timeouts(builder_cfg)
|
|
return trigger_task(
|
|
api,
|
|
'perf',
|
|
api.properties['buildername'],
|
|
api.properties['mastername'],
|
|
api.properties['slavename'],
|
|
api.properties['buildnumber'],
|
|
builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
extra_isolate_hashes=extra_hashes,
|
|
expiration=expiration,
|
|
hard_timeout=hard_timeout,
|
|
io_timeout=io_timeout,
|
|
cipd_packages=cipd_packages,
|
|
store_output=False)
|
|
|
|
|
|
def perf_steps_collect(api, task, builder_cfg, got_revision, infrabots_dir):
|
|
"""Wait for perf steps to finish and upload results."""
|
|
# Wait for nanobench to finish, download the results.
|
|
if not api.vars.upload_perf_results: # pragma: nocover
|
|
api.swarming.collect_swarming_task(task)
|
|
return
|
|
|
|
perf_hash = api.swarming.collect_swarming_task_isolate_hash(task)
|
|
|
|
# Upload the results.
|
|
task = trigger_task(
|
|
api,
|
|
'upload_nano_results',
|
|
api.properties['buildername'],
|
|
api.properties['mastername'],
|
|
api.properties['slavename'],
|
|
api.properties['buildnumber'],
|
|
builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
idempotent=True,
|
|
store_output=False,
|
|
cipd_packages=None,
|
|
extra_isolate_hashes=[perf_hash],
|
|
recipe_name='upload_nano_results',
|
|
isolate_file='upload_nano_results.isolate',
|
|
dimensions=UPLOAD_DIMENSIONS)
|
|
|
|
return api.swarming.collect_swarming_task(task)
|
|
|
|
|
|
def test_steps_trigger(api, builder_cfg, got_revision, infrabots_dir,
|
|
extra_hashes, cipd_packages):
|
|
"""Trigger DM via Swarming."""
|
|
expiration, hard_timeout, io_timeout = get_timeouts(builder_cfg)
|
|
return trigger_task(
|
|
api,
|
|
'test',
|
|
api.properties['buildername'],
|
|
api.properties['mastername'],
|
|
api.properties['slavename'],
|
|
api.properties['buildnumber'],
|
|
builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
extra_isolate_hashes=extra_hashes,
|
|
expiration=expiration,
|
|
hard_timeout=hard_timeout,
|
|
io_timeout=io_timeout,
|
|
cipd_packages=cipd_packages,
|
|
store_output=False)
|
|
|
|
|
|
def test_steps_collect(api, task, builder_cfg, got_revision, infrabots_dir):
|
|
"""Collect the test results from Swarming."""
|
|
# Wait for tests to finish, download the results.
|
|
if not api.vars.upload_dm_results: # pragma: nocover
|
|
api.swarming.collect_swarming_task(task)
|
|
return
|
|
|
|
dm_hash = api.swarming.collect_swarming_task_isolate_hash(task)
|
|
|
|
# Upload the results.
|
|
task = trigger_task(
|
|
api,
|
|
'upload_dm_results',
|
|
api.properties['buildername'],
|
|
api.properties['mastername'],
|
|
api.properties['slavename'],
|
|
api.properties['buildnumber'],
|
|
builder_cfg,
|
|
got_revision,
|
|
infrabots_dir,
|
|
idempotent=True,
|
|
store_output=False,
|
|
cipd_packages=None,
|
|
extra_isolate_hashes=[dm_hash],
|
|
recipe_name='upload_dm_results',
|
|
isolate_file='upload_dm_results.isolate',
|
|
dimensions=UPLOAD_DIMENSIONS)
|
|
|
|
return api.swarming.collect_swarming_task(task)
|
|
|
|
|
|
def cipd_pkg(api, infrabots_dir, asset_name):
|
|
"""Find and return the CIPD package info for the given asset."""
|
|
version_file = infrabots_dir.join('assets', asset_name, 'VERSION')
|
|
version = api.run.readfile(version_file,
|
|
name='read %s VERSION' % asset_name,
|
|
test_data='0').rstrip()
|
|
version = 'version:%s' % version
|
|
return (asset_name, 'skia/bots/%s' % asset_name, version)
|
|
|
|
|
|
def print_properties(api):
|
|
"""Dump out all properties for debugging purposes."""
|
|
props = {}
|
|
for k, v in api.properties.iteritems():
|
|
props[k] = v
|
|
api.python.inline(
|
|
'print properties',
|
|
'''
|
|
import json
|
|
import sys
|
|
|
|
with open(sys.argv[1]) as f:
|
|
content = json.load(f)
|
|
|
|
print json.dumps(content, indent=2)
|
|
''',
|
|
args=[api.json.input(props)])
|
|
|
|
|
|
def RunSteps(api):
|
|
# TODO(borenet): Remove this once SwarmBucket is working.
|
|
print_properties(api)
|
|
|
|
got_revision = checkout_steps(api)
|
|
infrabots_dir = api.path['checkout'].join('infra', 'bots')
|
|
api.swarming.setup(
|
|
infrabots_dir.join('tools', 'luci-go'),
|
|
swarming_rev='')
|
|
|
|
# Run gsutil.py to ensure that it's installed.
|
|
api.gsutil(['help'])
|
|
|
|
extra_hashes = []
|
|
|
|
builder_name = api.properties['buildername']
|
|
|
|
# Get ready to compile.
|
|
infrabots_dir = api.path['checkout'].join('infra', 'bots')
|
|
if 'Infra' in builder_name:
|
|
return infra_swarm(api, got_revision, infrabots_dir, extra_hashes)
|
|
|
|
builder_cfg = api.builder_name_schema.DictForBuilderName(builder_name)
|
|
|
|
if 'RecreateSKPs' in builder_name:
|
|
recreate_skps_swarm(api, builder_cfg, got_revision, infrabots_dir,
|
|
extra_hashes)
|
|
return
|
|
|
|
if '-CT_' in builder_name:
|
|
ct_skps_swarm(api, builder_cfg, got_revision, infrabots_dir, extra_hashes)
|
|
return
|
|
|
|
# Compile.
|
|
do_compile_steps = True
|
|
if 'Coverage' in builder_name:
|
|
do_compile_steps = False
|
|
if do_compile_steps:
|
|
extra_hashes.append(compile_steps_swarm(
|
|
api, builder_cfg, got_revision, infrabots_dir))
|
|
|
|
if builder_cfg['role'] == 'Housekeeper':
|
|
housekeeper_swarm(api, builder_cfg, got_revision, infrabots_dir,
|
|
extra_hashes)
|
|
return
|
|
|
|
# Get ready to test/perf.
|
|
|
|
# CIPD packages needed by test/perf.
|
|
cipd_packages = []
|
|
|
|
do_test_steps = (
|
|
builder_cfg['role'] == api.builder_name_schema.BUILDER_ROLE_TEST)
|
|
do_perf_steps = (
|
|
builder_cfg['role'] == api.builder_name_schema.BUILDER_ROLE_PERF)
|
|
|
|
if not (do_test_steps or do_perf_steps):
|
|
return
|
|
|
|
# SKPs, SkImages, SVGs.
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'skp'))
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'skimage'))
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'svg'))
|
|
|
|
# To find llvm-symbolizer and/or MSAN-compiled libc++.
|
|
if 'Ubuntu' in builder_name and 'SAN' in builder_name:
|
|
cipd_packages.append(cipd_pkg(api, infrabots_dir, 'clang_linux'))
|
|
|
|
# Trigger test and perf tasks.
|
|
test_task = None
|
|
perf_task = None
|
|
if do_test_steps:
|
|
test_task = test_steps_trigger(api, builder_cfg, got_revision,
|
|
infrabots_dir, extra_hashes, cipd_packages)
|
|
if do_perf_steps:
|
|
perf_task = perf_steps_trigger(api, builder_cfg, got_revision,
|
|
infrabots_dir, extra_hashes, cipd_packages)
|
|
|
|
# Wait for results, then upload them if necessary.
|
|
|
|
if test_task:
|
|
test_steps_collect(api, test_task, builder_cfg,
|
|
got_revision, infrabots_dir)
|
|
|
|
if perf_task:
|
|
perf_steps_collect(api, perf_task, builder_cfg,
|
|
got_revision, infrabots_dir)
|
|
|
|
|
|
def test_for_bot(api, builder, mastername, slavename, testname=None):
|
|
"""Generate a test for the given bot."""
|
|
testname = testname or builder
|
|
test = (
|
|
api.test(testname) +
|
|
api.properties(buildername=builder,
|
|
mastername=mastername,
|
|
slavename=slavename,
|
|
buildnumber=5,
|
|
path_config='kitchen',
|
|
revision='abc123')
|
|
)
|
|
paths = [
|
|
api.path['slave_build'].join('skia'),
|
|
api.path['slave_build'].join('tmp', 'uninteresting_hashes.txt'),
|
|
]
|
|
if 'Trybot' in builder:
|
|
test += api.properties(issue=500,
|
|
patchset=1,
|
|
rietveld='https://codereview.chromium.org')
|
|
if 'Android' in builder:
|
|
paths.append(api.path['slave_build'].join(
|
|
'skia', 'infra', 'bots', 'assets', 'android_sdk', 'VERSION'))
|
|
if 'Test' in builder and 'Coverage' not in builder:
|
|
test += api.step_data(
|
|
'upload new .isolated file for test_skia',
|
|
stdout=api.raw_io.output('def456 XYZ.isolated'))
|
|
if not 'Valgrind' in builder and not 'MSAN' in builder:
|
|
test += api.step_data(
|
|
'upload new .isolated file for upload_dm_results_skia',
|
|
stdout=api.raw_io.output('def456 XYZ.isolated'))
|
|
if 'Perf' in builder and '-CT_' not in builder:
|
|
test += api.step_data(
|
|
'upload new .isolated file for perf_skia',
|
|
stdout=api.raw_io.output('def456 XYZ.isolated'))
|
|
test += api.step_data(
|
|
'upload new .isolated file for upload_nano_results_skia',
|
|
stdout=api.raw_io.output('def456 XYZ.isolated'))
|
|
if 'Housekeeper' in builder and 'RecreateSKPs' not in builder:
|
|
test += api.step_data(
|
|
'upload new .isolated file for housekeeper_skia',
|
|
stdout=api.raw_io.output('def456 XYZ.isolated'))
|
|
if 'Win' in builder:
|
|
paths.append(api.path['slave_build'].join(
|
|
'skia', 'infra', 'bots', 'assets', 'win_toolchain', 'VERSION'))
|
|
paths.append(api.path['slave_build'].join(
|
|
'skia', 'infra', 'bots', 'assets', 'win_vulkan_sdk', 'VERSION'))
|
|
paths.append(api.path['slave_build'].join(
|
|
'skia', 'infra', 'bots', 'assets', 'skimage', 'VERSION'))
|
|
paths.append(api.path['slave_build'].join(
|
|
'skia', 'infra', 'bots', 'assets', 'skp', 'VERSION'))
|
|
paths.append(api.path['slave_build'].join(
|
|
'skia', 'infra', 'bots', 'assets', 'svg', 'VERSION'))
|
|
|
|
test += api.path.exists(*paths)
|
|
|
|
return test
|
|
|
|
|
|
def GenTests(api):
|
|
for mastername, slaves in TEST_BUILDERS.iteritems():
|
|
for slavename, builders_by_slave in slaves.iteritems():
|
|
for builder in builders_by_slave:
|
|
yield test_for_bot(api, builder, mastername, slavename)
|
|
|
|
gerrit_kwargs = {
|
|
'patch_storage': 'gerrit',
|
|
'repository': 'skia',
|
|
'event.patchSet.ref': 'refs/changes/00/2100/2',
|
|
'event.change.number': '2100',
|
|
}
|
|
yield (
|
|
api.test('recipe_with_gerrit_patch') +
|
|
api.properties(
|
|
buildername='Test-Ubuntu-GCC-GCE-CPU-AVX2-x86_64-Debug-Trybot',
|
|
mastername='client.skia',
|
|
slavename='skiabot-linux-swarm-000',
|
|
buildnumber=5,
|
|
path_config='kitchen',
|
|
revision='abc123',
|
|
**gerrit_kwargs) +
|
|
api.step_data(
|
|
'upload new .isolated file for test_skia',
|
|
stdout=api.raw_io.output('def456 XYZ.isolated')) +
|
|
api.step_data(
|
|
'upload new .isolated file for upload_dm_results_skia',
|
|
stdout=api.raw_io.output('def456 XYZ.isolated'))
|
|
)
|