dba0bc843a
Bug: skia:7394 Change-Id: I2e7ed586d81d19a0b7895ff81ce6670eb7cc5003 No-Try: true Reviewed-on: https://skia-review.googlesource.com/83445 Reviewed-by: Brian Osman <brianosman@google.com> Commit-Queue: Ben Wagner <benjaminwagner@google.com>
1324 lines
45 KiB
Go
1324 lines
45 KiB
Go
// Copyright 2016 The Chromium Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file.
|
|
|
|
package main
|
|
|
|
/*
|
|
Generate the tasks.json file.
|
|
*/
|
|
|
|
import (
|
|
"encoding/json"
|
|
"flag"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"os"
|
|
"path"
|
|
"path/filepath"
|
|
"regexp"
|
|
"runtime"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/skia-dev/glog"
|
|
"go.skia.org/infra/go/sklog"
|
|
"go.skia.org/infra/go/util"
|
|
"go.skia.org/infra/task_scheduler/go/specs"
|
|
)
|
|
|
|
const (
|
|
BUNDLE_RECIPES_NAME = "Housekeeper-PerCommit-BundleRecipes"
|
|
ISOLATE_SKIMAGE_NAME = "Housekeeper-PerCommit-IsolateSkImage"
|
|
ISOLATE_SKP_NAME = "Housekeeper-PerCommit-IsolateSKP"
|
|
ISOLATE_SVG_NAME = "Housekeeper-PerCommit-IsolateSVG"
|
|
ISOLATE_NDK_LINUX_NAME = "Housekeeper-PerCommit-IsolateAndroidNDKLinux"
|
|
ISOLATE_WIN_TOOLCHAIN_NAME = "Housekeeper-PerCommit-IsolateWinToolchain"
|
|
ISOLATE_WIN_VULKAN_SDK_NAME = "Housekeeper-PerCommit-IsolateWinVulkanSDK"
|
|
|
|
DEFAULT_OS_DEBIAN = "Debian-9.1"
|
|
DEFAULT_OS_LINUX_GCE = "Debian-9.2"
|
|
DEFAULT_OS_MAC = "Mac-10.13.1"
|
|
DEFAULT_OS_UBUNTU = "Ubuntu-14.04"
|
|
DEFAULT_OS_WIN = "Windows-2016Server-14393"
|
|
|
|
// Name prefix for upload jobs.
|
|
PREFIX_UPLOAD = "Upload"
|
|
)
|
|
|
|
var (
|
|
// "Constants"
|
|
|
|
// Top-level list of all jobs to run at each commit; loaded from
|
|
// jobs.json.
|
|
JOBS []string
|
|
|
|
// General configuration information.
|
|
CONFIG struct {
|
|
GsBucketCoverage string `json:"gs_bucket_coverage"`
|
|
GsBucketGm string `json:"gs_bucket_gm"`
|
|
GsBucketNano string `json:"gs_bucket_nano"`
|
|
GsBucketCalm string `json:"gs_bucket_calm"`
|
|
NoUpload []string `json:"no_upload"`
|
|
Pool string `json:"pool"`
|
|
}
|
|
|
|
// alternateSwarmDimensions can be set in an init function to override the default swarming bot
|
|
// dimensions for the given task.
|
|
alternateSwarmDimensions func(parts map[string]string) []string
|
|
|
|
// internalHardwareLabelFn can be set in an init function to provide an
|
|
// internal_hardware_label variable to the recipe.
|
|
internalHardwareLabelFn func(parts map[string]string) *int
|
|
|
|
// Defines the structure of job names.
|
|
jobNameSchema *JobNameSchema
|
|
|
|
// Git 2.13.
|
|
cipdGit1 = &specs.CipdPackage{
|
|
Name: fmt.Sprintf("infra/git/${platform}"),
|
|
Path: "git",
|
|
Version: fmt.Sprintf("version:2.13.0.chromium9"),
|
|
}
|
|
cipdGit2 = &specs.CipdPackage{
|
|
Name: fmt.Sprintf("infra/tools/git/${platform}"),
|
|
Path: "git",
|
|
Version: fmt.Sprintf("git_revision:a78b5f3658c0578a017db48df97d20ac09822bcd"),
|
|
}
|
|
|
|
// Flags.
|
|
builderNameSchemaFile = flag.String("builder_name_schema", "", "Path to the builder_name_schema.json file. If not specified, uses infra/bots/recipe_modules/builder_name_schema/builder_name_schema.json from this repo.")
|
|
assetsDir = flag.String("assets_dir", "", "Directory containing assets.")
|
|
cfgFile = flag.String("cfg_file", "", "JSON file containing general configuration information.")
|
|
jobsFile = flag.String("jobs", "", "JSON file containing jobs to run.")
|
|
)
|
|
|
|
// internalHardwareLabel returns the internal ID for the bot, if any.
|
|
func internalHardwareLabel(parts map[string]string) *int {
|
|
if internalHardwareLabelFn != nil {
|
|
return internalHardwareLabelFn(parts)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// linuxGceDimensions are the Swarming dimensions for Linux GCE
|
|
// instances.
|
|
func linuxGceDimensions() []string {
|
|
return []string{
|
|
// Specify CPU to avoid running builds on bots with a more unique CPU.
|
|
"cpu:x86-64-Haswell_GCE",
|
|
"gpu:none",
|
|
fmt.Sprintf("os:%s", DEFAULT_OS_LINUX_GCE),
|
|
fmt.Sprintf("pool:%s", CONFIG.Pool),
|
|
}
|
|
}
|
|
|
|
// deriveCompileTaskName returns the name of a compile task based on the given
|
|
// job name.
|
|
func deriveCompileTaskName(jobName string, parts map[string]string) string {
|
|
if jobName == "Housekeeper-Nightly-Bookmaker" {
|
|
return "Build-Debian9-GCC-x86_64-Release"
|
|
} else if parts["role"] == "Housekeeper" {
|
|
return "Build-Debian9-GCC-x86_64-Release-Shared"
|
|
} else if parts["role"] == "Test" || parts["role"] == "Perf" {
|
|
task_os := parts["os"]
|
|
ec := []string{}
|
|
if val := parts["extra_config"]; val != "" {
|
|
ec = strings.Split(val, "_")
|
|
ignore := []string{"Skpbench", "AbandonGpuContext", "PreAbandonGpuContext", "Valgrind", "ReleaseAndAbandonGpuContext", "CCPR", "FSAA", "FAAA", "FDAA", "NativeFonts", "GDI", "NoGPUThreads"}
|
|
keep := make([]string, 0, len(ec))
|
|
for _, part := range ec {
|
|
if !util.In(part, ignore) {
|
|
keep = append(keep, part)
|
|
}
|
|
}
|
|
ec = keep
|
|
}
|
|
if task_os == "Android" {
|
|
if !util.In("Android", ec) {
|
|
ec = append([]string{"Android"}, ec...)
|
|
}
|
|
task_os = "Debian9"
|
|
} else if task_os == "Chromecast" {
|
|
task_os = "Debian9"
|
|
ec = append([]string{"Chromecast"}, ec...)
|
|
} else if strings.Contains(task_os, "ChromeOS") {
|
|
ec = append([]string{"Chromebook", "GLES"}, ec...)
|
|
task_os = "Debian9"
|
|
} else if task_os == "iOS" {
|
|
ec = append([]string{task_os}, ec...)
|
|
task_os = "Mac"
|
|
} else if strings.Contains(task_os, "Win") {
|
|
task_os = "Win"
|
|
} else if strings.Contains(task_os, "Ubuntu") || strings.Contains(task_os, "Debian") {
|
|
task_os = "Debian9"
|
|
}
|
|
jobNameMap := map[string]string{
|
|
"role": "Build",
|
|
"os": task_os,
|
|
"compiler": parts["compiler"],
|
|
"target_arch": parts["arch"],
|
|
"configuration": parts["configuration"],
|
|
}
|
|
if len(ec) > 0 {
|
|
jobNameMap["extra_config"] = strings.Join(ec, "_")
|
|
}
|
|
name, err := jobNameSchema.MakeJobName(jobNameMap)
|
|
if err != nil {
|
|
glog.Fatal(err)
|
|
}
|
|
return name
|
|
} else {
|
|
return jobName
|
|
}
|
|
}
|
|
|
|
// swarmDimensions generates swarming bot dimensions for the given task.
|
|
func swarmDimensions(parts map[string]string) []string {
|
|
if alternateSwarmDimensions != nil {
|
|
return alternateSwarmDimensions(parts)
|
|
}
|
|
return defaultSwarmDimensions(parts)
|
|
}
|
|
|
|
// defaultSwarmDimensions generates default swarming bot dimensions for the given task.
|
|
func defaultSwarmDimensions(parts map[string]string) []string {
|
|
d := map[string]string{
|
|
"pool": CONFIG.Pool,
|
|
}
|
|
if os, ok := parts["os"]; ok {
|
|
d["os"], ok = map[string]string{
|
|
"Android": "Android",
|
|
"Chromecast": "Android",
|
|
"ChromeOS": "ChromeOS",
|
|
"Debian9": DEFAULT_OS_DEBIAN,
|
|
"Mac": DEFAULT_OS_MAC,
|
|
"Ubuntu14": DEFAULT_OS_UBUNTU,
|
|
"Ubuntu16": "Ubuntu-16.10",
|
|
"Ubuntu17": "Ubuntu-17.04",
|
|
"Win": DEFAULT_OS_WIN,
|
|
"Win10": "Windows-10-15063",
|
|
"Win2k8": "Windows-2008ServerR2-SP1",
|
|
"Win2016": DEFAULT_OS_WIN,
|
|
"Win7": "Windows-7-SP1",
|
|
"Win8": "Windows-8.1-SP0",
|
|
"iOS": "iOS-10.3.1",
|
|
}[os]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in OS mapping.", os)
|
|
}
|
|
} else {
|
|
d["os"] = DEFAULT_OS_DEBIAN
|
|
}
|
|
if parts["role"] == "Test" || parts["role"] == "Perf" || parts["role"] == "Calmbench" {
|
|
if strings.Contains(parts["os"], "Android") || strings.Contains(parts["os"], "Chromecast") {
|
|
// For Android, the device type is a better dimension
|
|
// than CPU or GPU.
|
|
deviceInfo, ok := map[string][]string{
|
|
"AndroidOne": {"sprout", "MOB30Q"},
|
|
"Chorizo": {"chorizo", "1.24_82923"},
|
|
"GalaxyS6": {"zerofltetmo", "NRD90M_G920TUVU5FQK1"},
|
|
"GalaxyS7_G930A": {"heroqlteatt", "NRD90M_G930AUCS4BQC2"},
|
|
"GalaxyS7_G930FD": {"herolte", "NRD90M_G930FXXU1DQAS"},
|
|
"MotoG4": {"athene", "NPJ25.93-14"},
|
|
"NVIDIA_Shield": {"foster", "NRD90M_1915764_848"},
|
|
"Nexus5": {"hammerhead", "M4B30Z_3437181"},
|
|
"Nexus5x": {"bullhead", "OPR6.170623.023"},
|
|
"Nexus7": {"grouper", "LMY47V_1836172"}, // 2012 Nexus 7
|
|
"NexusPlayer": {"fugu", "OPR6.170623.021"},
|
|
"Pixel": {"sailfish", "OPR3.170623.008"},
|
|
"Pixel2XL": {"taimen", "OPD1.170816.023"},
|
|
"PixelC": {"dragon", "OPR1.170623.034"},
|
|
"PixelXL": {"marlin", "OPR3.170623.008"},
|
|
}[parts["model"]]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in Android mapping.", parts["model"])
|
|
}
|
|
d["device_type"] = deviceInfo[0]
|
|
d["device_os"] = deviceInfo[1]
|
|
// TODO(kjlubick): Remove the python dimension after we have removed the
|
|
// Nexus5x devices from the local lab (on Monday, Dec 11, 2017 should be fine).
|
|
d["python"] = "2.7.9" // This indicates a RPI, e.g. in Skolo. Golo is 2.7.12
|
|
if parts["model"] == "Nexus5x" {
|
|
d["python"] = "2.7.12"
|
|
}
|
|
} else if strings.Contains(parts["os"], "iOS") {
|
|
device, ok := map[string]string{
|
|
"iPadMini4": "iPad5,1",
|
|
"iPhone6": "iPhone7,2",
|
|
"iPhone7": "iPhone9,1",
|
|
"iPadPro": "iPad6,3",
|
|
}[parts["model"]]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in iOS mapping.", parts["model"])
|
|
}
|
|
d["device"] = device
|
|
} else if parts["cpu_or_gpu"] == "CPU" {
|
|
modelMapping, ok := map[string]map[string]string{
|
|
"AVX": {
|
|
"MacMini7.1": "x86-64-E5-2697_v2",
|
|
"Golo": "x86-64-E5-2670",
|
|
},
|
|
"AVX2": {
|
|
"GCE": "x86-64-Haswell_GCE",
|
|
"NUC5i7RYH": "x86-64-i7-5557U",
|
|
},
|
|
"AVX512": {
|
|
"GCE": "x86-64-Skylake_GCE",
|
|
},
|
|
}[parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in CPU mapping.", parts["cpu_or_gpu_value"])
|
|
}
|
|
cpu, ok := modelMapping[parts["model"]]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in %q model mapping.", parts["model"], parts["cpu_or_gpu_value"])
|
|
}
|
|
d["cpu"] = cpu
|
|
if parts["model"] == "GCE" && d["os"] == DEFAULT_OS_DEBIAN {
|
|
d["os"] = DEFAULT_OS_LINUX_GCE
|
|
}
|
|
} else {
|
|
if strings.Contains(parts["os"], "Win") {
|
|
gpu, ok := map[string]string{
|
|
"GT610": "10de:104a-22.21.13.8205",
|
|
"GTX1070": "10de:1ba1-23.21.13.8813",
|
|
"GTX660": "10de:11c0-23.21.13.8813",
|
|
"GTX960": "10de:1401-23.21.13.8813",
|
|
"IntelHD530": "8086:1912-21.20.16.4590",
|
|
"IntelHD4400": "8086:0a16-20.19.15.4703",
|
|
"IntelHD4600": "8086:0412-20.19.15.4703",
|
|
"IntelIris540": "8086:1926-21.20.16.4590",
|
|
"IntelIris6100": "8086:162b-20.19.15.4703",
|
|
"RadeonHD7770": "1002:683d-22.19.165.512",
|
|
"RadeonR9M470X": "1002:6646-22.19.165.512",
|
|
"QuadroP400": "10de:1cb3-22.21.13.8205",
|
|
}[parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in Win GPU mapping.", parts["cpu_or_gpu_value"])
|
|
}
|
|
d["gpu"] = gpu
|
|
|
|
// Specify cpu dimension for NUCs and ShuttleCs. We temporarily have two
|
|
// types of machines with a GTX960.
|
|
cpu, ok := map[string]string{
|
|
"NUC6i7KYK": "x86-64-i7-6770HQ",
|
|
"ShuttleC": "x86-64-i7-6700K",
|
|
}[parts["model"]]
|
|
if ok {
|
|
d["cpu"] = cpu
|
|
}
|
|
} else if strings.Contains(parts["os"], "Ubuntu") || strings.Contains(parts["os"], "Debian") {
|
|
gpu, ok := map[string]string{
|
|
// Intel drivers come from CIPD, so no need to specify the version here.
|
|
"IntelBayTrail": "8086:0f31",
|
|
"IntelHD2000": "8086:0102",
|
|
"IntelHD405": "8086:22b1",
|
|
"IntelIris540": "8086:1926",
|
|
"QuadroP400": "10de:1cb3-384.59",
|
|
}[parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in Ubuntu GPU mapping.", parts["cpu_or_gpu_value"])
|
|
}
|
|
d["gpu"] = gpu
|
|
} else if strings.Contains(parts["os"], "Mac") {
|
|
gpu, ok := map[string]string{
|
|
"IntelIris5100": "8086:0a2e",
|
|
}[parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in Mac GPU mapping.", parts["cpu_or_gpu_value"])
|
|
}
|
|
d["gpu"] = gpu
|
|
} else if strings.Contains(parts["os"], "ChromeOS") {
|
|
version, ok := map[string]string{
|
|
"MaliT604": "9901.12.0",
|
|
"MaliT764": "10172.0.0",
|
|
"MaliT860": "10172.0.0",
|
|
"PowerVRGX6250": "10176.5.0",
|
|
"TegraK1": "10172.0.0",
|
|
"IntelHDGraphics615": "10032.17.0",
|
|
}[parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
glog.Fatalf("Entry %q not found in ChromeOS GPU mapping.", parts["cpu_or_gpu_value"])
|
|
}
|
|
d["gpu"] = parts["cpu_or_gpu_value"]
|
|
d["release_version"] = version
|
|
} else {
|
|
glog.Fatalf("Unknown GPU mapping for OS %q.", parts["os"])
|
|
}
|
|
}
|
|
} else {
|
|
d["gpu"] = "none"
|
|
if d["os"] == DEFAULT_OS_DEBIAN {
|
|
return linuxGceDimensions()
|
|
} else if d["os"] == DEFAULT_OS_WIN {
|
|
// Windows CPU bots.
|
|
d["cpu"] = "x86-64-Haswell_GCE"
|
|
} else if d["os"] == DEFAULT_OS_MAC {
|
|
// Mac CPU bots.
|
|
d["cpu"] = "x86-64-E5-2697_v2"
|
|
}
|
|
}
|
|
|
|
rv := make([]string, 0, len(d))
|
|
for k, v := range d {
|
|
rv = append(rv, fmt.Sprintf("%s:%s", k, v))
|
|
}
|
|
sort.Strings(rv)
|
|
return rv
|
|
}
|
|
|
|
// relpath returns the relative path to the given file from the config file.
|
|
func relpath(f string) string {
|
|
_, filename, _, _ := runtime.Caller(0)
|
|
dir := path.Dir(filename)
|
|
rel := dir
|
|
if *cfgFile != "" {
|
|
rel = path.Dir(*cfgFile)
|
|
}
|
|
rv, err := filepath.Rel(rel, path.Join(dir, f))
|
|
if err != nil {
|
|
sklog.Fatal(err)
|
|
}
|
|
return rv
|
|
}
|
|
|
|
// bundleRecipes generates the task to bundle and isolate the recipes.
|
|
func bundleRecipes(b *specs.TasksCfgBuilder) string {
|
|
b.MustAddTask(BUNDLE_RECIPES_NAME, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{cipdGit1, cipdGit2},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "bundle_recipes",
|
|
fmt.Sprintf("buildername=%s", BUNDLE_RECIPES_NAME),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
},
|
|
Isolate: relpath("bundle_recipes.isolate"),
|
|
Priority: 0.7,
|
|
})
|
|
return BUNDLE_RECIPES_NAME
|
|
}
|
|
|
|
// useBundledRecipes returns true iff the given bot should use bundled recipes
|
|
// instead of syncing recipe DEPS itself.
|
|
func useBundledRecipes(parts map[string]string) bool {
|
|
// Use bundled recipes for all test/perf tasks.
|
|
return true
|
|
}
|
|
|
|
type isolateAssetCfg struct {
|
|
isolateFile string
|
|
cipdPkg string
|
|
}
|
|
|
|
var ISOLATE_ASSET_MAPPING = map[string]isolateAssetCfg{
|
|
ISOLATE_SKIMAGE_NAME: {
|
|
isolateFile: "isolate_skimage.isolate",
|
|
cipdPkg: "skimage",
|
|
},
|
|
ISOLATE_SKP_NAME: {
|
|
isolateFile: "isolate_skp.isolate",
|
|
cipdPkg: "skp",
|
|
},
|
|
ISOLATE_SVG_NAME: {
|
|
isolateFile: "isolate_svg.isolate",
|
|
cipdPkg: "svg",
|
|
},
|
|
ISOLATE_NDK_LINUX_NAME: {
|
|
isolateFile: "isolate_ndk_linux.isolate",
|
|
cipdPkg: "android_ndk_linux",
|
|
},
|
|
ISOLATE_WIN_TOOLCHAIN_NAME: {
|
|
isolateFile: "isolate_win_toolchain.isolate",
|
|
cipdPkg: "win_toolchain",
|
|
},
|
|
ISOLATE_WIN_VULKAN_SDK_NAME: {
|
|
isolateFile: "isolate_win_vulkan_sdk.isolate",
|
|
cipdPkg: "win_vulkan_sdk",
|
|
},
|
|
}
|
|
|
|
// bundleRecipes generates the task to bundle and isolate the recipes.
|
|
func isolateCIPDAsset(b *specs.TasksCfgBuilder, name string) string {
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{
|
|
b.MustGetCipdPackageFromAsset(ISOLATE_ASSET_MAPPING[name].cipdPkg),
|
|
},
|
|
Dimensions: linuxGceDimensions(),
|
|
Isolate: relpath(ISOLATE_ASSET_MAPPING[name].isolateFile),
|
|
Priority: 0.7,
|
|
})
|
|
return name
|
|
}
|
|
|
|
// getIsolatedCIPDDeps returns the slice of Isolate_* tasks a given task needs.
|
|
// This allows us to save time on I/O bound bots, like the RPIs.
|
|
func getIsolatedCIPDDeps(parts map[string]string) []string {
|
|
deps := []string{}
|
|
// Only do this on the RPIs for now. Other, faster machines shouldn't see much
|
|
// benefit and we don't need the extra complexity, for now
|
|
rpiOS := []string{"Android", "ChromeOS", "iOS"}
|
|
|
|
if o := parts["os"]; strings.Contains(o, "Chromecast") {
|
|
// Chromecasts don't have enough disk space to fit all of the content,
|
|
// so we do a subset of the skps.
|
|
deps = append(deps, ISOLATE_SKP_NAME)
|
|
} else if e := parts["extra_config"]; strings.Contains(e, "Skpbench") {
|
|
// Skpbench only needs skps
|
|
deps = append(deps, ISOLATE_SKP_NAME)
|
|
} else if util.In(o, rpiOS) {
|
|
deps = append(deps, ISOLATE_SKP_NAME)
|
|
deps = append(deps, ISOLATE_SVG_NAME)
|
|
deps = append(deps, ISOLATE_SKIMAGE_NAME)
|
|
}
|
|
|
|
return deps
|
|
}
|
|
|
|
// compile generates a compile task. Returns the name of the last task in the
|
|
// generated chain of tasks, which the Job should add as a dependency.
|
|
func compile(b *specs.TasksCfgBuilder, name string, parts map[string]string) string {
|
|
// Collect the necessary CIPD packages.
|
|
pkgs := []*specs.CipdPackage{}
|
|
deps := []string{}
|
|
|
|
// Android bots require a toolchain.
|
|
if strings.Contains(name, "Android") {
|
|
if strings.Contains(name, "Mac") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("android_ndk_darwin"))
|
|
} else if strings.Contains(name, "Win") {
|
|
pkg := b.MustGetCipdPackageFromAsset("android_ndk_windows")
|
|
pkg.Path = "n"
|
|
pkgs = append(pkgs, pkg)
|
|
} else {
|
|
deps = append(deps, isolateCIPDAsset(b, ISOLATE_NDK_LINUX_NAME))
|
|
}
|
|
} else if strings.Contains(name, "Chromecast") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("cast_toolchain"))
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("chromebook_arm_gles"))
|
|
} else if strings.Contains(name, "Chromebook") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux"))
|
|
if parts["target_arch"] == "x86_64" {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("chromebook_x86_64_gles"))
|
|
} else if parts["target_arch"] == "arm" {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("armhf_sysroot"))
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("chromebook_arm_gles"))
|
|
}
|
|
} else if strings.Contains(name, "Debian") {
|
|
if strings.Contains(name, "Clang") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux"))
|
|
}
|
|
if strings.Contains(name, "Vulkan") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_sdk"))
|
|
}
|
|
if strings.Contains(name, "EMCC") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("emscripten_sdk"))
|
|
}
|
|
} else if strings.Contains(name, "Win") {
|
|
deps = append(deps, isolateCIPDAsset(b, ISOLATE_WIN_TOOLCHAIN_NAME))
|
|
if strings.Contains(name, "Clang") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_win"))
|
|
}
|
|
if strings.Contains(name, "Vulkan") {
|
|
deps = append(deps, isolateCIPDAsset(b, ISOLATE_WIN_VULKAN_SDK_NAME))
|
|
}
|
|
}
|
|
|
|
dimensions := swarmDimensions(parts)
|
|
|
|
// Add the task.
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: pkgs,
|
|
Dimensions: dimensions,
|
|
Dependencies: deps,
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "compile",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
Isolate: relpath("compile_skia.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
// All compile tasks are runnable as their own Job. Assert that the Job
|
|
// is listed in JOBS.
|
|
if !util.In(name, JOBS) {
|
|
glog.Fatalf("Job %q is missing from the JOBS list!", name)
|
|
}
|
|
return name
|
|
}
|
|
|
|
// recreateSKPs generates a RecreateSKPs task. Returns the name of the last
|
|
// task in the generated chain of tasks, which the Job should add as a
|
|
// dependency.
|
|
func recreateSKPs(b *specs.TasksCfgBuilder, name string) string {
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExecutionTimeout: 4 * time.Hour,
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "recreate_skps",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
IoTimeout: 40 * time.Minute,
|
|
Isolate: relpath("compile_skia.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return name
|
|
}
|
|
|
|
// updateMetaConfig generates a UpdateMetaConfig task. Returns the name of the
|
|
// last task in the generated chain of tasks, which the Job should add as a
|
|
// dependency.
|
|
func updateMetaConfig(b *specs.TasksCfgBuilder, name string) string {
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "update_meta_config",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
Isolate: relpath("meta_config.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return name
|
|
}
|
|
|
|
// ctSKPs generates a CT SKPs task. Returns the name of the last task in the
|
|
// generated chain of tasks, which the Job should add as a dependency.
|
|
func ctSKPs(b *specs.TasksCfgBuilder, name string) string {
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{},
|
|
Dimensions: []string{
|
|
"pool:SkiaCT",
|
|
fmt.Sprintf("os:%s", DEFAULT_OS_LINUX_GCE),
|
|
},
|
|
ExecutionTimeout: 24 * time.Hour,
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "ct_skps",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
IoTimeout: time.Hour,
|
|
Isolate: relpath("ct_skps_skia.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return name
|
|
}
|
|
|
|
// checkGeneratedFiles verifies that no generated SKSL files have been edited
|
|
// by hand.
|
|
func checkGeneratedFiles(b *specs.TasksCfgBuilder, name string) string {
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "check_generated_files",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
Isolate: relpath("compile_skia.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return name
|
|
}
|
|
|
|
// housekeeper generates a Housekeeper task. Returns the name of the last task
|
|
// in the generated chain of tasks, which the Job should add as a dependency.
|
|
func housekeeper(b *specs.TasksCfgBuilder, name, compileTaskName string) string {
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")},
|
|
Dependencies: []string{compileTaskName},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "housekeeper",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
Isolate: relpath("housekeeper_skia.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return name
|
|
}
|
|
|
|
// bookmaker generates a Bookmaker task. Returns the name of the last task
|
|
// in the generated chain of tasks, which the Job should add as a dependency.
|
|
func bookmaker(b *specs.TasksCfgBuilder, name, compileTaskName string) string {
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")},
|
|
Dependencies: []string{compileTaskName},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "bookmaker",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
Isolate: relpath("compile_skia.isolate"),
|
|
Priority: 0.8,
|
|
ExecutionTimeout: 2 * time.Hour,
|
|
IoTimeout: 2 * time.Hour,
|
|
})
|
|
return name
|
|
}
|
|
|
|
// infra generates an infra_tests task. Returns the name of the last task in the
|
|
// generated chain of tasks, which the Job should add as a dependency.
|
|
func infra(b *specs.TasksCfgBuilder, name string) string {
|
|
b.MustAddTask(name, &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "infra",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
Isolate: relpath("infra_skia.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return name
|
|
}
|
|
|
|
// calmbench generates a calmbench task. Returns the name of the last task in the
|
|
// generated chain of tasks, which the Job should add as a dependency.
|
|
func calmbench(b *specs.TasksCfgBuilder, name string, parts map[string]string) string {
|
|
s := &specs.TaskSpec{
|
|
CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("clang_linux")},
|
|
Dimensions: swarmDimensions(parts),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "calmbench",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
Isolate: relpath("infra_skia.isolate"),
|
|
Priority: 0.8,
|
|
}
|
|
|
|
s.Dependencies = append(s.Dependencies, ISOLATE_SKP_NAME, ISOLATE_SVG_NAME)
|
|
|
|
b.MustAddTask(name, s)
|
|
|
|
// Upload results if necessary.
|
|
if strings.Contains(name, "Release") && doUpload(name) {
|
|
uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name)
|
|
b.MustAddTask(uploadName, &specs.TaskSpec{
|
|
Dependencies: []string{name},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "upload_calmbench_results",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketCalm),
|
|
},
|
|
// We're using the same isolate as upload_nano_results
|
|
Isolate: relpath("upload_nano_results.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return uploadName
|
|
}
|
|
|
|
return name
|
|
}
|
|
|
|
// doUpload indicates whether the given Job should upload its results.
|
|
func doUpload(name string) bool {
|
|
for _, s := range CONFIG.NoUpload {
|
|
m, err := regexp.MatchString(s, name)
|
|
if err != nil {
|
|
glog.Fatal(err)
|
|
}
|
|
if m {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
// test generates a Test task. Returns the name of the last task in the
|
|
// generated chain of tasks, which the Job should add as a dependency.
|
|
func test(b *specs.TasksCfgBuilder, name string, parts map[string]string, compileTaskName string, pkgs []*specs.CipdPackage) string {
|
|
s := &specs.TaskSpec{
|
|
CipdPackages: pkgs,
|
|
Dependencies: []string{compileTaskName},
|
|
Dimensions: swarmDimensions(parts),
|
|
ExecutionTimeout: 4 * time.Hour,
|
|
Expiration: 20 * time.Hour,
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "test",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildbucket_build_id=%s", specs.PLACEHOLDER_BUILDBUCKET_BUILD_ID),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
IoTimeout: 40 * time.Minute,
|
|
Isolate: relpath("test_skia.isolate"),
|
|
MaxAttempts: 1,
|
|
Priority: 0.8,
|
|
}
|
|
if useBundledRecipes(parts) {
|
|
s.Dependencies = append(s.Dependencies, BUNDLE_RECIPES_NAME)
|
|
if strings.Contains(parts["os"], "Win") {
|
|
s.Isolate = relpath("test_skia_bundled_win.isolate")
|
|
} else {
|
|
s.Isolate = relpath("test_skia_bundled_unix.isolate")
|
|
}
|
|
}
|
|
if deps := getIsolatedCIPDDeps(parts); len(deps) > 0 {
|
|
s.Dependencies = append(s.Dependencies, deps...)
|
|
}
|
|
if strings.Contains(parts["extra_config"], "Valgrind") {
|
|
s.ExecutionTimeout = 9 * time.Hour
|
|
s.Expiration = 48 * time.Hour
|
|
s.IoTimeout = time.Hour
|
|
s.CipdPackages = append(s.CipdPackages, b.MustGetCipdPackageFromAsset("valgrind"))
|
|
s.Dimensions = append(s.Dimensions, "valgrind:1")
|
|
} else if strings.Contains(parts["extra_config"], "MSAN") {
|
|
s.ExecutionTimeout = 9 * time.Hour
|
|
} else if parts["arch"] == "x86" && parts["configuration"] == "Debug" {
|
|
// skia:6737
|
|
s.ExecutionTimeout = 6 * time.Hour
|
|
}
|
|
iid := internalHardwareLabel(parts)
|
|
if iid != nil {
|
|
s.ExtraArgs = append(s.ExtraArgs, fmt.Sprintf("internal_hardware_label=%d", *iid))
|
|
}
|
|
b.MustAddTask(name, s)
|
|
|
|
// Upload results if necessary. TODO(kjlubick): If we do coverage analysis at the same
|
|
// time as normal tests (which would be nice), cfg.json needs to have Coverage removed.
|
|
if doUpload(name) {
|
|
uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name)
|
|
b.MustAddTask(uploadName, &specs.TaskSpec{
|
|
Dependencies: []string{name},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "upload_dm_results",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketGm),
|
|
},
|
|
Isolate: relpath("upload_dm_results.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return uploadName
|
|
}
|
|
|
|
return name
|
|
}
|
|
|
|
func coverage(b *specs.TasksCfgBuilder, name string, parts map[string]string, compileTaskName string, pkgs []*specs.CipdPackage) string {
|
|
shards := 1
|
|
deps := []string{}
|
|
|
|
tf := parts["test_filter"]
|
|
if strings.Contains(tf, "Shard") {
|
|
// Expected Shard_NN
|
|
shardstr := strings.Split(tf, "_")[1]
|
|
var err error
|
|
shards, err = strconv.Atoi(shardstr)
|
|
if err != nil {
|
|
glog.Fatalf("Expected int for number of shards %q in %s: %s", shardstr, name, err)
|
|
}
|
|
}
|
|
for i := 0; i < shards; i++ {
|
|
n := strings.Replace(name, tf, fmt.Sprintf("shard_%02d_%02d", i, shards), 1)
|
|
s := &specs.TaskSpec{
|
|
CipdPackages: pkgs,
|
|
Dependencies: []string{compileTaskName},
|
|
Dimensions: swarmDimensions(parts),
|
|
ExecutionTimeout: 4 * time.Hour,
|
|
Expiration: 20 * time.Hour,
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "test",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", n),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
IoTimeout: 40 * time.Minute,
|
|
Isolate: relpath("test_skia.isolate"),
|
|
MaxAttempts: 1,
|
|
Priority: 0.8,
|
|
}
|
|
if useBundledRecipes(parts) {
|
|
s.Dependencies = append(s.Dependencies, BUNDLE_RECIPES_NAME)
|
|
if strings.Contains(parts["os"], "Win") {
|
|
s.Isolate = relpath("test_skia_bundled_win.isolate")
|
|
} else {
|
|
s.Isolate = relpath("test_skia_bundled_unix.isolate")
|
|
}
|
|
}
|
|
if deps := getIsolatedCIPDDeps(parts); len(deps) > 0 {
|
|
s.Dependencies = append(s.Dependencies, deps...)
|
|
}
|
|
b.MustAddTask(n, s)
|
|
deps = append(deps, n)
|
|
}
|
|
|
|
uploadName := fmt.Sprintf("%s%s%s", "Upload", jobNameSchema.Sep, name)
|
|
// We need clang_linux to get access to the llvm-profdata and llvm-cov binaries
|
|
// which are used to deal with the raw coverage data output by the Test step.
|
|
pkgs = append([]*specs.CipdPackage{}, b.MustGetCipdPackageFromAsset("clang_linux"))
|
|
deps = append(deps, compileTaskName)
|
|
|
|
b.MustAddTask(uploadName, &specs.TaskSpec{
|
|
// A dependency on compileTaskName makes the TaskScheduler link the
|
|
// isolated output of the compile step to the input of the upload step,
|
|
// which gives us access to the instrumented binary. The binary is
|
|
// needed to figure out symbol names and line numbers.
|
|
Dependencies: deps,
|
|
Dimensions: linuxGceDimensions(),
|
|
CipdPackages: pkgs,
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "upload_coverage_results",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketCoverage),
|
|
},
|
|
Isolate: relpath("upload_coverage_results.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return uploadName
|
|
}
|
|
|
|
// perf generates a Perf task. Returns the name of the last task in the
|
|
// generated chain of tasks, which the Job should add as a dependency.
|
|
func perf(b *specs.TasksCfgBuilder, name string, parts map[string]string, compileTaskName string, pkgs []*specs.CipdPackage) string {
|
|
recipe := "perf"
|
|
isolate := relpath("perf_skia.isolate")
|
|
if strings.Contains(parts["extra_config"], "Skpbench") {
|
|
recipe = "skpbench"
|
|
isolate = relpath("skpbench_skia.isolate")
|
|
if useBundledRecipes(parts) {
|
|
if strings.Contains(parts["os"], "Win") {
|
|
isolate = relpath("skpbench_skia_bundled_win.isolate")
|
|
} else {
|
|
isolate = relpath("skpbench_skia_bundled_unix.isolate")
|
|
}
|
|
}
|
|
} else if useBundledRecipes(parts) {
|
|
if strings.Contains(parts["os"], "Win") {
|
|
isolate = relpath("perf_skia_bundled_win.isolate")
|
|
} else {
|
|
isolate = relpath("perf_skia_bundled_unix.isolate")
|
|
}
|
|
}
|
|
s := &specs.TaskSpec{
|
|
CipdPackages: pkgs,
|
|
Dependencies: []string{compileTaskName},
|
|
Dimensions: swarmDimensions(parts),
|
|
ExecutionTimeout: 4 * time.Hour,
|
|
Expiration: 20 * time.Hour,
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", recipe,
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
},
|
|
IoTimeout: 40 * time.Minute,
|
|
Isolate: isolate,
|
|
MaxAttempts: 1,
|
|
Priority: 0.8,
|
|
}
|
|
if useBundledRecipes(parts) {
|
|
s.Dependencies = append(s.Dependencies, BUNDLE_RECIPES_NAME)
|
|
}
|
|
if deps := getIsolatedCIPDDeps(parts); len(deps) > 0 {
|
|
s.Dependencies = append(s.Dependencies, deps...)
|
|
}
|
|
|
|
if strings.Contains(parts["extra_config"], "Valgrind") {
|
|
s.ExecutionTimeout = 9 * time.Hour
|
|
s.Expiration = 48 * time.Hour
|
|
s.IoTimeout = time.Hour
|
|
s.CipdPackages = append(s.CipdPackages, b.MustGetCipdPackageFromAsset("valgrind"))
|
|
s.Dimensions = append(s.Dimensions, "valgrind:1")
|
|
} else if strings.Contains(parts["extra_config"], "MSAN") {
|
|
s.ExecutionTimeout = 9 * time.Hour
|
|
} else if parts["arch"] == "x86" && parts["configuration"] == "Debug" {
|
|
// skia:6737
|
|
s.ExecutionTimeout = 6 * time.Hour
|
|
}
|
|
iid := internalHardwareLabel(parts)
|
|
if iid != nil {
|
|
s.ExtraArgs = append(s.ExtraArgs, fmt.Sprintf("internal_hardware_label=%d", *iid))
|
|
}
|
|
b.MustAddTask(name, s)
|
|
|
|
// Upload results if necessary.
|
|
if strings.Contains(name, "Release") && doUpload(name) {
|
|
uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name)
|
|
b.MustAddTask(uploadName, &specs.TaskSpec{
|
|
Dependencies: []string{name},
|
|
Dimensions: linuxGceDimensions(),
|
|
ExtraArgs: []string{
|
|
"--workdir", "../../..", "upload_nano_results",
|
|
fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO),
|
|
fmt.Sprintf("buildername=%s", name),
|
|
fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR),
|
|
fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION),
|
|
fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO),
|
|
fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE),
|
|
fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE),
|
|
fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET),
|
|
fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketNano),
|
|
},
|
|
Isolate: relpath("upload_nano_results.isolate"),
|
|
Priority: 0.8,
|
|
})
|
|
return uploadName
|
|
}
|
|
return name
|
|
}
|
|
|
|
// process generates tasks and jobs for the given job name.
|
|
func process(b *specs.TasksCfgBuilder, name string) {
|
|
deps := []string{}
|
|
|
|
// Bundle Recipes.
|
|
if name == BUNDLE_RECIPES_NAME {
|
|
deps = append(deps, bundleRecipes(b))
|
|
}
|
|
|
|
// Isolate CIPD assets.
|
|
if _, ok := ISOLATE_ASSET_MAPPING[name]; ok {
|
|
deps = append(deps, isolateCIPDAsset(b, name))
|
|
}
|
|
|
|
parts, err := jobNameSchema.ParseJobName(name)
|
|
if err != nil {
|
|
glog.Fatal(err)
|
|
}
|
|
|
|
// RecreateSKPs.
|
|
if strings.Contains(name, "RecreateSKPs") {
|
|
deps = append(deps, recreateSKPs(b, name))
|
|
}
|
|
|
|
// UpdateMetaConfig bot.
|
|
if strings.Contains(name, "UpdateMetaConfig") {
|
|
deps = append(deps, updateMetaConfig(b, name))
|
|
}
|
|
|
|
// CT bots.
|
|
if strings.Contains(name, "-CT_") {
|
|
deps = append(deps, ctSKPs(b, name))
|
|
}
|
|
|
|
// Infra tests.
|
|
if name == "Housekeeper-PerCommit-InfraTests" {
|
|
deps = append(deps, infra(b, name))
|
|
}
|
|
|
|
// Compile bots.
|
|
if parts["role"] == "Build" {
|
|
deps = append(deps, compile(b, name, parts))
|
|
}
|
|
|
|
// Calmbench bots.
|
|
if parts["role"] == "Calmbench" {
|
|
deps = append(deps, calmbench(b, name, parts))
|
|
}
|
|
|
|
// Most remaining bots need a compile task.
|
|
compileTaskName := deriveCompileTaskName(name, parts)
|
|
compileTaskParts, err := jobNameSchema.ParseJobName(compileTaskName)
|
|
if err != nil {
|
|
glog.Fatal(err)
|
|
}
|
|
// These bots do not need a compile task.
|
|
if parts["role"] != "Build" && parts["role"] != "Calmbench" &&
|
|
name != "Housekeeper-PerCommit-BundleRecipes" &&
|
|
name != "Housekeeper-PerCommit-InfraTests" &&
|
|
name != "Housekeeper-PerCommit-CheckGeneratedFiles" &&
|
|
!strings.Contains(name, "RecreateSKPs") &&
|
|
!strings.Contains(name, "UpdateMetaConfig") &&
|
|
!strings.Contains(name, "-CT_") &&
|
|
!strings.Contains(name, "Housekeeper-PerCommit-Isolate") {
|
|
compile(b, compileTaskName, compileTaskParts)
|
|
}
|
|
|
|
// Housekeepers.
|
|
if name == "Housekeeper-PerCommit" {
|
|
deps = append(deps, housekeeper(b, name, compileTaskName))
|
|
}
|
|
if name == "Housekeeper-PerCommit-CheckGeneratedFiles" {
|
|
deps = append(deps, checkGeneratedFiles(b, name))
|
|
}
|
|
if name == "Housekeeper-Nightly-Bookmaker" {
|
|
deps = append(deps, bookmaker(b, name, compileTaskName))
|
|
}
|
|
|
|
// Common assets needed by the remaining bots.
|
|
|
|
pkgs := []*specs.CipdPackage{}
|
|
|
|
if deps := getIsolatedCIPDDeps(parts); len(deps) == 0 {
|
|
pkgs = []*specs.CipdPackage{
|
|
b.MustGetCipdPackageFromAsset("skimage"),
|
|
b.MustGetCipdPackageFromAsset("skp"),
|
|
b.MustGetCipdPackageFromAsset("svg"),
|
|
}
|
|
}
|
|
|
|
if strings.Contains(name, "Ubuntu") || strings.Contains(name, "Debian") {
|
|
if strings.Contains(name, "SAN") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux"))
|
|
}
|
|
if strings.Contains(name, "Vulkan") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_sdk"))
|
|
}
|
|
if strings.Contains(name, "Intel") && strings.Contains(name, "GPU") {
|
|
if strings.Contains(name, "Release") {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_intel_driver_release"))
|
|
} else {
|
|
pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_intel_driver_debug"))
|
|
}
|
|
}
|
|
}
|
|
|
|
// Test bots.
|
|
|
|
if parts["role"] == "Test" {
|
|
if strings.Contains(parts["extra_config"], "Coverage") {
|
|
deps = append(deps, coverage(b, name, parts, compileTaskName, pkgs))
|
|
} else if !strings.Contains(name, "-CT_") {
|
|
deps = append(deps, test(b, name, parts, compileTaskName, pkgs))
|
|
}
|
|
|
|
}
|
|
|
|
// Perf bots.
|
|
if parts["role"] == "Perf" && !strings.Contains(name, "-CT_") {
|
|
deps = append(deps, perf(b, name, parts, compileTaskName, pkgs))
|
|
}
|
|
|
|
// Add the Job spec.
|
|
j := &specs.JobSpec{
|
|
Priority: 0.8,
|
|
TaskSpecs: deps,
|
|
Trigger: specs.TRIGGER_ANY_BRANCH,
|
|
}
|
|
if strings.Contains(name, "-Nightly-") {
|
|
j.Trigger = specs.TRIGGER_NIGHTLY
|
|
} else if strings.Contains(name, "-Weekly-") || strings.Contains(name, "CT_DM_1m_SKPs") {
|
|
j.Trigger = specs.TRIGGER_WEEKLY
|
|
} else if strings.Contains(name, "Flutter") || strings.Contains(name, "PDFium") || strings.Contains(name, "CommandBuffer") {
|
|
j.Trigger = specs.TRIGGER_MASTER_ONLY
|
|
}
|
|
b.MustAddJob(name, j)
|
|
}
|
|
|
|
func loadJson(flag *string, defaultFlag string, val interface{}) {
|
|
if *flag == "" {
|
|
*flag = defaultFlag
|
|
}
|
|
b, err := ioutil.ReadFile(*flag)
|
|
if err != nil {
|
|
glog.Fatal(err)
|
|
}
|
|
if err := json.Unmarshal(b, val); err != nil {
|
|
glog.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// Regenerate the tasks.json file.
|
|
func main() {
|
|
b := specs.MustNewTasksCfgBuilder()
|
|
b.SetAssetsDir(*assetsDir)
|
|
infraBots := path.Join(b.CheckoutRoot(), "infra", "bots")
|
|
|
|
// Load the jobs from a JSON file.
|
|
loadJson(jobsFile, path.Join(infraBots, "jobs.json"), &JOBS)
|
|
|
|
// Load general config information from a JSON file.
|
|
loadJson(cfgFile, path.Join(infraBots, "cfg.json"), &CONFIG)
|
|
|
|
// Create the JobNameSchema.
|
|
if *builderNameSchemaFile == "" {
|
|
*builderNameSchemaFile = path.Join(b.CheckoutRoot(), "infra", "bots", "recipe_modules", "builder_name_schema", "builder_name_schema.json")
|
|
}
|
|
schema, err := NewJobNameSchema(*builderNameSchemaFile)
|
|
if err != nil {
|
|
glog.Fatal(err)
|
|
}
|
|
jobNameSchema = schema
|
|
|
|
// Create Tasks and Jobs.
|
|
for _, name := range JOBS {
|
|
process(b, name)
|
|
}
|
|
|
|
b.MustFinish()
|
|
}
|
|
|
|
// TODO(borenet): The below really belongs in its own file, probably next to the
|
|
// builder_name_schema.json file.
|
|
|
|
// JobNameSchema is a struct used for (de)constructing Job names in a
|
|
// predictable format.
|
|
type JobNameSchema struct {
|
|
Schema map[string][]string `json:"builder_name_schema"`
|
|
Sep string `json:"builder_name_sep"`
|
|
}
|
|
|
|
// NewJobNameSchema returns a JobNameSchema instance based on the given JSON
|
|
// file.
|
|
func NewJobNameSchema(jsonFile string) (*JobNameSchema, error) {
|
|
var rv JobNameSchema
|
|
f, err := os.Open(jsonFile)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
defer util.Close(f)
|
|
if err := json.NewDecoder(f).Decode(&rv); err != nil {
|
|
return nil, err
|
|
}
|
|
return &rv, nil
|
|
}
|
|
|
|
// ParseJobName splits the given Job name into its component parts, according
|
|
// to the schema.
|
|
func (s *JobNameSchema) ParseJobName(n string) (map[string]string, error) {
|
|
split := strings.Split(n, s.Sep)
|
|
if len(split) < 2 {
|
|
return nil, fmt.Errorf("Invalid job name: %q", n)
|
|
}
|
|
role := split[0]
|
|
split = split[1:]
|
|
keys, ok := s.Schema[role]
|
|
if !ok {
|
|
return nil, fmt.Errorf("Invalid job name; %q is not a valid role.", role)
|
|
}
|
|
extraConfig := ""
|
|
if len(split) == len(keys)+1 {
|
|
extraConfig = split[len(split)-1]
|
|
split = split[:len(split)-1]
|
|
}
|
|
if len(split) != len(keys) {
|
|
return nil, fmt.Errorf("Invalid job name; %q has incorrect number of parts.", n)
|
|
}
|
|
rv := make(map[string]string, len(keys)+2)
|
|
rv["role"] = role
|
|
if extraConfig != "" {
|
|
rv["extra_config"] = extraConfig
|
|
}
|
|
for i, k := range keys {
|
|
rv[k] = split[i]
|
|
}
|
|
return rv, nil
|
|
}
|
|
|
|
// MakeJobName assembles the given parts of a Job name, according to the schema.
|
|
func (s *JobNameSchema) MakeJobName(parts map[string]string) (string, error) {
|
|
role, ok := parts["role"]
|
|
if !ok {
|
|
return "", fmt.Errorf("Invalid job parts; jobs must have a role.")
|
|
}
|
|
keys, ok := s.Schema[role]
|
|
if !ok {
|
|
return "", fmt.Errorf("Invalid job parts; unknown role %q", role)
|
|
}
|
|
rvParts := make([]string, 0, len(parts))
|
|
rvParts = append(rvParts, role)
|
|
for _, k := range keys {
|
|
v, ok := parts[k]
|
|
if !ok {
|
|
return "", fmt.Errorf("Invalid job parts; missing %q", k)
|
|
}
|
|
rvParts = append(rvParts, v)
|
|
}
|
|
if _, ok := parts["extra_config"]; ok {
|
|
rvParts = append(rvParts, parts["extra_config"])
|
|
}
|
|
return strings.Join(rvParts, s.Sep), nil
|
|
}
|