// Copyright 2016 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. package main /* Generate the tasks.json file. */ import ( "encoding/json" "flag" "fmt" "io/ioutil" "os" "path" "path/filepath" "regexp" "runtime" "sort" "strings" "time" "github.com/skia-dev/glog" "go.skia.org/infra/go/sklog" "go.skia.org/infra/go/util" "go.skia.org/infra/task_scheduler/go/specs" ) const ( BUNDLE_RECIPES_NAME = "Housekeeper-PerCommit-BundleRecipes" ISOLATE_SKIMAGE_NAME = "Housekeeper-PerCommit-IsolateSkImage" ISOLATE_SKP_NAME = "Housekeeper-PerCommit-IsolateSKP" ISOLATE_SVG_NAME = "Housekeeper-PerCommit-IsolateSVG" DEFAULT_OS = DEFAULT_OS_LINUX DEFAULT_OS_LINUX = "Ubuntu-14.04" // Name prefix for upload jobs. PREFIX_UPLOAD = "Upload" ) var ( // "Constants" // Top-level list of all jobs to run at each commit; loaded from // jobs.json. JOBS []string // General configuration information. CONFIG struct { GsBucketGm string `json:"gs_bucket_gm"` GsBucketNano string `json:"gs_bucket_nano"` NoUpload []string `json:"no_upload"` Pool string `json:"pool"` } // alternateSwarmDimensions can be set in an init function to override the default swarming bot // dimensions for the given task. alternateSwarmDimensions func(parts map[string]string) []string // Defines the structure of job names. jobNameSchema *JobNameSchema // Git 2.13. cipdGit1 = &specs.CipdPackage{ Name: fmt.Sprintf("infra/git/${platform}"), Path: "git", Version: fmt.Sprintf("version:2.13.0.chromium9"), } cipdGit2 = &specs.CipdPackage{ Name: fmt.Sprintf("infra/tools/git/${platform}"), Path: "git", Version: fmt.Sprintf("git_revision:a78b5f3658c0578a017db48df97d20ac09822bcd"), } // Flags. builderNameSchemaFile = flag.String("builder_name_schema", "", "Path to the builder_name_schema.json file. If not specified, uses infra/bots/recipe_modules/builder_name_schema/builder_name_schema.json from this repo.") assetsDir = flag.String("assets_dir", "", "Directory containing assets.") cfgFile = flag.String("cfg_file", "", "JSON file containing general configuration information.") jobsFile = flag.String("jobs", "", "JSON file containing jobs to run.") ) // linuxGceDimensions are the Swarming dimensions for Linux GCE // instances. func linuxGceDimensions() []string { return []string{ "cpu:x86-64-avx2", "gpu:none", fmt.Sprintf("os:%s", DEFAULT_OS_LINUX), fmt.Sprintf("pool:%s", CONFIG.Pool), } } // deriveCompileTaskName returns the name of a compile task based on the given // job name. func deriveCompileTaskName(jobName string, parts map[string]string) string { if parts["role"] == "Housekeeper" { return "Build-Ubuntu-GCC-x86_64-Release-Shared" } else if parts["role"] == "Test" || parts["role"] == "Perf" { task_os := parts["os"] ec := []string{} if val := parts["extra_config"]; val != "" { ec = strings.Split(val, "_") ignore := []string{"Skpbench", "AbandonGpuContext", "PreAbandonGpuContext", "Valgrind", "ReleaseAndAbandonGpuContext"} keep := make([]string, 0, len(ec)) for _, part := range ec { if !util.In(part, ignore) { keep = append(keep, part) } } ec = keep } if task_os == "Android" { if !util.In("Android", ec) { ec = append([]string{"Android"}, ec...) } task_os = "Ubuntu" } else if task_os == "Chromecast" { task_os = "Ubuntu" ec = append([]string{"Chromecast"}, ec...) } else if strings.Contains(task_os, "ChromeOS") { ec = append([]string{"Chromebook", "ARM", "GLES"}, ec...) task_os = "Ubuntu" } else if task_os == "iOS" { ec = append([]string{task_os}, ec...) task_os = "Mac" } else if strings.Contains(task_os, "Win") { task_os = "Win" } else if strings.Contains(task_os, "Ubuntu") { task_os = "Ubuntu" } jobNameMap := map[string]string{ "role": "Build", "os": task_os, "compiler": parts["compiler"], "target_arch": parts["arch"], "configuration": parts["configuration"], } if len(ec) > 0 { jobNameMap["extra_config"] = strings.Join(ec, "_") } name, err := jobNameSchema.MakeJobName(jobNameMap) if err != nil { glog.Fatal(err) } return name } else { return jobName } } // swarmDimensions generates swarming bot dimensions for the given task. func swarmDimensions(parts map[string]string) []string { if alternateSwarmDimensions != nil { return alternateSwarmDimensions(parts) } return defaultSwarmDimensions(parts) } // defaultSwarmDimensions generates default swarming bot dimensions for the given task. func defaultSwarmDimensions(parts map[string]string) []string { d := map[string]string{ "pool": CONFIG.Pool, } if os, ok := parts["os"]; ok { d["os"], ok = map[string]string{ "Android": "Android", "Chromecast": "Android", "ChromeOS": "ChromeOS", "Mac": "Mac-10.11", "Ubuntu": DEFAULT_OS_LINUX, "Ubuntu16": "Ubuntu-16.10", "Win": "Windows-2008ServerR2-SP1", "Win10": "Windows-10-15063", "Win2k8": "Windows-2008ServerR2-SP1", "Win7": "Windows-7-SP1", "Win8": "Windows-8.1-SP0", "iOS": "iOS-10.3.1", }[os] if !ok { glog.Fatalf("Entry %q not found in OS mapping.", os) } // Chrome Golo has a different Windows image. if parts["model"] == "Golo" && os == "Win10" { d["os"] = "Windows-10-10586" } } else { d["os"] = DEFAULT_OS } if parts["role"] == "Test" || parts["role"] == "Perf" { if strings.Contains(parts["os"], "Android") || strings.Contains(parts["os"], "Chromecast") { // For Android, the device type is a better dimension // than CPU or GPU. deviceInfo, ok := map[string][]string{ "AndroidOne": {"sprout", "MOB30Q"}, "Chorizo": {"chorizo", "1.24_82923"}, "Ci20": {"ci20", "NRD90M"}, "GalaxyJ5": {"j5xnlte", "MMB29M"}, "GalaxyS6": {"zerofltetmo", "MMB29K"}, "GalaxyS7_G930A": {"heroqlteatt", "NRD90M_G930AUCS4BQC2"}, "GalaxyS7_G930FD": {"herolte", "NRD90M_G930FXXU1DQAS"}, "GalaxyTab3": {"goyawifi", "JDQ39"}, "MotoG4": {"athene", "NPJ25.93-14"}, "NVIDIA_Shield": {"foster", "NRD90M"}, "Nexus10": {"manta", "LMY49J"}, "Nexus5": {"hammerhead", "M4B30Z"}, "Nexus6": {"shamu", "M"}, "Nexus6p": {"angler", "OPP1.170223.012"}, "Nexus7": {"grouper", "LMY47V"}, "Nexus7v2": {"flo", "M"}, "NexusPlayer": {"fugu", "OPP2.170420.017"}, "Pixel": {"sailfish", "NMF26Q"}, "PixelC": {"dragon", "N2G47D"}, "PixelXL": {"marlin", "OPP3.170518.006"}, }[parts["model"]] if !ok { glog.Fatalf("Entry %q not found in Android mapping.", parts["model"]) } d["device_type"] = deviceInfo[0] d["device_os"] = deviceInfo[1] } else if strings.Contains(parts["os"], "iOS") { device, ok := map[string]string{ "iPadMini4": "iPad5,1", "iPhone6": "iPhone7,2", "iPhone7": "iPhone9,1", "iPadPro": "iPad6,3", }[parts["model"]] if !ok { glog.Fatalf("Entry %q not found in iOS mapping.", parts["model"]) } d["device"] = device } else if parts["cpu_or_gpu"] == "CPU" { d["gpu"] = "none" cpu, ok := map[string]string{ "AVX": "x86-64", "AVX2": "x86-64-avx2", "SSE4": "x86-64", }[parts["cpu_or_gpu_value"]] if !ok { glog.Fatalf("Entry %q not found in CPU mapping.", parts["cpu_or_gpu_value"]) } d["cpu"] = cpu if strings.Contains(parts["os"], "Win") && parts["cpu_or_gpu_value"] == "AVX2" { // AVX2 is not correctly detected on Windows. Fall back on other // dimensions to ensure that we correctly target machines which we know // have AVX2 support. d["cpu"] = "x86-64" if parts["model"] != "GCE" { glog.Fatalf("Please double-check that %q supports AVX2 and update this assertion.", parts["model"]) } } } else { if strings.Contains(parts["os"], "Win") { gpu, ok := map[string]string{ "AMDHD7770": "1002:683d-22.19.165.512", "GT610": "10de:104a-21.21.13.7619", "GTX1070": "10de:1ba1-22.21.13.8205", "GTX660": "10de:11c0-22.21.13.8205", "GTX960": "10de:1401-22.21.13.8205", "IntelHD530": "8086:1912-21.20.16.4590", "IntelHD4400": "8086:0a16-20.19.15.4624", "IntelHD4600": "8086:0412-20.19.15.4624", "IntelIris540": "8086:1926-21.20.16.4590", "IntelIris6100": "8086:162b-20.19.15.4624", "RadeonR9M470X": "1002:6646-22.19.165.512", }[parts["cpu_or_gpu_value"]] if !ok { glog.Fatalf("Entry %q not found in Win GPU mapping.", parts["cpu_or_gpu_value"]) } d["gpu"] = gpu // Specify cpu dimension for NUCs and ShuttleCs. We temporarily have two // types of machines with a GTX960. cpu, ok := map[string]string{ "NUC6i7KYK": "x86-64-i7-6770HQ", "ShuttleC": "x86-64-i7-6700K", }[parts["model"]] if ok { d["cpu"] = cpu } } else if strings.Contains(parts["os"], "Ubuntu") { gpu, ok := map[string]string{ "GT610": "10de:104a-340.96", "GTX550Ti": "10de:1244-340.76", "GTX660": "10de:11c0-367.57", "GTX960": "10de:1401-367.57", // Intel drivers come from CIPD, so no need to specify the version here. "IntelBayTrail": "8086:0f31", "IntelHD2000": "8086:0102", "IntelHD405": "8086:22b1", "IntelIris540": "8086:1926", }[parts["cpu_or_gpu_value"]] if !ok { glog.Fatalf("Entry %q not found in Ubuntu GPU mapping.", parts["cpu_or_gpu_value"]) } d["gpu"] = gpu } else if strings.Contains(parts["os"], "Mac") { gpu, ok := map[string]string{ // TODO(benjaminwagner): GPU name doesn't match device ID. "IntelHD4000": "8086:0a2e", }[parts["cpu_or_gpu_value"]] if !ok { glog.Fatalf("Entry %q not found in Mac GPU mapping.", parts["cpu_or_gpu_value"]) } d["gpu"] = gpu } else if strings.Contains(parts["os"], "ChromeOS") { gpu, ok := map[string]string{ "MaliT604": "MaliT604", "MaliT764": "MaliT764", "MaliT860": "MaliT860", "TegraK1": "TegraK1", }[parts["cpu_or_gpu_value"]] if !ok { glog.Fatalf("Entry %q not found in ChromeOS GPU mapping.", parts["cpu_or_gpu_value"]) } d["gpu"] = gpu } else { glog.Fatalf("Unknown GPU mapping for OS %q.", parts["os"]) } } } else { d["gpu"] = "none" if d["os"] == DEFAULT_OS_LINUX { return linuxGceDimensions() } } rv := make([]string, 0, len(d)) for k, v := range d { rv = append(rv, fmt.Sprintf("%s:%s", k, v)) } sort.Strings(rv) return rv } // relpath returns the relative path to the given file from the config file. func relpath(f string) string { _, filename, _, _ := runtime.Caller(0) dir := path.Dir(filename) rel := dir if *cfgFile != "" { rel = path.Dir(*cfgFile) } rv, err := filepath.Rel(rel, path.Join(dir, f)) if err != nil { sklog.Fatal(err) } return rv } // bundleRecipes generates the task to bundle and isolate the recipes. func bundleRecipes(b *specs.TasksCfgBuilder) string { b.MustAddTask(BUNDLE_RECIPES_NAME, &specs.TaskSpec{ CipdPackages: []*specs.CipdPackage{cipdGit1, cipdGit2}, Dimensions: linuxGceDimensions(), ExtraArgs: []string{ "--workdir", "../../..", "bundle_recipes", fmt.Sprintf("buildername=%s", BUNDLE_RECIPES_NAME), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), }, Isolate: relpath("bundle_recipes.isolate"), Priority: 0.7, }) return BUNDLE_RECIPES_NAME } // useBundledRecipes returns true iff the given bot should use bundled recipes // instead of syncing recipe DEPS itself. func useBundledRecipes(parts map[string]string) bool { // Use bundled recipes for all test/perf tasks. return true } type isolateAssetCfg struct { isolateFile string cipdPkg string } var ISOLATE_ASSET_MAPPING = map[string]isolateAssetCfg{ ISOLATE_SKIMAGE_NAME: { isolateFile: "isolate_skimage.isolate", cipdPkg: "skimage", }, ISOLATE_SKP_NAME: { isolateFile: "isolate_skp.isolate", cipdPkg: "skp", }, ISOLATE_SVG_NAME: { isolateFile: "isolate_svg.isolate", cipdPkg: "svg", }, } // bundleRecipes generates the task to bundle and isolate the recipes. func isolateCIPDAsset(b *specs.TasksCfgBuilder, name string) string { b.MustAddTask(name, &specs.TaskSpec{ CipdPackages: []*specs.CipdPackage{ b.MustGetCipdPackageFromAsset(ISOLATE_ASSET_MAPPING[name].cipdPkg), }, Dimensions: linuxGceDimensions(), Isolate: relpath(ISOLATE_ASSET_MAPPING[name].isolateFile), Priority: 0.7, }) return name } // getIsolatedCIPDDeps returns the slice of Isolate_* tasks a given task needs. // This allows us to save time on I/O bound bots, like the RPIs. func getIsolatedCIPDDeps(parts map[string]string) []string { deps := []string{} // Only do this on the RPIs for now. Other, faster machines shouldn't see much // benefit and we don't need the extra complexity, for now rpiOS := []string{"Android", "ChromeOS", "iOS"} if o := parts["os"]; strings.Contains(o, "Chromecast") { // Chromecasts don't have enough disk space to fit all of the content, // so we do a subset of the skps. deps = append(deps, ISOLATE_SKP_NAME) } else if e := parts["extra_config"]; strings.Contains(e, "Skpbench") { // Skpbench only needs skps deps = append(deps, ISOLATE_SKP_NAME) } else if util.In(o, rpiOS) { deps = append(deps, ISOLATE_SKP_NAME) deps = append(deps, ISOLATE_SVG_NAME) deps = append(deps, ISOLATE_SKIMAGE_NAME) } return deps } // compile generates a compile task. Returns the name of the last task in the // generated chain of tasks, which the Job should add as a dependency. func compile(b *specs.TasksCfgBuilder, name string, parts map[string]string) string { // Collect the necessary CIPD packages. pkgs := []*specs.CipdPackage{} // Android bots require a toolchain. if strings.Contains(name, "Android") { if strings.Contains(name, "Mac") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("android_ndk_darwin")) } else if strings.Contains(name, "Win") { pkg := b.MustGetCipdPackageFromAsset("android_ndk_windows") pkg.Path = "n" pkgs = append(pkgs, pkg) } else { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("android_ndk_linux")) } } else if strings.Contains(name, "Chromecast") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("cast_toolchain")) pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("chromebook_arm_gles")) } else if strings.Contains(name, "Chromebook") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux")) pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("armhf_sysroot")) pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("chromebook_arm_gles")) } else if strings.Contains(name, "Ubuntu") { if strings.Contains(name, "Clang") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux")) } if strings.Contains(name, "Vulkan") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_sdk")) } } else if strings.Contains(name, "Win") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("win_toolchain")) if strings.Contains(name, "Vulkan") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("win_vulkan_sdk")) } } // TODO(stephana): Remove this once all Mac machines are on the same // OS version again. Move the call to swarmDimensions back to the // creation of the TaskSpec struct below. dimensions := swarmDimensions(parts) if strings.Contains(name, "Mac") { for idx, dim := range dimensions { if strings.HasPrefix(dim, "os") { dimensions[idx] = "os:Mac-10.12" break } } } // Add the task. b.MustAddTask(name, &specs.TaskSpec{ CipdPackages: pkgs, Dimensions: dimensions, ExtraArgs: []string{ "--workdir", "../../..", "compile", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), }, Isolate: relpath("compile_skia.isolate"), Priority: 0.8, }) // All compile tasks are runnable as their own Job. Assert that the Job // is listed in JOBS. if !util.In(name, JOBS) { glog.Fatalf("Job %q is missing from the JOBS list!", name) } return name } // recreateSKPs generates a RecreateSKPs task. Returns the name of the last // task in the generated chain of tasks, which the Job should add as a // dependency. func recreateSKPs(b *specs.TasksCfgBuilder, name string) string { b.MustAddTask(name, &specs.TaskSpec{ CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")}, Dimensions: linuxGceDimensions(), ExecutionTimeout: 4 * time.Hour, ExtraArgs: []string{ "--workdir", "../../..", "recreate_skps", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), }, IoTimeout: 40 * time.Minute, Isolate: relpath("compile_skia.isolate"), Priority: 0.8, }) return name } // updateMetaConfig generates a UpdateMetaConfig task. Returns the name of the // last task in the generated chain of tasks, which the Job should add as a // dependency. func updateMetaConfig(b *specs.TasksCfgBuilder, name string) string { b.MustAddTask(name, &specs.TaskSpec{ CipdPackages: []*specs.CipdPackage{}, Dimensions: linuxGceDimensions(), ExtraArgs: []string{ "--workdir", "../../..", "update_meta_config", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), }, Isolate: relpath("meta_config.isolate"), Priority: 0.8, }) return name } // ctSKPs generates a CT SKPs task. Returns the name of the last task in the // generated chain of tasks, which the Job should add as a dependency. func ctSKPs(b *specs.TasksCfgBuilder, name string) string { b.MustAddTask(name, &specs.TaskSpec{ CipdPackages: []*specs.CipdPackage{}, Dimensions: []string{"pool:SkiaCT"}, ExecutionTimeout: 24 * time.Hour, ExtraArgs: []string{ "--workdir", "../../..", "ct_skps", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), }, IoTimeout: time.Hour, Isolate: relpath("ct_skps_skia.isolate"), Priority: 0.8, }) return name } // housekeeper generates a Housekeeper task. Returns the name of the last task // in the generated chain of tasks, which the Job should add as a dependency. func housekeeper(b *specs.TasksCfgBuilder, name, compileTaskName string) string { b.MustAddTask(name, &specs.TaskSpec{ CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")}, Dependencies: []string{compileTaskName}, Dimensions: linuxGceDimensions(), ExtraArgs: []string{ "--workdir", "../../..", "housekeeper", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), }, Isolate: relpath("housekeeper_skia.isolate"), Priority: 0.8, }) return name } // infra generates an infra_tests task. Returns the name of the last task in the // generated chain of tasks, which the Job should add as a dependency. func infra(b *specs.TasksCfgBuilder, name string) string { b.MustAddTask(name, &specs.TaskSpec{ CipdPackages: []*specs.CipdPackage{b.MustGetCipdPackageFromAsset("go")}, Dimensions: linuxGceDimensions(), ExtraArgs: []string{ "--workdir", "../../..", "infra", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), }, Isolate: relpath("infra_skia.isolate"), Priority: 0.8, }) return name } // doUpload indicates whether the given Job should upload its results. func doUpload(name string) bool { for _, s := range CONFIG.NoUpload { m, err := regexp.MatchString(s, name) if err != nil { glog.Fatal(err) } if m { return false } } return true } // test generates a Test task. Returns the name of the last task in the // generated chain of tasks, which the Job should add as a dependency. func test(b *specs.TasksCfgBuilder, name string, parts map[string]string, compileTaskName string, pkgs []*specs.CipdPackage) string { s := &specs.TaskSpec{ CipdPackages: pkgs, Dependencies: []string{compileTaskName}, Dimensions: swarmDimensions(parts), ExecutionTimeout: 4 * time.Hour, Expiration: 20 * time.Hour, ExtraArgs: []string{ "--workdir", "../../..", "test", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), }, IoTimeout: 40 * time.Minute, Isolate: relpath("test_skia.isolate"), MaxAttempts: 1, Priority: 0.8, } if useBundledRecipes(parts) { s.Dependencies = append(s.Dependencies, BUNDLE_RECIPES_NAME) if strings.Contains(parts["os"], "Win") { s.Isolate = relpath("test_skia_bundled_win.isolate") } else { s.Isolate = relpath("test_skia_bundled_unix.isolate") } } if deps := getIsolatedCIPDDeps(parts); len(deps) > 0 { s.Dependencies = append(s.Dependencies, deps...) } if strings.Contains(parts["extra_config"], "Valgrind") { s.ExecutionTimeout = 9 * time.Hour s.Expiration = 48 * time.Hour s.IoTimeout = time.Hour s.CipdPackages = append(s.CipdPackages, b.MustGetCipdPackageFromAsset("valgrind")) } else if strings.Contains(parts["extra_config"], "MSAN") { s.ExecutionTimeout = 9 * time.Hour } else if parts["arch"] == "x86" && parts["configuration"] == "Debug" { // skia:6737 s.ExecutionTimeout = 6 * time.Hour } b.MustAddTask(name, s) // Upload results if necessary. if doUpload(name) { uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name) b.MustAddTask(uploadName, &specs.TaskSpec{ Dependencies: []string{name}, Dimensions: linuxGceDimensions(), ExtraArgs: []string{ "--workdir", "../../..", "upload_dm_results", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketGm), }, Isolate: relpath("upload_dm_results.isolate"), Priority: 0.8, }) return uploadName } return name } // perf generates a Perf task. Returns the name of the last task in the // generated chain of tasks, which the Job should add as a dependency. func perf(b *specs.TasksCfgBuilder, name string, parts map[string]string, compileTaskName string, pkgs []*specs.CipdPackage) string { recipe := "perf" isolate := relpath("perf_skia.isolate") if strings.Contains(parts["extra_config"], "Skpbench") { recipe = "skpbench" isolate = relpath("skpbench_skia.isolate") if useBundledRecipes(parts) { if strings.Contains(parts["os"], "Win") { isolate = relpath("skpbench_skia_bundled_win.isolate") } else { isolate = relpath("skpbench_skia_bundled_unix.isolate") } } } else if useBundledRecipes(parts) { if strings.Contains(parts["os"], "Win") { isolate = relpath("perf_skia_bundled_win.isolate") } else { isolate = relpath("perf_skia_bundled_unix.isolate") } } s := &specs.TaskSpec{ CipdPackages: pkgs, Dependencies: []string{compileTaskName}, Dimensions: swarmDimensions(parts), ExecutionTimeout: 4 * time.Hour, Expiration: 20 * time.Hour, ExtraArgs: []string{ "--workdir", "../../..", recipe, fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), }, IoTimeout: 40 * time.Minute, Isolate: isolate, MaxAttempts: 1, Priority: 0.8, } if useBundledRecipes(parts) { s.Dependencies = append(s.Dependencies, BUNDLE_RECIPES_NAME) } if deps := getIsolatedCIPDDeps(parts); len(deps) > 0 { s.Dependencies = append(s.Dependencies, deps...) } if strings.Contains(parts["extra_config"], "Valgrind") { s.ExecutionTimeout = 9 * time.Hour s.Expiration = 48 * time.Hour s.IoTimeout = time.Hour s.CipdPackages = append(s.CipdPackages, b.MustGetCipdPackageFromAsset("valgrind")) } else if strings.Contains(parts["extra_config"], "MSAN") { s.ExecutionTimeout = 9 * time.Hour } else if parts["arch"] == "x86" && parts["configuration"] == "Debug" { // skia:6737 s.ExecutionTimeout = 6 * time.Hour } b.MustAddTask(name, s) // Upload results if necessary. if strings.Contains(name, "Release") && doUpload(name) { uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, jobNameSchema.Sep, name) b.MustAddTask(uploadName, &specs.TaskSpec{ Dependencies: []string{name}, Dimensions: linuxGceDimensions(), ExtraArgs: []string{ "--workdir", "../../..", "upload_nano_results", fmt.Sprintf("repository=%s", specs.PLACEHOLDER_REPO), fmt.Sprintf("buildername=%s", name), fmt.Sprintf("swarm_out_dir=%s", specs.PLACEHOLDER_ISOLATED_OUTDIR), fmt.Sprintf("revision=%s", specs.PLACEHOLDER_REVISION), fmt.Sprintf("patch_repo=%s", specs.PLACEHOLDER_PATCH_REPO), fmt.Sprintf("patch_storage=%s", specs.PLACEHOLDER_PATCH_STORAGE), fmt.Sprintf("patch_issue=%s", specs.PLACEHOLDER_ISSUE), fmt.Sprintf("patch_set=%s", specs.PLACEHOLDER_PATCHSET), fmt.Sprintf("gs_bucket=%s", CONFIG.GsBucketNano), }, Isolate: relpath("upload_nano_results.isolate"), Priority: 0.8, }) return uploadName } return name } // process generates tasks and jobs for the given job name. func process(b *specs.TasksCfgBuilder, name string) { deps := []string{} // Bundle Recipes. if name == BUNDLE_RECIPES_NAME { deps = append(deps, bundleRecipes(b)) } // Isolate CIPD assets. if _, ok := ISOLATE_ASSET_MAPPING[name]; ok { deps = append(deps, isolateCIPDAsset(b, name)) } parts, err := jobNameSchema.ParseJobName(name) if err != nil { glog.Fatal(err) } // RecreateSKPs. if strings.Contains(name, "RecreateSKPs") { deps = append(deps, recreateSKPs(b, name)) } // UpdateMetaConfig bot. if strings.Contains(name, "UpdateMetaConfig") { deps = append(deps, updateMetaConfig(b, name)) } // CT bots. if strings.Contains(name, "-CT_") { deps = append(deps, ctSKPs(b, name)) } // Infra tests. if name == "Housekeeper-PerCommit-InfraTests" { deps = append(deps, infra(b, name)) } // Compile bots. if parts["role"] == "Build" { deps = append(deps, compile(b, name, parts)) } // Most remaining bots need a compile task. compileTaskName := deriveCompileTaskName(name, parts) compileTaskParts, err := jobNameSchema.ParseJobName(compileTaskName) if err != nil { glog.Fatal(err) } // These bots do not need a compile task. if parts["role"] != "Build" && name != "Housekeeper-PerCommit-BundleRecipes" && name != "Housekeeper-PerCommit-InfraTests" && !strings.Contains(name, "RecreateSKPs") && !strings.Contains(name, "UpdateMetaConfig") && !strings.Contains(name, "-CT_") && !strings.Contains(name, "Housekeeper-PerCommit-Isolate") { compile(b, compileTaskName, compileTaskParts) } // Housekeeper. if name == "Housekeeper-PerCommit" { deps = append(deps, housekeeper(b, name, compileTaskName)) } // Common assets needed by the remaining bots. pkgs := []*specs.CipdPackage{} if deps := getIsolatedCIPDDeps(parts); len(deps) == 0 { pkgs = []*specs.CipdPackage{ b.MustGetCipdPackageFromAsset("skimage"), b.MustGetCipdPackageFromAsset("skp"), b.MustGetCipdPackageFromAsset("svg"), } } if strings.Contains(name, "Ubuntu") && strings.Contains(name, "SAN") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("clang_linux")) } if strings.Contains(name, "Ubuntu16") { if strings.Contains(name, "Vulkan") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_sdk")) } if strings.Contains(name, "Release") { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_intel_driver_release")) } else { pkgs = append(pkgs, b.MustGetCipdPackageFromAsset("linux_vulkan_intel_driver_debug")) } } // Test bots. if parts["role"] == "Test" && !strings.Contains(name, "-CT_") { deps = append(deps, test(b, name, parts, compileTaskName, pkgs)) } // Perf bots. if parts["role"] == "Perf" && !strings.Contains(name, "-CT_") { deps = append(deps, perf(b, name, parts, compileTaskName, pkgs)) } // Add the Job spec. j := &specs.JobSpec{ Priority: 0.8, TaskSpecs: deps, } if name == "Housekeeper-Nightly-RecreateSKPs_Canary" { j.Trigger = "nightly" } if name == "Housekeeper-Nightly-UpdateMetaConfig" { j.Trigger = "nightly" } if name == "Housekeeper-Weekly-RecreateSKPs" { j.Trigger = "weekly" } if name == "Test-Ubuntu-GCC-GCE-CPU-AVX2-x86_64-Debug-CT_DM_1m_SKPs" { j.Trigger = "weekly" } b.MustAddJob(name, j) } func loadJson(flag *string, defaultFlag string, val interface{}) { if *flag == "" { *flag = defaultFlag } b, err := ioutil.ReadFile(*flag) if err != nil { glog.Fatal(err) } if err := json.Unmarshal(b, val); err != nil { glog.Fatal(err) } } // Regenerate the tasks.json file. func main() { b := specs.MustNewTasksCfgBuilder() b.SetAssetsDir(*assetsDir) infraBots := path.Join(b.CheckoutRoot(), "infra", "bots") // Load the jobs from a JSON file. loadJson(jobsFile, path.Join(infraBots, "jobs.json"), &JOBS) // Load general config information from a JSON file. loadJson(cfgFile, path.Join(infraBots, "cfg.json"), &CONFIG) // Create the JobNameSchema. if *builderNameSchemaFile == "" { *builderNameSchemaFile = path.Join(b.CheckoutRoot(), "infra", "bots", "recipe_modules", "builder_name_schema", "builder_name_schema.json") } schema, err := NewJobNameSchema(*builderNameSchemaFile) if err != nil { glog.Fatal(err) } jobNameSchema = schema // Create Tasks and Jobs. for _, name := range JOBS { process(b, name) } b.MustFinish() } // TODO(borenet): The below really belongs in its own file, probably next to the // builder_name_schema.json file. // JobNameSchema is a struct used for (de)constructing Job names in a // predictable format. type JobNameSchema struct { Schema map[string][]string `json:"builder_name_schema"` Sep string `json:"builder_name_sep"` } // NewJobNameSchema returns a JobNameSchema instance based on the given JSON // file. func NewJobNameSchema(jsonFile string) (*JobNameSchema, error) { var rv JobNameSchema f, err := os.Open(jsonFile) if err != nil { return nil, err } defer util.Close(f) if err := json.NewDecoder(f).Decode(&rv); err != nil { return nil, err } return &rv, nil } // ParseJobName splits the given Job name into its component parts, according // to the schema. func (s *JobNameSchema) ParseJobName(n string) (map[string]string, error) { split := strings.Split(n, s.Sep) if len(split) < 2 { return nil, fmt.Errorf("Invalid job name: %q", n) } role := split[0] split = split[1:] keys, ok := s.Schema[role] if !ok { return nil, fmt.Errorf("Invalid job name; %q is not a valid role.", role) } extraConfig := "" if len(split) == len(keys)+1 { extraConfig = split[len(split)-1] split = split[:len(split)-1] } if len(split) != len(keys) { return nil, fmt.Errorf("Invalid job name; %q has incorrect number of parts.", n) } rv := make(map[string]string, len(keys)+2) rv["role"] = role if extraConfig != "" { rv["extra_config"] = extraConfig } for i, k := range keys { rv[k] = split[i] } return rv, nil } // MakeJobName assembles the given parts of a Job name, according to the schema. func (s *JobNameSchema) MakeJobName(parts map[string]string) (string, error) { role, ok := parts["role"] if !ok { return "", fmt.Errorf("Invalid job parts; jobs must have a role.") } keys, ok := s.Schema[role] if !ok { return "", fmt.Errorf("Invalid job parts; unknown role %q", role) } rvParts := make([]string, 0, len(parts)) rvParts = append(rvParts, role) for _, k := range keys { v, ok := parts[k] if !ok { return "", fmt.Errorf("Invalid job parts; missing %q", k) } rvParts = append(rvParts, v) } if _, ok := parts["extra_config"]; ok { rvParts = append(rvParts, parts["extra_config"]) } return strings.Join(rvParts, s.Sep), nil }