9b395f55ea
Bug: skia:11082 Change-Id: I5ba2f9ae85ce7e008c2fc656be605c022466720d Reviewed-on: https://skia-review.googlesource.com/c/skia/+/343938 Reviewed-by: Joe Gregorio <jcgregorio@google.com> Reviewed-by: Tyler Denniston <tdenniston@google.com> Commit-Queue: Weston Tracey <westont@google.com>
1779 lines
57 KiB
Go
1779 lines
57 KiB
Go
// Copyright 2016 The Chromium Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file.
|
|
|
|
package gen_tasks_logic
|
|
|
|
/*
|
|
Generate the tasks.json file.
|
|
*/
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"log"
|
|
"path"
|
|
"path/filepath"
|
|
"regexp"
|
|
"runtime"
|
|
"sort"
|
|
"strconv"
|
|
"strings"
|
|
"time"
|
|
|
|
"go.skia.org/infra/go/cipd"
|
|
"go.skia.org/infra/task_scheduler/go/specs"
|
|
)
|
|
|
|
const (
|
|
BUILD_TASK_DRIVERS_NAME = "Housekeeper-PerCommit-BuildTaskDrivers"
|
|
BUNDLE_RECIPES_NAME = "Housekeeper-PerCommit-BundleRecipes"
|
|
ISOLATE_GCLOUD_LINUX_NAME = "Housekeeper-PerCommit-IsolateGCloudLinux"
|
|
ISOLATE_SKIMAGE_NAME = "Housekeeper-PerCommit-IsolateSkImage"
|
|
ISOLATE_SKP_NAME = "Housekeeper-PerCommit-IsolateSKP"
|
|
ISOLATE_MSKP_NAME = "Housekeeper-PerCommit-IsolateMSKP"
|
|
ISOLATE_SVG_NAME = "Housekeeper-PerCommit-IsolateSVG"
|
|
ISOLATE_NDK_LINUX_NAME = "Housekeeper-PerCommit-IsolateAndroidNDKLinux"
|
|
ISOLATE_SDK_LINUX_NAME = "Housekeeper-PerCommit-IsolateAndroidSDKLinux"
|
|
ISOLATE_WIN_TOOLCHAIN_NAME = "Housekeeper-PerCommit-IsolateWinToolchain"
|
|
|
|
DEFAULT_OS_DEBIAN = "Debian-10.3"
|
|
DEFAULT_OS_LINUX_GCE = "Debian-10.3"
|
|
OLD_OS_LINUX_GCE = "Debian-9.8"
|
|
COMPILE_TASK_NAME_OS_LINUX = "Debian10"
|
|
COMPILE_TASK_NAME_OS_LINUX_OLD = "Debian9"
|
|
DEFAULT_OS_MAC = "Mac-10.14.6"
|
|
DEFAULT_OS_WIN = "Windows-Server-17763"
|
|
|
|
// Small is a 2-core machine.
|
|
// TODO(dogben): Would n1-standard-1 or n1-standard-2 be sufficient?
|
|
MACHINE_TYPE_SMALL = "n1-highmem-2"
|
|
// Medium is a 16-core machine
|
|
MACHINE_TYPE_MEDIUM = "n1-standard-16"
|
|
// Large is a 64-core machine. (We use "highcpu" because we don't need more than 57GB memory for
|
|
// any of our tasks.)
|
|
MACHINE_TYPE_LARGE = "n1-highcpu-64"
|
|
|
|
// Swarming output dirs.
|
|
OUTPUT_NONE = "output_ignored" // This will result in outputs not being isolated.
|
|
OUTPUT_BUILD = "build"
|
|
OUTPUT_TEST = "test"
|
|
OUTPUT_PERF = "perf"
|
|
|
|
// Name prefix for upload jobs.
|
|
PREFIX_UPLOAD = "Upload"
|
|
)
|
|
|
|
var (
|
|
// "Constants"
|
|
|
|
// Named caches used by tasks.
|
|
CACHES_GIT = []*specs.Cache{
|
|
&specs.Cache{
|
|
Name: "git",
|
|
Path: "cache/git",
|
|
},
|
|
&specs.Cache{
|
|
Name: "git_cache",
|
|
Path: "cache/git_cache",
|
|
},
|
|
}
|
|
CACHES_GO = []*specs.Cache{
|
|
&specs.Cache{
|
|
Name: "go_cache",
|
|
Path: "cache/go_cache",
|
|
},
|
|
&specs.Cache{
|
|
Name: "gopath",
|
|
Path: "cache/gopath",
|
|
},
|
|
}
|
|
CACHES_WORKDIR = []*specs.Cache{
|
|
&specs.Cache{
|
|
Name: "work",
|
|
Path: "cache/work",
|
|
},
|
|
}
|
|
CACHES_CCACHE = []*specs.Cache{
|
|
&specs.Cache{
|
|
Name: "ccache",
|
|
Path: "cache/ccache",
|
|
},
|
|
}
|
|
// The "docker" cache is used as a persistent working directory for
|
|
// tasks which use Docker. It is not to be confused with Docker's own
|
|
// cache, which stores images. We do not currently use a named Swarming
|
|
// cache for the latter.
|
|
// TODO(borenet): We should ensure that any task which uses Docker does
|
|
// not also use the normal "work" cache, to prevent issues like
|
|
// https://bugs.chromium.org/p/skia/issues/detail?id=9749.
|
|
CACHES_DOCKER = []*specs.Cache{
|
|
&specs.Cache{
|
|
Name: "docker",
|
|
Path: "cache/docker",
|
|
},
|
|
}
|
|
|
|
// TODO(borenet): This hacky and bad.
|
|
CIPD_PKG_LUCI_AUTH = cipd.MustGetPackage("infra/tools/luci-auth/${platform}")
|
|
|
|
CIPD_PKGS_GOLDCTL = []*specs.CipdPackage{cipd.MustGetPackage("skia/tools/goldctl/${platform}")}
|
|
|
|
CIPD_PKGS_XCODE = []*specs.CipdPackage{
|
|
// https://chromium.googlesource.com/chromium/tools/build/+/e19b7d9390e2bb438b566515b141ed2b9ed2c7c2/scripts/slave/recipe_modules/ios/api.py#317
|
|
// This package is really just an installer for XCode.
|
|
&specs.CipdPackage{
|
|
Name: "infra/tools/mac_toolchain/${platform}",
|
|
Path: "mac_toolchain",
|
|
// When this is updated, also update
|
|
// https://skia.googlesource.com/skcms.git/+/f1e2b45d18facbae2dece3aca673fe1603077846/infra/bots/gen_tasks.go#56
|
|
Version: "git_revision:796d2b92cff93fc2059623ce0a66284373ceea0a",
|
|
},
|
|
}
|
|
|
|
// These properties are required by some tasks, eg. for running
|
|
// bot_update, but they prevent de-duplication, so they should only be
|
|
// used where necessary.
|
|
EXTRA_PROPS = map[string]string{
|
|
"buildbucket_build_id": specs.PLACEHOLDER_BUILDBUCKET_BUILD_ID,
|
|
"patch_issue": specs.PLACEHOLDER_ISSUE_INT,
|
|
"patch_ref": specs.PLACEHOLDER_PATCH_REF,
|
|
"patch_repo": specs.PLACEHOLDER_PATCH_REPO,
|
|
"patch_set": specs.PLACEHOLDER_PATCHSET_INT,
|
|
"patch_storage": specs.PLACEHOLDER_PATCH_STORAGE,
|
|
"repository": specs.PLACEHOLDER_REPO,
|
|
"revision": specs.PLACEHOLDER_REVISION,
|
|
"task_id": specs.PLACEHOLDER_TASK_ID,
|
|
}
|
|
|
|
// ISOLATE_ASSET_MAPPING maps the name of an asset to the configuration
|
|
// for how the CIPD package should be installed for a given task.
|
|
ISOLATE_ASSET_MAPPING = map[string]isolateAssetCfg{
|
|
"gcloud_linux": {
|
|
isolateTaskName: ISOLATE_GCLOUD_LINUX_NAME,
|
|
path: "gcloud_linux",
|
|
},
|
|
"skimage": {
|
|
isolateTaskName: ISOLATE_SKIMAGE_NAME,
|
|
path: "skimage",
|
|
},
|
|
"skp": {
|
|
isolateTaskName: ISOLATE_SKP_NAME,
|
|
path: "skp",
|
|
},
|
|
"svg": {
|
|
isolateTaskName: ISOLATE_SVG_NAME,
|
|
path: "svg",
|
|
},
|
|
"mskp": {
|
|
isolateTaskName: ISOLATE_MSKP_NAME,
|
|
path: "mskp",
|
|
},
|
|
"android_ndk_linux": {
|
|
isolateTaskName: ISOLATE_NDK_LINUX_NAME,
|
|
path: "android_ndk_linux",
|
|
},
|
|
"android_sdk_linux": {
|
|
isolateTaskName: ISOLATE_SDK_LINUX_NAME,
|
|
path: "android_sdk_linux",
|
|
},
|
|
"win_toolchain": {
|
|
alwaysIsolate: true,
|
|
isolateTaskName: ISOLATE_WIN_TOOLCHAIN_NAME,
|
|
path: "win_toolchain",
|
|
},
|
|
}
|
|
)
|
|
|
|
// Config contains general configuration information.
|
|
type Config struct {
|
|
// Directory containing assets. Assumed to be relative to the directory
|
|
// which contains the calling gen_tasks.go file. If not specified, uses
|
|
// the infra/bots/assets from this repo.
|
|
AssetsDir string `json:"assets_dir"`
|
|
|
|
// Path to the builder name schema JSON file. Assumed to be relative to
|
|
// the directory which contains the calling gen_tasks.go file. If not
|
|
// specified, uses infra/bots/recipe_modules/builder_name_schema/builder_name_schema.json
|
|
// from this repo.
|
|
BuilderNameSchemaFile string `json:"builder_name_schema"`
|
|
|
|
// URL of the Skia Gold known hashes endpoint.
|
|
GoldHashesURL string `json:"gold_hashes_url"`
|
|
|
|
// GCS bucket used for GM results.
|
|
GsBucketGm string `json:"gs_bucket_gm"`
|
|
|
|
// GCS bucket used for Nanobench results.
|
|
GsBucketNano string `json:"gs_bucket_nano"`
|
|
|
|
// Optional function which returns a bot ID for internal devices.
|
|
InternalHardwareLabel func(parts map[string]string) *int `json:"-"`
|
|
|
|
// List of task names for which we'll never upload results.
|
|
NoUpload []string `json:"no_upload"`
|
|
|
|
// Swarming pool used for triggering tasks.
|
|
Pool string `json:"pool"`
|
|
|
|
// LUCI project associated with this repo.
|
|
Project string `json:"project"`
|
|
|
|
// Service accounts.
|
|
ServiceAccountCanary string `json:"service_account_canary"`
|
|
ServiceAccountCompile string `json:"service_account_compile"`
|
|
ServiceAccountHousekeeper string `json:"service_account_housekeeper"`
|
|
ServiceAccountRecreateSKPs string `json:"service_account_recreate_skps"`
|
|
ServiceAccountUploadBinary string `json:"service_account_upload_binary"`
|
|
ServiceAccountUploadGM string `json:"service_account_upload_gm"`
|
|
ServiceAccountUploadNano string `json:"service_account_upload_nano"`
|
|
|
|
// Optional override function which derives Swarming bot dimensions
|
|
// from parts of task names.
|
|
SwarmDimensions func(parts map[string]string) []string `json:"-"`
|
|
}
|
|
|
|
// LoadConfig loads the Config from a cfg.json file which is the sibling of the
|
|
// calling gen_tasks.go file.
|
|
func LoadConfig() *Config {
|
|
cfgDir := getCallingDirName()
|
|
var cfg Config
|
|
LoadJson(filepath.Join(cfgDir, "cfg.json"), &cfg)
|
|
return &cfg
|
|
}
|
|
|
|
// CheckoutRoot is a wrapper around specs.GetCheckoutRoot which prevents the
|
|
// caller from needing a dependency on the specs package.
|
|
func CheckoutRoot() string {
|
|
root, err := specs.GetCheckoutRoot()
|
|
if err != nil {
|
|
log.Fatal(err)
|
|
}
|
|
return root
|
|
}
|
|
|
|
// LoadJson loads JSON from the given file and unmarshals it into the given
|
|
// destination.
|
|
func LoadJson(filename string, dest interface{}) {
|
|
b, err := ioutil.ReadFile(filename)
|
|
if err != nil {
|
|
log.Fatalf("Unable to read %q: %s", filename, err)
|
|
}
|
|
if err := json.Unmarshal(b, dest); err != nil {
|
|
log.Fatalf("Unable to parse %q: %s", filename, err)
|
|
}
|
|
}
|
|
|
|
// In returns true if |s| is *in* |a| slice.
|
|
// TODO(borenet): This is copied from go.skia.org/infra/go/util to avoid the
|
|
// huge set of additional dependencies added by that package.
|
|
func In(s string, a []string) bool {
|
|
for _, x := range a {
|
|
if x == s {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
// GenTasks regenerates the tasks.json file. Loads the job list from a jobs.json
|
|
// file which is the sibling of the calling gen_tasks.go file. If cfg is nil, it
|
|
// is similarly loaded from a cfg.json file which is the sibling of the calling
|
|
// gen_tasks.go file.
|
|
func GenTasks(cfg *Config) {
|
|
b := specs.MustNewTasksCfgBuilder()
|
|
|
|
// Find the paths to the infra/bots directories in this repo and the
|
|
// repo of the calling file.
|
|
relpathTargetDir := getThisDirName()
|
|
relpathBaseDir := getCallingDirName()
|
|
|
|
var jobs []string
|
|
LoadJson(filepath.Join(relpathBaseDir, "jobs.json"), &jobs)
|
|
|
|
if cfg == nil {
|
|
cfg = new(Config)
|
|
LoadJson(filepath.Join(relpathBaseDir, "cfg.json"), cfg)
|
|
}
|
|
|
|
// Create the JobNameSchema.
|
|
builderNameSchemaFile := filepath.Join(relpathTargetDir, "recipe_modules", "builder_name_schema", "builder_name_schema.json")
|
|
if cfg.BuilderNameSchemaFile != "" {
|
|
builderNameSchemaFile = filepath.Join(relpathBaseDir, cfg.BuilderNameSchemaFile)
|
|
}
|
|
schema, err := NewJobNameSchema(builderNameSchemaFile)
|
|
if err != nil {
|
|
log.Fatal(err)
|
|
}
|
|
|
|
// Set the assets dir.
|
|
assetsDir := filepath.Join(relpathTargetDir, "assets")
|
|
if cfg.AssetsDir != "" {
|
|
assetsDir = filepath.Join(relpathBaseDir, cfg.AssetsDir)
|
|
}
|
|
b.SetAssetsDir(assetsDir)
|
|
|
|
// Create Tasks and Jobs.
|
|
builder := &builder{
|
|
TasksCfgBuilder: b,
|
|
cfg: cfg,
|
|
jobNameSchema: schema,
|
|
jobs: jobs,
|
|
relpathBaseDir: relpathBaseDir,
|
|
relpathTargetDir: relpathTargetDir,
|
|
}
|
|
for _, name := range jobs {
|
|
jb := newJobBuilder(builder, name)
|
|
jb.genTasksForJob()
|
|
jb.finish()
|
|
}
|
|
builder.MustFinish()
|
|
}
|
|
|
|
// getThisDirName returns the infra/bots directory which is an ancestor of this
|
|
// file.
|
|
func getThisDirName() string {
|
|
_, thisFileName, _, ok := runtime.Caller(0)
|
|
if !ok {
|
|
log.Fatal("Unable to find path to current file.")
|
|
}
|
|
return filepath.Dir(filepath.Dir(thisFileName))
|
|
}
|
|
|
|
// getCallingDirName returns the infra/bots directory which is an ancestor of
|
|
// the calling gen_tasks.go file. WARNING: assumes that the calling gen_tasks.go
|
|
// file appears two steps up the stack; do not call from a function which is not
|
|
// directly called by gen_tasks.go.
|
|
func getCallingDirName() string {
|
|
_, callingFileName, _, ok := runtime.Caller(2)
|
|
if !ok {
|
|
log.Fatal("Unable to find path to calling file.")
|
|
}
|
|
return filepath.Dir(callingFileName)
|
|
}
|
|
|
|
// builder is a wrapper for specs.TasksCfgBuilder.
|
|
type builder struct {
|
|
*specs.TasksCfgBuilder
|
|
cfg *Config
|
|
jobNameSchema *JobNameSchema
|
|
jobs []string
|
|
relpathBaseDir string
|
|
relpathTargetDir string
|
|
}
|
|
|
|
// marshalJson encodes the given data as JSON and fixes escaping of '<' which Go
|
|
// does by default.
|
|
func marshalJson(data interface{}) string {
|
|
j, err := json.Marshal(data)
|
|
if err != nil {
|
|
log.Fatal(err)
|
|
}
|
|
return strings.Replace(string(j), "\\u003c", "<", -1)
|
|
}
|
|
|
|
// kitchenTaskNoBundle sets up the task to run a recipe via Kitchen, without the
|
|
// recipe bundle.
|
|
func (b *taskBuilder) kitchenTaskNoBundle(recipe string, outputDir string) {
|
|
b.cipd(CIPD_PKG_LUCI_AUTH)
|
|
b.cipd(cipd.MustGetPackage("infra/tools/luci/kitchen/${platform}"))
|
|
b.usesPython()
|
|
b.recipeProp("swarm_out_dir", outputDir)
|
|
if outputDir != OUTPUT_NONE {
|
|
b.output(outputDir)
|
|
}
|
|
python := "cipd_bin_packages/vpython${EXECUTABLE_SUFFIX}"
|
|
b.cmd(python, "-u", "skia/infra/bots/run_recipe.py", "${ISOLATED_OUTDIR}", recipe, b.getRecipeProps(), b.cfg.Project)
|
|
// Most recipes want this isolate; they can override if necessary.
|
|
b.isolate("swarm_recipe.isolate")
|
|
b.timeout(time.Hour)
|
|
b.addToPATH("cipd_bin_packages", "cipd_bin_packages/bin")
|
|
b.Spec.ExtraTags = map[string]string{
|
|
"log_location": fmt.Sprintf("logdog://logs.chromium.org/%s/${SWARMING_TASK_ID}/+/annotations", b.cfg.Project),
|
|
}
|
|
|
|
// Attempts.
|
|
if !b.role("Build", "Upload") && b.extraConfig("ASAN", "MSAN", "TSAN", "Valgrind") {
|
|
// Sanitizers often find non-deterministic issues that retries would hide.
|
|
b.attempts(1)
|
|
} else {
|
|
// Retry by default to hide random bot/hardware failures.
|
|
b.attempts(2)
|
|
}
|
|
}
|
|
|
|
// kitchenTask sets up the task to run a recipe via Kitchen.
|
|
func (b *taskBuilder) kitchenTask(recipe string, outputDir string) {
|
|
b.kitchenTaskNoBundle(recipe, outputDir)
|
|
b.dep(b.bundleRecipes())
|
|
}
|
|
|
|
// internalHardwareLabel returns the internal ID for the bot, if any.
|
|
func (b *taskBuilder) internalHardwareLabel() *int {
|
|
if b.cfg.InternalHardwareLabel != nil {
|
|
return b.cfg.InternalHardwareLabel(b.parts)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// linuxGceDimensions adds the Swarming bot dimensions for Linux GCE instances.
|
|
func (b *taskBuilder) linuxGceDimensions(machineType string) {
|
|
b.dimension(
|
|
// Specify CPU to avoid running builds on bots with a more unique CPU.
|
|
"cpu:x86-64-Haswell_GCE",
|
|
"gpu:none",
|
|
// Currently all Linux GCE tasks run on 16-CPU machines.
|
|
fmt.Sprintf("machine_type:%s", machineType),
|
|
fmt.Sprintf("os:%s", DEFAULT_OS_LINUX_GCE),
|
|
fmt.Sprintf("pool:%s", b.cfg.Pool),
|
|
)
|
|
}
|
|
|
|
// deriveCompileTaskName returns the name of a compile task based on the given
|
|
// job name.
|
|
func (b *jobBuilder) deriveCompileTaskName() string {
|
|
if b.role("Test", "Perf", "FM") {
|
|
task_os := b.parts["os"]
|
|
ec := []string{}
|
|
if val := b.parts["extra_config"]; val != "" {
|
|
ec = strings.Split(val, "_")
|
|
ignore := []string{
|
|
"Skpbench", "AbandonGpuContext", "PreAbandonGpuContext", "Valgrind",
|
|
"ReleaseAndAbandonGpuContext", "CCPR", "FSAA", "FAAA", "FDAA", "NativeFonts", "GDI",
|
|
"NoGPUThreads", "ProcDump", "DDL1", "DDL3", "OOPRDDL", "T8888",
|
|
"DDLTotal", "DDLRecord", "9x9", "BonusConfigs", "SkottieTracing", "SkottieWASM",
|
|
"GpuTess", "NonNVPR", "Mskp", "Docker", "PDF", "SkVM", "Puppeteer",
|
|
"SkottieFrames", "RenderSKP", "CanvasPerf", "AllPathsVolatile", "WebGL2",
|
|
"ReduceOpsTaskSplitting"}
|
|
keep := make([]string, 0, len(ec))
|
|
for _, part := range ec {
|
|
if !In(part, ignore) {
|
|
keep = append(keep, part)
|
|
}
|
|
}
|
|
ec = keep
|
|
}
|
|
if b.os("Android") {
|
|
if !In("Android", ec) {
|
|
ec = append([]string{"Android"}, ec...)
|
|
}
|
|
task_os = COMPILE_TASK_NAME_OS_LINUX
|
|
} else if b.os("ChromeOS") {
|
|
ec = append([]string{"Chromebook", "GLES"}, ec...)
|
|
task_os = COMPILE_TASK_NAME_OS_LINUX
|
|
if b.model("Pixelbook") {
|
|
task_os = COMPILE_TASK_NAME_OS_LINUX_OLD
|
|
ec = append(ec, "Docker")
|
|
}
|
|
} else if b.os("iOS") {
|
|
ec = append([]string{task_os}, ec...)
|
|
task_os = "Mac"
|
|
// iPhone11 requires xcode 11.4.1 which requires >10.15.2.
|
|
if b.parts["model"] == "iPhone11" {
|
|
// TODO(westont): Make this Mac10.15.7. We only have a single 10.15.5.
|
|
task_os = "Mac10.15.5"
|
|
}
|
|
} else if b.matchOs("Win") {
|
|
task_os = "Win"
|
|
} else if b.compiler("GCC") {
|
|
// GCC compiles are now on a Docker container. We use the same OS and
|
|
// version to compile as to test.
|
|
ec = append(ec, "Docker")
|
|
} else if b.matchOs("Ubuntu", "Debian") {
|
|
task_os = COMPILE_TASK_NAME_OS_LINUX
|
|
} else if b.matchOs("Mac") {
|
|
task_os = "Mac"
|
|
if b.extraConfig("CommandBuffer") {
|
|
task_os = "Mac10.15.7"
|
|
}
|
|
}
|
|
jobNameMap := map[string]string{
|
|
"role": "Build",
|
|
"os": task_os,
|
|
"compiler": b.parts["compiler"],
|
|
"target_arch": b.parts["arch"],
|
|
"configuration": b.parts["configuration"],
|
|
}
|
|
if b.extraConfig("PathKit") {
|
|
ec = []string{"PathKit"}
|
|
}
|
|
if b.extraConfig("CanvasKit", "SkottieWASM", "Puppeteer") {
|
|
if b.cpu() {
|
|
ec = []string{"CanvasKit_CPU"}
|
|
} else {
|
|
ec = []string{"CanvasKit"}
|
|
}
|
|
|
|
}
|
|
if len(ec) > 0 {
|
|
jobNameMap["extra_config"] = strings.Join(ec, "_")
|
|
}
|
|
name, err := b.jobNameSchema.MakeJobName(jobNameMap)
|
|
if err != nil {
|
|
log.Fatal(err)
|
|
}
|
|
return name
|
|
} else if b.parts["role"] == "BuildStats" {
|
|
return strings.Replace(b.Name, "BuildStats", "Build", 1)
|
|
} else {
|
|
return b.Name
|
|
}
|
|
}
|
|
|
|
// swarmDimensions generates swarming bot dimensions for the given task.
|
|
func (b *taskBuilder) swarmDimensions() {
|
|
if b.cfg.SwarmDimensions != nil {
|
|
dims := b.cfg.SwarmDimensions(b.parts)
|
|
if dims != nil {
|
|
b.dimension(dims...)
|
|
return
|
|
}
|
|
}
|
|
b.defaultSwarmDimensions()
|
|
}
|
|
|
|
// defaultSwarmDimensions generates default swarming bot dimensions for the given task.
|
|
func (b *taskBuilder) defaultSwarmDimensions() {
|
|
d := map[string]string{
|
|
"pool": b.cfg.Pool,
|
|
}
|
|
if os, ok := b.parts["os"]; ok {
|
|
d["os"], ok = map[string]string{
|
|
"Android": "Android",
|
|
"ChromeOS": "ChromeOS",
|
|
"Debian9": DEFAULT_OS_LINUX_GCE, // Runs in Deb9 Docker.
|
|
"Debian10": DEFAULT_OS_LINUX_GCE,
|
|
"Mac": DEFAULT_OS_MAC,
|
|
"Mac10.13": "Mac-10.13.6",
|
|
"Mac10.14": "Mac-10.14.3",
|
|
"Mac10.15": "Mac-10.15.1",
|
|
"Mac10.15.5": "Mac-10.15.5", // We have some builders at 10.15.5 to run Xcode 11.4.1
|
|
"Mac10.15.7": "Mac-10.15.7", // We have some Golo machines at 10.15.7 to run Xcode 12.2
|
|
"Ubuntu18": "Ubuntu-18.04",
|
|
"Win": DEFAULT_OS_WIN,
|
|
"Win10": "Windows-10-18363",
|
|
"Win2019": DEFAULT_OS_WIN,
|
|
"Win7": "Windows-7-SP1",
|
|
"Win8": "Windows-8.1-SP0",
|
|
"iOS": "iOS-13.3.1",
|
|
}[os]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in OS mapping.", os)
|
|
}
|
|
if os == "Win10" && b.parts["model"] == "Golo" {
|
|
// ChOps-owned machines have Windows 10 v1709.
|
|
d["os"] = "Windows-10-16299"
|
|
}
|
|
if os == "Mac10.14" && b.parts["model"] == "VMware7.1" {
|
|
// ChOps VMs are at a newer version of MacOS.
|
|
d["os"] = "Mac-10.14.6"
|
|
}
|
|
if os == "Mac10.15" && b.parts["model"] == "VMware7.1" {
|
|
// ChOps VMs are at a newer version of MacOS.
|
|
d["os"] = "Mac-10.15.7"
|
|
}
|
|
if b.parts["model"] == "LenovoYogaC630" {
|
|
// This is currently a unique snowflake.
|
|
d["os"] = "Windows-10"
|
|
}
|
|
if b.parts["model"] == "iPhone6" {
|
|
// This is the latest iOS that supports iPhone6.
|
|
d["os"] = "iOS-12.4.5"
|
|
}
|
|
if b.parts["model"] == "iPhone11" {
|
|
d["os"] = "iOS-13.6"
|
|
}
|
|
if b.parts["model"] == "iPadPro" {
|
|
d["os"] = "iOS-13.6"
|
|
}
|
|
} else {
|
|
d["os"] = DEFAULT_OS_DEBIAN
|
|
}
|
|
if b.role("Test", "Perf") {
|
|
if b.os("Android") {
|
|
// For Android, the device type is a better dimension
|
|
// than CPU or GPU.
|
|
deviceInfo, ok := map[string][]string{
|
|
"AndroidOne": {"sprout", "MOB30Q"},
|
|
"GalaxyS6": {"zerofltetmo", "NRD90M_G920TUVS6FRC1"},
|
|
"GalaxyS7_G930FD": {"herolte", "R16NW_G930FXXS2ERH6"}, // This is Oreo.
|
|
"GalaxyS9": {"starlte", "QP1A.190711.020"}, // This is Android10.
|
|
"GalaxyS20": {"exynos990", "QP1A.190711.020"},
|
|
"MotoG4": {"athene", "NPJS25.93-14.7-8"},
|
|
"NVIDIA_Shield": {"foster", "OPR6.170623.010_3507953_1441.7411"},
|
|
"Nexus5": {"hammerhead", "M4B30Z_3437181"},
|
|
"Nexus5x": {"bullhead", "OPR6.170623.023"},
|
|
"Nexus7": {"grouper", "LMY47V_1836172"}, // 2012 Nexus 7
|
|
"P30": {"HWELE", "HUAWEIELE-L29"},
|
|
"Pixel": {"sailfish", "PPR1.180610.009"},
|
|
"Pixel2XL": {"taimen", "PPR1.180610.009"},
|
|
"Pixel3": {"blueline", "PQ1A.190105.004"},
|
|
"Pixel3a": {"sargo", "QP1A.190711.020"},
|
|
"Pixel4": {"flame", "RPB2.200611.009"}, // R Preview
|
|
"Pixel4XL": {"coral", "QD1A.190821.011.C4"},
|
|
"TecnoSpark3Pro": {"TECNO-KB8", "PPR1.180610.011"},
|
|
}[b.parts["model"]]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in Android mapping.", b.parts["model"])
|
|
}
|
|
d["device_type"] = deviceInfo[0]
|
|
d["device_os"] = deviceInfo[1]
|
|
} else if b.os("iOS") {
|
|
device, ok := map[string]string{
|
|
"iPadMini4": "iPad5,1",
|
|
"iPhone6": "iPhone7,2",
|
|
"iPhone7": "iPhone9,1",
|
|
"iPhone8": "iPhone10,1",
|
|
"iPhone11": "iPhone12,1",
|
|
"iPadPro": "iPad6,3",
|
|
}[b.parts["model"]]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in iOS mapping.", b.parts["model"])
|
|
}
|
|
d["device_type"] = device
|
|
// Temporarily use this dimension to ensure we only use the new libimobiledevice, since the
|
|
// old version won't work with current recipes.
|
|
d["libimobiledevice"] = "1582155448"
|
|
} else if b.extraConfig("SKQP") && b.cpu("Emulator") {
|
|
if !b.model("NUC7i5BNK") || d["os"] != DEFAULT_OS_DEBIAN {
|
|
log.Fatalf("Please update defaultSwarmDimensions for SKQP::Emulator %s %s.", b.parts["os"], b.parts["model"])
|
|
}
|
|
d["cpu"] = "x86-64-i5-7260U"
|
|
d["os"] = DEFAULT_OS_DEBIAN
|
|
// KVM means Kernel-based Virtual Machine, that is, can this vm virtualize commands
|
|
// For us, this means, can we run an x86 android emulator on it.
|
|
// kjlubick tried running this on GCE, but it was a bit too slow on the large install.
|
|
// So, we run on bare metal machines in the Skolo (that should also have KVM).
|
|
d["kvm"] = "1"
|
|
d["docker_installed"] = "true"
|
|
} else if b.cpu() || b.extraConfig("CanvasKit", "Docker", "SwiftShader") {
|
|
modelMapping, ok := map[string]map[string]string{
|
|
"AVX": {
|
|
"VMware7.1": "x86-64-E5-2697_v2",
|
|
},
|
|
"AVX2": {
|
|
"GCE": "x86-64-Haswell_GCE",
|
|
"MacBookAir7.2": "x86-64-i5-5350U",
|
|
"MacBookPro11.5": "x86-64-i7-4870HQ",
|
|
"NUC5i7RYH": "x86-64-i7-5557U",
|
|
},
|
|
"AVX512": {
|
|
"GCE": "x86-64-Skylake_GCE",
|
|
"Golo": "Intel64_Family_6_Model_85_Stepping_7__GenuineIntel",
|
|
},
|
|
"Rome": {
|
|
"GCE": "x86-64-AMD_Rome_GCE",
|
|
},
|
|
"Snapdragon850": {
|
|
"LenovoYogaC630": "arm64-64-Snapdragon850",
|
|
},
|
|
"SwiftShader": {
|
|
"GCE": "x86-64-Haswell_GCE",
|
|
},
|
|
}[b.parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in CPU mapping.", b.parts["cpu_or_gpu_value"])
|
|
}
|
|
cpu, ok := modelMapping[b.parts["model"]]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in %q model mapping.", b.parts["model"], b.parts["cpu_or_gpu_value"])
|
|
}
|
|
d["cpu"] = cpu
|
|
if b.model("GCE") && b.matchOs("Debian") {
|
|
d["os"] = DEFAULT_OS_LINUX_GCE
|
|
}
|
|
if b.model("GCE") && d["cpu"] == "x86-64-Haswell_GCE" {
|
|
d["machine_type"] = MACHINE_TYPE_MEDIUM
|
|
}
|
|
} else {
|
|
if b.matchOs("Win") {
|
|
gpu, ok := map[string]string{
|
|
// At some point this might use the device ID, but for now it's like Chromebooks.
|
|
"Adreno630": "Adreno630",
|
|
"GT610": "10de:104a-23.21.13.9101",
|
|
"GTX660": "10de:11c0-26.21.14.4120",
|
|
"GTX960": "10de:1401-26.21.14.4120",
|
|
"IntelHD4400": "8086:0a16-20.19.15.4963",
|
|
"IntelIris540": "8086:1926-26.20.100.7463",
|
|
"IntelIris6100": "8086:162b-20.19.15.4963",
|
|
"IntelIris655": "8086:3ea5-26.20.100.7463",
|
|
"RadeonHD7770": "1002:683d-26.20.13031.18002",
|
|
"RadeonR9M470X": "1002:6646-26.20.13031.18002",
|
|
"QuadroP400": "10de:1cb3-25.21.14.1678",
|
|
}[b.parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in Win GPU mapping.", b.parts["cpu_or_gpu_value"])
|
|
}
|
|
d["gpu"] = gpu
|
|
} else if b.isLinux() {
|
|
gpu, ok := map[string]string{
|
|
// Intel drivers come from CIPD, so no need to specify the version here.
|
|
"IntelBayTrail": "8086:0f31",
|
|
"IntelHD2000": "8086:0102",
|
|
"IntelHD405": "8086:22b1",
|
|
"IntelIris640": "8086:5926",
|
|
"QuadroP400": "10de:1cb3-430.14",
|
|
}[b.parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in Ubuntu GPU mapping.", b.parts["cpu_or_gpu_value"])
|
|
}
|
|
d["gpu"] = gpu
|
|
} else if b.matchOs("Mac") {
|
|
gpu, ok := map[string]string{
|
|
"IntelHD6000": "8086:1626",
|
|
"IntelHD615": "8086:591e",
|
|
"IntelIris5100": "8086:0a2e",
|
|
"RadeonHD8870M": "1002:6821-4.0.20-3.2.8",
|
|
}[b.parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in Mac GPU mapping.", b.parts["cpu_or_gpu_value"])
|
|
}
|
|
d["gpu"] = gpu
|
|
// Yuck. We have two different types of MacMini7,1 with the same GPU but different CPUs.
|
|
if b.gpu("IntelIris5100") {
|
|
// Run all tasks on Golo machines for now.
|
|
d["cpu"] = "x86-64-i7-4578U"
|
|
}
|
|
} else if b.os("ChromeOS") {
|
|
version, ok := map[string]string{
|
|
"MaliT604": "10575.22.0",
|
|
"MaliT764": "10575.22.0",
|
|
"MaliT860": "10575.22.0",
|
|
"PowerVRGX6250": "10575.22.0",
|
|
"TegraK1": "10575.22.0",
|
|
"IntelHDGraphics615": "10575.22.0",
|
|
}[b.parts["cpu_or_gpu_value"]]
|
|
if !ok {
|
|
log.Fatalf("Entry %q not found in ChromeOS GPU mapping.", b.parts["cpu_or_gpu_value"])
|
|
}
|
|
d["gpu"] = b.parts["cpu_or_gpu_value"]
|
|
d["release_version"] = version
|
|
} else {
|
|
log.Fatalf("Unknown GPU mapping for OS %q.", b.parts["os"])
|
|
}
|
|
}
|
|
} else {
|
|
d["gpu"] = "none"
|
|
if d["os"] == DEFAULT_OS_LINUX_GCE {
|
|
if b.extraConfig("CanvasKit", "CMake", "Docker", "PathKit") || b.role("BuildStats") {
|
|
b.linuxGceDimensions(MACHINE_TYPE_MEDIUM)
|
|
return
|
|
}
|
|
// Use many-core machines for Build tasks.
|
|
b.linuxGceDimensions(MACHINE_TYPE_LARGE)
|
|
return
|
|
} else if d["os"] == DEFAULT_OS_WIN {
|
|
// Windows CPU bots.
|
|
d["cpu"] = "x86-64-Haswell_GCE"
|
|
// Use many-core machines for Build tasks.
|
|
d["machine_type"] = MACHINE_TYPE_LARGE
|
|
} else if d["os"] == DEFAULT_OS_MAC {
|
|
// Mac CPU bots.
|
|
d["cpu"] = "x86-64-E5-2697_v2"
|
|
}
|
|
}
|
|
|
|
dims := make([]string, 0, len(d))
|
|
for k, v := range d {
|
|
dims = append(dims, fmt.Sprintf("%s:%s", k, v))
|
|
}
|
|
sort.Strings(dims)
|
|
b.dimension(dims...)
|
|
}
|
|
|
|
// relpath returns the relative path to the given file from the config file.
|
|
func (b *builder) relpath(f string) string {
|
|
target := filepath.Join(b.relpathTargetDir, f)
|
|
rv, err := filepath.Rel(b.relpathBaseDir, target)
|
|
if err != nil {
|
|
log.Fatal(err)
|
|
}
|
|
return rv
|
|
}
|
|
|
|
// bundleRecipes generates the task to bundle and isolate the recipes. Returns
|
|
// the name of the task, which may be added as a dependency.
|
|
func (b *jobBuilder) bundleRecipes() string {
|
|
b.addTask(BUNDLE_RECIPES_NAME, func(b *taskBuilder) {
|
|
b.cipd(specs.CIPD_PKGS_GIT_LINUX_AMD64...)
|
|
b.cipd(specs.CIPD_PKGS_PYTHON_LINUX_AMD64...)
|
|
b.cmd("/bin/bash", "skia/infra/bots/bundle_recipes.sh", specs.PLACEHOLDER_ISOLATED_OUTDIR)
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.addToPATH("cipd_bin_packages", "cipd_bin_packages/bin")
|
|
b.idempotent()
|
|
b.isolate("recipes.isolate")
|
|
})
|
|
return BUNDLE_RECIPES_NAME
|
|
}
|
|
|
|
// buildTaskDrivers generates the task to compile the task driver code to run on
|
|
// all platforms. Returns the name of the task, which may be added as a
|
|
// dependency.
|
|
func (b *jobBuilder) buildTaskDrivers() string {
|
|
b.addTask(BUILD_TASK_DRIVERS_NAME, func(b *taskBuilder) {
|
|
b.usesGo()
|
|
b.cmd("/bin/bash", "skia/infra/bots/build_task_drivers.sh", specs.PLACEHOLDER_ISOLATED_OUTDIR)
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.addToPATH("cipd_bin_packages", "cipd_bin_packages/bin", "go/go/bin")
|
|
b.idempotent()
|
|
b.isolate("task_drivers.isolate")
|
|
})
|
|
return BUILD_TASK_DRIVERS_NAME
|
|
}
|
|
|
|
// updateGoDeps generates the task to update Go dependencies.
|
|
func (b *jobBuilder) updateGoDeps() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.usesGo()
|
|
b.asset("protoc")
|
|
b.cmd(
|
|
"./update_go_deps",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--workdir", ".",
|
|
"--gerrit_project", "skia",
|
|
"--gerrit_url", "https://skia-review.googlesource.com",
|
|
"--repo", specs.PLACEHOLDER_REPO,
|
|
"--revision", specs.PLACEHOLDER_REVISION,
|
|
"--patch_issue", specs.PLACEHOLDER_ISSUE,
|
|
"--patch_set", specs.PLACEHOLDER_PATCHSET,
|
|
"--patch_server", specs.PLACEHOLDER_CODEREVIEW_SERVER,
|
|
"--alsologtostderr",
|
|
)
|
|
b.dep(b.buildTaskDrivers())
|
|
b.linuxGceDimensions(MACHINE_TYPE_MEDIUM)
|
|
b.addToPATH("cipd_bin_packages", "cipd_bin_packages/bin", "go/go/bin")
|
|
b.isolate("empty.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountRecreateSKPs)
|
|
})
|
|
}
|
|
|
|
// createDockerImage creates the specified docker image. Returns the name of the
|
|
// generated task.
|
|
func (b *jobBuilder) createDockerImage(wasm bool) string {
|
|
// First, derive the name of the task.
|
|
imageName := "skia-release"
|
|
taskName := "Housekeeper-PerCommit-CreateDockerImage_Skia_Release"
|
|
if wasm {
|
|
imageName = "skia-wasm-release"
|
|
taskName = "Housekeeper-PerCommit-CreateDockerImage_Skia_WASM_Release"
|
|
}
|
|
imageDir := path.Join("docker", imageName)
|
|
|
|
// Add the task.
|
|
b.addTask(taskName, func(b *taskBuilder) {
|
|
// TODO(borenet): Make this task not use Git.
|
|
b.usesGit()
|
|
b.cmd(
|
|
"./build_push_docker_image",
|
|
"--image_name", fmt.Sprintf("gcr.io/skia-public/%s", imageName),
|
|
"--dockerfile_dir", imageDir,
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--workdir", ".",
|
|
"--gerrit_project", "skia",
|
|
"--gerrit_url", "https://skia-review.googlesource.com",
|
|
"--repo", specs.PLACEHOLDER_REPO,
|
|
"--revision", specs.PLACEHOLDER_REVISION,
|
|
"--patch_issue", specs.PLACEHOLDER_ISSUE,
|
|
"--patch_set", specs.PLACEHOLDER_PATCHSET,
|
|
"--patch_server", specs.PLACEHOLDER_CODEREVIEW_SERVER,
|
|
"--swarm_out_dir", specs.PLACEHOLDER_ISOLATED_OUTDIR,
|
|
"--alsologtostderr",
|
|
)
|
|
b.dep(b.buildTaskDrivers())
|
|
b.addToPATH("cipd_bin_packages", "cipd_bin_packages/bin", "go/go/bin")
|
|
b.isolate("empty.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.linuxGceDimensions(MACHINE_TYPE_MEDIUM)
|
|
b.usesDocker()
|
|
b.cache(CACHES_DOCKER...)
|
|
})
|
|
return taskName
|
|
}
|
|
|
|
// createPushAppsFromSkiaDockerImage creates and pushes docker images of some apps
|
|
// (eg: fiddler, debugger, api) using the skia-release docker image.
|
|
func (b *jobBuilder) createPushAppsFromSkiaDockerImage() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
// TODO(borenet): Make this task not use Git.
|
|
b.usesGit()
|
|
b.cmd(
|
|
"./push_apps_from_skia_image",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--workdir", ".",
|
|
"--gerrit_project", "buildbot",
|
|
"--gerrit_url", "https://skia-review.googlesource.com",
|
|
"--repo", specs.PLACEHOLDER_REPO,
|
|
"--revision", specs.PLACEHOLDER_REVISION,
|
|
"--patch_issue", specs.PLACEHOLDER_ISSUE,
|
|
"--patch_set", specs.PLACEHOLDER_PATCHSET,
|
|
"--patch_server", specs.PLACEHOLDER_CODEREVIEW_SERVER,
|
|
"--alsologtostderr",
|
|
)
|
|
b.dep(b.buildTaskDrivers())
|
|
b.dep(b.createDockerImage(false))
|
|
b.addToPATH("cipd_bin_packages", "cipd_bin_packages/bin", "go/go/bin")
|
|
b.isolate("empty.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.linuxGceDimensions(MACHINE_TYPE_MEDIUM)
|
|
b.usesDocker()
|
|
b.cache(CACHES_DOCKER...)
|
|
})
|
|
}
|
|
|
|
// createPushAppsFromWASMDockerImage creates and pushes docker images of some apps
|
|
// (eg: jsfiddle, skottie, particles) using the skia-wasm-release docker image.
|
|
func (b *jobBuilder) createPushAppsFromWASMDockerImage() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
// TODO(borenet): Make this task not use Git.
|
|
b.usesGit()
|
|
b.cmd(
|
|
"./push_apps_from_wasm_image",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--workdir", ".",
|
|
"--gerrit_project", "buildbot",
|
|
"--gerrit_url", "https://skia-review.googlesource.com",
|
|
"--repo", specs.PLACEHOLDER_REPO,
|
|
"--revision", specs.PLACEHOLDER_REVISION,
|
|
"--patch_issue", specs.PLACEHOLDER_ISSUE,
|
|
"--patch_set", specs.PLACEHOLDER_PATCHSET,
|
|
"--patch_server", specs.PLACEHOLDER_CODEREVIEW_SERVER,
|
|
"--alsologtostderr",
|
|
)
|
|
b.dep(b.buildTaskDrivers())
|
|
b.dep(b.createDockerImage(true))
|
|
b.addToPATH("cipd_bin_packages", "cipd_bin_packages/bin", "go/go/bin")
|
|
b.isolate("empty.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.linuxGceDimensions(MACHINE_TYPE_MEDIUM)
|
|
b.usesDocker()
|
|
b.cache(CACHES_DOCKER...)
|
|
})
|
|
}
|
|
|
|
var iosRegex = regexp.MustCompile(`os:iOS-(.*)`)
|
|
|
|
func (b *taskBuilder) maybeAddIosDevImage() {
|
|
for _, dim := range b.Spec.Dimensions {
|
|
if m := iosRegex.FindStringSubmatch(dim); len(m) >= 2 {
|
|
var asset string
|
|
switch m[1] {
|
|
// Other patch versions can be added to the same case.
|
|
case "11.4.1":
|
|
asset = "ios-dev-image-11.4"
|
|
case "12.4.5":
|
|
asset = "ios-dev-image-12.4"
|
|
case "13.3.1":
|
|
asset = "ios-dev-image-13.3"
|
|
case "13.4.1":
|
|
asset = "ios-dev-image-13.4"
|
|
case "13.5.1":
|
|
asset = "ios-dev-image-13.5"
|
|
case "13.6":
|
|
asset = "ios-dev-image-13.6"
|
|
default:
|
|
log.Fatalf("Unable to determine correct ios-dev-image asset for %s. If %s is a new iOS release, you must add a CIPD package containing the corresponding iOS dev image; see ios-dev-image-11.4 for an example.", b.Name, m[1])
|
|
}
|
|
b.asset(asset)
|
|
break
|
|
} else if strings.Contains(dim, "iOS") {
|
|
log.Fatalf("Must specify iOS version for %s to obtain correct dev image; os dimension is missing version: %s", b.Name, dim)
|
|
}
|
|
}
|
|
}
|
|
|
|
// compile generates a compile task. Returns the name of the compile task.
|
|
func (b *jobBuilder) compile() string {
|
|
name := b.deriveCompileTaskName()
|
|
if b.extraConfig("WasmGMTests") {
|
|
b.compileWasmGMTests(name)
|
|
} else {
|
|
b.addTask(name, func(b *taskBuilder) {
|
|
recipe := "compile"
|
|
isolate := "compile.isolate"
|
|
if b.extraConfig("NoDEPS", "CMake", "CommandBuffer", "Flutter", "SKQP") {
|
|
recipe = "sync_and_compile"
|
|
isolate = "swarm_recipe.isolate"
|
|
b.recipeProps(EXTRA_PROPS)
|
|
b.usesGit()
|
|
if !b.extraConfig("NoDEPS") {
|
|
b.cache(CACHES_WORKDIR...)
|
|
}
|
|
} else {
|
|
b.idempotent()
|
|
}
|
|
b.kitchenTask(recipe, OUTPUT_BUILD)
|
|
b.isolate(isolate)
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.swarmDimensions()
|
|
if b.extraConfig("Docker", "LottieWeb", "SKQP", "CMake") || b.compiler("EMCC") {
|
|
b.usesDocker()
|
|
b.cache(CACHES_DOCKER...)
|
|
}
|
|
|
|
// Android bots require a toolchain.
|
|
if b.extraConfig("Android") {
|
|
if b.matchOs("Mac") {
|
|
b.asset("android_ndk_darwin")
|
|
} else if b.matchOs("Win") {
|
|
pkg := b.MustGetCipdPackageFromAsset("android_ndk_windows")
|
|
pkg.Path = "n"
|
|
b.cipd(pkg)
|
|
} else if !b.extraConfig("SKQP") {
|
|
b.asset("android_ndk_linux")
|
|
}
|
|
} else if b.extraConfig("Chromebook") {
|
|
b.asset("clang_linux")
|
|
if b.arch("x86_64") {
|
|
b.asset("chromebook_x86_64_gles")
|
|
} else if b.arch("arm") {
|
|
b.asset("armhf_sysroot")
|
|
b.asset("chromebook_arm_gles")
|
|
}
|
|
} else if b.isLinux() {
|
|
if b.compiler("Clang") {
|
|
b.asset("clang_linux")
|
|
}
|
|
if b.extraConfig("SwiftShader") {
|
|
b.asset("cmake_linux")
|
|
}
|
|
if b.extraConfig("OpenCL") {
|
|
b.asset("opencl_headers", "opencl_ocl_icd_linux")
|
|
}
|
|
b.asset("ccache_linux")
|
|
b.usesCCache()
|
|
} else if b.matchOs("Win") {
|
|
b.asset("win_toolchain")
|
|
if b.compiler("Clang") {
|
|
b.asset("clang_win")
|
|
}
|
|
if b.extraConfig("OpenCL") {
|
|
b.asset("opencl_headers")
|
|
}
|
|
} else if b.matchOs("Mac") {
|
|
b.cipd(CIPD_PKGS_XCODE...)
|
|
b.Spec.Caches = append(b.Spec.Caches, &specs.Cache{
|
|
Name: "xcode",
|
|
Path: "cache/Xcode.app",
|
|
})
|
|
b.asset("ccache_mac")
|
|
b.usesCCache()
|
|
if b.extraConfig("CommandBuffer") {
|
|
b.timeout(2 * time.Hour)
|
|
}
|
|
if b.extraConfig("iOS") {
|
|
b.asset("provisioning_profile_ios")
|
|
}
|
|
}
|
|
})
|
|
}
|
|
|
|
// All compile tasks are runnable as their own Job. Assert that the Job
|
|
// is listed in jobs.
|
|
if !In(name, b.jobs) {
|
|
log.Fatalf("Job %q is missing from the jobs list! Derived from: %q", name, b.Name)
|
|
}
|
|
|
|
return name
|
|
}
|
|
|
|
// recreateSKPs generates a RecreateSKPs task.
|
|
func (b *jobBuilder) recreateSKPs() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.recipeProps(EXTRA_PROPS)
|
|
b.kitchenTask("recreate_skps", OUTPUT_NONE)
|
|
b.serviceAccount(b.cfg.ServiceAccountRecreateSKPs)
|
|
b.dimension(
|
|
"pool:SkiaCT",
|
|
fmt.Sprintf("os:%s", DEFAULT_OS_LINUX_GCE),
|
|
)
|
|
b.usesGo()
|
|
b.cache(CACHES_WORKDIR...)
|
|
b.timeout(4 * time.Hour)
|
|
})
|
|
}
|
|
|
|
// checkGeneratedFiles verifies that no generated SKSL files have been edited
|
|
// by hand.
|
|
func (b *jobBuilder) checkGeneratedFiles() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.recipeProps(EXTRA_PROPS)
|
|
b.kitchenTask("check_generated_files", OUTPUT_NONE)
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.linuxGceDimensions(MACHINE_TYPE_LARGE)
|
|
b.usesGo()
|
|
b.asset("clang_linux")
|
|
b.asset("ccache_linux")
|
|
b.usesCCache()
|
|
b.cache(CACHES_WORKDIR...)
|
|
})
|
|
}
|
|
|
|
// checkGnToBp verifies that the gn_to_bp.py script continues to work.
|
|
func (b *jobBuilder) checkGnToBp() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.isolate("compile.isolate")
|
|
b.dep(b.buildTaskDrivers())
|
|
b.cmd("./run_gn_to_bp",
|
|
"--local=false",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--alsologtostderr")
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.usesPython()
|
|
b.serviceAccount(b.cfg.ServiceAccountHousekeeper)
|
|
})
|
|
}
|
|
|
|
// housekeeper generates a Housekeeper task.
|
|
func (b *jobBuilder) housekeeper() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.recipeProps(EXTRA_PROPS)
|
|
b.kitchenTask("housekeeper", OUTPUT_NONE)
|
|
b.serviceAccount(b.cfg.ServiceAccountHousekeeper)
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.usesGit()
|
|
b.cache(CACHES_WORKDIR...)
|
|
})
|
|
}
|
|
|
|
// g3FrameworkCanary generates a G3 Framework Canary task. Returns
|
|
// the name of the last task in the generated chain of tasks, which the Job
|
|
// should add as a dependency.
|
|
func (b *jobBuilder) g3FrameworkCanary() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.isolate("empty.isolate")
|
|
b.dep(b.buildTaskDrivers())
|
|
b.cmd("./g3_canary",
|
|
"--local=false",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--repo", specs.PLACEHOLDER_REPO,
|
|
"--revision", specs.PLACEHOLDER_REVISION,
|
|
"--patch_issue", specs.PLACEHOLDER_ISSUE,
|
|
"--patch_set", specs.PLACEHOLDER_PATCHSET,
|
|
"--patch_server", specs.PLACEHOLDER_CODEREVIEW_SERVER,
|
|
"--alsologtostderr")
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.cipd(CIPD_PKG_LUCI_AUTH)
|
|
b.serviceAccount("skia-g3-framework-compile@skia-swarming-bots.iam.gserviceaccount.com")
|
|
b.timeout(3 * time.Hour)
|
|
b.attempts(1)
|
|
})
|
|
}
|
|
|
|
// infra generates an infra_tests task.
|
|
func (b *jobBuilder) infra() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
if b.matchOs("Win") || b.matchExtraConfig("Win") {
|
|
b.dimension(
|
|
// Specify CPU to avoid running builds on bots with a more unique CPU.
|
|
"cpu:x86-64-Haswell_GCE",
|
|
"gpu:none",
|
|
fmt.Sprintf("machine_type:%s", MACHINE_TYPE_MEDIUM), // We don't have any small Windows instances.
|
|
fmt.Sprintf("os:%s", DEFAULT_OS_WIN),
|
|
fmt.Sprintf("pool:%s", b.cfg.Pool),
|
|
)
|
|
} else {
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
}
|
|
b.recipeProp("repository", specs.PLACEHOLDER_REPO)
|
|
b.kitchenTask("infra", OUTPUT_NONE)
|
|
b.isolate("infra_tests.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.cipd(specs.CIPD_PKGS_GSUTIL...)
|
|
b.idempotent()
|
|
// Repos which call into Skia's gen_tasks.go should define their own
|
|
// infra_tests.isolate and therefore should not use relpath().
|
|
b.Spec.Isolate = "infra_tests.isolate"
|
|
b.usesGo()
|
|
})
|
|
}
|
|
|
|
// buildstats generates a builtstats task, which compiles code and generates
|
|
// statistics about the build.
|
|
func (b *jobBuilder) buildstats() {
|
|
compileTaskName := b.compile()
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.recipeProps(EXTRA_PROPS)
|
|
b.kitchenTask("compute_buildstats", OUTPUT_PERF)
|
|
b.dep(compileTaskName)
|
|
b.asset("bloaty")
|
|
b.linuxGceDimensions(MACHINE_TYPE_MEDIUM)
|
|
b.usesDocker()
|
|
b.usesGit()
|
|
b.cache(CACHES_WORKDIR...)
|
|
})
|
|
// Upload release results (for tracking in perf)
|
|
// We have some jobs that are FYI (e.g. Debug-CanvasKit, tree-map generator)
|
|
if b.release() && !b.arch("x86_64") {
|
|
uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, b.jobNameSchema.Sep, b.Name)
|
|
depName := b.Name
|
|
b.addTask(uploadName, func(b *taskBuilder) {
|
|
b.recipeProp("gs_bucket", b.cfg.GsBucketNano)
|
|
b.recipeProps(EXTRA_PROPS)
|
|
// TODO(borenet): I'm not sure why the upload task is
|
|
// using the BuildStats task name, but I've done this
|
|
// to maintain existing behavior.
|
|
b.Name = depName
|
|
b.kitchenTask("upload_buildstats_results", OUTPUT_NONE)
|
|
b.Name = uploadName
|
|
b.serviceAccount(b.cfg.ServiceAccountUploadNano)
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.cipd(specs.CIPD_PKGS_GSUTIL...)
|
|
b.dep(depName)
|
|
})
|
|
}
|
|
}
|
|
|
|
// doUpload indicates whether the given Job should upload its results.
|
|
func (b *jobBuilder) doUpload() bool {
|
|
for _, s := range b.cfg.NoUpload {
|
|
m, err := regexp.MatchString(s, b.Name)
|
|
if err != nil {
|
|
log.Fatal(err)
|
|
}
|
|
if m {
|
|
return false
|
|
}
|
|
}
|
|
return true
|
|
}
|
|
|
|
// commonTestPerfAssets adds the assets needed by Test and Perf tasks.
|
|
func (b *taskBuilder) commonTestPerfAssets() {
|
|
// Docker-based tests don't need the standard CIPD assets
|
|
if b.extraConfig("CanvasKit", "PathKit") || (b.role("Test") && b.extraConfig("LottieWeb")) {
|
|
return
|
|
}
|
|
if b.extraConfig("Skpbench") {
|
|
// Skpbench only needs skps
|
|
b.asset("skp", "mskp")
|
|
} else if b.os("Android", "ChromeOS", "iOS") {
|
|
b.asset("skp", "svg", "skimage")
|
|
} else {
|
|
// for desktop machines
|
|
b.asset("skimage", "skp", "svg")
|
|
}
|
|
|
|
if b.isLinux() && b.matchExtraConfig("SAN") {
|
|
b.asset("clang_linux")
|
|
}
|
|
|
|
if b.isLinux() {
|
|
if b.extraConfig("Vulkan") {
|
|
b.asset("linux_vulkan_sdk")
|
|
}
|
|
if b.matchGpu("Intel") {
|
|
b.asset("mesa_intel_driver_linux")
|
|
}
|
|
if b.extraConfig("OpenCL") {
|
|
b.asset("opencl_ocl_icd_linux", "opencl_intel_neo_linux")
|
|
}
|
|
}
|
|
if b.matchOs("Win") && b.extraConfig("ProcDump") {
|
|
b.asset("procdump_win")
|
|
}
|
|
}
|
|
|
|
// dm generates a Test task using dm.
|
|
func (b *jobBuilder) dm() {
|
|
compileTaskName := ""
|
|
// LottieWeb doesn't require anything in Skia to be compiled.
|
|
if !b.extraConfig("LottieWeb") {
|
|
compileTaskName = b.compile()
|
|
}
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
isolate := "test_skia_bundled.isolate"
|
|
recipe := "test"
|
|
if b.extraConfig("SKQP") {
|
|
isolate = "skqp.isolate"
|
|
recipe = "skqp_test"
|
|
if b.cpu("Emulator") {
|
|
recipe = "test_skqp_emulator"
|
|
}
|
|
} else if b.extraConfig("OpenCL") {
|
|
// TODO(dogben): Longer term we may not want this to be called a "Test" task, but until we start
|
|
// running hs_bench or kx, it will be easier to fit into the current job name schema.
|
|
recipe = "compute_test"
|
|
} else if b.extraConfig("PathKit") {
|
|
isolate = "pathkit.isolate"
|
|
recipe = "test_pathkit"
|
|
} else if b.extraConfig("CanvasKit") {
|
|
isolate = "canvaskit.isolate"
|
|
recipe = "test_canvaskit"
|
|
} else if b.extraConfig("LottieWeb") {
|
|
// lottie_ci.isolate differs from lottie_web.isolate in that it includes more of the files,
|
|
// especially those brought in via DEPS in the lottie-ci repo. The main difference between
|
|
// Perf.+LottieWeb and Test.+LottieWeb is that the former pulls in the lottie build via
|
|
// npm and the latter always tests at lottie's ToT.
|
|
isolate = "lottie_ci.isolate"
|
|
recipe = "test_lottie_web"
|
|
}
|
|
b.recipeProp("gold_hashes_url", b.cfg.GoldHashesURL)
|
|
b.recipeProps(EXTRA_PROPS)
|
|
iid := b.internalHardwareLabel()
|
|
iidStr := ""
|
|
if iid != nil {
|
|
iidStr = strconv.Itoa(*iid)
|
|
}
|
|
if recipe == "test" {
|
|
b.dmFlags(iidStr)
|
|
}
|
|
b.kitchenTask(recipe, OUTPUT_TEST)
|
|
b.isolate(isolate)
|
|
b.swarmDimensions()
|
|
if b.extraConfig("CanvasKit", "Docker", "LottieWeb", "PathKit", "SKQP") {
|
|
b.usesDocker()
|
|
}
|
|
if compileTaskName != "" {
|
|
b.dep(compileTaskName)
|
|
}
|
|
if b.os("Android") && b.extraConfig("ASAN") {
|
|
b.asset("android_ndk_linux")
|
|
}
|
|
b.commonTestPerfAssets()
|
|
if b.matchExtraConfig("Lottie") {
|
|
b.asset("lottie-samples")
|
|
}
|
|
if b.extraConfig("SKQP") {
|
|
if !b.cpu("Emulator") {
|
|
b.asset("gcloud_linux")
|
|
}
|
|
}
|
|
b.expiration(20 * time.Hour)
|
|
|
|
b.timeout(4 * time.Hour)
|
|
if b.extraConfig("Valgrind") {
|
|
b.timeout(9 * time.Hour)
|
|
b.expiration(48 * time.Hour)
|
|
b.asset("valgrind")
|
|
// Since Valgrind runs on the same bots as the CQ, we restrict Valgrind to a subset of the bots
|
|
// to ensure there are always bots free for CQ tasks.
|
|
b.dimension("valgrind:1")
|
|
} else if b.extraConfig("MSAN") {
|
|
b.timeout(9 * time.Hour)
|
|
} else if b.arch("x86") && b.debug() {
|
|
// skia:6737
|
|
b.timeout(6 * time.Hour)
|
|
}
|
|
b.maybeAddIosDevImage()
|
|
})
|
|
|
|
// Upload results if necessary. TODO(kjlubick): If we do coverage analysis at the same
|
|
// time as normal tests (which would be nice), cfg.json needs to have Coverage removed.
|
|
if b.doUpload() {
|
|
uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, b.jobNameSchema.Sep, b.Name)
|
|
depName := b.Name
|
|
b.addTask(uploadName, func(b *taskBuilder) {
|
|
b.recipeProp("gs_bucket", b.cfg.GsBucketGm)
|
|
b.recipeProps(EXTRA_PROPS)
|
|
// TODO(borenet): I'm not sure why the upload task is
|
|
// using the Test task name, but I've done this
|
|
// to maintain existing behavior.
|
|
b.Name = depName
|
|
b.kitchenTask("upload_dm_results", OUTPUT_NONE)
|
|
b.Name = uploadName
|
|
b.serviceAccount(b.cfg.ServiceAccountUploadGM)
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.cipd(specs.CIPD_PKGS_GSUTIL...)
|
|
b.dep(depName)
|
|
})
|
|
}
|
|
}
|
|
|
|
func (b *jobBuilder) fm() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.isolate("test_skia_bundled.isolate")
|
|
b.dep(b.buildTaskDrivers(), b.compile())
|
|
b.cmd("./fm_driver",
|
|
"--local=false",
|
|
"--resources=skia/resources",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"build/fm")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.swarmDimensions()
|
|
b.expiration(15 * time.Minute)
|
|
b.attempts(1)
|
|
})
|
|
}
|
|
|
|
// canary generates a task that uses TaskDrivers to trigger canary manual rolls on autorollers.
|
|
// Canary-G3 does not use this path because it is very different from other autorollers.
|
|
func (b *jobBuilder) canary(rollerName string) {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.isolate("empty.isolate")
|
|
b.dep(b.buildTaskDrivers())
|
|
b.cmd("./canary",
|
|
"--local=false",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--roller_name", rollerName,
|
|
"--repo", specs.PLACEHOLDER_REPO,
|
|
"--revision", specs.PLACEHOLDER_REVISION,
|
|
"--patch_issue", specs.PLACEHOLDER_ISSUE,
|
|
"--patch_set", specs.PLACEHOLDER_PATCHSET,
|
|
"--patch_server", specs.PLACEHOLDER_CODEREVIEW_SERVER,
|
|
"--alsologtostderr")
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.cipd(CIPD_PKG_LUCI_AUTH)
|
|
b.serviceAccount(b.cfg.ServiceAccountCanary)
|
|
b.timeout(3 * time.Hour)
|
|
b.attempts(1)
|
|
})
|
|
}
|
|
|
|
// puppeteer generates a task that uses TaskDrivers combined with a node script and puppeteer to
|
|
// benchmark something using Chromium (e.g. CanvasKit, LottieWeb).
|
|
func (b *jobBuilder) puppeteer() {
|
|
compileTaskName := b.compile()
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.defaultSwarmDimensions()
|
|
b.usesNode()
|
|
b.cipd(CIPD_PKG_LUCI_AUTH)
|
|
b.dep(b.buildTaskDrivers(), compileTaskName)
|
|
b.output(OUTPUT_PERF)
|
|
b.timeout(20 * time.Minute)
|
|
b.isolate("perf_puppeteer.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
|
|
webglversion := "2"
|
|
if b.extraConfig("WebGL1") {
|
|
webglversion = "1"
|
|
}
|
|
|
|
if b.extraConfig("SkottieFrames") {
|
|
b.cmd(
|
|
"./perf_puppeteer_skottie_frames",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--git_hash", specs.PLACEHOLDER_REVISION,
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--canvaskit_bin_path", "./build",
|
|
"--lotties_path", "./lotties_with_assets",
|
|
"--node_bin_path", "./node/node/bin",
|
|
"--benchmark_path", "./tools/perf-canvaskit-puppeteer",
|
|
"--output_path", OUTPUT_PERF,
|
|
"--os_trace", b.parts["os"],
|
|
"--model_trace", b.parts["model"],
|
|
"--cpu_or_gpu_trace", b.parts["cpu_or_gpu"],
|
|
"--cpu_or_gpu_value_trace", b.parts["cpu_or_gpu_value"],
|
|
"--webgl_version", webglversion, // ignore when running with cpu backend
|
|
"--alsologtostderr",
|
|
)
|
|
// This CIPD package was made by hand with the following invocation:
|
|
// cipd create -name skia/internal/lotties_with_assets -in ./lotties/ -tag version:0
|
|
// cipd acl-edit skia/internal/lotties_with_assets -reader group:project-skia-external-task-accounts
|
|
// cipd acl-edit skia/internal/lotties_with_assets -reader user:pool-skia@chromium-swarm.iam.gserviceaccount.com
|
|
// Where lotties is a hand-selected set of lottie animations and (optionally) assets used in
|
|
// them (e.g. fonts, images).
|
|
b.cipd(&specs.CipdPackage{
|
|
Name: "skia/internal/lotties_with_assets",
|
|
Path: "lotties_with_assets",
|
|
Version: "version:0",
|
|
})
|
|
} else if b.extraConfig("RenderSKP") {
|
|
b.cmd(
|
|
"./perf_puppeteer_render_skps",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--git_hash", specs.PLACEHOLDER_REVISION,
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--canvaskit_bin_path", "./build",
|
|
"--skps_path", "./skp",
|
|
"--node_bin_path", "./node/node/bin",
|
|
"--benchmark_path", "./tools/perf-canvaskit-puppeteer",
|
|
"--output_path", OUTPUT_PERF,
|
|
"--os_trace", b.parts["os"],
|
|
"--model_trace", b.parts["model"],
|
|
"--cpu_or_gpu_trace", b.parts["cpu_or_gpu"],
|
|
"--cpu_or_gpu_value_trace", b.parts["cpu_or_gpu_value"],
|
|
"--webgl_version", webglversion,
|
|
"--alsologtostderr",
|
|
)
|
|
b.asset("skp")
|
|
} else if b.extraConfig("CanvasPerf") { // refers to the canvas_perf.js test suite
|
|
b.cmd(
|
|
"./perf_puppeteer_canvas",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--git_hash", specs.PLACEHOLDER_REVISION,
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--canvaskit_bin_path", "./build",
|
|
"--node_bin_path", "./node/node/bin",
|
|
"--benchmark_path", "./tools/perf-canvaskit-puppeteer",
|
|
"--output_path", OUTPUT_PERF,
|
|
"--os_trace", b.parts["os"],
|
|
"--model_trace", b.parts["model"],
|
|
"--cpu_or_gpu_trace", b.parts["cpu_or_gpu"],
|
|
"--cpu_or_gpu_value_trace", b.parts["cpu_or_gpu_value"],
|
|
"--webgl_version", webglversion,
|
|
"--alsologtostderr",
|
|
)
|
|
b.asset("skp")
|
|
}
|
|
|
|
})
|
|
|
|
// Upload results to Perf after.
|
|
// TODO(kjlubick,borenet) deduplicate this with the logic in perf().
|
|
uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, b.jobNameSchema.Sep, b.Name)
|
|
depName := b.Name
|
|
b.addTask(uploadName, func(b *taskBuilder) {
|
|
b.recipeProp("gs_bucket", b.cfg.GsBucketNano)
|
|
b.recipeProps(EXTRA_PROPS)
|
|
// TODO(borenet): I'm not sure why the upload task is
|
|
// using the Perf task name, but I've done this to
|
|
// maintain existing behavior.
|
|
b.Name = depName
|
|
b.kitchenTask("upload_nano_results", OUTPUT_NONE)
|
|
b.Name = uploadName
|
|
b.serviceAccount(b.cfg.ServiceAccountUploadNano)
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.cipd(specs.CIPD_PKGS_GSUTIL...)
|
|
b.dep(depName)
|
|
})
|
|
}
|
|
|
|
func (b *jobBuilder) cifuzz() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.attempts(1)
|
|
b.usesDocker()
|
|
b.linuxGceDimensions(MACHINE_TYPE_MEDIUM)
|
|
b.cipd(CIPD_PKG_LUCI_AUTH)
|
|
b.cipd(specs.CIPD_PKGS_GIT_LINUX_AMD64...)
|
|
b.dep(b.buildTaskDrivers())
|
|
b.output("cifuzz_out")
|
|
b.timeout(60 * time.Minute)
|
|
b.isolate("whole_repo.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.cmd(
|
|
"./cifuzz",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--git_exe_path", "./cipd_bin_packages/git",
|
|
"--out_path", "./cifuzz_out",
|
|
"--skia_path", "./skia",
|
|
"--work_path", "./cifuzz_work",
|
|
"--alsologtostderr",
|
|
)
|
|
})
|
|
}
|
|
|
|
// perf generates a Perf task.
|
|
func (b *jobBuilder) perf() {
|
|
compileTaskName := ""
|
|
// LottieWeb doesn't require anything in Skia to be compiled.
|
|
if !b.extraConfig("LottieWeb") {
|
|
compileTaskName = b.compile()
|
|
}
|
|
doUpload := b.release() && b.doUpload()
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
recipe := "perf"
|
|
isolate := "perf_skia_bundled.isolate"
|
|
if b.extraConfig("Skpbench") {
|
|
recipe = "skpbench"
|
|
isolate = "skpbench_skia_bundled.isolate"
|
|
} else if b.extraConfig("PathKit") {
|
|
isolate = "pathkit.isolate"
|
|
recipe = "perf_pathkit"
|
|
} else if b.extraConfig("CanvasKit") {
|
|
isolate = "canvaskit.isolate"
|
|
recipe = "perf_canvaskit"
|
|
} else if b.extraConfig("SkottieTracing") {
|
|
recipe = "perf_skottietrace"
|
|
} else if b.extraConfig("SkottieWASM") {
|
|
recipe = "perf_skottiewasm_lottieweb"
|
|
isolate = "skottie_wasm.isolate"
|
|
} else if b.extraConfig("LottieWeb") {
|
|
recipe = "perf_skottiewasm_lottieweb"
|
|
isolate = "lottie_web.isolate"
|
|
}
|
|
b.recipeProps(EXTRA_PROPS)
|
|
if recipe == "perf" {
|
|
b.nanobenchFlags(doUpload)
|
|
}
|
|
b.kitchenTask(recipe, OUTPUT_PERF)
|
|
b.isolate(isolate)
|
|
b.swarmDimensions()
|
|
if b.extraConfig("CanvasKit", "Docker", "PathKit") {
|
|
b.usesDocker()
|
|
}
|
|
if compileTaskName != "" {
|
|
b.dep(compileTaskName)
|
|
}
|
|
b.commonTestPerfAssets()
|
|
b.expiration(20 * time.Hour)
|
|
b.timeout(4 * time.Hour)
|
|
|
|
if b.extraConfig("Valgrind") {
|
|
b.timeout(9 * time.Hour)
|
|
b.expiration(48 * time.Hour)
|
|
b.asset("valgrind")
|
|
// Since Valgrind runs on the same bots as the CQ, we restrict Valgrind to a subset of the bots
|
|
// to ensure there are always bots free for CQ tasks.
|
|
b.dimension("valgrind:1")
|
|
} else if b.extraConfig("MSAN") {
|
|
b.timeout(9 * time.Hour)
|
|
} else if b.parts["arch"] == "x86" && b.parts["configuration"] == "Debug" {
|
|
// skia:6737
|
|
b.timeout(6 * time.Hour)
|
|
} else if b.extraConfig("LottieWeb", "SkottieWASM") {
|
|
b.asset("node", "lottie-samples")
|
|
} else if b.matchExtraConfig("Skottie") {
|
|
b.asset("lottie-samples")
|
|
}
|
|
|
|
if b.os("Android") && b.cpu() {
|
|
b.asset("text_blob_traces")
|
|
}
|
|
b.maybeAddIosDevImage()
|
|
|
|
iid := b.internalHardwareLabel()
|
|
if iid != nil {
|
|
b.Spec.Command = append(b.Spec.Command, fmt.Sprintf("internal_hardware_label=%d", *iid))
|
|
}
|
|
})
|
|
|
|
// Upload results if necessary.
|
|
if doUpload {
|
|
uploadName := fmt.Sprintf("%s%s%s", PREFIX_UPLOAD, b.jobNameSchema.Sep, b.Name)
|
|
depName := b.Name
|
|
b.addTask(uploadName, func(b *taskBuilder) {
|
|
b.recipeProp("gs_bucket", b.cfg.GsBucketNano)
|
|
b.recipeProps(EXTRA_PROPS)
|
|
// TODO(borenet): I'm not sure why the upload task is
|
|
// using the Perf task name, but I've done this to
|
|
// maintain existing behavior.
|
|
b.Name = depName
|
|
b.kitchenTask("upload_nano_results", OUTPUT_NONE)
|
|
b.Name = uploadName
|
|
b.serviceAccount(b.cfg.ServiceAccountUploadNano)
|
|
b.linuxGceDimensions(MACHINE_TYPE_SMALL)
|
|
b.cipd(specs.CIPD_PKGS_GSUTIL...)
|
|
b.dep(depName)
|
|
})
|
|
}
|
|
}
|
|
|
|
// presubmit generates a task which runs the presubmit for this repo.
|
|
func (b *jobBuilder) presubmit() {
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.recipeProps(map[string]string{
|
|
"category": "cq",
|
|
"patch_gerrit_url": "https://skia-review.googlesource.com",
|
|
"patch_project": "skia",
|
|
"patch_ref": specs.PLACEHOLDER_PATCH_REF,
|
|
"reason": "CQ",
|
|
"repo_name": "skia",
|
|
})
|
|
b.recipeProps(EXTRA_PROPS)
|
|
b.kitchenTaskNoBundle("run_presubmit", OUTPUT_NONE)
|
|
b.isolate("run_recipe.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
// Use MACHINE_TYPE_LARGE because it seems to save time versus
|
|
// MEDIUM and we want presubmit to be fast.
|
|
b.linuxGceDimensions(MACHINE_TYPE_LARGE)
|
|
b.usesGit()
|
|
b.cipd(&specs.CipdPackage{
|
|
Name: "infra/recipe_bundles/chromium.googlesource.com/chromium/tools/build",
|
|
Path: "recipe_bundle",
|
|
Version: "git_revision:a8bcedad6768e206c4d2bd1718caa849f29cd42d",
|
|
})
|
|
})
|
|
}
|
|
|
|
// compileWasmGMTests uses a task driver to compile the GMs and unit tests for Web Assembly (WASM).
|
|
// We can use the same build for both CPU and GPU tests since the latter requires the code for the
|
|
// former anyway.
|
|
func (b *jobBuilder) compileWasmGMTests(compileName string) {
|
|
b.addTask(compileName, func(b *taskBuilder) {
|
|
b.attempts(1)
|
|
b.usesDocker()
|
|
b.linuxGceDimensions(MACHINE_TYPE_MEDIUM)
|
|
b.cipd(CIPD_PKG_LUCI_AUTH)
|
|
b.dep(b.buildTaskDrivers())
|
|
b.output("wasm_out")
|
|
b.timeout(60 * time.Minute)
|
|
b.isolate("compile.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountCompile)
|
|
b.cache(CACHES_DOCKER...)
|
|
// For now, we only have one compile mode - a GPU release mode. This should be sufficient to
|
|
// run CPU, WebGL1, and WebGL2 tests. Debug mode is not needed for the waterfall because
|
|
// when using puppeteer, stacktraces from exceptions are hard to get access to, so we do not
|
|
// even bother.
|
|
b.cmd(
|
|
"./compile_wasm_gm_tests",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", compileName,
|
|
"--out_path", "./wasm_out",
|
|
"--skia_path", "./skia",
|
|
"--work_path", "./cache/docker/wasm_gm",
|
|
"--alsologtostderr",
|
|
)
|
|
})
|
|
}
|
|
|
|
// compileWasmGMTests uses a task driver to compile the GMs and unit tests for Web Assembly (WASM).
|
|
// We can use the same build for both CPU and GPU tests since the latter requires the code for the
|
|
// former anyway.
|
|
func (b *jobBuilder) runWasmGMTests() {
|
|
compileTaskName := b.compile()
|
|
|
|
b.addTask(b.Name, func(b *taskBuilder) {
|
|
b.attempts(1)
|
|
b.usesNode()
|
|
b.swarmDimensions()
|
|
b.cipd(CIPD_PKG_LUCI_AUTH)
|
|
b.cipd(CIPD_PKGS_GOLDCTL...)
|
|
b.dep(b.buildTaskDrivers())
|
|
b.dep(compileTaskName)
|
|
b.timeout(60 * time.Minute)
|
|
b.isolate("wasm_gm_tests.isolate")
|
|
b.serviceAccount(b.cfg.ServiceAccountUploadGM)
|
|
b.cmd(
|
|
"./run_wasm_gm_tests",
|
|
"--project_id", "skia-swarming-bots",
|
|
"--task_id", specs.PLACEHOLDER_TASK_ID,
|
|
"--task_name", b.Name,
|
|
"--test_harness_path", "./tools/run-wasm-gm-tests",
|
|
"--built_path", "./wasm_out",
|
|
"--node_bin_path", "./node/node/bin",
|
|
"--resource_path", "./resources",
|
|
"--work_path", "./wasm_gm/work",
|
|
"--gold_ctl_path", "./cipd_bin_packages/goldctl",
|
|
"--git_commit", specs.PLACEHOLDER_REVISION,
|
|
"--changelist_id", specs.PLACEHOLDER_ISSUE,
|
|
"--patchset_order", specs.PLACEHOLDER_PATCHSET,
|
|
"--tryjob_id", specs.PLACEHOLDER_BUILDBUCKET_BUILD_ID,
|
|
// TODO(kjlubick, nifong) Make these not hard coded if we change the configs we test on.
|
|
"--webgl_version", "2", // 0 means CPU ; this flag controls cpu_or_gpu and extra_config
|
|
"--gold_key", "alpha_type:Premul",
|
|
"--gold_key", "arch:wasm",
|
|
"--gold_key", "browser:Chrome",
|
|
"--gold_key", "color_depth:8888",
|
|
"--gold_key", "config:gles",
|
|
"--gold_key", "configuration:Release",
|
|
"--gold_key", "cpu_or_gpu_value:QuadroP400",
|
|
"--gold_key", "model:Golo",
|
|
"--gold_key", "os:Ubuntu18",
|
|
"--alsologtostderr",
|
|
)
|
|
})
|
|
}
|