blob: 15484e8344d80181e204e3c9c4d0e1b937eb4fef [file] [log] [blame]
/*
Pulls data from multiple sources and funnels into metrics.
*/
package main
import (
"context"
"flag"
"fmt"
"os"
"path"
"path/filepath"
"regexp"
"time"
"cloud.google.com/go/bigtable"
"cloud.google.com/go/storage"
"go.skia.org/infra/datahopper/go/bot_metrics"
"go.skia.org/infra/datahopper/go/swarming_metrics"
"go.skia.org/infra/go/auth"
"go.skia.org/infra/go/common"
"go.skia.org/infra/go/gcs"
"go.skia.org/infra/go/git/repograph"
"go.skia.org/infra/go/httputils"
"go.skia.org/infra/go/metrics2"
"go.skia.org/infra/go/sklog"
"go.skia.org/infra/go/swarming"
"go.skia.org/infra/go/taskname"
"go.skia.org/infra/perf/go/perfclient"
"go.skia.org/infra/task_scheduler/go/db"
"go.skia.org/infra/task_scheduler/go/db/firestore"
"go.skia.org/infra/task_scheduler/go/db/pubsub"
"go.skia.org/infra/task_scheduler/go/db/remote_db"
"google.golang.org/api/option"
)
// flags
var (
firestoreInstance = flag.String("firestore_instance", "", "Firestore instance to use, eg. \"prod\"")
local = flag.Bool("local", false, "Running locally if true. As opposed to in production.")
promPort = flag.String("prom_port", ":20000", "Metrics service address (e.g., ':10110')")
recipesCfgFile = flag.String("recipes_cfg", "", "Path to the recipes.cfg file.")
taskSchedulerDbUrl = flag.String("task_db_url", "http://skia-task-scheduler:8008/db/", "Where the Skia task scheduler database is hosted.")
workdir = flag.String("workdir", ".", "Working directory used by data processors.")
perfBucket = flag.String("perf_bucket", "skia-perf", "The GCS bucket that should be used for writing into perf")
perfPrefix = flag.String("perf_duration_prefix", "task-duration", "The folder name in the bucket that task duration metric shoudl be written.")
pubsubTopicSet = flag.String("pubsub_topic_set", "", fmt.Sprintf("Pubsub topic set; one of: %v", pubsub.VALID_TOPIC_SETS))
tasksCfgProject = flag.String("tasks_cfg_project", "", "GCE project to use for tasks cfg cache.")
tasksCfgInstance = flag.String("tasks_cfg_instance", "", "BigTable instance to use for tasks cfg cache.")
)
var (
// Regexp matching non-alphanumeric characters.
re = regexp.MustCompile("[^A-Za-z0-9]+")
BUILDSLAVE_OFFLINE_BLACKLIST = []string{
"build3-a3",
"build4-a3",
"vm255-m3",
}
)
func main() {
common.InitWithMust(
"datahopper",
common.PrometheusOpt(promPort),
common.CloudLoggingOpt(),
)
ctx := context.Background()
// Absolutify the workdir.
w, err := filepath.Abs(*workdir)
if err != nil {
sklog.Fatal(w)
}
sklog.Infof("Workdir is %s", w)
// Authenticated HTTP client.
oauthCacheFile := path.Join(w, "google_storage_token.data")
legacyTs, err := auth.NewLegacyTokenSource(*local, oauthCacheFile, "", swarming.AUTH_SCOPE)
if err != nil {
sklog.Fatal(err)
}
httpClient := httputils.DefaultClientConfig().WithTokenSource(legacyTs).With2xxOnly().Client()
// Swarming API client.
swarm, err := swarming.NewApiClient(httpClient, swarming.SWARMING_SERVER)
if err != nil {
sklog.Fatal(err)
}
swarmInternal, err := swarming.NewApiClient(httpClient, swarming.SWARMING_SERVER_PRIVATE)
if err != nil {
sklog.Fatal(err)
}
jwtTs, err := auth.NewDefaultJWTServiceAccountTokenSource(auth.SCOPE_READ_WRITE)
if err != nil {
sklog.Fatal(err)
}
authClient := httputils.DefaultClientConfig().WithTokenSource(jwtTs).With2xxOnly().Client()
gsClient, err := storage.NewClient(ctx, option.WithHTTPClient(authClient))
if err != nil {
sklog.Fatal(err)
}
storageClient := gcs.NewGCSClient(gsClient, *perfBucket)
pc := perfclient.New(*perfPrefix, storageClient)
tnp := taskname.DefaultTaskNameParser()
// Shared repo objects.
reposDir := path.Join(w, "repos")
if err := os.MkdirAll(reposDir, os.ModePerm); err != nil {
sklog.Fatal(err)
}
repos, err := repograph.NewMap(ctx, []string{common.REPO_SKIA, common.REPO_SKIA_INFRA}, reposDir)
if err != nil {
sklog.Fatal(err)
}
if err := repos.Update(ctx); err != nil {
sklog.Fatal(err)
}
// Data generation goroutines.
// Swarming bots.
swarmingClients := map[string]swarming.ApiClient{
swarming.SWARMING_SERVER: swarm,
swarming.SWARMING_SERVER_PRIVATE: swarmInternal,
}
swarmingPools := map[string][]string{
swarming.SWARMING_SERVER: swarming.POOLS_PUBLIC,
swarming.SWARMING_SERVER_PRIVATE: swarming.POOLS_PRIVATE,
}
swarming_metrics.StartSwarmingBotMetrics(swarmingClients, swarmingPools, metrics2.GetDefaultClient())
// Swarming tasks.
if err := swarming_metrics.StartSwarmingTaskMetrics(w, swarm, ctx, pc, tnp); err != nil {
sklog.Fatal(err)
}
// Number of commits in the repo.
go func() {
skiaGauge := metrics2.GetInt64Metric("repo_commits", map[string]string{"repo": "skia"})
infraGauge := metrics2.GetInt64Metric("repo_commits", map[string]string{"repo": "infra"})
for range time.Tick(5 * time.Minute) {
nSkia, err := repos[common.REPO_SKIA].Repo().NumCommits(ctx)
if err != nil {
sklog.Errorf("Failed to get number of commits for Skia: %s", err)
} else {
skiaGauge.Update(nSkia)
}
nInfra, err := repos[common.REPO_SKIA_INFRA].Repo().NumCommits(ctx)
if err != nil {
sklog.Errorf("Failed to get number of commits for Infra: %s", err)
} else {
infraGauge.Update(nInfra)
}
}
}()
// Tasks metrics.
newTs, err := auth.NewDefaultTokenSource(*local, pubsub.AUTH_SCOPE, bigtable.Scope)
if err != nil {
sklog.Fatal(err)
}
var d db.RemoteDB
if *firestoreInstance != "" {
label := "datahopper"
mod, err := pubsub.NewModifiedData(*pubsubTopicSet, label, newTs)
if err != nil {
sklog.Fatal(err)
}
d, err = firestore.NewDB(ctx, firestore.FIRESTORE_PROJECT, *firestoreInstance, newTs, mod)
if err != nil {
sklog.Fatalf("Failed to create Firestore DB client: %s", err)
}
} else {
d, err = remote_db.NewClient(*taskSchedulerDbUrl, *pubsubTopicSet, "datahopper", newTs)
if err != nil {
sklog.Fatal(err)
}
}
if err := StartTaskMetrics(ctx, d); err != nil {
sklog.Fatal(err)
}
// Jobs metrics.
if err := StartJobMetrics(ctx, d); err != nil {
sklog.Fatal(err)
}
// Generate "time to X% bot coverage" metrics.
if *recipesCfgFile == "" {
*recipesCfgFile = path.Join(*workdir, "recipes.cfg")
}
if err := bot_metrics.Start(ctx, d, *workdir, *recipesCfgFile, *tasksCfgProject, *tasksCfgInstance, newTs); err != nil {
sklog.Fatal(err)
}
// Wait while the above goroutines generate data.
select {}
}