2017-01-20 23:10:01 +01:00
|
|
|
// Copyright 2017 Google Inc. All rights reserved.
|
|
|
|
//
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
|
|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"flag"
|
|
|
|
"fmt"
|
2017-05-20 01:39:04 +02:00
|
|
|
"io/ioutil"
|
2017-01-20 23:10:01 +01:00
|
|
|
"os"
|
|
|
|
"path/filepath"
|
|
|
|
"runtime"
|
|
|
|
"strings"
|
|
|
|
"sync"
|
2017-12-11 23:35:23 +01:00
|
|
|
"syscall"
|
2017-03-30 04:26:09 +02:00
|
|
|
"time"
|
2017-01-20 23:10:01 +01:00
|
|
|
|
|
|
|
"android/soong/ui/build"
|
|
|
|
"android/soong/ui/logger"
|
2016-08-22 00:17:17 +02:00
|
|
|
"android/soong/ui/tracer"
|
2017-11-07 20:23:27 +01:00
|
|
|
"android/soong/zip"
|
2017-01-20 23:10:01 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
// We default to number of cpus / 4, which seems to be the sweet spot for my
|
|
|
|
// system. I suspect this is mostly due to memory or disk bandwidth though, and
|
|
|
|
// may depend on the size ofthe source tree, so this probably isn't a great
|
|
|
|
// default.
|
|
|
|
func detectNumJobs() int {
|
|
|
|
if runtime.NumCPU() < 4 {
|
|
|
|
return 1
|
|
|
|
}
|
|
|
|
return runtime.NumCPU() / 4
|
|
|
|
}
|
|
|
|
|
|
|
|
var numJobs = flag.Int("j", detectNumJobs(), "number of parallel kati jobs")
|
|
|
|
|
2017-11-07 20:23:27 +01:00
|
|
|
var keepArtifacts = flag.Bool("keep", false, "keep archives of artifacts")
|
2017-01-20 23:10:01 +01:00
|
|
|
|
|
|
|
var outDir = flag.String("out", "", "path to store output directories (defaults to tmpdir under $OUT when empty)")
|
2017-05-20 01:39:04 +02:00
|
|
|
var alternateResultDir = flag.Bool("dist", false, "write select results to $DIST_DIR (or <out>/dist when empty)")
|
2017-01-20 23:10:01 +01:00
|
|
|
|
|
|
|
var onlyConfig = flag.Bool("only-config", false, "Only run product config (not Soong or Kati)")
|
|
|
|
var onlySoong = flag.Bool("only-soong", false, "Only run product config and Soong (not Kati)")
|
|
|
|
|
2017-05-07 20:40:30 +02:00
|
|
|
var buildVariant = flag.String("variant", "eng", "build variant to use")
|
|
|
|
|
2017-10-07 00:05:05 +02:00
|
|
|
var skipProducts = flag.String("skip-products", "", "comma-separated list of products to skip (known failures, etc)")
|
2017-10-26 00:02:45 +02:00
|
|
|
var includeProducts = flag.String("products", "", "comma-separated list of products to build")
|
2017-10-07 00:05:05 +02:00
|
|
|
|
2017-05-20 01:39:04 +02:00
|
|
|
const errorLeadingLines = 20
|
|
|
|
const errorTrailingLines = 20
|
|
|
|
|
2017-01-20 23:10:01 +01:00
|
|
|
type Product struct {
|
2017-05-20 01:39:04 +02:00
|
|
|
ctx build.Context
|
|
|
|
config build.Config
|
|
|
|
logFile string
|
2017-01-20 23:10:01 +01:00
|
|
|
}
|
|
|
|
|
2017-05-07 01:58:26 +02:00
|
|
|
type Status struct {
|
|
|
|
cur int
|
|
|
|
total int
|
|
|
|
failed int
|
|
|
|
|
|
|
|
ctx build.Context
|
|
|
|
haveBlankLine bool
|
|
|
|
smartTerminal bool
|
|
|
|
|
|
|
|
lock sync.Mutex
|
|
|
|
}
|
|
|
|
|
|
|
|
func NewStatus(ctx build.Context) *Status {
|
|
|
|
return &Status{
|
|
|
|
ctx: ctx,
|
|
|
|
haveBlankLine: true,
|
|
|
|
smartTerminal: ctx.IsTerminal(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Status) SetTotal(total int) {
|
|
|
|
s.total = total
|
|
|
|
}
|
|
|
|
|
2017-05-20 01:39:04 +02:00
|
|
|
func (s *Status) Fail(product string, err error, logFile string) {
|
2017-05-07 01:58:26 +02:00
|
|
|
s.Finish(product)
|
|
|
|
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
|
|
|
if s.smartTerminal && !s.haveBlankLine {
|
|
|
|
fmt.Fprintln(s.ctx.Stdout())
|
|
|
|
s.haveBlankLine = true
|
|
|
|
}
|
|
|
|
|
|
|
|
s.failed++
|
|
|
|
fmt.Fprintln(s.ctx.Stderr(), "FAILED:", product)
|
|
|
|
s.ctx.Verboseln("FAILED:", product)
|
2017-05-20 01:39:04 +02:00
|
|
|
|
|
|
|
if logFile != "" {
|
|
|
|
data, err := ioutil.ReadFile(logFile)
|
|
|
|
if err == nil {
|
|
|
|
lines := strings.Split(strings.TrimSpace(string(data)), "\n")
|
|
|
|
if len(lines) > errorLeadingLines+errorTrailingLines+1 {
|
|
|
|
lines[errorLeadingLines] = fmt.Sprintf("... skipping %d lines ...",
|
|
|
|
len(lines)-errorLeadingLines-errorTrailingLines)
|
|
|
|
|
|
|
|
lines = append(lines[:errorLeadingLines+1],
|
|
|
|
lines[len(lines)-errorTrailingLines:]...)
|
|
|
|
}
|
|
|
|
for _, line := range lines {
|
|
|
|
fmt.Fprintln(s.ctx.Stderr(), "> ", line)
|
|
|
|
s.ctx.Verboseln(line)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
s.ctx.Print(err)
|
2017-05-07 01:58:26 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Status) Finish(product string) {
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
|
|
|
s.cur++
|
|
|
|
line := fmt.Sprintf("[%d/%d] %s", s.cur, s.total, product)
|
|
|
|
|
|
|
|
if s.smartTerminal {
|
|
|
|
if max, ok := s.ctx.TermWidth(); ok {
|
|
|
|
if len(line) > max {
|
|
|
|
line = line[:max]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fmt.Fprint(s.ctx.Stdout(), "\r", line, "\x1b[K")
|
|
|
|
s.haveBlankLine = false
|
|
|
|
} else {
|
|
|
|
s.ctx.Println(line)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Status) Finished() int {
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
|
|
|
if !s.haveBlankLine {
|
|
|
|
fmt.Fprintln(s.ctx.Stdout())
|
|
|
|
s.haveBlankLine = true
|
|
|
|
}
|
|
|
|
return s.failed
|
|
|
|
}
|
|
|
|
|
2017-12-11 23:35:23 +01:00
|
|
|
// TODO(b/70370883): This tool uses a lot of open files -- over the default
|
|
|
|
// soft limit of 1024 on some systems. So bump up to the hard limit until I fix
|
|
|
|
// the algorithm.
|
|
|
|
func setMaxFiles(log logger.Logger) {
|
|
|
|
var limits syscall.Rlimit
|
|
|
|
|
|
|
|
err := syscall.Getrlimit(syscall.RLIMIT_NOFILE, &limits)
|
|
|
|
if err != nil {
|
|
|
|
log.Println("Failed to get file limit:", err)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Verbosef("Current file limits: %d soft, %d hard", limits.Cur, limits.Max)
|
|
|
|
if limits.Cur == limits.Max {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
limits.Cur = limits.Max
|
|
|
|
err = syscall.Setrlimit(syscall.RLIMIT_NOFILE, &limits)
|
|
|
|
if err != nil {
|
|
|
|
log.Println("Failed to increase file limit:", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-10-26 00:02:45 +02:00
|
|
|
func inList(str string, list []string) bool {
|
|
|
|
for _, other := range list {
|
|
|
|
if str == other {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2017-01-20 23:10:01 +01:00
|
|
|
func main() {
|
|
|
|
log := logger.New(os.Stderr)
|
|
|
|
defer log.Cleanup()
|
|
|
|
|
|
|
|
flag.Parse()
|
|
|
|
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
|
2016-08-22 00:17:17 +02:00
|
|
|
trace := tracer.New(log)
|
|
|
|
defer trace.Close()
|
2017-01-20 23:10:01 +01:00
|
|
|
|
2016-08-22 00:17:17 +02:00
|
|
|
build.SetupSignals(log, cancel, func() {
|
|
|
|
trace.Close()
|
|
|
|
log.Cleanup()
|
|
|
|
})
|
|
|
|
|
|
|
|
buildCtx := build.Context{&build.ContextImpl{
|
2017-01-20 23:10:01 +01:00
|
|
|
Context: ctx,
|
|
|
|
Logger: log,
|
2016-08-22 00:17:17 +02:00
|
|
|
Tracer: trace,
|
2017-01-20 23:10:01 +01:00
|
|
|
StdioInterface: build.StdioImpl{},
|
2016-08-22 00:17:17 +02:00
|
|
|
}}
|
2017-01-20 23:10:01 +01:00
|
|
|
|
2017-05-07 01:58:26 +02:00
|
|
|
status := NewStatus(buildCtx)
|
2017-01-20 23:10:01 +01:00
|
|
|
|
|
|
|
config := build.NewConfig(buildCtx)
|
|
|
|
if *outDir == "" {
|
2017-03-30 04:26:09 +02:00
|
|
|
name := "multiproduct-" + time.Now().Format("20060102150405")
|
|
|
|
|
|
|
|
*outDir = filepath.Join(config.OutDir(), name)
|
|
|
|
|
2017-05-20 01:39:04 +02:00
|
|
|
// Ensure the empty files exist in the output directory
|
|
|
|
// containing our output directory too. This is mostly for
|
|
|
|
// safety, but also triggers the ninja_build file so that our
|
|
|
|
// build servers know that they can parse the output as if it
|
|
|
|
// was ninja output.
|
|
|
|
build.SetupOutDir(buildCtx, config)
|
|
|
|
|
2017-03-30 04:26:09 +02:00
|
|
|
if err := os.MkdirAll(*outDir, 0777); err != nil {
|
2017-01-20 23:10:01 +01:00
|
|
|
log.Fatalf("Failed to create tempdir: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
config.Environment().Set("OUT_DIR", *outDir)
|
|
|
|
log.Println("Output directory:", *outDir)
|
|
|
|
|
2017-11-07 20:23:27 +01:00
|
|
|
logsDir := filepath.Join(config.OutDir(), "logs")
|
|
|
|
os.MkdirAll(logsDir, 0777)
|
|
|
|
|
2017-01-20 23:10:01 +01:00
|
|
|
build.SetupOutDir(buildCtx, config)
|
2017-05-20 01:39:04 +02:00
|
|
|
if *alternateResultDir {
|
2017-11-07 20:23:27 +01:00
|
|
|
distLogsDir := filepath.Join(config.DistDir(), "logs")
|
|
|
|
os.MkdirAll(distLogsDir, 0777)
|
|
|
|
log.SetOutput(filepath.Join(distLogsDir, "soong.log"))
|
|
|
|
trace.SetOutput(filepath.Join(distLogsDir, "build.trace"))
|
2017-05-20 01:39:04 +02:00
|
|
|
} else {
|
|
|
|
log.SetOutput(filepath.Join(config.OutDir(), "soong.log"))
|
|
|
|
trace.SetOutput(filepath.Join(config.OutDir(), "build.trace"))
|
|
|
|
}
|
2017-01-20 23:10:01 +01:00
|
|
|
|
2017-12-11 23:35:23 +01:00
|
|
|
setMaxFiles(log)
|
|
|
|
|
2017-07-14 02:24:44 +02:00
|
|
|
vars, err := build.DumpMakeVars(buildCtx, config, nil, []string{"all_named_products"})
|
2017-01-20 23:10:01 +01:00
|
|
|
if err != nil {
|
|
|
|
log.Fatal(err)
|
|
|
|
}
|
2017-10-26 00:02:45 +02:00
|
|
|
var productsList []string
|
|
|
|
allProducts := strings.Fields(vars["all_named_products"])
|
|
|
|
|
|
|
|
if *includeProducts != "" {
|
|
|
|
missingProducts := []string{}
|
|
|
|
for _, product := range strings.Split(*includeProducts, ",") {
|
|
|
|
if inList(product, allProducts) {
|
|
|
|
productsList = append(productsList, product)
|
|
|
|
} else {
|
|
|
|
missingProducts = append(missingProducts, product)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(missingProducts) > 0 {
|
|
|
|
log.Fatalf("Products don't exist: %s\n", missingProducts)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
productsList = allProducts
|
|
|
|
}
|
2017-10-07 00:05:05 +02:00
|
|
|
|
|
|
|
products := make([]string, 0, len(productsList))
|
|
|
|
skipList := strings.Split(*skipProducts, ",")
|
|
|
|
skipProduct := func(p string) bool {
|
|
|
|
for _, s := range skipList {
|
|
|
|
if p == s {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
for _, product := range productsList {
|
|
|
|
if !skipProduct(product) {
|
|
|
|
products = append(products, product)
|
|
|
|
} else {
|
|
|
|
log.Verbose("Skipping: ", product)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Verbose("Got product list: ", products)
|
2017-01-20 23:10:01 +01:00
|
|
|
|
2017-05-07 01:58:26 +02:00
|
|
|
status.SetTotal(len(products))
|
|
|
|
|
2017-01-20 23:10:01 +01:00
|
|
|
var wg sync.WaitGroup
|
|
|
|
productConfigs := make(chan Product, len(products))
|
|
|
|
|
2017-08-04 21:30:12 +02:00
|
|
|
finder := build.NewSourceFinder(buildCtx, config)
|
|
|
|
defer finder.Shutdown()
|
|
|
|
|
2017-01-20 23:10:01 +01:00
|
|
|
// Run the product config for every product in parallel
|
|
|
|
for _, product := range products {
|
|
|
|
wg.Add(1)
|
|
|
|
go func(product string) {
|
2017-05-20 01:39:04 +02:00
|
|
|
var stdLog string
|
|
|
|
|
2017-01-20 23:10:01 +01:00
|
|
|
defer wg.Done()
|
|
|
|
defer logger.Recover(func(err error) {
|
2017-05-20 01:39:04 +02:00
|
|
|
status.Fail(product, err, stdLog)
|
2017-01-20 23:10:01 +01:00
|
|
|
})
|
|
|
|
|
|
|
|
productOutDir := filepath.Join(config.OutDir(), product)
|
2017-11-07 20:23:27 +01:00
|
|
|
productLogDir := filepath.Join(logsDir, product)
|
2017-01-20 23:10:01 +01:00
|
|
|
|
|
|
|
if err := os.MkdirAll(productOutDir, 0777); err != nil {
|
|
|
|
log.Fatalf("Error creating out directory: %v", err)
|
|
|
|
}
|
2017-11-07 20:23:27 +01:00
|
|
|
if err := os.MkdirAll(productLogDir, 0777); err != nil {
|
|
|
|
log.Fatalf("Error creating log directory: %v", err)
|
|
|
|
}
|
2017-01-20 23:10:01 +01:00
|
|
|
|
2017-05-20 01:39:04 +02:00
|
|
|
stdLog = filepath.Join(productLogDir, "std.log")
|
|
|
|
f, err := os.Create(stdLog)
|
2017-01-20 23:10:01 +01:00
|
|
|
if err != nil {
|
|
|
|
log.Fatalf("Error creating std.log: %v", err)
|
|
|
|
}
|
|
|
|
|
2017-11-12 00:44:51 +01:00
|
|
|
productLog := logger.New(f)
|
2017-05-20 01:39:04 +02:00
|
|
|
productLog.SetOutput(filepath.Join(productLogDir, "soong.log"))
|
2017-01-20 23:10:01 +01:00
|
|
|
|
2016-08-22 00:17:17 +02:00
|
|
|
productCtx := build.Context{&build.ContextImpl{
|
2017-01-20 23:10:01 +01:00
|
|
|
Context: ctx,
|
|
|
|
Logger: productLog,
|
2016-08-22 00:17:17 +02:00
|
|
|
Tracer: trace,
|
2017-01-20 23:10:01 +01:00
|
|
|
StdioInterface: build.NewCustomStdio(nil, f, f),
|
2016-08-22 00:17:17 +02:00
|
|
|
Thread: trace.NewThread(product),
|
|
|
|
}}
|
2017-01-20 23:10:01 +01:00
|
|
|
|
|
|
|
productConfig := build.NewConfig(productCtx)
|
|
|
|
productConfig.Environment().Set("OUT_DIR", productOutDir)
|
2017-08-17 23:09:23 +02:00
|
|
|
build.FindSources(productCtx, productConfig, finder)
|
2017-05-07 20:40:30 +02:00
|
|
|
productConfig.Lunch(productCtx, product, *buildVariant)
|
2017-01-20 23:10:01 +01:00
|
|
|
|
|
|
|
build.Build(productCtx, productConfig, build.BuildProductConfig)
|
2017-05-20 01:39:04 +02:00
|
|
|
productConfigs <- Product{productCtx, productConfig, stdLog}
|
2017-01-20 23:10:01 +01:00
|
|
|
}(product)
|
|
|
|
}
|
|
|
|
go func() {
|
|
|
|
defer close(productConfigs)
|
|
|
|
wg.Wait()
|
|
|
|
}()
|
|
|
|
|
|
|
|
var wg2 sync.WaitGroup
|
|
|
|
// Then run up to numJobs worth of Soong and Kati
|
|
|
|
for i := 0; i < *numJobs; i++ {
|
|
|
|
wg2.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg2.Done()
|
|
|
|
for product := range productConfigs {
|
|
|
|
func() {
|
|
|
|
defer logger.Recover(func(err error) {
|
2017-05-20 01:39:04 +02:00
|
|
|
status.Fail(product.config.TargetProduct(), err, product.logFile)
|
2017-01-20 23:10:01 +01:00
|
|
|
})
|
|
|
|
|
2017-11-07 20:23:27 +01:00
|
|
|
defer func() {
|
|
|
|
if *keepArtifacts {
|
|
|
|
args := zip.ZipArgs{
|
|
|
|
FileArgs: []zip.FileArg{
|
|
|
|
{
|
|
|
|
GlobDir: product.config.OutDir(),
|
|
|
|
SourcePrefixToStrip: product.config.OutDir(),
|
|
|
|
},
|
|
|
|
},
|
|
|
|
OutputFilePath: filepath.Join(config.OutDir(), product.config.TargetProduct()+".zip"),
|
|
|
|
NumParallelJobs: runtime.NumCPU(),
|
|
|
|
CompressionLevel: 5,
|
|
|
|
}
|
|
|
|
if err := zip.Run(args); err != nil {
|
|
|
|
log.Fatalf("Error zipping artifacts: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
os.RemoveAll(product.config.OutDir())
|
|
|
|
}()
|
|
|
|
|
2017-01-20 23:10:01 +01:00
|
|
|
buildWhat := 0
|
|
|
|
if !*onlyConfig {
|
|
|
|
buildWhat |= build.BuildSoong
|
|
|
|
if !*onlySoong {
|
|
|
|
buildWhat |= build.BuildKati
|
|
|
|
}
|
|
|
|
}
|
|
|
|
build.Build(product.ctx, product.config, buildWhat)
|
2017-05-07 01:58:26 +02:00
|
|
|
status.Finish(product.config.TargetProduct())
|
2017-01-20 23:10:01 +01:00
|
|
|
}()
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
}
|
2017-05-07 01:58:26 +02:00
|
|
|
wg2.Wait()
|
2017-01-20 23:10:01 +01:00
|
|
|
|
2017-11-07 20:23:27 +01:00
|
|
|
if *alternateResultDir {
|
|
|
|
args := zip.ZipArgs{
|
|
|
|
FileArgs: []zip.FileArg{
|
|
|
|
{GlobDir: logsDir, SourcePrefixToStrip: logsDir},
|
|
|
|
},
|
|
|
|
OutputFilePath: filepath.Join(config.DistDir(), "logs.zip"),
|
|
|
|
NumParallelJobs: runtime.NumCPU(),
|
|
|
|
CompressionLevel: 5,
|
|
|
|
}
|
|
|
|
if err := zip.Run(args); err != nil {
|
|
|
|
log.Fatalf("Error zipping logs: %v", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-05-07 01:58:26 +02:00
|
|
|
if count := status.Finished(); count > 0 {
|
|
|
|
log.Fatalln(count, "products failed")
|
2017-01-20 23:10:01 +01:00
|
|
|
}
|
|
|
|
}
|