diff options
| author | Daniel Morsing <daniel.morsing@gmail.com> | 2025-12-01 09:40:16 +0000 |
|---|---|---|
| committer | Gopher Robot <gobot@golang.org> | 2026-01-26 08:07:51 -0800 |
| commit | cc6923e839ff10581223a600fca805203d56acd5 (patch) | |
| tree | dac8a1902a0714cb672eef61a3712f2ca99fc430 /src/cmd/compile | |
| parent | f809faeb8e2534da29e980ebd98dfbd4e8a7f4ba (diff) | |
| download | go-cc6923e839ff10581223a600fca805203d56acd5.tar.xz | |
cmd/compile: reduce lock/scheduler contention
During the parallel section of compilation, we limit the amount of
parallelism to prevent scheduler churn. We do this with a worker
scheduler, but it does not have insight on when a compilation is blocked
due to lock contention.
The symbol table was protected with a lock. While most lookups were
quick lock->read->unlock affairs, sometimes there would be
initialization logic involved. This caused every lookup to stall,
waiting for init. Since our worker scheduler couldn't see this, it would
not launch new goroutine to "cover" the gap.
Fix by splitting the symbol lock into 2 cases, initialization and
lookup. If symbols need initialization simultaneously, they will wait
for each other, but the common case of looking up a symbol will be
handled by a syncmap. In practice, I have yet to see this lock being
blocked on.
Additionally, get rid of the scheduler goroutine and have each
compilation goroutine grab work from a central queue. When multiple
compilations finished at the same time, the work scheduler would
sometime not get run immediately. This ended up starving the system of
work.
These 2 changes together cuts -1.37% off the build time of typescriptgo
on systems with a lot of cores (specifically, the c3h88 perf builder).
Updates #73044.
Change-Id: I6d4b3be56fd00a4fdd4df132bcbd52e4b2a3e91f
Reviewed-on: https://go-review.googlesource.com/c/go/+/724623
Reviewed-by: Keith Randall <khr@google.com>
Reviewed-by: Dmitri Shuralyov <dmitshur@google.com>
LUCI-TryBot-Result: Go LUCI <golang-scoped@luci-project-accounts.iam.gserviceaccount.com>
Reviewed-by: Keith Randall <khr@golang.org>
Auto-Submit: Keith Randall <khr@golang.org>
Diffstat (limited to 'src/cmd/compile')
| -rw-r--r-- | src/cmd/compile/internal/gc/compile.go | 76 |
1 files changed, 25 insertions, 51 deletions
diff --git a/src/cmd/compile/internal/gc/compile.go b/src/cmd/compile/internal/gc/compile.go index 1eb4b8cc37..ab9dac2f70 100644 --- a/src/cmd/compile/internal/gc/compile.go +++ b/src/cmd/compile/internal/gc/compile.go @@ -142,73 +142,47 @@ func compileFunctions(profile *pgoir.Profile) { // Compile the longest functions first, // since they're most likely to be the slowest. // This helps avoid stragglers. + // Since we remove from the end of the slice queue, + // that means shortest to longest. slices.SortFunc(compilequeue, func(a, b *ir.Func) int { - return cmp.Compare(len(b.Body), len(a.Body)) + return cmp.Compare(len(a.Body), len(b.Body)) }) } - // By default, we perform work right away on the current goroutine - // as the solo worker. - queue := func(work func(int)) { - work(0) - } + var mu sync.Mutex + var wg sync.WaitGroup + mu.Lock() - if nWorkers := base.Flag.LowerC; nWorkers > 1 { - // For concurrent builds, we allow the work queue - // to grow arbitrarily large, but only nWorkers work items - // can be running concurrently. - workq := make(chan func(int)) - done := make(chan int) + for workerId := range base.Flag.LowerC { + // TODO: replace with wg.Go when the oldest bootstrap has it. + // With the current policy, that'd be go1.27. + wg.Add(1) go func() { - ids := make([]int, nWorkers) - for i := range ids { - ids[i] = i - } - var pending []func(int) + defer wg.Done() + var closures []*ir.Func for { - select { - case work := <-workq: - pending = append(pending, work) - case id := <-done: - ids = append(ids, id) - } - for len(pending) > 0 && len(ids) > 0 { - work := pending[len(pending)-1] - id := ids[len(ids)-1] - pending = pending[:len(pending)-1] - ids = ids[:len(ids)-1] - go func() { - work(id) - done <- id - }() + mu.Lock() + compilequeue = append(compilequeue, closures...) + remaining := len(compilequeue) + if remaining == 0 { + mu.Unlock() + return } + fn := compilequeue[len(compilequeue)-1] + compilequeue = compilequeue[:len(compilequeue)-1] + mu.Unlock() + ssagen.Compile(fn, workerId, profile) + closures = fn.Closures } }() - queue = func(work func(int)) { - workq <- work - } - } - - var wg sync.WaitGroup - var compile func([]*ir.Func) - compile = func(fns []*ir.Func) { - wg.Add(len(fns)) - for _, fn := range fns { - fn := fn - queue(func(worker int) { - ssagen.Compile(fn, worker, profile) - compile(fn.Closures) - wg.Done() - }) - } } types.CalcSizeDisabled = true // not safe to calculate sizes concurrently base.Ctxt.InParallel = true - compile(compilequeue) - compilequeue = nil + mu.Unlock() wg.Wait() + compilequeue = nil base.Ctxt.InParallel = false types.CalcSizeDisabled = false |
