aboutsummaryrefslogtreecommitdiff
path: root/src/cmd/compile
diff options
context:
space:
mode:
authorDaniel Morsing <daniel.morsing@gmail.com>2025-12-01 09:40:16 +0000
committerGopher Robot <gobot@golang.org>2026-01-26 08:07:51 -0800
commitcc6923e839ff10581223a600fca805203d56acd5 (patch)
treedac8a1902a0714cb672eef61a3712f2ca99fc430 /src/cmd/compile
parentf809faeb8e2534da29e980ebd98dfbd4e8a7f4ba (diff)
downloadgo-cc6923e839ff10581223a600fca805203d56acd5.tar.xz
cmd/compile: reduce lock/scheduler contention
During the parallel section of compilation, we limit the amount of parallelism to prevent scheduler churn. We do this with a worker scheduler, but it does not have insight on when a compilation is blocked due to lock contention. The symbol table was protected with a lock. While most lookups were quick lock->read->unlock affairs, sometimes there would be initialization logic involved. This caused every lookup to stall, waiting for init. Since our worker scheduler couldn't see this, it would not launch new goroutine to "cover" the gap. Fix by splitting the symbol lock into 2 cases, initialization and lookup. If symbols need initialization simultaneously, they will wait for each other, but the common case of looking up a symbol will be handled by a syncmap. In practice, I have yet to see this lock being blocked on. Additionally, get rid of the scheduler goroutine and have each compilation goroutine grab work from a central queue. When multiple compilations finished at the same time, the work scheduler would sometime not get run immediately. This ended up starving the system of work. These 2 changes together cuts -1.37% off the build time of typescriptgo on systems with a lot of cores (specifically, the c3h88 perf builder). Updates #73044. Change-Id: I6d4b3be56fd00a4fdd4df132bcbd52e4b2a3e91f Reviewed-on: https://go-review.googlesource.com/c/go/+/724623 Reviewed-by: Keith Randall <khr@google.com> Reviewed-by: Dmitri Shuralyov <dmitshur@google.com> LUCI-TryBot-Result: Go LUCI <golang-scoped@luci-project-accounts.iam.gserviceaccount.com> Reviewed-by: Keith Randall <khr@golang.org> Auto-Submit: Keith Randall <khr@golang.org>
Diffstat (limited to 'src/cmd/compile')
-rw-r--r--src/cmd/compile/internal/gc/compile.go76
1 files changed, 25 insertions, 51 deletions
diff --git a/src/cmd/compile/internal/gc/compile.go b/src/cmd/compile/internal/gc/compile.go
index 1eb4b8cc37..ab9dac2f70 100644
--- a/src/cmd/compile/internal/gc/compile.go
+++ b/src/cmd/compile/internal/gc/compile.go
@@ -142,73 +142,47 @@ func compileFunctions(profile *pgoir.Profile) {
// Compile the longest functions first,
// since they're most likely to be the slowest.
// This helps avoid stragglers.
+ // Since we remove from the end of the slice queue,
+ // that means shortest to longest.
slices.SortFunc(compilequeue, func(a, b *ir.Func) int {
- return cmp.Compare(len(b.Body), len(a.Body))
+ return cmp.Compare(len(a.Body), len(b.Body))
})
}
- // By default, we perform work right away on the current goroutine
- // as the solo worker.
- queue := func(work func(int)) {
- work(0)
- }
+ var mu sync.Mutex
+ var wg sync.WaitGroup
+ mu.Lock()
- if nWorkers := base.Flag.LowerC; nWorkers > 1 {
- // For concurrent builds, we allow the work queue
- // to grow arbitrarily large, but only nWorkers work items
- // can be running concurrently.
- workq := make(chan func(int))
- done := make(chan int)
+ for workerId := range base.Flag.LowerC {
+ // TODO: replace with wg.Go when the oldest bootstrap has it.
+ // With the current policy, that'd be go1.27.
+ wg.Add(1)
go func() {
- ids := make([]int, nWorkers)
- for i := range ids {
- ids[i] = i
- }
- var pending []func(int)
+ defer wg.Done()
+ var closures []*ir.Func
for {
- select {
- case work := <-workq:
- pending = append(pending, work)
- case id := <-done:
- ids = append(ids, id)
- }
- for len(pending) > 0 && len(ids) > 0 {
- work := pending[len(pending)-1]
- id := ids[len(ids)-1]
- pending = pending[:len(pending)-1]
- ids = ids[:len(ids)-1]
- go func() {
- work(id)
- done <- id
- }()
+ mu.Lock()
+ compilequeue = append(compilequeue, closures...)
+ remaining := len(compilequeue)
+ if remaining == 0 {
+ mu.Unlock()
+ return
}
+ fn := compilequeue[len(compilequeue)-1]
+ compilequeue = compilequeue[:len(compilequeue)-1]
+ mu.Unlock()
+ ssagen.Compile(fn, workerId, profile)
+ closures = fn.Closures
}
}()
- queue = func(work func(int)) {
- workq <- work
- }
- }
-
- var wg sync.WaitGroup
- var compile func([]*ir.Func)
- compile = func(fns []*ir.Func) {
- wg.Add(len(fns))
- for _, fn := range fns {
- fn := fn
- queue(func(worker int) {
- ssagen.Compile(fn, worker, profile)
- compile(fn.Closures)
- wg.Done()
- })
- }
}
types.CalcSizeDisabled = true // not safe to calculate sizes concurrently
base.Ctxt.InParallel = true
- compile(compilequeue)
- compilequeue = nil
+ mu.Unlock()
wg.Wait()
+ compilequeue = nil
base.Ctxt.InParallel = false
types.CalcSizeDisabled = false