// Copyright 2009 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. package runtime import ( "runtime/internal/atomic" "unsafe" ) //go:generate go run wincallback.go //go:generate go run mkduff.go //go:generate go run mkfastlog2table.go //go:generate go run mklockrank.go -o lockrank.go var ticks ticksType type ticksType struct { // lock protects access to start* and val. lock mutex startTicks int64 startTime int64 val atomic.Int64 } // init initializes ticks to maximize the chance that we have a good ticksPerSecond reference. // // Must not run concurrently with ticksPerSecond. func (t *ticksType) init() { lock(&ticks.lock) t.startTime = nanotime() t.startTicks = cputicks() unlock(&ticks.lock) } // minTimeForTicksPerSecond is the minimum elapsed time we require to consider our ticksPerSecond // measurement to be of decent enough quality for profiling. // // There's a linear relationship here between minimum time and error from the true value. // The error from the true ticks-per-second in a linux/amd64 VM seems to be: // - 1 ms -> ~0.02% error // - 5 ms -> ~0.004% error // - 10 ms -> ~0.002% error // - 50 ms -> ~0.0003% error // - 100 ms -> ~0.0001% error // // We're willing to take 0.004% error here, because ticksPerSecond is intended to be used for // converting durations, not timestamps. Durations are usually going to be much larger, and so // the tiny error doesn't matter. The error is definitely going to be a problem when trying to // use this for timestamps, as it'll make those timestamps much less likely to line up. const minTimeForTicksPerSecond = 5_000_000*(1-osHasLowResClockInt) + 100_000_000*osHasLowResClockInt // ticksPerSecond returns a conversion rate between the cputicks clock and the nanotime clock. // // Note: Clocks are hard. Using this as an actual conversion rate for timestamps is ill-advised // and should be avoided when possible. Use only for durations, where a tiny error term isn't going // to make a meaningful difference in even a 1ms duration. If an accurate timestamp is needed, // use nanotime instead. (The entire Windows platform is a broad exception to this rule, where nanotime // produces timestamps on such a coarse granularity that the error from this conversion is actually // preferable.) // // The strategy for computing the conversion rate is to write down nanotime and cputicks as // early in process startup as possible. From then, we just need to wait until we get values // from nanotime that we can use (some platforms have a really coarse system time granularity). // We require some amount of time to pass to ensure that the conversion rate is fairly accurate // in aggregate. But because we compute this rate lazily, there's a pretty good chance a decent // amount of time has passed by the time we get here. // // Must be called from a normal goroutine context (running regular goroutine with a P). // // Called by runtime/pprof in addition to runtime code. // // TODO(mknyszek): This doesn't account for things like CPU frequency scaling. Consider // a more sophisticated and general approach in the future. func ticksPerSecond() int64 { // Get the conversion rate if we've already computed it. r := ticks.val.Load() if r != 0 { return r } // Compute the conversion rate. for { lock(&ticks.lock) r = ticks.val.Load() if r != 0 { unlock(&ticks.lock) return r } // Grab the current time in both clocks. nowTime := nanotime() nowTicks := cputicks() // See if we can use these times. if nowTicks > ticks.startTicks && nowTime-ticks.startTime > minTimeForTicksPerSecond { // Perform the calculation with floats. We don't want to risk overflow. r = int64(float64(nowTicks-ticks.startTicks) * 1e9 / float64(nowTime-ticks.startTime)) if r == 0 { // Zero is both a sentinel value and it would be bad if callers used this as // a divisor. We tried out best, so just make it 1. r++ } ticks.val.Store(r) unlock(&ticks.lock) break } unlock(&ticks.lock) // Sleep in one millisecond increments until we have a reliable time. timeSleep(1_000_000) } return r } var envs []string var argslice []string //go:linkname syscall_runtime_envs syscall.runtime_envs func syscall_runtime_envs() []string { return append([]string{}, envs...) } //go:linkname syscall_Getpagesize syscall.Getpagesize func syscall_Getpagesize() int { return int(physPageSize) } //go:linkname os_runtime_args os.runtime_args func os_runtime_args() []string { return append([]string{}, argslice...) } //go:linkname syscall_Exit syscall.Exit //go:nosplit func syscall_Exit(code int) { exit(int32(code)) } var godebugDefault string var godebugUpdate atomic.Pointer[func(string, string)] var godebugEnv atomic.Pointer[string] // set by parsedebugvars var godebugNewIncNonDefault atomic.Pointer[func(string) func()] //go:linkname godebug_setUpdate internal/godebug.setUpdate func godebug_setUpdate(update func(string, string)) { p := new(func(string, string)) *p = update godebugUpdate.Store(p) godebugNotify(false) } //go:linkname godebug_setNewIncNonDefault internal/godebug.setNewIncNonDefault func godebug_setNewIncNonDefault(newIncNonDefault func(string) func()) { p := new(func(string) func()) *p = newIncNonDefault godebugNewIncNonDefault.Store(p) } // A godebugInc provides access to internal/godebug's IncNonDefault function // for a given GODEBUG setting. // Calls before internal/godebug registers itself are dropped on the floor. type godebugInc struct { name string inc atomic.Pointer[func()] } func (g *godebugInc) IncNonDefault() { inc := g.inc.Load() if inc == nil { newInc := godebugNewIncNonDefault.Load() if newInc == nil { return } inc = new(func()) *inc = (*newInc)(g.name) if raceenabled { racereleasemerge(unsafe.Pointer(&g.inc)) } if !g.inc.CompareAndSwap(nil, inc) { inc = g.inc.Load() } } if raceenabled { raceacquire(unsafe.Pointer(&g.inc)) } (*inc)() } func godebugNotify(envChanged bool) { update := godebugUpdate.Load() var env string if p := godebugEnv.Load(); p != nil { env = *p } if envChanged { reparsedebugvars(env) } if update != nil { (*update)(godebugDefault, env) } } //go:linkname syscall_runtimeSetenv syscall.runtimeSetenv func syscall_runtimeSetenv(key, value string) { setenv_c(key, value) if key == "GODEBUG" { p := new(string) *p = value godebugEnv.Store(p) godebugNotify(true) } } //go:linkname syscall_runtimeUnsetenv syscall.runtimeUnsetenv func syscall_runtimeUnsetenv(key string) { unsetenv_c(key) if key == "GODEBUG" { godebugEnv.Store(nil) godebugNotify(true) } } // writeErrStr writes a string to descriptor 2. // //go:nosplit func writeErrStr(s string) { write(2, unsafe.Pointer(unsafe.StringData(s)), int32(len(s))) } // auxv is populated on relevant platforms but defined here for all platforms // so x/sys/cpu can assume the getAuxv symbol exists without keeping its list // of auxv-using GOOS build tags in sync. // // It contains an even number of elements, (tag, value) pairs. var auxv []uintptr func getAuxv() []uintptr { return auxv } // accessed from x/sys/cpu; see issue 57336