internal/runtime/atomic.Int32.Load (method)
56 uses
internal/runtime/atomic (current package)
types.go#L20: func (i *Int32) Load() int32 {
runtime
chan.go#L822: async := debug.asynctimerchan.Load() != 0
chan.go#L840: async := debug.asynctimerchan.Load() != 0
cpuprof.go#L113: if prof.hz.Load() != 0 { // implies cpuprof.log != nil
metrics.go#L289: out.scalar = uint64(gcController.gcPercent.Load())
mgc.go#L625: if gcController.gcPercent.Load() < 0 {
mgcpacer.go#L493: gcPercent := c.gcPercent.Load()
mgcpacer.go#L716: if sched.npidle.Load() != 0 && sched.nmspinning.Load() == 0 {
mgcpacer.go#L1237: if gcPercent := c.gcPercent.Load(); gcPercent >= 0 {
mgcpacer.go#L1282: out := c.gcPercent.Load()
panic.go#L735: if debug.panicnil.Load() != 1 {
proc.go#L409: if !gp.preempt && sched.npidle.Load() > 0 {
proc.go#L3157: if sched.nmspinning.Load()+sched.npidle.Load() == 0 && sched.nmspinning.CompareAndSwap(0, 1) { // TODO: fast atomic
proc.go#L3187: if sched.npidle.Load() == gomaxprocs-1 && sched.lastpoll.Load() != 0 {
proc.go#L3220: if sched.nmspinning.Load() != 0 || !sched.nmspinning.CompareAndSwap(0, 1) {
proc.go#L3505: if mp.spinning || 2*sched.nmspinning.Load() < gomaxprocs-sched.npidle.Load() {
proc.go#L4086: npidle := sched.npidle.Load()
proc.go#L5122: for pendingPreemptSignals.Load() > 0 {
proc.go#L5555: n := int32(atomic.Loaduintptr(&allglen)) - sched.gFree.stack.size - sched.gFree.noStack.size - sched.ngsys.Load()
proc.go#L5593: if prof.hz.Load() == 0 {
proc.go#L5685: if prof.hz.Load() != 0 {
proc.go#L5731: if prof.hz.Load() != hz {
proc.go#L6266: if debug.schedtrace <= 0 && (sched.gcwaiting.Load() || sched.npidle.Load() == gomaxprocs) {
proc.go#L6268: if sched.gcwaiting.Load() || sched.npidle.Load() == gomaxprocs {
proc.go#L6437: if runqempty(pp) && sched.nmspinning.Load()+sched.npidle.Load() > 0 && pd.syscallwhen+10*1000*1000 > now {
proc.go#L6531: print("SCHED ", (now-starttime)/1e6, "ms: gomaxprocs=", gomaxprocs, " idleprocs=", sched.npidle.Load(), " threads=", mcount(), " spinningthreads=", sched.nmspinning.Load(), " needspinning=", sched.needspinning.Load(), " idlethreads=", sched.nmidle, " runqueue=", sched.runq.size)
proc.go#L6799: for ; n != 0 && sched.npidle.Load() != 0; n-- {
proc.go#L7511: if i >= active_spin || numCPUStartup <= 1 || gomaxprocs <= sched.npidle.Load()+sched.nmspinning.Load()+1 {
signal_unix.go#L516: if prof.hz.Load() != 0 {
signal_unix.go#L539: if prof.hz.Load() != 0 {
signal_unix.go#L758: if crashing.Load() == 0 {
signal_unix.go#L768: if crashing.Load() > 0 && gp != mp.curg && mp.curg != nil && readgstatus(mp.curg)&^_Gscan == _Grunning {
signal_unix.go#L772: } else if crashing.Load() == 0 {
signal_unix.go#L789: if crashing.Load() < mcount()-int32(extraMLength.Load()) {
signal_unix.go#L810: maxCrashing := crashing.Load()
signal_unix.go#L811: for timeout > 0 && (crashing.Load() < mcount()-int32(extraMLength.Load())) {
signal_unix.go#L815: if c := crashing.Load(); c > maxCrashing {
signal_unix.go#L823: c := crashing.Load()
signal_unix.go#L827: c = crashing.Load()
synctest.go#L172: if debug.asynctimerchan.Load() != 0 {
time.go#L498: async := debug.asynctimerchan.Load() != 0
time.go#L528: if t.period == 0 && t.isSending.Load() > 0 {
time.go#L579: async := debug.asynctimerchan.Load() != 0
time.go#L655: if oldPeriod == 0 && t.isSending.Load() > 0 {
time.go#L1023: zombies := ts.zombies.Load()
time.go#L1053: force = ts == &getg().m.p.ptr().timers && int(ts.zombies.Load()) > int(ts.len.Load())/4
time.go#L1168: async := debug.asynctimerchan.Load() != 0
trace.go#L260: if debug.traceallocfree.Load() != 0 {
trace.go#L281: for trace.exitingSyscall.Load() != 0 {
traceruntime.go#L579: if heapGoal == ^uint64(0) || gcController.gcPercent.Load() < 0 {