internal/runtime/atomic.Int32.Load (method)

56 uses

	internal/runtime/atomic (current package)
		types.go#L20: func (i *Int32) Load() int32 {

	runtime
		chan.go#L822: 	async := debug.asynctimerchan.Load() != 0
		chan.go#L840: 		async := debug.asynctimerchan.Load() != 0
		cpuprof.go#L113: 	if prof.hz.Load() != 0 { // implies cpuprof.log != nil
		metrics.go#L289: 				out.scalar = uint64(gcController.gcPercent.Load())
		mgc.go#L625: 		if gcController.gcPercent.Load() < 0 {
		mgcpacer.go#L493: 	gcPercent := c.gcPercent.Load()
		mgcpacer.go#L716: 		if sched.npidle.Load() != 0 && sched.nmspinning.Load() == 0 {
		mgcpacer.go#L1237: 	if gcPercent := c.gcPercent.Load(); gcPercent >= 0 {
		mgcpacer.go#L1282: 	out := c.gcPercent.Load()
		panic.go#L735: 		if debug.panicnil.Load() != 1 {
		proc.go#L409: 	if !gp.preempt && sched.npidle.Load() > 0 {
		proc.go#L3157: 	if sched.nmspinning.Load()+sched.npidle.Load() == 0 && sched.nmspinning.CompareAndSwap(0, 1) { // TODO: fast atomic
		proc.go#L3187: 	if sched.npidle.Load() == gomaxprocs-1 && sched.lastpoll.Load() != 0 {
		proc.go#L3220: 	if sched.nmspinning.Load() != 0 || !sched.nmspinning.CompareAndSwap(0, 1) {
		proc.go#L3505: 	if mp.spinning || 2*sched.nmspinning.Load() < gomaxprocs-sched.npidle.Load() {
		proc.go#L4086: 	npidle := sched.npidle.Load()
		proc.go#L5122: 		for pendingPreemptSignals.Load() > 0 {
		proc.go#L5555: 	n := int32(atomic.Loaduintptr(&allglen)) - sched.gFree.stack.size - sched.gFree.noStack.size - sched.ngsys.Load()
		proc.go#L5593: 	if prof.hz.Load() == 0 {
		proc.go#L5685: 	if prof.hz.Load() != 0 {
		proc.go#L5731: 	if prof.hz.Load() != hz {
		proc.go#L6266: 		if debug.schedtrace <= 0 && (sched.gcwaiting.Load() || sched.npidle.Load() == gomaxprocs) {
		proc.go#L6268: 			if sched.gcwaiting.Load() || sched.npidle.Load() == gomaxprocs {
		proc.go#L6437: 			if runqempty(pp) && sched.nmspinning.Load()+sched.npidle.Load() > 0 && pd.syscallwhen+10*1000*1000 > now {
		proc.go#L6531: 	print("SCHED ", (now-starttime)/1e6, "ms: gomaxprocs=", gomaxprocs, " idleprocs=", sched.npidle.Load(), " threads=", mcount(), " spinningthreads=", sched.nmspinning.Load(), " needspinning=", sched.needspinning.Load(), " idlethreads=", sched.nmidle, " runqueue=", sched.runq.size)
		proc.go#L6799: 		for ; n != 0 && sched.npidle.Load() != 0; n-- {
		proc.go#L7511: 	if i >= active_spin || numCPUStartup <= 1 || gomaxprocs <= sched.npidle.Load()+sched.nmspinning.Load()+1 {
		signal_unix.go#L516: 	if prof.hz.Load() != 0 {
		signal_unix.go#L539: 	if prof.hz.Load() != 0 {
		signal_unix.go#L758: 	if crashing.Load() == 0 {
		signal_unix.go#L768: 		if crashing.Load() > 0 && gp != mp.curg && mp.curg != nil && readgstatus(mp.curg)&^_Gscan == _Grunning {
		signal_unix.go#L772: 		} else if crashing.Load() == 0 {
		signal_unix.go#L789: 		if crashing.Load() < mcount()-int32(extraMLength.Load()) {
		signal_unix.go#L810: 			maxCrashing := crashing.Load()
		signal_unix.go#L811: 			for timeout > 0 && (crashing.Load() < mcount()-int32(extraMLength.Load())) {
		signal_unix.go#L815: 				if c := crashing.Load(); c > maxCrashing {
		signal_unix.go#L823: 			c := crashing.Load()
		signal_unix.go#L827: 				c = crashing.Load()
		synctest.go#L172: 	if debug.asynctimerchan.Load() != 0 {
		time.go#L498: 	async := debug.asynctimerchan.Load() != 0
		time.go#L528: 		if t.period == 0 && t.isSending.Load() > 0 {
		time.go#L579: 	async := debug.asynctimerchan.Load() != 0
		time.go#L655: 		if oldPeriod == 0 && t.isSending.Load() > 0 {
		time.go#L1023: 	zombies := ts.zombies.Load()
		time.go#L1053: 		force = ts == &getg().m.p.ptr().timers && int(ts.zombies.Load()) > int(ts.len.Load())/4
		time.go#L1168: 	async := debug.asynctimerchan.Load() != 0
		trace.go#L260: 	if debug.traceallocfree.Load() != 0 {
		trace.go#L281: 	for trace.exitingSyscall.Load() != 0 {
		traceruntime.go#L579: 	if heapGoal == ^uint64(0) || gcController.gcPercent.Load() < 0 {