internal/runtime/atomic.Uint32.Load (method)
97 uses
internal/runtime/atomic (current package)
types.go#L193: func (u *Uint32) Load() uint32 {
runtime
cgocall.go#L385: if gp.m.needextram || extraMWaiters.Load() > 0 {
debuglog.go#L88: if l1.owned.Load() == 0 && l1.owned.CompareAndSwap(0, 1) {
mcache.go#L331: flushGen := c.flushGen.Load()
mcleanup.go#L506: return q.workUnits.Load() > 0 && (q.asleep.Load() > 0 || q.ng.Load() < maxCleanupGs())
mcleanup.go#L519: asleep := uint64(q.asleep.Load())
mcleanup.go#L535: newg := min(extra, maxCleanupGs()-q.ng.Load())
mcleanup.go#L564: have := q.ng.Load()
mcleanup.go#L572: return q.needg.Load() > 0
mcleanup.go#L577: have := q.ng.Load()
mcleanup.go#L691: if q.flushed.Load() && q.full.empty() && uint32(q.sleeping.size) == q.ng.Load() {
metrics.go#L316: out.scalar = uint64(gcCPULimiter.lastEnabledCycle.Load())
mfinal.go#L171: if fingStatus.Load() == fingUninitialized && fingStatus.CompareAndSwap(fingUninitialized, fingCreated) {
mgc.go#L508: n := work.cycles.Load()
mgc.go#L523: for work.cycles.Load() == n+1 && sweepone() != ^uintptr(0) {
mgc.go#L538: for work.cycles.Load() == n+1 && !isSweepDone() {
mgc.go#L546: cycle := work.cycles.Load()
mgc.go#L559: nMarks := work.cycles.Load()
mgc.go#L617: if !memstats.enablegc || panicking.Load() != 0 || gcphase != _GCoff {
mgc.go#L632: return int32(t.n-work.cycles.Load()) > 0
mgc.go#L716: if fg := p.mcache.flushGen.Load(); fg != mheap_.sweepgen {
mgcpacer.go#L1305: gcWaitOnMark(work.cycles.Load())
mgcsweep.go#L150: state := a.state.Load()
mgcsweep.go#L167: state := a.state.Load()
mgcsweep.go#L195: state := a.state.Load()
mgcsweep.go#L207: return a.state.Load() &^ sweepDrainedMask
mgcsweep.go#L214: return a.state.Load() == sweepDrainedMask
mprof.go#L191: v := c.value.Load()
mprof.go#L200: prev := c.value.Load()
mprof.go#L217: prev := c.value.Load()
mprof.go#L1294: return goroutineProfileState((*atomic.Uint32)(p).Load())
mprof.go#L1328: if fingStatus.Load()&fingRunningFinalizer != 0 {
mprof.go#L1331: n += int(gcCleanups.running.Load())
mspanset.go#L256: if block.popped.Load() == 0 {
mspanset.go#L262: if block.popped.Load() == spanSetBlockEntries {
mstats.go#L787: gen := m.gen.Load() % 3
mstats.go#L856: currGen := m.gen.Load()
mstats.go#L879: for p.statsSeq.Load()%2 != 0 {
netpoll.go#L143: return pollInfo(pd.atomicInfo.Load())
netpoll.go#L167: x := pd.atomicInfo.Load()
netpoll.go#L169: x = pd.atomicInfo.Load()
netpoll.go#L178: x := pd.atomicInfo.Load()
netpoll.go#L184: x = pd.atomicInfo.Load()
netpoll.go#L219: if netpollInited.Load() == 0 {
netpoll.go#L223: if netpollInited.Load() == 0 {
netpoll.go#L232: return netpollInited.Load() != 0
netpoll.go#L678: return netpollWaiters.Load() > 0
preempt.go#L195: if gp.preemptStop && gp.preempt && gp.stackguard0 == stackPreempt && asyncM == gp.m && asyncM.preemptGen.Load() == asyncGen {
preempt.go#L211: asyncGen2 := asyncM2.preemptGen.Load()
print.go#L42: if panicking.Load() == 0 {
proc.go#L311: if runningPanicDefers.Load() != 0 {
proc.go#L314: if runningPanicDefers.Load() == 0 {
proc.go#L320: if panicking.Load() != 0 {
proc.go#L973: count := mcount() - int32(extraMInUse.Load()) - int32(extraMLength.Load())
proc.go#L1205: return gp.atomicstatus.Load()
proc.go#L1289: if oldval == _Gwaiting && gp.atomicstatus.Load() == _Grunnable {
proc.go#L1300: for x := 0; x < 10 && gp.atomicstatus.Load() != oldval; x++ {
proc.go#L2079: if mp.signalPending.Load() != 0 {
proc.go#L2304: wait := freem.freeWait.Load()
proc.go#L2503: } else if extraMLength.Load() == 0 {
proc.go#L3441: if fingStatus.Load()&(fingWait|fingWake) == fingWait|fingWake {
proc.go#L3598: if !mp.spinning && sched.needspinning.Load() == 1 {
proc.go#L5644: if mp.cgoCallersUse.Load() == 0 && mp.cgoCallers != nil && mp.cgoCallers[0] != 0 {
proc.go#L6121: if panicking.Load() > 0 {
proc.go#L6130: if !iscgo && cgoHasExtraM && extraMLength.Load() > 0 {
proc.go#L6353: if scavenger.sysmonWake.Load() != 0 {
proc.go#L6531: print("SCHED ", (now-starttime)/1e6, "ms: gomaxprocs=", gomaxprocs, " idleprocs=", sched.npidle.Load(), " threads=", mcount(), " spinningthreads=", sched.nmspinning.Load(), " needspinning=", sched.needspinning.Load(), " idlethreads=", sched.nmidle, " runqueue=", sched.runq.size)
proc.go#L6962: if pp.timers.len.Load() == 0 {
profbuf.go#L395: if b.eof.Load() > 0 {
profbuf.go#L474: if b.eof.Load() > 0 {
runtime.go#L238: gp == nil && panicking.Load() > 0 {
runtime.go#L260: if panicking.Load() > 0 {
runtime.go#L287: if panicking.Load() > 0 {
sema.go#L214: if root.nwait.Load() == 0 {
sema.go#L220: if root.nwait.Load() == 0 {
sema.go#L619: if l.wait.Load() == atomic.Load(&l.notify) {
sema.go#L634: atomic.Store(&l.notify, l.wait.Load())
sema.go#L656: if l.wait.Load() == atomic.Load(&l.notify) {
sema.go#L664: if t == l.wait.Load() {
signal_unix.go#L789: if crashing.Load() < mcount()-int32(extraMLength.Load()) {
signal_unix.go#L811: for timeout > 0 && (crashing.Load() < mcount()-int32(extraMLength.Load())) {
sigqueue.go#L100: switch sig.state.Load() {
sigqueue.go#L143: switch sig.state.Load() {
sigqueue.go#L185: for sig.delivering.Load() != 0 {
sigqueue.go#L192: for sig.state.Load() != sigReceiving {
symtab.go#L1061: if strict && panicking.Load() == 0 {
symtab.go#L1116: if panicking.Load() != 0 || !strict {
time.go#L454: if netpollInited.Load() == 0 {
time.go#L1027: force := ts == &getg().m.p.ptr().timers && int(zombies) > int(ts.len.Load())/4
time.go#L1053: force = ts == &getg().m.p.ptr().timers && int(ts.zombies.Load()) > int(ts.len.Load())/4
time.go#L1285: if n := int(ts.len.Load()); len(ts.heap) != n {
traceback.go#L1384: return fingStatus.Load()&fingRunningFinalizer == 0
traceback.go#L1669: if panicking.Load() > 0 || getg().m.curg != getg() {
traceback.go#L1689: if panicking.Load() > 0 || getg().m.curg != getg() {
tracestatus.go#L189: return r.statusTraced[gen%3].Load() != 0