func unsafe.Sizeof
227 uses
unsafe (current package)
unsafe.go#L198: func Sizeof(x ArbitraryType) uintptr
golang.org/x/sys/unix
affinity_linux.go#L20: _, _, e := RawSyscall(trap, uintptr(pid), uintptr(unsafe.Sizeof(*set)), uintptr(unsafe.Pointer(set)))
syscall_linux.go#L840: copy(out[:], (*(*[unsafe.Sizeof(TIPCSocketAddr{})]byte)(unsafe.Pointer(sa)))[:])
syscall_linux.go#L848: copy(out[:], (*(*[unsafe.Sizeof(TIPCServiceRange{})]byte)(unsafe.Pointer(sa)))[:])
syscall_linux.go#L856: copy(out[:], (*(*[unsafe.Sizeof(TIPCServiceName{})]byte)(unsafe.Pointer(sa)))[:])
syscall_linux.go#L1374: return setsockopt(fd, level, opt, unsafe.Pointer(mreq), unsafe.Sizeof(*mreq))
syscall_linux.go#L1378: return setsockopt(fd, level, opt, unsafe.Pointer(mreq), unsafe.Sizeof(*mreq))
syscall_linux.go#L1384: return setsockopt(fd, level, opt, unsafe.Pointer(fprog), unsafe.Sizeof(*fprog))
syscall_linux.go#L1396: return setsockopt(fd, level, opt, unsafe.Pointer(tp), unsafe.Sizeof(*tp))
syscall_linux.go#L1400: return setsockopt(fd, level, opt, unsafe.Pointer(tp), unsafe.Sizeof(*tp))
syscall_linux.go#L1411: return setsockopt(fd, level, opt, unsafe.Pointer(s), unsafe.Sizeof(*s))
syscall_linux.go#L1753: iov.SetLen(int(unsafe.Sizeof(*regsout)))
syscall_linux.go#L1760: iov.SetLen(int(unsafe.Sizeof(*regs)))
syscall_linux.go#L1800: return readInt(buf, unsafe.Offsetof(Dirent{}.Ino), unsafe.Sizeof(Dirent{}.Ino))
syscall_linux.go#L1804: return readInt(buf, unsafe.Offsetof(Dirent{}.Reclen), unsafe.Sizeof(Dirent{}.Reclen))
syscall_linux.go#L1837: return mountSetattr(dirfd, pathname, flags, attr, unsafe.Sizeof(*attr))
syscall_linux.go#L2433: const hdrSize = unsafe.Sizeof(fileHandle{})
syscall_linux.go#L2458: size := uint32(32 + unsafe.Sizeof(fileHandle{}))
syscall_linux.go#L2463: fh.Bytes = size - uint32(unsafe.Sizeof(fileHandle{}))
syscall_linux.go#L2471: size = fh.Bytes + uint32(unsafe.Sizeof(fileHandle{}))
syscall_unix.go#L326: vallen := _Socklen(unsafe.Sizeof(tv))
syscall_unix.go#L526: return setsockopt(fd, level, opt, unsafe.Pointer(tv), unsafe.Sizeof(*tv))
google.golang.org/protobuf/internal/impl
pointer_unsafe.go#L187: const _ = uint(unsafe.Sizeof(unsafe.Pointer(nil)) - unsafe.Sizeof(MessageState{}))
presence.go#L24: siz = unsafe.Sizeof(*ret)
presence.go#L73: o := uintptr(num/32) * unsafe.Sizeof(uint32(0))
presence.go#L85: o := j * unsafe.Sizeof(uint32(0))
presence.go#L137: o := (uintptr(unsafe.Pointer(&template.a)) - uintptr(unsafe.Pointer(&template.d))) + uintptr(num/32)*unsafe.Sizeof(uint32(0))
internal/abi
type.go#L520: uadd := unsafe.Sizeof(*t)
type.go#L522: uadd += unsafe.Sizeof(UncommonType{})
type.go#L534: uadd := unsafe.Sizeof(*t)
type.go#L536: uadd += unsafe.Sizeof(UncommonType{})
internal/runtime/maps
group.go#L255: ctrlGroupsSize = unsafe.Sizeof(ctrlGroup(0))
internal/saferio
io.go#L130: size := uint64(unsafe.Sizeof(v))
net
cgo_unix.go#L321: resStateSize := unsafe.Sizeof(_C_struct___res_state{})
os
dirent_linux.go#L13: return readInt(buf, unsafe.Offsetof(syscall.Dirent{}.Ino), unsafe.Sizeof(syscall.Dirent{}.Ino))
dirent_linux.go#L17: return readInt(buf, unsafe.Offsetof(syscall.Dirent{}.Reclen), unsafe.Sizeof(syscall.Dirent{}.Reclen))
reflect
type.go#L1132: if unsafe.Sizeof(int(0)) == 4 && goarch.BigEndian {
type.go#L2417: ut.Moff = uint32(unsafe.Sizeof(uncommonType{}))
runtime
alg.go#L40: size := *(*uintptr)(unsafe.Pointer(ptr + unsafe.Sizeof(h)))
arena.go#L231: return userArenaChunkBytes/goarch.PtrSize/8 + unsafe.Sizeof(_type{})
arena.go#L1119: *(*uintptr)(unsafe.Pointer(&s.largeType.GCData)) = s.limit + unsafe.Sizeof(_type{})
cgo_sigaction.go#L29: msanwrite(unsafe.Pointer(new), unsafe.Sizeof(*new))
cgo_sigaction.go#L32: asanwrite(unsafe.Pointer(new), unsafe.Sizeof(*new))
cgo_sigaction.go#L83: msanread(unsafe.Pointer(old), unsafe.Sizeof(*old))
cgo_sigaction.go#L86: asanread(unsafe.Pointer(old), unsafe.Sizeof(*old))
chan.go#L30: hchanSize = unsafe.Sizeof(hchan{}) + uintptr(-int(unsafe.Sizeof(hchan{}))&(maxAlign-1))
covercounter.go#L15: u32sz := unsafe.Sizeof(uint32(0))
debuglog.go#L99: l = (*dloggerImpl)(sysAllocOS(unsafe.Sizeof(dloggerImpl{}), "debug log"))
debuglog.go#L768: state1 := sysAllocOS(unsafe.Sizeof(readState{})*uintptr(n), "debug log")
defs_linux_amd64.go#L136: _ [_si_max_size - unsafe.Sizeof(siginfoFields{})]byte
defs_linux_amd64.go#L161: _ [_sigev_max_size - unsafe.Sizeof(sigeventFields{})]byte
heapdump.go#L681: memclrNoHeapPointers(unsafe.Pointer(&typecache), unsafe.Sizeof(typecache))
iface.go#L77: m = (*itab)(persistentalloc(unsafe.Sizeof(itab{})+uintptr(len(inter.Methods)-1)*goarch.PtrSize, 0, &memstats.other_sys))
iface.go#L423: x = mallocgc(unsafe.Sizeof(val), stringType, true)
iface.go#L443: x = mallocgc(unsafe.Sizeof(val), sliceType, true)
iface.go#L525: newSize := unsafe.Sizeof(abi.TypeAssertCache{}) + uintptr(newN-1)*unsafe.Sizeof(abi.TypeAssertCacheEntry{})
iface.go#L631: newSize := unsafe.Sizeof(abi.InterfaceSwitchCache{}) + uintptr(newN-1)*unsafe.Sizeof(abi.InterfaceSwitchCacheEntry{})
lock_spinbit.go#L95: size := roundupsize(unsafe.Sizeof(mPadded{}), false) + gc.MallocHeaderSize
malloc.go#L574: const arenaMetaSize = (1 << arenaBits) * unsafe.Sizeof(heapArena{})
malloc.go#L774: l2 = (*[1 << arenaL2Bits]*heapArena)(sysAllocOS(unsafe.Sizeof(*l2), "heap index"))
malloc.go#L779: sysHugePage(unsafe.Pointer(l2), unsafe.Sizeof(*l2))
malloc.go#L781: sysNoHugePage(unsafe.Pointer(l2), unsafe.Sizeof(*l2))
malloc.go#L790: r = (*heapArena)(h.heapArenaAlloc.alloc(unsafe.Sizeof(*r), goarch.PtrSize, &memstats.gcMiscSys, "heap metadata"))
malloc.go#L792: r = (*heapArena)(persistentalloc(unsafe.Sizeof(*r), goarch.PtrSize, &memstats.gcMiscSys))
malloc.go#L922: sysHugePage(unsafe.Pointer(l2), unsafe.Sizeof(*l2))
mbitmap.go#L574: base -= unsafe.Sizeof(spanInlineMarkBits{})
mcheckmark.go#L48: bitmap = (*checkmarksMap)(persistentalloc(unsafe.Sizeof(*bitmap), 0, &memstats.gcMiscSys))
mcleanup.go#L219: cleanups [(cleanupBlockSize - unsafe.Sizeof(cleanupBlockHeader{})) / goarch.PtrSize]*funcval
mcleanup.go#L286: _ [cpu.CacheLinePadSize - unsafe.Sizeof(lfstack(0)) - unsafe.Sizeof(atomic.Uint64{})]byte
mcleanup.go#L297: _ [cpu.CacheLinePadSize - unsafe.Sizeof(lfstack(0)) - 1]byte
mcleanup.go#L301: _ [cpu.CacheLinePadSize - unsafe.Sizeof(atomic.UnsafePointer{})]byte
mfinal.go#L32: fin [(finBlockSize - 2*goarch.PtrSize - 2*4) / unsafe.Sizeof(finalizer{})]finalizer
mfinal.go#L122: if (unsafe.Sizeof(finalizer{}) != 5*goarch.PtrSize ||
mfinal.go#L232: framesz := unsafe.Sizeof((any)(nil)) + f.nret
mfixalloc.go#L60: size = max(size, unsafe.Sizeof(mlink{}))
mgc.go#L180: if unsafe.Sizeof(workbuf{}) != _WorkbufSize {
mgc.go#L1423: pad [tagAlign - unsafe.Sizeof(gcBgMarkWorkerNode{}) - gcBgMarkWorkerNodeRedZoneSize]byte
mgcmark.go#L183: scanblock(uintptr(unsafe.Pointer(&fb.fin[0])), cnt*unsafe.Sizeof(fb.fin[0]), &finptrmask[0], gcw, nil)
mgcstack.go#L110: obj [(_WorkbufSize - unsafe.Sizeof(stackWorkBufHdr{})) / goarch.PtrSize]uintptr
mgcstack.go#L128: obj [(_WorkbufSize - unsafe.Sizeof(stackObjectBufHdr{})) / unsafe.Sizeof(stackObject{})]stackObject
mgcstack.go#L138: if unsafe.Sizeof(stackWorkBuf{}) > unsafe.Sizeof(workbuf{}) {
mgcstack.go#L141: if unsafe.Sizeof(stackObjectBuf{}) > unsafe.Sizeof(workbuf{}) {
mgcwork.go#L386: obj [(_WorkbufSize - unsafe.Sizeof(workbufhdr{})) / goarch.PtrSize]uintptr
mgcwork.go#L501: memmove(unsafe.Pointer(&b1.obj[0]), unsafe.Pointer(&b.obj[b.nobj]), uintptr(n)*unsafe.Sizeof(b1.obj[0]))
mheap.go#L213: pad [(cpu.CacheLinePadSize - unsafe.Sizeof(mcentral{})%cpu.CacheLinePadSize) % cpu.CacheLinePadSize]byte
mheap.go#L573: sysFree(unsafe.Pointer(&oldAllspans[0]), uintptr(cap(oldAllspans))*unsafe.Sizeof(oldAllspans[0]), &memstats.other_sys)
mheap.go#L795: h.spanalloc.init(unsafe.Sizeof(mspan{}), recordspan, unsafe.Pointer(h), &memstats.mspan_sys)
mheap.go#L796: h.cachealloc.init(unsafe.Sizeof(mcache{}), nil, nil, &memstats.mcache_sys)
mheap.go#L797: h.specialfinalizeralloc.init(unsafe.Sizeof(specialfinalizer{}), nil, nil, &memstats.other_sys)
mheap.go#L798: h.specialCleanupAlloc.init(unsafe.Sizeof(specialCleanup{}), nil, nil, &memstats.other_sys)
mheap.go#L799: h.specialCheckFinalizerAlloc.init(unsafe.Sizeof(specialCheckFinalizer{}), nil, nil, &memstats.other_sys)
mheap.go#L800: h.specialTinyBlockAlloc.init(unsafe.Sizeof(specialTinyBlock{}), nil, nil, &memstats.other_sys)
mheap.go#L801: h.specialprofilealloc.init(unsafe.Sizeof(specialprofile{}), nil, nil, &memstats.other_sys)
mheap.go#L802: h.specialReachableAlloc.init(unsafe.Sizeof(specialReachable{}), nil, nil, &memstats.other_sys)
mheap.go#L803: h.specialPinCounterAlloc.init(unsafe.Sizeof(specialPinCounter{}), nil, nil, &memstats.other_sys)
mheap.go#L804: h.specialWeakHandleAlloc.init(unsafe.Sizeof(specialWeakHandle{}), nil, nil, &memstats.gcMiscSys)
mheap.go#L805: h.specialBubbleAlloc.init(unsafe.Sizeof(specialBubble{}), nil, nil, &memstats.other_sys)
mheap.go#L806: h.arenaHintAlloc.init(unsafe.Sizeof(arenaHint{}), nil, nil, &memstats.other_sys)
mheap.go#L1465: reserve += unsafe.Sizeof(spanInlineMarkBits{})
mheap.go#L2722: newNode = (*immortalWeakHandle)(persistentalloc(unsafe.Sizeof(immortalWeakHandle{}), goarch.PtrSize, &memstats.gcMiscSys))
mheap.go#L2880: const gcBitsHeaderBytes = unsafe.Sizeof(gcBitsHeader{})
minmax.go#L55: switch unsafe.Sizeof(x) {
minmax.go#L65: switch unsafe.Sizeof(x) {
mpagealloc.go#L406: const l2Size = unsafe.Sizeof(*p.chunks[0])
mpagealloc.go#L475: sysHugePage(unsafe.Pointer(p.chunks[i]), unsafe.Sizeof(*p.chunks[0]))
mpagealloc.go#L976: pallocSumBytes = unsafe.Sizeof(pallocSum(0))
mpagealloc_64bit.go#L196: scSize := unsafe.Sizeof(atomicScavChunkData{})
mpagealloc_64bit.go#L250: nbytes := n * unsafe.Sizeof(atomicScavChunkData{})
mpagecache.go#L12: const pageCachePages = 8 * unsafe.Sizeof(pageCache{}.cache)
mprof.go#L229: size := unsafe.Sizeof(bucket{}) + uintptr(nstk)*unsafe.Sizeof(uintptr(0))
mprof.go#L234: size += unsafe.Sizeof(memRecord{})
mprof.go#L236: size += unsafe.Sizeof(blockRecord{})
mprof.go#L248: stk := (*[maxProfStackDepth]uintptr)(add(unsafe.Pointer(b), unsafe.Sizeof(*b)))
mprof.go#L261: data := add(unsafe.Pointer(b), unsafe.Sizeof(*b)+b.nstk*unsafe.Sizeof(uintptr(0)))
mprof.go#L270: data := add(unsafe.Pointer(b), unsafe.Sizeof(*b)+b.nstk*unsafe.Sizeof(uintptr(0)))
mprof.go#L282: bh = (*buckhashArray)(sysAlloc(unsafe.Sizeof(buckhashArray{}), &memstats.buckhash_sys, "profiler hash buckets"))
mprof.go#L1019: racewriterangepc(unsafe.Pointer(&dst.Stack0[0]), unsafe.Sizeof(dst.Stack0), sys.GetCallerPC(), abi.FuncPCABIInternal(MemProfile))
mprof.go#L1022: msanwrite(unsafe.Pointer(&dst.Stack0[0]), unsafe.Sizeof(dst.Stack0))
mprof.go#L1025: asanwrite(unsafe.Pointer(&dst.Stack0[0]), unsafe.Sizeof(dst.Stack0))
mprof.go#L1133: racewriterangepc(unsafe.Pointer(&dst.Stack0[0]), unsafe.Sizeof(dst.Stack0), sys.GetCallerPC(), abi.FuncPCABIInternal(BlockProfile))
mprof.go#L1136: msanwrite(unsafe.Pointer(&dst.Stack0[0]), unsafe.Sizeof(dst.Stack0))
mprof.go#L1139: asanwrite(unsafe.Pointer(&dst.Stack0[0]), unsafe.Sizeof(dst.Stack0))
mranges.go#L258: ranges.array = (*notInHeap)(persistentalloc(unsafe.Sizeof(addrRange{})*uintptr(ranges.cap), goarch.PtrSize, sysStat))
mranges.go#L385: ranges.array = (*notInHeap)(persistentalloc(unsafe.Sizeof(addrRange{})*uintptr(ranges.cap), goarch.PtrSize, a.sysStat))
mranges.go#L455: ranges.array = (*notInHeap)(persistentalloc(unsafe.Sizeof(addrRange{})*uintptr(ranges.cap), goarch.PtrSize, b.sysStat))
mspanset.go#L71: pad [tagAlign - unsafe.Sizeof(spanSetBlockHeader{})]byte
mspanset.go#L325: return (*spanSetBlock)(persistentalloc(unsafe.Sizeof(spanSetBlock{}), max(cpu.CacheLineSize, tagAlign), &memstats.gcMiscSys))
mstats.go#L344: if size := unsafe.Sizeof(heapStatsDelta{}); size%8 != 0 {
mwbbuf.go#L87: b.end = start + uintptr(len(b.buf))*unsafe.Sizeof(b.buf[0])
mwbbuf.go#L90: if (b.end-b.next)%unsafe.Sizeof(b.buf[0]) != 0 {
mwbbuf.go#L198: n := (pp.wbBuf.next - start) / unsafe.Sizeof(pp.wbBuf.buf[0])
netpoll.go#L693: pad [tagAlign - unsafe.Sizeof(pollDesc{})]byte
netpoll.go#L695: const pdSize = unsafe.Sizeof(pollDescPadded{})
netpoll_epoll.go#L74: oneSize := int32(unsafe.Sizeof(one))
netpoll_epoll.go#L152: read(int32(netpollEventFd), noescape(unsafe.Pointer(&one)), int32(unsafe.Sizeof(one)))
os_linux.go#L114: r := sched_getaffinity(0, unsafe.Sizeof(buf), &buf[0])
os_linux.go#L286: n = read(fd, noescape(unsafe.Pointer(&auxvreadbuf[0])), int32(unsafe.Sizeof(auxvreadbuf)))
os_linux.go#L454: rtsigprocmask(how, new, old, int32(unsafe.Sizeof(*new)))
os_linux.go#L540: if rt_sigaction(uintptr(sig), new, old, unsafe.Sizeof(sigactiont{}.sa_mask)) != 0 {
panic.go#L688: fd = add(fd, unsafe.Sizeof(b))
pinner.go#L83: pinnerRefStoreSize = (pinnerSize - unsafe.Sizeof([]unsafe.Pointer{})) / unsafe.Sizeof(unsafe.Pointer(nil))
print.go#L272: minhexdigits = int(unsafe.Sizeof(uintptr(0)) * 2)
proc.go#L2921: msanwrite(unsafe.Pointer(&ts), unsafe.Sizeof(ts))
proc.go#L2924: asanwrite(unsafe.Pointer(&ts), unsafe.Sizeof(ts))
proc.go#L5210: memclrNoHeapPointers(unsafe.Pointer(&newg.sched), unsafe.Sizeof(newg.sched))
runtime1.go#L169: if unsafe.Sizeof(a) != 1 {
runtime1.go#L172: if unsafe.Sizeof(b) != 1 {
runtime1.go#L175: if unsafe.Sizeof(c) != 2 {
runtime1.go#L178: if unsafe.Sizeof(d) != 2 {
runtime1.go#L181: if unsafe.Sizeof(e) != 4 {
runtime1.go#L184: if unsafe.Sizeof(f) != 4 {
runtime1.go#L187: if unsafe.Sizeof(g) != 8 {
runtime1.go#L190: if unsafe.Sizeof(h) != 8 {
runtime1.go#L193: if unsafe.Sizeof(i) != 4 {
runtime1.go#L196: if unsafe.Sizeof(j) != 8 {
runtime1.go#L199: if unsafe.Sizeof(k) != goarch.PtrSize {
runtime1.go#L202: if unsafe.Sizeof(l) != goarch.PtrSize {
runtime1.go#L205: if unsafe.Sizeof(x1) != 1 {
runtime1.go#L211: if unsafe.Sizeof(y1) != 2 {
runtime2.go#L639: _ [(1 - goarch.IsWasm) * (2048 - mallocHeaderSize - mRedZoneSize - unsafe.Sizeof(m{}))]byte
sema.go#L53: pad [cpu.CacheLinePadSize - unsafe.Sizeof(semaRoot{})]byte
sema.go#L711: if sz != unsafe.Sizeof(notifyList{}) {
sema.go#L712: print("runtime: bad notifyList size - sync=", sz, " runtime=", unsafe.Sizeof(notifyList{}), "\n")
stack.go#L153: _ [(cpu.CacheLinePadSize - unsafe.Sizeof(stackpoolItem{})%cpu.CacheLinePadSize) % cpu.CacheLinePadSize]byte
stack.go#L219: valgrindMalloc(unsafe.Pointer(x.ptr()), unsafe.Sizeof(x.ptr()))
stack.go#L514: valgrindMalloc(unsafe.Pointer(x.ptr()), unsafe.Sizeof(x.ptr()))
stack.go#L527: valgrindMalloc(unsafe.Pointer(x.ptr()), unsafe.Sizeof(x.ptr()))
stack.go#L624: valgrindMakeMemDefined(unsafe.Pointer(&p), unsafe.Sizeof(&p))
stkframe.go#L238: unsafe.Sizeof(abi.RegArgs{}) > 0 && isReflect {
string.go#L257: uintptr(len(a))*unsafe.Sizeof(a[0]),
string.go#L262: msanread(unsafe.Pointer(&a[0]), uintptr(len(a))*unsafe.Sizeof(a[0]))
string.go#L265: asanread(unsafe.Pointer(&a[0]), uintptr(len(a))*unsafe.Sizeof(a[0]))
symtab.go#L936: ffb := (*findfuncbucket)(add(unsafe.Pointer(datap.findfunctab), b*unsafe.Sizeof(findfuncbucket{})))
symtab.go#L1229: return *(*uint32)(add(unsafe.Pointer(&f.nfuncdata), unsafe.Sizeof(f.nfuncdata)+uintptr(table)*4))
symtab.go#L1263: p := uintptr(unsafe.Pointer(&f.nfuncdata)) + unsafe.Sizeof(f.nfuncdata) + uintptr(f.npcdata)*4 + uintptr(i)*4
trace.go#L680: sysFree(unsafe.Pointer(buf), unsafe.Sizeof(*buf), &memstats.other_sys)
traceback.go#L1675: msanwrite(unsafe.Pointer(arg), unsafe.Sizeof(cgoSymbolizerArg{}))
traceback.go#L1678: asanwrite(unsafe.Pointer(arg), unsafe.Sizeof(cgoSymbolizerArg{}))
traceback.go#L1700: msanwrite(unsafe.Pointer(&arg), unsafe.Sizeof(arg))
traceback.go#L1703: asanwrite(unsafe.Pointer(&arg), unsafe.Sizeof(arg))
tracebuf.go#L168: w.traceBuf = (*traceBuf)(sysAlloc(unsafe.Sizeof(traceBuf{}), &memstats.other_sys, "trace buffer"))
tracebuf.go#L272: arr [tracev2.MaxBatchSize - unsafe.Sizeof(traceBufHeader{})]byte // underlying buffer for traceBufHeader.buf
tracemap.go#L125: meta := (*traceMapNode)(unsafe.Pointer(tab.mem.alloc(unsafe.Sizeof(traceMapNode{}))))
traceregion.go#L40: const traceRegionAllocBlockData = 64<<10 - unsafe.Sizeof(traceRegionAllocBlockHeader{})
traceregion.go#L85: block = (*traceRegionAllocBlock)(sysAlloc(unsafe.Sizeof(traceRegionAllocBlock{}), &memstats.other_sys, "trace arena alloc"))
traceregion.go#L111: sysFree(unsafe.Pointer(block), unsafe.Sizeof(traceRegionAllocBlock{}), &memstats.other_sys)
traceregion.go#L114: sysFree(current, unsafe.Sizeof(traceRegionAllocBlock{}), &memstats.other_sys)
tracestack.go#L151: id, _ := t.tab.put(noescape(unsafe.Pointer(&pcs[0])), uintptr(len(pcs))*unsafe.Sizeof(uintptr(0)))
tracestack.go#L169: stack := unsafe.Slice((*uintptr)(unsafe.Pointer(&node.data[0])), uintptr(len(node.data))/unsafe.Sizeof(uintptr(0)))
vdso_linux.go#L53: vdsoSymTabSize = vdsoArrayMax / unsafe.Sizeof(elfSym{})
vdso_linux.go#L54: vdsoDynSize = vdsoArrayMax / unsafe.Sizeof(elfDyn{})
vdso_linux.go#L61: vdsoBloomSizeScale = unsafe.Sizeof(uintptr(0)) / 4 // uint32
vdso_linux.go#L115: pt := (*elfPhdr)(add(pt, uintptr(i)*unsafe.Sizeof(elfPhdr{})))
runtime/pprof
map.go#L32: h = h<<8 | (h >> (8 * (unsafe.Sizeof(h) - 1)))
map.go#L35: h = h<<8 | (h >> (8 * (unsafe.Sizeof(h) - 1)))
slices
slices.go#L457: elemSize := unsafe.Sizeof(a[0])
sync
pool.go#L77: pad [128 - unsafe.Sizeof(poolLocalInternal{})%128]byte
pool.go#L301: lp := unsafe.Pointer(uintptr(l) + uintptr(i)*unsafe.Sizeof(poolLocal{}))
runtime.go#L55: runtime_notifyListCheck(unsafe.Sizeof(n))
syscall
exec_linux.go#L165: RawSyscall(SYS_WRITE, uintptr(mapPipe[1]), uintptr(unsafe.Pointer(&err2)), unsafe.Sizeof(err2))
exec_linux.go#L332: pid, err1 = rawVforkSyscall(_SYS_clone3, uintptr(unsafe.Pointer(clone3)), unsafe.Sizeof(*clone3), 0)
exec_linux.go#L368: pid, _, err1 = RawSyscall(SYS_READ, uintptr(mapPipe[0]), uintptr(unsafe.Pointer(&err2)), unsafe.Sizeof(err2))
exec_linux.go#L372: if pid != unsafe.Sizeof(err2) {
exec_linux.go#L673: RawSyscall(SYS_WRITE, uintptr(pipe), uintptr(unsafe.Pointer(&err1)), unsafe.Sizeof(err1))
exec_unix.go#L220: n, err = readlen(p[0], (*byte)(unsafe.Pointer(&err1)), int(unsafe.Sizeof(err1)))
exec_unix.go#L227: if n == int(unsafe.Sizeof(err1)) {
lsf_linux.go#L79: return setsockopt(fd, SOL_SOCKET, SO_ATTACH_FILTER, unsafe.Pointer(&p), unsafe.Sizeof(p))
lsf_linux.go#L85: return setsockopt(fd, SOL_SOCKET, SO_DETACH_FILTER, unsafe.Pointer(&dummy), unsafe.Sizeof(dummy))
sockcmsg_linux.go#L34: if uintptr(len(m.Data)) < unsafe.Sizeof(Ucred{}) {
syscall_linux.go#L767: return setsockopt(fd, level, opt, unsafe.Pointer(mreq), unsafe.Sizeof(*mreq))
syscall_linux.go#L965: iov.SetLen(int(unsafe.Sizeof(*regsout)))
syscall_linux.go#L972: iov.SetLen(int(unsafe.Sizeof(*regs)))
syscall_linux.go#L1012: return readInt(buf, unsafe.Offsetof(Dirent{}.Ino), unsafe.Sizeof(Dirent{}.Ino))
syscall_linux.go#L1016: return readInt(buf, unsafe.Offsetof(Dirent{}.Reclen), unsafe.Sizeof(Dirent{}.Reclen))
syscall_unix.go#L496: return setsockopt(fd, level, opt, unsafe.Pointer(tv), unsafe.Sizeof(*tv))