func runtime.alignUp

29 uses

	runtime (current package)
		malloc.go#L597: 		p = alignUp(p+(256<<10), heapArenaBytes)
		malloc.go#L632: 	n = alignUp(n, heapArenaBytes)
		malloc.go#L815: 		p = alignUp(p, align)
		malloc.go#L829: 		pAligned := alignUp(p, align)
		malloc.go#L1018: 				off = alignUp(off, 8)
		malloc.go#L1026: 				off = alignUp(off, 8)
		malloc.go#L1028: 				off = alignUp(off, 4)
		malloc.go#L1030: 				off = alignUp(off, 2)
		malloc.go#L1353: 	persistent.off = alignUp(persistent.off, align)
		malloc.go#L1371: 		persistent.off = alignUp(sys.PtrSize, align)
		malloc.go#L1424: 	p := alignUp(l.next, align)
		malloc.go#L1429: 	if pEnd := alignUp(l.next-1, physPageSize); pEnd > l.mapped {
		mfinal.go#L410: 		nret = alignUp(nret, uintptr(t.align)) + uintptr(t.size)
		mfinal.go#L412: 	nret = alignUp(nret, sys.PtrSize)
		mgcscavenge.go#L482: 	p.scav.reservationBytes = alignUp(p.inUse.totalBytes, pallocChunkBytes) / scavengeReservationShards
		mgcscavenge.go#L871: 		max = alignUp(max, min)
		mgcscavenge.go#L934: 		hugePageAbove := uint(alignUp(uintptr(start), pagesPerHugePage))
		mheap.go#L1191: 		base = alignUp(allocBase, physPageSize)
		mheap.go#L1335: 	ask := alignUp(npage, pallocChunkPages) * pageSize
		mheap.go#L1341: 	nBase := alignUp(end, physPageSize)
		mheap.go#L1384: 		nBase = alignUp(h.curArena.base+ask, physPageSize)
		mpagealloc.go#L172: 	return int(alignDown(uintptr(lo), e)), int(alignUp(uintptr(hi), e))
		mpagealloc.go#L356: 	limit := alignUp(base+size, pallocChunkBytes)
		mpagealloc_64bit.go#L77: 		b := alignUp(uintptr(entries)*pallocSumBytes, physPageSize)
		mpagealloc_64bit.go#L118: 		limitOffset := alignUp(uintptr(sumIdxLimit)*pallocSumBytes, physPageSize)
		msize.go#L24: 	return alignUp(size, _PageSize)
		stack.go#L343: 		n = uint32(alignUp(uintptr(n), physPageSize))
		stubs.go#L318: func alignUp(n, a uintptr) uintptr {
		trace.go#L942: 	n = alignUp(n, sys.PtrSize)