Copyright 2019 The Go Authors. All rights reserved. Use of this source code is governed by a BSD-style license that can be found in the LICENSE file.
Page allocator. The page allocator manages mapped pages (defined by pageSize, NOT physPageSize) for allocation and re-use. It is embedded into mheap. Pages are managed using a bitmap that is sharded into chunks. In the bitmap, 1 means in-use, and 0 means free. The bitmap spans the process's address space. Chunks are managed in a sparse-array-style structure similar to mheap.arenas, since the bitmap may be large on some systems. The bitmap is efficiently searched by using a radix tree in combination with fast bit-wise intrinsics. Allocation is performed using an address-ordered first-fit approach. Each entry in the radix tree is a summary that describes three properties of a particular region of the address space: the number of contiguous free pages at the start and end of the region it represents, and the maximum number of contiguous free pages found anywhere in that region. Each level of the radix tree is stored as one contiguous array, which represents a different granularity of subdivision of the processes' address space. Thus, this radix tree is actually implicit in these large arrays, as opposed to having explicit dynamically-allocated pointer-based node structures. Naturally, these arrays may be quite large for system with large address spaces, so in these cases they are mapped into memory as needed. The leaf summaries of the tree correspond to a bitmap chunk. The root level (referred to as L0 and index 0 in pageAlloc.summary) has each summary represent the largest section of address space (16 GiB on 64-bit systems), with each subsequent level representing successively smaller subsections until we reach the finest granularity at the leaves, a chunk. More specifically, each summary in each level (except for leaf summaries) represents some number of entries in the following level. For example, each summary in the root level may represent a 16 GiB region of address space, and in the next level there could be 8 corresponding entries which represent 2 GiB subsections of that 16 GiB region, each of which could correspond to 8 entries in the next level which each represent 256 MiB regions, and so on. Thus, this design only scales to heaps so large, but can always be extended to larger heaps by simply adding levels to the radix tree, which mostly costs additional virtual address space. The choice of managing large arrays also means that a large amount of virtual address space may be reserved by the runtime.

package runtime

import (
	
	
)

The size of a bitmap chunk, i.e. the amount of bits (that is, pages) to consider in the bitmap at once.
The number of radix bits for each level. The value of 3 is chosen such that the block of summaries we need to scan at each level fits in 64 bytes (2^3 summaries * 8 bytes per summary), which is close to the L1 cache line width on many systems. Also, a value of 3 fits 4 tree levels perfectly into the 21-bit pallocBits summary field at the root level. The following equation explains how each of the constants relate: summaryL0Bits + (summaryLevels-1)*summaryLevelBits + logPallocChunkBytes = heapAddrBits summaryLevels is an architecture-dependent value defined in mpagealloc_*.go.
pallocChunksL2Bits is the number of bits of the chunk index number covered by the second level of the chunks map. See (*pageAlloc).chunks for more details. Update the documentation there should this change.
Maximum searchAddr value, which indicates that the heap has no free space. We alias maxOffAddr just to make it clear that this is the maximum address for the page allocator's search space. See maxOffAddr for details.
Global chunk index. Represents an index into the leaf level of the radix tree. Similar to arenaIndex, except instead of arenas, it divides the address space into chunks.
chunkIndex returns the global index of the palloc chunk containing the pointer p.
chunkIndex returns the base address of the palloc chunk at index ci.
chunkPageIndex computes the index of the page that contains p, relative to the chunk which contains p.
l1 returns the index into the first level of (*pageAlloc).chunks.
func ( chunkIdx) () uint {
Let the compiler optimize this away if there's no L1 map.
		return 0
	} else {
		return uint() >> pallocChunksL1Shift
	}
}
l2 returns the index into the second level of (*pageAlloc).chunks.
func ( chunkIdx) () uint {
	if pallocChunksL1Bits == 0 {
		return uint()
	} else {
		return uint() & (1<<pallocChunksL2Bits - 1)
	}
}
offAddrToLevelIndex converts an address in the offset address space to the index into summary[level] containing addr.
func ( int,  offAddr) int {
	return int((.a - arenaBaseOffset) >> levelShift[])
}
levelIndexToOffAddr converts an index into summary[level] into the corresponding address in the offset address space.
func (,  int) offAddr {
	return offAddr{(uintptr() << levelShift[]) + arenaBaseOffset}
}
addrsToSummaryRange converts base and limit pointers into a range of entries for the given summary level. The returned range is inclusive on the lower bound and exclusive on the upper bound.
This is slightly more nuanced than just a shift for the exclusive upper-bound. Note that the exclusive upper bound may be within a summary at this level, meaning if we just do the obvious computation hi will end up being an inclusive upper bound. Unfortunately, just adding 1 to that is too broad since we might be on the very edge of of a summary's max page count boundary for this level (1 << levelLogPages[level]). So, make limit an inclusive upper bound then shift, then add 1, so we get an exclusive upper bound at the end.
	 = int(( - arenaBaseOffset) >> levelShift[])
	 = int(((-1)-arenaBaseOffset)>>levelShift[]) + 1
	return
}
blockAlignSummaryRange aligns indices into the given level to that level's block width (1 << levelBits[level]). It assumes lo is inclusive and hi is exclusive, and so aligns them down and up respectively.
func ( int, ,  int) (int, int) {
	 := uintptr(1) << levelBits[]
	return int(alignDown(uintptr(), )), int(alignUp(uintptr(), ))
}

Radix tree of summaries. Each slice's cap represents the whole memory reservation. Each slice's len reflects the allocator's maximum known mapped heap address for that level. The backing store of each summary level is reserved in init and may or may not be committed in grow (small address spaces may commit all the memory in init). The purpose of keeping len <= cap is to enforce bounds checks on the top end of the slice so that instead of an unknown runtime segmentation fault, we get a much friendlier out-of-bounds error. To iterate over a summary level, use inUse to determine which ranges are currently available. Otherwise one might try to access memory which is only Reserved which may result in a hard fault. We may still get segmentation faults < len since some of that memory may not be committed yet.
chunks is a slice of bitmap chunks. The total size of chunks is quite large on most 64-bit platforms (O(GiB) or more) if flattened, so rather than making one large mapping (which has problems on some platforms, even when PROT_NONE) we use a two-level sparse array approach similar to the arena index in mheap. To find the chunk containing a memory address `a`, do: chunkOf(chunkIndex(a)) Below is a table describing the configuration for chunks for various heapAddrBits supported by the runtime. heapAddrBits | L1 Bits | L2 Bits | L2 Entry Size ------------------------------------------------ 32 | 0 | 10 | 128 KiB 33 (iOS) | 0 | 11 | 256 KiB 48 | 13 | 13 | 1 MiB There's no reason to use the L1 part of chunks on 32-bit, the address space is small so the L2 is small. For platforms with a 48-bit address space, we pick the L1 such that the L2 is 1 MiB in size, which is a good balance between low granularity without making the impact on BSS too high (note the L1 is stored directly in pageAlloc). To iterate over the bitmap, use inUse to determine which ranges are currently available. Otherwise one might iterate over unused ranges. TODO(mknyszek): Consider changing the definition of the bitmap such that 1 means free and 0 means in-use so that summaries and the bitmaps align better on zero-values.
The address to start an allocation search with. It must never point to any memory that is not contained in inUse, i.e. inUse.contains(searchAddr.addr()) must always be true. The one exception to this rule is that it may take on the value of maxOffAddr to indicate that the heap is exhausted. We guarantee that all valid heap addresses below this value are allocated and not worth searching.
start and end represent the chunk indices which pageAlloc knows about. It assumes chunks in the range [start, end) are currently ready to use.
inUse is a slice of ranges of address space which are known by the page allocator to be currently in-use (passed to grow). This field is currently unused on 32-bit architectures but is harmless to track. We care much more about having a contiguous heap in these cases and take additional measures to ensure that, so in nearly all cases this should have just 1 element. All access is protected by the mheapLock.
scav stores the scavenger state. All fields are protected by mheapLock.
inUse is a slice of ranges of address space which have not yet been looked at by the scavenger.
		inUse addrRanges
gen is the scavenge generation number.
		gen uint32
reservationBytes is how large of a reservation should be made in bytes of address space for each scavenge iteration.
		reservationBytes uintptr
released is the amount of memory released this generation.
		released uintptr
scavLWM is the lowest (offset) address that the scavenger reached this scavenge generation.
		scavLWM offAddr
freeHWM is the highest (offset) address of a page that was freed to the page allocator this scavenge generation.
		freeHWM offAddr
	}
mheap_.lock. This level of indirection makes it possible to test pageAlloc indepedently of the runtime allocator.
sysStat is the runtime memstat to update when new system memory is committed by the pageAlloc for allocation metadata.
Whether or not this struct is being used in tests.
	test bool
}

func ( *pageAlloc) ( *mutex,  *sysMemStat) {
We can't represent 1<<levelLogPages[0] pages, the maximum number of pages we need to represent at the root level, in a summary, which is a big problem. Throw.
		print("runtime: root level max pages = ", 1<<levelLogPages[0], "\n")
		print("runtime: summary max pages = ", maxPackedValue, "\n")
		throw("root level max pages doesn't fit in summary")
	}
	.sysStat = 
Initialize p.inUse.
	.inUse.init()
System-dependent initialization.
	.sysInit()
Start with the searchAddr in a state indicating there's no free memory.
Set the mheapLock.
	.mheapLock = 
Initialize scavenge tracking state.
tryChunkOf returns the bitmap data for the given chunk. Returns nil if the chunk data has not been mapped.
func ( *pageAlloc) ( chunkIdx) *pallocData {
	 := .chunks[.l1()]
	if  == nil {
		return nil
	}
	return &[.l2()]
}
chunkOf returns the chunk at the given chunk index. The chunk index must be valid or this method may throw.
func ( *pageAlloc) ( chunkIdx) *pallocData {
	return &.chunks[.l1()][.l2()]
}
grow sets up the metadata for the address range [base, base+size). It may allocate metadata, in which case *p.sysStat will be updated. p.mheapLock must be held.
func ( *pageAlloc) (,  uintptr) {
	assertLockHeld(.mheapLock)
Round up to chunks, since we can't deal with increments smaller than chunks. Also, sysGrow expects aligned values.
	 := alignUp(+, pallocChunkBytes)
	 = alignDown(, pallocChunkBytes)
Grow the summary levels in a system-dependent manner. We just update a bunch of additional metadata here.
	.sysGrow(, )
Update p.start and p.end. If no growth happened yet, start == 0. This is generally safe since the zero page is unmapped.
	 := .start == 0
	,  := chunkIndex(), chunkIndex()
	if  ||  < .start {
		.start = 
	}
	if  > .end {
		.end = 
Note that [base, limit) will never overlap with any existing range inUse because grow only ever adds never-used memory regions to the page allocator.
	.inUse.add(makeAddrRange(, ))
A grow operation is a lot like a free operation, so if our chunk ends up below p.searchAddr, update p.searchAddr to the new address, just like in free.
	if  := (offAddr{}); .lessThan(.searchAddr) {
		.searchAddr = 
	}
Add entries into chunks, which is sparse, if needed. Then, initialize the bitmap. Newly-grown memory is always considered scavenged. Set all the bits in the scavenged bitmaps high.
	for  := chunkIndex();  < chunkIndex(); ++ {
Create the necessary l2 entry. Store it atomically to avoid races with readers which don't acquire the heap lock.
Update summaries accordingly. The grow acts like a free, so we need to ensure this newly-free memory is visible in the summaries.
	.update(, /pageSize, true, false)
}
update updates heap metadata. It must be called each time the bitmap is updated. If contig is true, update does some optimizations assuming that there was a contiguous allocation or free between addr and addr+npages. alloc indicates whether the operation performed was an allocation or a free. p.mheapLock must be held.
func ( *pageAlloc) (,  uintptr, ,  bool) {
	assertLockHeld(.mheapLock)
base, limit, start, and end are inclusive.
	 :=  + *pageSize - 1
	,  := chunkIndex(), chunkIndex()
Handle updating the lowest level first.
Fast path: the allocation doesn't span more than one chunk, so update this one and if the summary didn't change, return.
		 := .summary[len(.summary)-1][]
		 := .chunkOf().summarize()
		if  ==  {
			return
		}
		.summary[len(.summary)-1][] = 
Slow contiguous path: the allocation spans more than one chunk and at least one summary is guaranteed to change.
		 := .summary[len(.summary)-1]
Update the summary for chunk sc.
		[] = .chunkOf().summarize()
Update the summaries for chunks in between, which are either totally allocated or freed.
		 := .summary[len(.summary)-1][+1 : ]
Should optimize into a memclr.
			for  := range  {
				[] = 0
			}
		} else {
			for  := range  {
				[] = freeChunkSum
			}
		}
Update the summary for chunk ec.
		[] = .chunkOf().summarize()
Slow general path: the allocation spans more than one chunk and at least one summary is guaranteed to change. We can't assume a contiguous allocation happened, so walk over every chunk in the range and manually recompute the summary.
		 := .summary[len(.summary)-1]
		for  := ;  <= ; ++ {
			[] = .chunkOf().summarize()
		}
	}
Walk up the radix tree and update the summaries appropriately.
	 := true
Update summaries at level l from summaries at level l+1.
		 = false
"Constants" for the previous level which we need to compute the summary from that level.
		 := levelBits[+1]
		 := levelLogPages[+1]
lo and hi describe all the parts of the level we need to look at.
		,  := addrsToSummaryRange(, , +1)
Iterate over each block, updating the corresponding summary in the less-granular level.
		for  := ;  < ; ++ {
			 := .summary[+1][<< : (+1)<<]
			 := mergeSummaries(, )
			 := .summary[][]
			if  !=  {
				 = true
				.summary[][] = 
			}
		}
	}
}
allocRange marks the range of memory [base, base+npages*pageSize) as allocated. It also updates the summaries to reflect the newly-updated bitmap. Returns the amount of scavenged memory in bytes present in the allocated range. p.mheapLock must be held.
func ( *pageAlloc) (,  uintptr) uintptr {
	assertLockHeld(.mheapLock)

	 :=  + *pageSize - 1
	,  := chunkIndex(), chunkIndex()
	,  := chunkPageIndex(), chunkPageIndex()

	 := uint(0)
The range doesn't cross any chunk boundaries.
		 := .chunkOf()
		 += .scavenged.popcntRange(, +1-)
		.allocRange(, +1-)
The range crosses at least one chunk boundary.
		 := .chunkOf()
		 += .scavenged.popcntRange(, pallocChunkPages-)
		.allocRange(, pallocChunkPages-)
		for  :=  + 1;  < ; ++ {
			 := .chunkOf()
			 += .scavenged.popcntRange(0, pallocChunkPages)
			.allocAll()
		}
		 = .chunkOf()
		 += .scavenged.popcntRange(0, +1)
		.allocRange(0, +1)
	}
	.update(, , true, true)
	return uintptr() * pageSize
}
findMappedAddr returns the smallest mapped offAddr that is >= addr. That is, if addr refers to mapped memory, then it is returned. If addr is higher than any mapped region, then it returns maxOffAddr. p.mheapLock must be held.
If we're not in a test, validate first by checking mheap_.arenas. This is a fast path which is only safe to use outside of testing.
	 := arenaIndex(.addr())
	if .test || mheap_.arenas[.l1()] == nil || mheap_.arenas[.l1()][.l2()] == nil {
		,  := .inUse.findAddrGreaterEqual(.addr())
		if  {
			return offAddr{}
The candidate search address is greater than any known address, which means we definitely have no free memory left.
			return maxOffAddr
		}
	}
	return 
}
find searches for the first (address-ordered) contiguous free region of npages in size and returns a base address for that region. It uses p.searchAddr to prune its search and assumes that no palloc chunks below chunkIndex(p.searchAddr) contain any free memory at all. find also computes and returns a candidate p.searchAddr, which may or may not prune more of the address space than p.searchAddr already does. This candidate is always a valid p.searchAddr. find represents the slow path and the full radix tree search. Returns a base address of 0 on failure, in which case the candidate searchAddr returned is invalid and must be ignored. p.mheapLock must be held.
Search algorithm. This algorithm walks each level l of the radix tree from the root level to the leaf level. It iterates over at most 1 << levelBits[l] of entries in a given level in the radix tree, and uses the summary information to find either: 1) That a given subtree contains a large enough contiguous region, at which point it continues iterating on the next level, or 2) That there are enough contiguous boundary-crossing bits to satisfy the allocation, at which point it knows exactly where to start allocating from. i tracks the index into the current level l's structure for the contiguous 1 << levelBits[l] entries we're actually interested in. NOTE: Technically this search could allocate a region which crosses the arenaBaseOffset boundary, which when arenaBaseOffset != 0, is a discontinuity. However, the only way this could happen is if the page at the zero address is mapped, and this is impossible on every system we support where arenaBaseOffset != 0. So, the discontinuity is already encoded in the fact that the OS will never map the zero page for us, and this function doesn't try to handle this case in any way.
i is the beginning of the block of entries we're searching at the current level.
	 := 0
firstFree is the region of address space that we are certain to find the first free page in the heap. base and bound are the inclusive bounds of this window, and both are addresses in the linearized, contiguous view of the address space (with arenaBaseOffset pre-added). At each level, this window is narrowed as we find the memory region containing the first free page of memory. To begin with, the range reflects the full process address space. firstFree is updated by calling foundFree each time free space in the heap is discovered. At the end of the search, base.addr() is the best new searchAddr we could deduce in this search.
	 := struct {
		,  offAddr
	}{
		:  minOffAddr,
		: maxOffAddr,
foundFree takes the given address range [addr, addr+size) and updates firstFree if it is a narrower range. The input range must either be fully contained within firstFree or not overlap with it at all. This way, we'll record the first summary we find with any free pages on the root level and narrow that down if we descend into that summary. But as soon as we need to iterate beyond that summary in a level to find a large enough range, we'll stop narrowing.
	 := func( offAddr,  uintptr) {
This range fits within the current firstFree window, so narrow down the firstFree window to the base and bound of this range.
			. = 
			. = .add( - 1)
This range only partially overlaps with the firstFree range, so throw.
			print("runtime: addr = ", hex(.addr()), ", size = ", , "\n")
			print("runtime: base = ", hex(..addr()), ", bound = ", hex(..addr()), "\n")
			throw("range partially overlaps")
		}
	}
lastSum is the summary which we saw on the previous level that made us move on to the next level. Used to print additional information in the case of a catastrophic failure. lastSumIdx is that summary's index in the previous level.
	 := packPallocSum(0, 0, 0)
	 := -1

:
For the root level, entriesPerBlock is the whole level.
		 := 1 << levelBits[]
		 := levelLogPages[]
We've moved into a new level, so let's update i to our new starting index. This is a no-op for level 0.
		 <<= levelBits[]
Slice out the block of entries we care about.
		 := .summary[][ : +]
Determine j0, the first index we should start iterating from. The searchAddr may help us eliminate iterations if we followed the searchAddr on the previous level or we're on the root leve, in which case the searchAddr should be the same as i after levelShift.
		 := 0
		if  := offAddrToLevelIndex(, .searchAddr); &^(-1) ==  {
			 =  & ( - 1)
		}
Run over the level entries looking for a contiguous run of at least npages either within an entry or across entries. base contains the page index (relative to the first entry's first page) of the currently considered run of consecutive pages. size contains the size of the currently considered run of consecutive pages.
		var ,  uint
		for  := ;  < len(); ++ {
			 := []
A full entry means we broke any streak and that we should skip it altogether.
				 = 0
				continue
			}
We've encountered a non-zero summary which means free memory, so update firstFree.
			(levelIndexToOffAddr(, +), (uintptr(1)<<)*pageSize)

			 := .start()
If size == 0 we don't have a run yet, which means base isn't valid. So, set base to the first page in this block.
				if  == 0 {
					 = uint() << 
We hit npages; we're done!
				 += 
				break
			}
The entry itself contains npages contiguous free pages, so continue on the next level to find that run.
				 += 
				 = 
				 = 
				continue 
			}
We either don't have a current run started, or this entry isn't totally free (meaning we can't continue the current one), so try to begin a new run by setting size and base based on sum.end.
				 = .end()
				 = uint(+1)<< - 
				continue
The entry is completely free, so continue the run.
			 += 1 << 
		}
We found a sufficiently large run of free pages straddling some boundary, so compute the address and return it.
			 := levelIndexToOffAddr(, ).add(uintptr() * pageSize).addr()
			return , .findMappedAddr(.)
		}
We're at level zero, so that means we've exhausted our search.
			return 0, maxSearchAddr
		}
We're not at level zero, and we exhausted the level we were looking in. This means that either our calculations were wrong or the level above lied to us. In either case, dump some useful state and throw.
		print("runtime: summary[", -1, "][", , "] = ", .start(), ", ", .max(), ", ", .end(), "\n")
		print("runtime: level = ", , ", npages = ", , ", j0 = ", , "\n")
		print("runtime: p.searchAddr = ", hex(.searchAddr.addr()), ", i = ", , "\n")
		print("runtime: levelShift[level] = ", levelShift[], ", levelBits[level] = ", levelBits[], "\n")
		for  := 0;  < len(); ++ {
			 := []
			print("runtime: summary[", , "][", +, "] = (", .start(), ", ", .max(), ", ", .end(), ")\n")
		}
		throw("bad summary data")
	}
Since we've gotten to this point, that means we haven't found a sufficiently-sized free region straddling some boundary (chunk or larger). This means the last summary we inspected must have had a large enough "max" value, so look inside the chunk to find a suitable run. After iterating over all levels, i must contain a chunk index which is what the final level represents.
	 := chunkIdx()
	,  := .chunkOf().find(, 0)
We couldn't find any space in this chunk despite the summaries telling us it should be there. There's likely a bug, so dump some state and throw.
		 := .summary[len(.summary)-1][]
		print("runtime: summary[", len(.summary)-1, "][", , "] = (", .start(), ", ", .max(), ", ", .end(), ")\n")
		print("runtime: npages = ", , "\n")
		throw("bad summary data")
	}
Compute the address at which the free space starts.
	 := chunkBase() + uintptr()*pageSize
Since we actually searched the chunk, we may have found an even narrower free window.
	 := chunkBase() + uintptr()*pageSize
	(offAddr{}, chunkBase(+1)-)
	return , .findMappedAddr(.)
}
alloc allocates npages worth of memory from the page heap, returning the base address for the allocation and the amount of scavenged memory in bytes contained in the region [base address, base address + npages*pageSize). Returns a 0 base address on failure, in which case other returned values should be ignored. p.mheapLock must be held. Must run on the system stack because p.mheapLock must be held.go:systemstack
func ( *pageAlloc) ( uintptr) ( uintptr,  uintptr) {
	assertLockHeld(.mheapLock)
If the searchAddr refers to a region which has a higher address than any known chunk, then we know we're out of memory.
	if chunkIndex(.searchAddr.addr()) >= .end {
		return 0, 0
	}
If npages has a chance of fitting in the chunk where the searchAddr is, search it directly.
	 := minOffAddr
npages is guaranteed to be no greater than pallocChunkPages here.
		 := chunkIndex(.searchAddr.addr())
		if  := .summary[len(.summary)-1][].max();  >= uint() {
			,  := .chunkOf().find(, chunkPageIndex(.searchAddr.addr()))
			if  == ^uint(0) {
				print("runtime: max = ", , ", npages = ", , "\n")
				print("runtime: searchIdx = ", chunkPageIndex(.searchAddr.addr()), ", p.searchAddr = ", hex(.searchAddr.addr()), "\n")
				throw("bad summary data")
			}
			 = chunkBase() + uintptr()*pageSize
			 = offAddr{chunkBase() + uintptr()*pageSize}
			goto 
		}
We failed to use a searchAddr for one reason or another, so try the slow path.
	,  = .find()
	if  == 0 {
We failed to find a single free page, the smallest unit of allocation. This means we know the heap is completely exhausted. Otherwise, the heap still might have free space in it, just not enough contiguous space to accommodate npages.
			.searchAddr = maxSearchAddr
		}
		return 0, 0
	}
Go ahead and actually mark the bits now that we have an address.
	 = .allocRange(, )
If we found a higher searchAddr, we know that all the heap memory before that searchAddr in an offset address space is allocated, so bump p.searchAddr up to the new one.
	if .searchAddr.lessThan() {
		.searchAddr = 
	}
	return , 
}
free returns npages worth of memory starting at base back to the page heap. p.mheapLock must be held. Must run on the system stack because p.mheapLock must be held.go:systemstack
func ( *pageAlloc) (,  uintptr) {
	assertLockHeld(.mheapLock)
If we're freeing pages below the p.searchAddr, update searchAddr.
	if  := (offAddr{}); .lessThan(.searchAddr) {
		.searchAddr = 
Update the free high watermark for the scavenger.
	 :=  + *pageSize - 1
	if  := (offAddr{}); .scav.freeHWM.lessThan() {
		.scav.freeHWM = 
	}
Fast path: we're clearing a single bit, and we know exactly where it is, so mark it directly.
		 := chunkIndex()
		.chunkOf().free1(chunkPageIndex())
Slow path: we're clearing more bits so we may need to iterate.
		,  := chunkIndex(), chunkIndex()
		,  := chunkPageIndex(), chunkPageIndex()

The range doesn't cross any chunk boundaries.
			.chunkOf().free(, +1-)
The range crosses at least one chunk boundary.
			.chunkOf().free(, pallocChunkPages-)
			for  :=  + 1;  < ; ++ {
				.chunkOf().freeAll()
			}
			.chunkOf().free(0, +1)
		}
	}
	.update(, , true, false)
}

const (
	pallocSumBytes = unsafe.Sizeof(pallocSum(0))
maxPackedValue is the maximum value that any of the three fields in the pallocSum may take on.
pallocSum is a packed summary type which packs three numbers: start, max, and end into a single 8-byte value. Each of these values are a summary of a bitmap and are thus counts, each of which may have a maximum value of 2^21 - 1, or all three may be equal to 2^21. The latter case is represented by just setting the 64th bit.
packPallocSum takes a start, max, and end value and produces a pallocSum.
func (, ,  uint) pallocSum {
	if  == maxPackedValue {
		return pallocSum(uint64(1 << 63))
	}
	return pallocSum((uint64() & (maxPackedValue - 1)) |
		((uint64() & (maxPackedValue - 1)) << logMaxPackedValue) |
		((uint64() & (maxPackedValue - 1)) << (2 * logMaxPackedValue)))
}
start extracts the start value from a packed sum.
func ( pallocSum) () uint {
	if uint64()&uint64(1<<63) != 0 {
		return maxPackedValue
	}
	return uint(uint64() & (maxPackedValue - 1))
}
max extracts the max value from a packed sum.
func ( pallocSum) () uint {
	if uint64()&uint64(1<<63) != 0 {
		return maxPackedValue
	}
	return uint((uint64() >> logMaxPackedValue) & (maxPackedValue - 1))
}
end extracts the end value from a packed sum.
func ( pallocSum) () uint {
	if uint64()&uint64(1<<63) != 0 {
		return maxPackedValue
	}
	return uint((uint64() >> (2 * logMaxPackedValue)) & (maxPackedValue - 1))
}
unpack unpacks all three values from the summary.
func ( pallocSum) () (uint, uint, uint) {
	if uint64()&uint64(1<<63) != 0 {
		return maxPackedValue, maxPackedValue, maxPackedValue
	}
	return uint(uint64() & (maxPackedValue - 1)),
		uint((uint64() >> logMaxPackedValue) & (maxPackedValue - 1)),
		uint((uint64() >> (2 * logMaxPackedValue)) & (maxPackedValue - 1))
}
mergeSummaries merges consecutive summaries which may each represent at most 1 << logMaxPagesPerSum pages each together into one.
Merge the summaries in sums into one. We do this by keeping a running summary representing the merged summaries of sums[:i] in start, max, and end.
	, ,  := [0].unpack()
Merge in sums[i].
		, ,  := [].unpack()
Merge in sums[i].start only if the running summary is completely free, otherwise this summary's start plays no role in the combined sum.
		if  == uint()<< {
			 += 
		}
Recompute the max value of the running sum by looking across the boundary between the running sum and sums[i] and at the max sums[i], taking the greatest of those two and the max of the running sum.
		if + >  {
			 =  + 
		}
		if  >  {
			 = 
		}
Merge in end by checking if this new summary is totally free. If it is, then we want to extend the running sum's end by the new summary. If not, then we have some alloc'd pages in there and we just want to take the end value in sums[i].
		if  == 1<< {
			 += 1 << 
		} else {
			 = 
		}
	}
	return packPallocSum(, , )