sizes              87 kern/vm/kmalloc.c static const size_t sizes[NSIZES] = { 16, 32, 64, 128, 256, 512, 1024, 2048 };
sizes             230 kern/vm/kmalloc.c 	KASSERT(pr->freelist_offset % sizes[blktype] == 0);
sizes             238 kern/vm/kmalloc.c 		KASSERT((fla-prpage) % sizes[blktype] == 0);
sizes             304 kern/vm/kmalloc.c 	n = PAGE_SIZE / sizes[blktype];
sizes             313 kern/vm/kmalloc.c 			index = (fla-prpage) / sizes[blktype];
sizes             320 kern/vm/kmalloc.c 		(unsigned long)prpage, (unsigned long) sizes[blktype],
sizes             383 kern/vm/kmalloc.c 		if (sz <= sizes[i]) {
sizes             410 kern/vm/kmalloc.c 	sz = sizes[blktype];
sizes             481 kern/vm/kmalloc.c 	pr->nfree = PAGE_SIZE / sizes[blktype];
sizes             493 kern/vm/kmalloc.c 		fl = (struct freelist *)(fla + i*sizes[blktype]);
sizes             494 kern/vm/kmalloc.c 		fl->next = (struct freelist *)(fla + (i-1)*sizes[blktype]);
sizes             499 kern/vm/kmalloc.c 	KASSERT(pr->freelist_offset == (pr->nfree-1)*sizes[blktype]);
sizes             551 kern/vm/kmalloc.c 	if (offset >= PAGE_SIZE || offset % sizes[blktype] != 0) {
sizes             559 kern/vm/kmalloc.c 	fill_deadbeef(ptr, sizes[blktype]);
sizes             576 kern/vm/kmalloc.c 	KASSERT(pr->nfree <= PAGE_SIZE / sizes[blktype]);
sizes             577 kern/vm/kmalloc.c 	if (pr->nfree == PAGE_SIZE / sizes[blktype]) {
sizes             520 user/testbin/malloctest/malloctest.c 	static const int sizes[8] = { 13, 17, 69, 176, 433, 871, 1150, 6060 };
sizes             537 user/testbin/malloctest/malloctest.c 			size = sizes[random()%8];