sizes 87 kern/vm/kmalloc.c static const size_t sizes[NSIZES] = { 16, 32, 64, 128, 256, 512, 1024, 2048 };
sizes 230 kern/vm/kmalloc.c KASSERT(pr->freelist_offset % sizes[blktype] == 0);
sizes 238 kern/vm/kmalloc.c KASSERT((fla-prpage) % sizes[blktype] == 0);
sizes 304 kern/vm/kmalloc.c n = PAGE_SIZE / sizes[blktype];
sizes 313 kern/vm/kmalloc.c index = (fla-prpage) / sizes[blktype];
sizes 320 kern/vm/kmalloc.c (unsigned long)prpage, (unsigned long) sizes[blktype],
sizes 383 kern/vm/kmalloc.c if (sz <= sizes[i]) {
sizes 410 kern/vm/kmalloc.c sz = sizes[blktype];
sizes 481 kern/vm/kmalloc.c pr->nfree = PAGE_SIZE / sizes[blktype];
sizes 493 kern/vm/kmalloc.c fl = (struct freelist *)(fla + i*sizes[blktype]);
sizes 494 kern/vm/kmalloc.c fl->next = (struct freelist *)(fla + (i-1)*sizes[blktype]);
sizes 499 kern/vm/kmalloc.c KASSERT(pr->freelist_offset == (pr->nfree-1)*sizes[blktype]);
sizes 551 kern/vm/kmalloc.c if (offset >= PAGE_SIZE || offset % sizes[blktype] != 0) {
sizes 559 kern/vm/kmalloc.c fill_deadbeef(ptr, sizes[blktype]);
sizes 576 kern/vm/kmalloc.c KASSERT(pr->nfree <= PAGE_SIZE / sizes[blktype]);
sizes 577 kern/vm/kmalloc.c if (pr->nfree == PAGE_SIZE / sizes[blktype]) {
sizes 520 user/testbin/malloctest/malloctest.c static const int sizes[8] = { 13, 17, 69, 176, 433, 871, 1150, 6060 };
sizes 537 user/testbin/malloctest/malloctest.c size = sizes[random()%8];