heap_allocated_pages 1339 gc.c for (i = 0; i < heap_allocated_pages; ++i) { heap_allocated_pages 1343 gc.c heap_allocated_pages = 0; heap_allocated_pages 1443 gc.c heap_allocated_pages--; heap_allocated_pages 1455 gc.c for (i = j = 1; j < heap_allocated_pages; i++) { heap_allocated_pages 1469 gc.c if (RGENGC_CHECK_MODE) assert(j == heap_allocated_pages); heap_allocated_pages 1506 gc.c hi = heap_allocated_pages; heap_allocated_pages 1522 gc.c if (hi < heap_allocated_pages) { heap_allocated_pages 1523 gc.c MEMMOVE(&heap_pages_sorted[hi+1], &heap_pages_sorted[hi], struct heap_page_header*, heap_allocated_pages - hi); heap_allocated_pages 1528 gc.c heap_allocated_pages++; heap_allocated_pages 1531 gc.c if (RGENGC_CHECK_MODE) assert(heap_allocated_pages <= heap_pages_sorted_length); heap_allocated_pages 1571 gc.c method, page, (int)heap_pages_sorted_length, (int)heap_allocated_pages, (int)heap_tomb->total_pages); heap_allocated_pages 1611 gc.c size_t used = heap_allocated_pages + heap_allocatable_pages; heap_allocated_pages 1652 gc.c if (next_used_limit == heap_allocated_pages) next_used_limit++; heap_allocated_pages 2048 gc.c hi = heap_allocated_pages; heap_allocated_pages 2348 gc.c while (i < heap_allocated_pages) { heap_allocated_pages 2350 gc.c while (i < heap_allocated_pages && heap_pages_sorted[i]->start <= pstart) i++; heap_allocated_pages 2351 gc.c if (heap_allocated_pages <= i) break; heap_allocated_pages 2862 gc.c for (i = 0; i < heap_allocated_pages; i++) { heap_allocated_pages 3299 gc.c for (i = 0; i < heap_allocated_pages; i++) { heap_allocated_pages 6276 gc.c if (!heap_allocated_pages) return FALSE; /* heap is not ready */ heap_allocated_pages 6340 gc.c objspace->profile.heap_used_at_gc_start = heap_allocated_pages; heap_allocated_pages 6855 gc.c S(heap_allocated_pages); heap_allocated_pages 6926 gc.c rb_hash_aset(table, OLD_SYM(gc_stat_heap_used), NEW_SYM(heap_allocated_pages)); heap_allocated_pages 7027 gc.c SET(heap_allocated_pages, heap_allocated_pages);