/* * Copyright (C) 2009 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include <errno.h> #include <limits.h> #include <sys/mman.h> #include "Dalvik.h" #include "alloc/Heap.h" #include "alloc/HeapBitmap.h" #include "alloc/HeapInternal.h" #include "alloc/HeapSource.h" #include "alloc/Verify.h" #include "alloc/clz.h" /* * A "mostly copying", generational, garbage collector. * * TODO: we allocate our own contiguous tract of page frames to back * object allocations. To cooperate with other heaps active in the * virtual machine we need to move the responsibility of allocating * pages someplace outside of this code. * * The other major data structures that maintain the state of the heap * are the block space table and the block queue. * * The block space table records the state of a block. We must track * whether a block is: * * - Free or allocated in some space. * * - If the block holds part of a large object allocation, whether the * block is the initial or a continued block of the allocation. * * - Whether the block is pinned, that is to say whether at least one * object in the block must remain stationary. Only needed during a * GC. * * - Which space the object belongs to. At present this means * from-space or to-space. * * The block queue is used during garbage collection. Unlike Cheney's * algorithm, from-space and to-space are not contiguous. Therefore, * one cannot maintain the state of the copy with just two pointers. * The block queue exists to thread lists of blocks from the various * spaces together. * * Additionally, we record the free space frontier of the heap, as * well as the address of the first object within a block, which is * required to copy objects following a large object (not currently * implemented). This is stored in the heap source structure. This * should be moved elsewhere to support in-line allocations from Java * threads. * * Allocation requests are satisfied by reserving storage from one or * more contiguous blocks. Objects that are small enough to fit * inside a block are packed together within a block. Objects that * are larger than a block are allocated from contiguous sequences of * blocks. When half the available blocks are filled, a garbage * collection occurs. We "flip" spaces (exchange from- and to-space), * copy live objects into to space, and perform pointer adjustment. * * Copying is made more complicated by the requirement that some * objects must not be moved. This property is known as "pinning". * These objects must be dealt with specially. We use Bartlett's * scheme; blocks containing such objects are grayed (promoted) at the * start of a garbage collection. By virtue of this trick, tracing * from the roots proceeds as usual but all objects on those pages are * considered promoted and therefore not moved. * * TODO: there is sufficient information within the garbage collector * to implement Attardi's scheme for evacuating unpinned objects from * a page that is otherwise pinned. This would eliminate false * retention caused by the large pinning granularity. * * We need a scheme for medium and large objects. Ignore that for * now, we can return to this later. * * Eventually we need to worry about promoting objects out of the * copy-collected heap (tenuring) into a less volatile space. Copying * may not always be the best policy for such spaces. We should * consider a variant of mark, sweep, compact. * * The block scheme allows us to use VM page faults to maintain a * write barrier. Consider having a special leaf state for a page. * * Bibliography: * * C. J. Cheney. 1970. A non-recursive list compacting * algorithm. CACM. 13-11 pp677--678. * * Joel F. Bartlett. 1988. Compacting Garbage Collection with * Ambiguous Roots. Digital Equipment Corporation. * * Joel F. Bartlett. 1989. Mostly-Copying Garbage Collection Picks Up * Generations and C++. Digital Equipment Corporation. * * G. May Yip. 1991. Incremental, Generational Mostly-Copying Garbage * Collection in Uncooperative Environments. Digital Equipment * Corporation. * * Giuseppe Attardi, Tito Flagella. 1994. A Customisable Memory * Management Framework. TR-94-010 * * Giuseppe Attardi, Tito Flagella, Pietro Iglio. 1998. A customisable * memory management framework for C++. Software -- Practice and * Experience. 28(11), 1143-1183. * */ #define ARRAYSIZE(x) (sizeof(x) / sizeof(x[0])) #if 0 #define LOG_ALLOC LOGI #define LOG_PIN LOGI #define LOG_PROM LOGI #define LOG_REF LOGI #define LOG_SCAV LOGI #define LOG_TRAN LOGI #define LOG_VER LOGI #else #define LOG_ALLOC(...) ((void)0) #define LOG_PIN(...) ((void)0) #define LOG_PROM(...) ((void)0) #define LOG_REF(...) ((void)0) #define LOG_SCAV(...) ((void)0) #define LOG_TRAN(...) ((void)0) #define LOG_VER(...) ((void)0) #endif static void enqueueBlock(HeapSource *heapSource, size_t block); static void scavengeReference(Object **obj); static bool toSpaceContains(const void *addr); static bool fromSpaceContains(const void *addr); static size_t sumHeapBitmap(const HeapBitmap *bitmap); static size_t objectSize(const Object *obj); static void scavengeDataObject(Object *obj); static void scavengeBlockQueue(void); /* * We use 512-byte blocks. */ enum { BLOCK_SHIFT = 9 }; enum { BLOCK_SIZE = 1 << BLOCK_SHIFT }; /* * Space identifiers, stored into the blockSpace array. */ enum { BLOCK_FREE = 0, BLOCK_FROM_SPACE = 1, BLOCK_TO_SPACE = 2, BLOCK_CONTINUED = 7 }; /* * Alignment for all allocations, in bytes. */ enum { ALLOC_ALIGNMENT = 8 }; /* * Sentinel value for the queue end. */ #define QUEUE_TAIL (~(size_t)0) struct HeapSource { /* The base address of backing store. */ u1 *blockBase; /* Total number of blocks available for allocation. */ size_t totalBlocks; size_t allocBlocks; /* * The scavenger work queue. Implemented as an array of index * values into the queue. */ size_t *blockQueue; /* * Base and limit blocks. Basically the shifted start address of * the block. We convert blocks to a relative number when * indexing in the block queue. TODO: make the block queue base * relative rather than the index into the block queue. */ size_t baseBlock, limitBlock; size_t queueHead; size_t queueTail; size_t queueSize; /* The space of the current block 0 (free), 1 or 2. */ char *blockSpace; /* Start of free space in the current block. */ u1 *allocPtr; /* Exclusive limit of free space in the current block. */ u1 *allocLimit; HeapBitmap allocBits; /* * The starting size of the heap. This value is the same as the * value provided to the -Xms flag. */ size_t minimumSize; /* * The maximum size of the heap. This value is the same as the * -Xmx flag. */ size_t maximumSize; /* * The current, committed size of the heap. At present, this is * equivalent to the maximumSize. */ size_t currentSize; size_t bytesAllocated; }; static unsigned long alignDown(unsigned long x, unsigned long n) { return x & -n; } static unsigned long alignUp(unsigned long x, unsigned long n) { return alignDown(x + (n - 1), n); } static void describeBlocks(const HeapSource *heapSource) { size_t i; for (i = 0; i < heapSource->totalBlocks; ++i) { if ((i % 32) == 0) putchar('\n'); printf("%d ", heapSource->blockSpace[i]); } putchar('\n'); } /* * Virtual memory interface. */ static void *virtualAlloc(size_t length) { void *addr; int flags, prot; flags = MAP_PRIVATE | MAP_ANONYMOUS; prot = PROT_READ | PROT_WRITE; addr = mmap(NULL, length, prot, flags, -1, 0); if (addr == MAP_FAILED) { LOGE_HEAP("mmap: %s", strerror(errno)); addr = NULL; } return addr; } static void virtualFree(void *addr, size_t length) { int res; assert(addr != NULL); assert((uintptr_t)addr % SYSTEM_PAGE_SIZE == 0); res = munmap(addr, length); if (res == -1) { LOGE_HEAP("munmap: %s", strerror(errno)); } } #ifndef NDEBUG static int isValidAddress(const HeapSource *heapSource, const u1 *addr) { size_t block; block = (uintptr_t)addr >> BLOCK_SHIFT; return heapSource->baseBlock <= block && heapSource->limitBlock > block; } #endif /* * Iterate over the block map looking for a contiguous run of free * blocks. */ static void *allocateBlocks(HeapSource *heapSource, size_t blocks) { void *addr; size_t allocBlocks, totalBlocks; size_t i, j; allocBlocks = heapSource->allocBlocks; totalBlocks = heapSource->totalBlocks; /* Check underflow. */ assert(blocks != 0); /* Check overflow. */ if (allocBlocks + blocks > totalBlocks / 2) { return NULL; } /* Scan block map. */ for (i = 0; i < totalBlocks; ++i) { /* Check fit. */ for (j = 0; j < blocks; ++j) { /* runs over totalBlocks */ if (heapSource->blockSpace[i+j] != BLOCK_FREE) { break; } } /* No fit? */ if (j != blocks) { i += j; continue; } /* Fit, allocate. */ heapSource->blockSpace[i] = BLOCK_TO_SPACE; /* why to-space? */ for (j = 1; j < blocks; ++j) { heapSource->blockSpace[i+j] = BLOCK_CONTINUED; } heapSource->allocBlocks += blocks; addr = &heapSource->blockBase[i*BLOCK_SIZE]; memset(addr, 0, blocks*BLOCK_SIZE); /* Collecting? */ if (heapSource->queueHead != QUEUE_TAIL) { LOG_ALLOC("allocateBlocks allocBlocks=%zu,block#=%zu", heapSource->allocBlocks, i); /* * This allocated was on behalf of the transporter when it * shaded a white object gray. We enqueue the block so * the scavenger can further shade the gray objects black. */ enqueueBlock(heapSource, i); } return addr; } /* Insufficient space, fail. */ LOGE("Insufficient space, %zu blocks, %zu blocks allocated and %zu bytes allocated", heapSource->totalBlocks, heapSource->allocBlocks, heapSource->bytesAllocated); return NULL; } /* Converts an absolute address to a relative block number. */ static size_t addressToBlock(const HeapSource *heapSource, const void *addr) { assert(heapSource != NULL); assert(isValidAddress(heapSource, addr)); return (((uintptr_t)addr) >> BLOCK_SHIFT) - heapSource->baseBlock; } /* Converts a relative block number to an absolute address. */ static u1 *blockToAddress(const HeapSource *heapSource, size_t block) { u1 *addr; addr = (u1 *) (((uintptr_t) heapSource->baseBlock + block) * BLOCK_SIZE); assert(isValidAddress(heapSource, addr)); return addr; } static void clearBlock(HeapSource *heapSource, size_t block) { u1 *addr; size_t i; assert(heapSource != NULL); assert(block < heapSource->totalBlocks); addr = heapSource->blockBase + block*BLOCK_SIZE; memset(addr, 0xCC, BLOCK_SIZE); for (i = 0; i < BLOCK_SIZE; i += 8) { dvmHeapBitmapClearObjectBit(&heapSource->allocBits, addr + i); } } static void clearFromSpace(HeapSource *heapSource) { size_t i, count; assert(heapSource != NULL); i = count = 0; while (i < heapSource->totalBlocks) { if (heapSource->blockSpace[i] != BLOCK_FROM_SPACE) { ++i; continue; } heapSource->blockSpace[i] = BLOCK_FREE; clearBlock(heapSource, i); ++i; ++count; while (i < heapSource->totalBlocks && heapSource->blockSpace[i] == BLOCK_CONTINUED) { heapSource->blockSpace[i] = BLOCK_FREE; clearBlock(heapSource, i); ++i; ++count; } } LOG_SCAV("freed %zu blocks (%zu bytes)", count, count*BLOCK_SIZE); } /* * Appends the given block to the block queue. The block queue is * processed in-order by the scavenger. */ static void enqueueBlock(HeapSource *heapSource, size_t block) { assert(heapSource != NULL); assert(block < heapSource->totalBlocks); if (heapSource->queueHead != QUEUE_TAIL) { heapSource->blockQueue[heapSource->queueTail] = block; } else { heapSource->queueHead = block; } heapSource->blockQueue[block] = QUEUE_TAIL; heapSource->queueTail = block; ++heapSource->queueSize; } /* * Grays all objects within the block corresponding to the given * address. */ static void promoteBlockByAddr(HeapSource *heapSource, const void *addr) { size_t block; block = addressToBlock(heapSource, (const u1 *)addr); if (heapSource->blockSpace[block] != BLOCK_TO_SPACE) { // LOG_PROM("promoting block %zu %d @ %p", block, heapSource->blockSpace[block], obj); heapSource->blockSpace[block] = BLOCK_TO_SPACE; enqueueBlock(heapSource, block); /* TODO(cshapiro): count continued blocks?*/ heapSource->allocBlocks += 1; } else { // LOG_PROM("NOT promoting block %zu %d @ %p", block, heapSource->blockSpace[block], obj); } } GcHeap *dvmHeapSourceStartup(size_t startSize, size_t absoluteMaxSize) { GcHeap* gcHeap; HeapSource *heapSource; assert(startSize <= absoluteMaxSize); heapSource = malloc(sizeof(*heapSource)); assert(heapSource != NULL); memset(heapSource, 0, sizeof(*heapSource)); heapSource->minimumSize = alignUp(startSize, BLOCK_SIZE); heapSource->maximumSize = alignUp(absoluteMaxSize, BLOCK_SIZE); heapSource->currentSize = heapSource->maximumSize; /* Allocate underlying storage for blocks. */ heapSource->blockBase = virtualAlloc(heapSource->maximumSize); assert(heapSource->blockBase != NULL); heapSource->baseBlock = (uintptr_t) heapSource->blockBase >> BLOCK_SHIFT; heapSource->limitBlock = ((uintptr_t) heapSource->blockBase + heapSource->maximumSize) >> BLOCK_SHIFT; heapSource->allocBlocks = 0; heapSource->totalBlocks = (heapSource->limitBlock - heapSource->baseBlock); assert(heapSource->totalBlocks = heapSource->maximumSize / BLOCK_SIZE); { size_t size = sizeof(heapSource->blockQueue[0]); heapSource->blockQueue = malloc(heapSource->totalBlocks*size); assert(heapSource->blockQueue != NULL); memset(heapSource->blockQueue, 0xCC, heapSource->totalBlocks*size); heapSource->queueHead = QUEUE_TAIL; } /* Byte indicating space residence or free status of block. */ { size_t size = sizeof(heapSource->blockSpace[0]); heapSource->blockSpace = malloc(heapSource->totalBlocks*size); assert(heapSource->blockSpace != NULL); memset(heapSource->blockSpace, 0, heapSource->totalBlocks*size); } dvmHeapBitmapInit(&heapSource->allocBits, heapSource->blockBase, heapSource->maximumSize, "blockBase"); /* Initialize allocation pointers. */ heapSource->allocPtr = allocateBlocks(heapSource, 1); heapSource->allocLimit = heapSource->allocPtr + BLOCK_SIZE; gcHeap = malloc(sizeof(*gcHeap)); assert(gcHeap != NULL); memset(gcHeap, 0, sizeof(*gcHeap)); gcHeap->heapSource = heapSource; return gcHeap; } /* * Perform any required heap initializations after forking from the * zygote process. This is a no-op for the time being. Eventually * this will demarcate the shared region of the heap. */ bool dvmHeapSourceStartupAfterZygote(void) { return true; } bool dvmHeapSourceStartupBeforeFork(void) { assert(!"implemented"); return false; } void dvmHeapSourceShutdown(GcHeap **gcHeap) { if (*gcHeap == NULL || (*gcHeap)->heapSource == NULL) return; free((*gcHeap)->heapSource->blockQueue); free((*gcHeap)->heapSource->blockSpace); virtualFree((*gcHeap)->heapSource->blockBase, (*gcHeap)->heapSource->maximumSize); free((*gcHeap)->heapSource); (*gcHeap)->heapSource = NULL; free(*gcHeap); *gcHeap = NULL; } size_t dvmHeapSourceGetValue(enum HeapSourceValueSpec spec, size_t perHeapStats[], size_t arrayLen) { HeapSource *heapSource; size_t value; heapSource = gDvm.gcHeap->heapSource; switch (spec) { case HS_EXTERNAL_BYTES_ALLOCATED: value = 0; break; case HS_EXTERNAL_LIMIT: value = 0; break; case HS_FOOTPRINT: value = heapSource->maximumSize; break; case HS_ALLOWED_FOOTPRINT: value = heapSource->maximumSize; break; case HS_BYTES_ALLOCATED: value = heapSource->bytesAllocated; break; case HS_OBJECTS_ALLOCATED: value = sumHeapBitmap(&heapSource->allocBits); break; default: assert(!"implemented"); value = 0; } if (perHeapStats) { *perHeapStats = value; } return value; } /* * Performs a shallow copy of the allocation bitmap into the given * vector of heap bitmaps. */ void dvmHeapSourceGetObjectBitmaps(HeapBitmap objBits[], HeapBitmap markBits[], size_t numHeaps) { assert(!"implemented"); } HeapBitmap *dvmHeapSourceGetLiveBits(void) { return &gDvm.gcHeap->heapSource->allocBits; } /* * Allocate the specified number of bytes from the heap. The * allocation cursor points into a block of free storage. If the * given allocation fits in the remaining space of the block, we * advance the cursor and return a pointer to the free storage. If * the allocation cannot fit in the current block but is smaller than * a block we request a new block and allocate from it instead. If * the allocation is larger than a block we must allocate from a span * of contiguous blocks. */ void *dvmHeapSourceAlloc(size_t length) { HeapSource *heapSource; unsigned char *addr; size_t aligned, available, blocks; heapSource = gDvm.gcHeap->heapSource; assert(heapSource != NULL); assert(heapSource->allocPtr != NULL); assert(heapSource->allocLimit != NULL); aligned = alignUp(length, ALLOC_ALIGNMENT); available = heapSource->allocLimit - heapSource->allocPtr; /* Try allocating inside the current block. */ if (aligned <= available) { addr = heapSource->allocPtr; heapSource->allocPtr += aligned; heapSource->bytesAllocated += aligned; dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr); return addr; } /* Try allocating in a new block. */ if (aligned <= BLOCK_SIZE) { addr = allocateBlocks(heapSource, 1); if (addr != NULL) { heapSource->allocLimit = addr + BLOCK_SIZE; heapSource->allocPtr = addr + aligned; heapSource->bytesAllocated += aligned; dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr); /* TODO(cshapiro): pad out the current block. */ } return addr; } /* Try allocating in a span of blocks. */ blocks = alignUp(aligned, BLOCK_SIZE) / BLOCK_SIZE; addr = allocateBlocks(heapSource, blocks); /* Propagate failure upward. */ if (addr != NULL) { heapSource->bytesAllocated += aligned; dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr); /* TODO(cshapiro): pad out free space in the last block. */ } return addr; } void *dvmHeapSourceAllocAndGrow(size_t size) { return dvmHeapSourceAlloc(size); } /* TODO: refactor along with dvmHeapSourceAlloc */ void *allocateGray(size_t size) { HeapSource *heapSource; void *addr; size_t block; /* TODO: add a check that we are in a GC. */ heapSource = gDvm.gcHeap->heapSource; addr = dvmHeapSourceAlloc(size); assert(addr != NULL); block = addressToBlock(heapSource, (const u1 *)addr); if (heapSource->queueHead == QUEUE_TAIL) { /* * Forcibly append the underlying block to the queue. This * condition occurs when referents are transported following * the initial trace. */ enqueueBlock(heapSource, block); LOG_PROM("forced promoting block %zu %d @ %p", block, heapSource->blockSpace[block], addr); } return addr; } bool dvmHeapSourceContainsAddress(const void *ptr) { HeapSource *heapSource = gDvm.gcHeap->heapSource; return dvmHeapBitmapCoversAddress(&heapSource->allocBits, ptr); } /* * Returns true if the given address is within the heap and points to * the header of a live object. */ bool dvmHeapSourceContains(const void *addr) { HeapSource *heapSource; HeapBitmap *bitmap; heapSource = gDvm.gcHeap->heapSource; bitmap = &heapSource->allocBits; if (!dvmHeapBitmapCoversAddress(bitmap, addr)) { return false; } else { return dvmHeapBitmapIsObjectBitSet(bitmap, addr); } } bool dvmHeapSourceGetPtrFlag(const void *ptr, enum HeapSourcePtrFlag flag) { assert(!"implemented"); return false; } size_t dvmHeapSourceChunkSize(const void *ptr) { assert(!"implemented"); return 0; } size_t dvmHeapSourceFootprint(void) { assert(!"implemented"); return 0; } /* * Returns the "ideal footprint" which appears to be the number of * bytes currently committed to the heap. This starts out at the * start size of the heap and grows toward the maximum size. */ size_t dvmHeapSourceGetIdealFootprint(void) { return gDvm.gcHeap->heapSource->currentSize; } float dvmGetTargetHeapUtilization(void) { return 0.5f; } void dvmSetTargetHeapUtilization(float newTarget) { assert(newTarget > 0.0f && newTarget < 1.0f); } size_t dvmMinimumHeapSize(size_t size, bool set) { return gDvm.gcHeap->heapSource->minimumSize; } /* * Expands the size of the heap after a collection. At present we * commit the pages for maximum size of the heap so this routine is * just a no-op. Eventually, we will either allocate or commit pages * on an as-need basis. */ void dvmHeapSourceGrowForUtilization(void) { /* do nothing */ } void dvmHeapSourceTrim(size_t bytesTrimmed[], size_t arrayLen) { /* do nothing */ } void dvmHeapSourceWalk(void (*callback)(const void *chunkptr, size_t chunklen, const void *userptr, size_t userlen, void *arg), void *arg) { assert(!"implemented"); } size_t dvmHeapSourceGetNumHeaps(void) { return 1; } bool dvmTrackExternalAllocation(size_t n) { /* do nothing */ return true; } void dvmTrackExternalFree(size_t n) { /* do nothing */ } size_t dvmGetExternalBytesAllocated(void) { assert(!"implemented"); return 0; } void dvmHeapSourceFlip(void) { HeapSource *heapSource; size_t i; heapSource = gDvm.gcHeap->heapSource; /* Reset the block queue. */ heapSource->allocBlocks = 0; heapSource->queueSize = 0; heapSource->queueHead = QUEUE_TAIL; /* TODO(cshapiro): pad the current (prev) block. */ heapSource->allocPtr = NULL; heapSource->allocLimit = NULL; /* Whiten all allocated blocks. */ for (i = 0; i < heapSource->totalBlocks; ++i) { if (heapSource->blockSpace[i] == BLOCK_TO_SPACE) { heapSource->blockSpace[i] = BLOCK_FROM_SPACE; } } } static void room(size_t *alloc, size_t *avail, size_t *total) { HeapSource *heapSource; heapSource = gDvm.gcHeap->heapSource; *total = heapSource->totalBlocks*BLOCK_SIZE; *alloc = heapSource->allocBlocks*BLOCK_SIZE; *avail = *total - *alloc; } static bool isSpaceInternal(u1 *addr, int space) { HeapSource *heapSource; u1 *base, *limit; size_t offset; char space2; heapSource = gDvm.gcHeap->heapSource; base = heapSource->blockBase; assert(addr >= base); limit = heapSource->blockBase + heapSource->maximumSize; assert(addr < limit); offset = addr - base; space2 = heapSource->blockSpace[offset >> BLOCK_SHIFT]; return space == space2; } static bool fromSpaceContains(const void *addr) { return isSpaceInternal((u1 *)addr, BLOCK_FROM_SPACE); } static bool toSpaceContains(const void *addr) { return isSpaceInternal((u1 *)addr, BLOCK_TO_SPACE); } /* * Notifies the collector that the object at the given address must * remain stationary during the current collection. */ static void pinObject(const Object *obj) { promoteBlockByAddr(gDvm.gcHeap->heapSource, obj); } static size_t sumHeapBitmap(const HeapBitmap *bitmap) { size_t i, sum; sum = 0; for (i = 0; i < bitmap->bitsLen >> 2; ++i) { sum += CLZ(bitmap->bits[i]); } return sum; } /* * Miscellaneous functionality. */ static int isForward(const void *addr) { return (uintptr_t)addr & 0x1; } static void setForward(const void *toObj, void *fromObj) { *(unsigned long *)fromObj = (uintptr_t)toObj | 0x1; } static void* getForward(const void *fromObj) { return (void *)((uintptr_t)fromObj & ~0x1); } /* Beware, uses the same encoding as a forwarding pointers! */ static int isPermanentString(const StringObject *obj) { return (uintptr_t)obj & 0x1; } static void* getPermanentString(const StringObject *obj) { return (void *)((uintptr_t)obj & ~0x1); } /* * Scavenging and transporting routines follow. A transporter grays * an object. A scavenger blackens an object. We define these * routines for each fundamental object type. Dispatch is performed * in scavengeObject. */ /* * Class object scavenging. */ static void scavengeClassObject(ClassObject *obj) { int i; LOG_SCAV("scavengeClassObject(obj=%p)", obj); assert(obj != NULL); assert(obj->obj.clazz != NULL); assert(obj->obj.clazz->descriptor != NULL); assert(!strcmp(obj->obj.clazz->descriptor, "Ljava/lang/Class;")); assert(obj->descriptor != NULL); LOG_SCAV("scavengeClassObject: descriptor='%s',vtableCount=%zu", obj->descriptor, obj->vtableCount); /* Delegate class object and instance field scavenging. */ scavengeDataObject((Object *)obj); /* Scavenge the array element class object. */ if (IS_CLASS_FLAG_SET(obj, CLASS_ISARRAY)) { scavengeReference((Object **)(void *)&obj->elementClass); } /* Scavenge the superclass. */ scavengeReference((Object **)(void *)&obj->super); /* Scavenge the class loader. */ scavengeReference(&obj->classLoader); /* Scavenge static fields. */ for (i = 0; i < obj->sfieldCount; ++i) { char ch = obj->sfields[i].field.signature[0]; if (ch == '[' || ch == 'L') { scavengeReference((Object **)(void *)&obj->sfields[i].value.l); } } /* Scavenge interface class objects. */ for (i = 0; i < obj->interfaceCount; ++i) { scavengeReference((Object **) &obj->interfaces[i]); } } /* * Array object scavenging. */ static size_t scavengeArrayObject(ArrayObject *array) { size_t i, length; LOG_SCAV("scavengeArrayObject(array=%p)", array); /* Scavenge the class object. */ assert(toSpaceContains(array)); assert(array != NULL); assert(array->obj.clazz != NULL); scavengeReference((Object **) array); length = dvmArrayObjectSize(array); /* Scavenge the array contents. */ if (IS_CLASS_FLAG_SET(array->obj.clazz, CLASS_ISOBJECTARRAY)) { Object **contents = (Object **)array->contents; for (i = 0; i < array->length; ++i) { scavengeReference(&contents[i]); } } return length; } /* * Reference object scavenging. */ static int getReferenceFlags(const Object *obj) { int flags; flags = CLASS_ISREFERENCE | CLASS_ISWEAKREFERENCE | CLASS_ISPHANTOMREFERENCE; return GET_CLASS_FLAG_GROUP(obj->clazz, flags); } static int isSoftReference(const Object *obj) { return getReferenceFlags(obj) == CLASS_ISREFERENCE; } static int isWeakReference(const Object *obj) { return getReferenceFlags(obj) & CLASS_ISWEAKREFERENCE; } #ifndef NDEBUG static bool isPhantomReference(const Object *obj) { return getReferenceFlags(obj) & CLASS_ISPHANTOMREFERENCE; } #endif /* * Returns true if the reference was registered with a reference queue * but has not yet been appended to it. */ static bool isReferenceEnqueuable(const Object *ref) { Object *queue, *queueNext; queue = dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queue); queueNext = dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queueNext); if (queue == NULL || queueNext != NULL) { /* * There is no queue, or the reference has already * been enqueued. The Reference.enqueue() method * will do nothing even if we call it. */ return false; } /* * We need to call enqueue(), but if we called it from * here we'd probably deadlock. Schedule a call. */ return true; } /* * Schedules a reference to be appended to its reference queue. */ static void enqueueReference(Object *ref) { assert(ref != NULL); assert(dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queue) != NULL); assert(dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queueNext) == NULL); if (!dvmHeapAddRefToLargeTable(&gDvm.gcHeap->referenceOperations, ref)) { LOGE("no room for any more reference operations"); dvmAbort(); } } /* * Sets the referent field of a reference object to NULL. */ static void clearReference(Object *obj) { dvmSetFieldObject(obj, gDvm.offJavaLangRefReference_referent, NULL); } /* * Clears reference objects with white referents. */ void clearWhiteReferences(Object **list) { size_t referentOffset, queueNextOffset; bool doSignal; queueNextOffset = gDvm.offJavaLangRefReference_queueNext; referentOffset = gDvm.offJavaLangRefReference_referent; doSignal = false; while (*list != NULL) { Object *ref = *list; JValue *field = dvmFieldPtr(ref, referentOffset); Object *referent = field->l; *list = dvmGetFieldObject(ref, queueNextOffset); dvmSetFieldObject(ref, queueNextOffset, NULL); assert(referent != NULL); if (isForward(referent->clazz)) { field->l = referent = getForward(referent->clazz); continue; } if (fromSpaceContains(referent)) { /* Referent is white, clear it. */ clearReference(ref); if (isReferenceEnqueuable(ref)) { enqueueReference(ref); doSignal = true; } } } /* * If we cleared a reference with a reference queue we must notify * the heap worker to append the reference. */ if (doSignal) { dvmSignalHeapWorker(false); } assert(*list == NULL); } /* * Blackens referents subject to the soft reference preservation * policy. */ void preserveSoftReferences(Object **list) { Object *ref; Object *prev, *next; size_t referentOffset, queueNextOffset; unsigned counter; bool white; queueNextOffset = gDvm.offJavaLangRefReference_queueNext; referentOffset = gDvm.offJavaLangRefReference_referent; counter = 0; prev = next = NULL; ref = *list; while (ref != NULL) { JValue *field = dvmFieldPtr(ref, referentOffset); Object *referent = field->l; next = dvmGetFieldObject(ref, queueNextOffset); assert(referent != NULL); if (isForward(referent->clazz)) { /* Referent is black. */ field->l = referent = getForward(referent->clazz); white = false; } else { white = fromSpaceContains(referent); } if (!white && ((++counter) & 1)) { /* Referent is white and biased toward saving, gray it. */ scavengeReference((Object **)(void *)&field->l); white = true; } if (white) { /* Referent is black, unlink it. */ if (prev != NULL) { dvmSetFieldObject(ref, queueNextOffset, NULL); dvmSetFieldObject(prev, queueNextOffset, next); } } else { /* Referent is white, skip over it. */ prev = ref; } ref = next; } /* * Restart the trace with the newly gray references added to the * root set. */ scavengeBlockQueue(); } void processFinalizableReferences(void) { HeapRefTable newPendingRefs; LargeHeapRefTable *finRefs = gDvm.gcHeap->finalizableRefs; Object **ref; Object **lastRef; size_t totalPendCount; /* * All strongly, reachable objects are black. * Any white finalizable objects need to be finalized. */ /* Create a table that the new pending refs will * be added to. */ if (!dvmHeapInitHeapRefTable(&newPendingRefs)) { //TODO: mark all finalizable refs and hope that // we can schedule them next time. Watch out, // because we may be expecting to free up space // by calling finalizers. LOG_REF("no room for pending finalizations\n"); dvmAbort(); } /* * Walk through finalizableRefs and move any white references to * the list of new pending refs. */ totalPendCount = 0; while (finRefs != NULL) { Object **gapRef; size_t newPendCount = 0; gapRef = ref = finRefs->refs.table; lastRef = finRefs->refs.nextEntry; while (ref < lastRef) { if (fromSpaceContains(*ref)) { if (!dvmHeapAddToHeapRefTable(&newPendingRefs, *ref)) { //TODO: add the current table and allocate // a new, smaller one. LOG_REF("no room for any more pending finalizations: %zd\n", dvmHeapNumHeapRefTableEntries(&newPendingRefs)); dvmAbort(); } newPendCount++; } else { /* This ref is black, so will remain on finalizableRefs. */ if (newPendCount > 0) { /* Copy it up to fill the holes. */ *gapRef++ = *ref; } else { /* No holes yet; don't bother copying. */ gapRef++; } } ref++; } finRefs->refs.nextEntry = gapRef; //TODO: if the table is empty when we're done, free it. totalPendCount += newPendCount; finRefs = finRefs->next; } LOG_REF("%zd finalizers triggered.\n", totalPendCount); if (totalPendCount == 0) { /* No objects required finalization. * Free the empty temporary table. */ dvmClearReferenceTable(&newPendingRefs); return; } /* Add the new pending refs to the main list. */ if (!dvmHeapAddTableToLargeTable(&gDvm.gcHeap->pendingFinalizationRefs, &newPendingRefs)) { LOG_REF("can't insert new pending finalizations\n"); dvmAbort(); } //TODO: try compacting the main list with a memcpy loop /* Blacken the refs we just moved; we don't want them or their * children to get swept yet. */ ref = newPendingRefs.table; lastRef = newPendingRefs.nextEntry; assert(ref < lastRef); HPROF_SET_GC_SCAN_STATE(HPROF_ROOT_FINALIZING, 0); while (ref < lastRef) { scavengeReference(ref); ref++; } HPROF_CLEAR_GC_SCAN_STATE(); scavengeBlockQueue(); dvmSignalHeapWorker(false); } /* * If a reference points to from-space and has been forwarded, we snap * the pointer to its new to-space address. If the reference points * to an unforwarded from-space address we must enqueue the reference * for later processing. TODO: implement proper reference processing * and move the referent scavenging elsewhere. */ static void scavengeReferenceObject(Object *obj) { Object *referent; Object **queue; size_t referentOffset, queueNextOffset; assert(obj != NULL); LOG_SCAV("scavengeReferenceObject(obj=%p),'%s'", obj, obj->clazz->descriptor); scavengeDataObject(obj); referentOffset = gDvm.offJavaLangRefReference_referent; referent = dvmGetFieldObject(obj, referentOffset); if (referent == NULL || toSpaceContains(referent)) { return; } if (isSoftReference(obj)) { queue = &gDvm.gcHeap->softReferences; } else if (isWeakReference(obj)) { queue = &gDvm.gcHeap->weakReferences; } else { assert(isPhantomReference(obj)); queue = &gDvm.gcHeap->phantomReferences; } queueNextOffset = gDvm.offJavaLangRefReference_queueNext; dvmSetFieldObject(obj, queueNextOffset, *queue); *queue = obj; LOG_SCAV("scavengeReferenceObject: enqueueing %p", obj); } /* * Data object scavenging. */ static void scavengeDataObject(Object *obj) { ClassObject *clazz; int i; // LOG_SCAV("scavengeDataObject(obj=%p)", obj); assert(obj != NULL); assert(obj->clazz != NULL); assert(obj->clazz->objectSize != 0); assert(toSpaceContains(obj)); /* Scavenge the class object. */ clazz = obj->clazz; scavengeReference((Object **) obj); /* Scavenge instance fields. */ if (clazz->refOffsets != CLASS_WALK_SUPER) { size_t refOffsets = clazz->refOffsets; while (refOffsets != 0) { size_t rshift = CLZ(refOffsets); size_t offset = CLASS_OFFSET_FROM_CLZ(rshift); Object **ref = (Object **)((u1 *)obj + offset); scavengeReference(ref); refOffsets &= ~(CLASS_HIGH_BIT >> rshift); } } else { for (; clazz != NULL; clazz = clazz->super) { InstField *field = clazz->ifields; for (i = 0; i < clazz->ifieldRefCount; ++i, ++field) { size_t offset = field->byteOffset; Object **ref = (Object **)((u1 *)obj + offset); scavengeReference(ref); } } } } static Object *transportObject(const Object *fromObj) { Object *toObj; size_t allocSize, copySize; LOG_TRAN("transportObject(fromObj=%p) allocBlocks=%zu", fromObj, gDvm.gcHeap->heapSource->allocBlocks); assert(fromObj != NULL); assert(fromSpaceContains(fromObj)); allocSize = copySize = objectSize(fromObj); if (LW_HASH_STATE(fromObj->lock) != LW_HASH_STATE_UNHASHED) { /* * The object has been hashed or hashed and moved. We must * reserve an additional word for a hash code. */ allocSize += sizeof(u4); } if (LW_HASH_STATE(fromObj->lock) == LW_HASH_STATE_HASHED_AND_MOVED) { /* * The object has its hash code allocated. Ensure the hash * code is copied along with the instance data. */ copySize += sizeof(u4); } /* TODO(cshapiro): don't copy, re-map large data objects. */ assert(copySize <= allocSize); toObj = allocateGray(allocSize); assert(toObj != NULL); assert(toSpaceContains(toObj)); memcpy(toObj, fromObj, copySize); if (LW_HASH_STATE(fromObj->lock) == LW_HASH_STATE_HASHED) { /* * The object has had its hash code exposed. Append it to the * instance and set a bit so we know to look for it there. */ *(u4 *)(((char *)toObj) + copySize) = (u4)fromObj >> 3; toObj->lock |= LW_HASH_STATE_HASHED_AND_MOVED << LW_HASH_STATE_SHIFT; } LOG_TRAN("transportObject: from %p/%zu to %p/%zu (%zu,%zu) %s", fromObj, addressToBlock(gDvm.gcHeap->heapSource,fromObj), toObj, addressToBlock(gDvm.gcHeap->heapSource,toObj), copySize, allocSize, copySize < allocSize ? "DIFFERENT" : ""); return toObj; } /* * Generic reference scavenging. */ /* * Given a reference to an object, the scavenge routine will gray the * reference. Any objects pointed to by the scavenger object will be * transported to new space and a forwarding pointer will be installed * in the header of the object. */ /* * Blacken the given pointer. If the pointer is in from space, it is * transported to new space. If the object has a forwarding pointer * installed it has already been transported and the referent is * snapped to the new address. */ static void scavengeReference(Object **obj) { ClassObject *clazz; Object *fromObj, *toObj; assert(obj); if (*obj == NULL) return; assert(dvmIsValidObject(*obj)); /* The entire block is black. */ if (toSpaceContains(*obj)) { LOG_SCAV("scavengeReference skipping pinned object @ %p", *obj); return; } LOG_SCAV("scavengeReference(*obj=%p)", *obj); assert(fromSpaceContains(*obj)); clazz = (*obj)->clazz; if (isForward(clazz)) { // LOG_SCAV("forwarding %p @ %p to %p", *obj, obj, (void *)((uintptr_t)clazz & ~0x1)); *obj = (Object *)getForward(clazz); return; } fromObj = *obj; if (clazz == NULL) { // LOG_SCAV("scavangeReference %p has a NULL class object", fromObj); assert(!"implemented"); toObj = NULL; } else { toObj = transportObject(fromObj); } setForward(toObj, fromObj); *obj = (Object *)toObj; } /* * Generic object scavenging. */ static void scavengeObject(Object *obj) { ClassObject *clazz; assert(obj != NULL); assert(obj->clazz != NULL); assert(!((uintptr_t)obj->clazz & 0x1)); clazz = obj->clazz; if (clazz == gDvm.classJavaLangClass) { scavengeClassObject((ClassObject *)obj); } else if (IS_CLASS_FLAG_SET(clazz, CLASS_ISARRAY)) { scavengeArrayObject((ArrayObject *)obj); } else if (IS_CLASS_FLAG_SET(clazz, CLASS_ISREFERENCE)) { scavengeReferenceObject(obj); } else { scavengeDataObject(obj); } } /* * External root scavenging routines. */ static void pinHashTableEntries(HashTable *table) { HashEntry *entry; void *obj; int i; LOG_PIN(">>> pinHashTableEntries(table=%p)", table); if (table == NULL) { return; } dvmHashTableLock(table); for (i = 0; i < table->tableSize; ++i) { entry = &table->pEntries[i]; obj = entry->data; if (obj == NULL || obj == HASH_TOMBSTONE) { continue; } pinObject(entry->data); } dvmHashTableUnlock(table); LOG_PIN("<<< pinHashTableEntries(table=%p)", table); } static void pinPrimitiveClasses(void) { size_t length; size_t i; length = ARRAYSIZE(gDvm.primitiveClass); for (i = 0; i < length; i++) { if (gDvm.primitiveClass[i] != NULL) { pinObject((Object *)gDvm.primitiveClass[i]); } } } /* * Scavenge interned strings. Permanent interned strings will have * been pinned and are therefore ignored. Non-permanent strings that * have been forwarded are snapped. All other entries are removed. */ static void scavengeInternedStrings(void) { HashTable *table; HashEntry *entry; Object *obj; int i; table = gDvm.internedStrings; if (table == NULL) { return; } dvmHashTableLock(table); for (i = 0; i < table->tableSize; ++i) { entry = &table->pEntries[i]; obj = (Object *)entry->data; if (obj == NULL || obj == HASH_TOMBSTONE) { continue; } else if (!isPermanentString((StringObject *)obj)) { // LOG_SCAV("entry->data=%p", entry->data); LOG_SCAV(">>> string obj=%p", entry->data); /* TODO(cshapiro): detach white string objects */ scavengeReference((Object **)(void *)&entry->data); LOG_SCAV("<<< string obj=%p", entry->data); } } dvmHashTableUnlock(table); } static void pinInternedStrings(void) { HashTable *table; HashEntry *entry; Object *obj; int i; table = gDvm.internedStrings; if (table == NULL) { return; } dvmHashTableLock(table); for (i = 0; i < table->tableSize; ++i) { entry = &table->pEntries[i]; obj = (Object *)entry->data; if (obj == NULL || obj == HASH_TOMBSTONE) { continue; } else if (isPermanentString((StringObject *)obj)) { obj = (Object *)getPermanentString((StringObject*)obj); LOG_PROM(">>> pin string obj=%p", obj); pinObject(obj); LOG_PROM("<<< pin string obj=%p", obj); } } dvmHashTableUnlock(table); } /* * At present, reference tables contain references that must not be * moved by the collector. Instead of scavenging each reference in * the table we pin each referenced object. */ static void pinReferenceTable(const ReferenceTable *table) { Object **entry; assert(table != NULL); assert(table->table != NULL); assert(table->nextEntry != NULL); for (entry = table->table; entry < table->nextEntry; ++entry) { assert(entry != NULL); assert(!isForward(*entry)); pinObject(*entry); } } static void scavengeLargeHeapRefTable(LargeHeapRefTable *table) { for (; table != NULL; table = table->next) { Object **ref = table->refs.table; for (; ref < table->refs.nextEntry; ++ref) { scavengeReference(ref); } } } /* This code was copied from Thread.c */ static void scavengeThreadStack(Thread *thread) { const u4 *framePtr; #if WITH_EXTRA_GC_CHECKS > 1 bool first = true; #endif framePtr = (const u4 *)thread->curFrame; while (framePtr != NULL) { const StackSaveArea *saveArea; const Method *method; saveArea = SAVEAREA_FROM_FP(framePtr); method = saveArea->method; if (method != NULL && !dvmIsNativeMethod(method)) { #ifdef COUNT_PRECISE_METHODS /* the GC is running, so no lock required */ if (dvmPointerSetAddEntry(gDvm.preciseMethods, method)) LOG_SCAV("PGC: added %s.%s %p\n", method->clazz->descriptor, method->name, method); #endif #if WITH_EXTRA_GC_CHECKS > 1 /* * May also want to enable the memset() in the "invokeMethod" * goto target in the portable interpreter. That sets the stack * to a pattern that makes referring to uninitialized data * very obvious. */ if (first) { /* * First frame, isn't native, check the "alternate" saved PC * as a sanity check. * * It seems like we could check the second frame if the first * is native, since the PCs should be the same. It turns out * this doesn't always work. The problem is that we could * have calls in the sequence: * interp method #2 * native method * interp method #1 * * and then GC while in the native method after returning * from interp method #2. The currentPc on the stack is * for interp method #1, but thread->currentPc2 is still * set for the last thing interp method #2 did. * * This can also happen in normal execution: * - sget-object on not-yet-loaded class * - class init updates currentPc2 * - static field init is handled by parsing annotations; * static String init requires creation of a String object, * which can cause a GC * * Essentially, any pattern that involves executing * interpreted code and then causes an allocation without * executing instructions in the original method will hit * this. These are rare enough that the test still has * some value. */ if (saveArea->xtra.currentPc != thread->currentPc2) { LOGW("PGC: savedPC(%p) != current PC(%p), %s.%s ins=%p\n", saveArea->xtra.currentPc, thread->currentPc2, method->clazz->descriptor, method->name, method->insns); if (saveArea->xtra.currentPc != NULL) LOGE(" pc inst = 0x%04x\n", *saveArea->xtra.currentPc); if (thread->currentPc2 != NULL) LOGE(" pc2 inst = 0x%04x\n", *thread->currentPc2); dvmDumpThread(thread, false); } } else { /* * It's unusual, but not impossible, for a non-first frame * to be at something other than a method invocation. For * example, if we do a new-instance on a nonexistent class, * we'll have a lot of class loader activity on the stack * above the frame with the "new" operation. Could also * happen while we initialize a Throwable when an instruction * fails. * * So there's not much we can do here to verify the PC, * except to verify that it's a GC point. */ } assert(saveArea->xtra.currentPc != NULL); #endif const RegisterMap* pMap; const u1* regVector; int i; Method* nonConstMethod = (Method*) method; // quiet gcc pMap = dvmGetExpandedRegisterMap(nonConstMethod); //LOG_SCAV("PGC: %s.%s\n", method->clazz->descriptor, method->name); if (pMap != NULL) { /* found map, get registers for this address */ int addr = saveArea->xtra.currentPc - method->insns; regVector = dvmRegisterMapGetLine(pMap, addr); /* if (regVector == NULL) { LOG_SCAV("PGC: map but no entry for %s.%s addr=0x%04x\n", method->clazz->descriptor, method->name, addr); } else { LOG_SCAV("PGC: found map for %s.%s 0x%04x (t=%d)\n", method->clazz->descriptor, method->name, addr, thread->threadId); } */ } else { /* * No map found. If precise GC is disabled this is * expected -- we don't create pointers to the map data even * if it's present -- but if it's enabled it means we're * unexpectedly falling back on a conservative scan, so it's * worth yelling a little. */ if (gDvm.preciseGc) { LOG_SCAV("PGC: no map for %s.%s\n", method->clazz->descriptor, method->name); } regVector = NULL; } if (regVector == NULL) { /* * There are no roots to scavenge. Skip over the entire frame. */ framePtr += method->registersSize; } else { /* * Precise scan. v0 is at the lowest address on the * interpreted stack, and is the first bit in the register * vector, so we can walk through the register map and * memory in the same direction. * * A '1' bit indicates a live reference. */ u2 bits = 1 << 1; for (i = method->registersSize - 1; i >= 0; i--) { u4 rval = *framePtr; bits >>= 1; if (bits == 1) { /* set bit 9 so we can tell when we're empty */ bits = *regVector++ | 0x0100; } if (rval != 0 && (bits & 0x01) != 0) { /* * Non-null, register marked as live reference. This * should always be a valid object. */ #if WITH_EXTRA_GC_CHECKS > 0 if ((rval & 0x3) != 0 || !dvmIsValidObject((Object*) rval)) { /* this is very bad */ LOGE("PGC: invalid ref in reg %d: 0x%08x\n", method->registersSize-1 - i, rval); } else #endif { // LOG_SCAV("stack reference %u@%p", *framePtr, framePtr); /* dvmMarkObjectNonNull((Object *)rval); */ scavengeReference((Object **) framePtr); } } else { /* * Null or non-reference, do nothing at all. */ #if WITH_EXTRA_GC_CHECKS > 1 if (dvmIsValidObject((Object*) rval)) { /* this is normal, but we feel chatty */ LOGD("PGC: ignoring valid ref in reg %d: 0x%08x\n", method->registersSize-1 - i, rval); } #endif } ++framePtr; } dvmReleaseRegisterMapLine(pMap, regVector); } } /* else this is a break frame and there is nothing to gray, or * this is a native method and the registers are just the "ins", * copied from various registers in the caller's set. */ #if WITH_EXTRA_GC_CHECKS > 1 first = false; #endif /* Don't fall into an infinite loop if things get corrupted. */ assert((uintptr_t)saveArea->prevFrame > (uintptr_t)framePtr || saveArea->prevFrame == NULL); framePtr = saveArea->prevFrame; } } static void scavengeThread(Thread *thread) { // LOG_SCAV("scavengeThread(thread=%p)", thread); // LOG_SCAV("Scavenging threadObj=%p", thread->threadObj); scavengeReference(&thread->threadObj); // LOG_SCAV("Scavenging exception=%p", thread->exception); scavengeReference(&thread->exception); scavengeThreadStack(thread); } static void scavengeThreadList(void) { Thread *thread; dvmLockThreadList(dvmThreadSelf()); thread = gDvm.threadList; while (thread) { scavengeThread(thread); thread = thread->next; } dvmUnlockThreadList(); } static void pinThreadStack(const Thread *thread) { const u4 *framePtr; const StackSaveArea *saveArea; Method *method; const char *shorty; Object *obj; int i; saveArea = NULL; framePtr = (const u4 *)thread->curFrame; for (; framePtr != NULL; framePtr = saveArea->prevFrame) { saveArea = SAVEAREA_FROM_FP(framePtr); method = (Method *)saveArea->method; if (method != NULL && dvmIsNativeMethod(method)) { /* * This is native method, pin its arguments. * * For purposes of graying references, we don't need to do * anything here, because all of the native "ins" were copied * from registers in the caller's stack frame and won't be * changed (an interpreted method can freely use registers * with parameters like any other register, but natives don't * work that way). * * However, we need to ensure that references visible to * native methods don't move around. We can do a precise scan * of the arguments by examining the method signature. */ LOG_PIN("+++ native scan %s.%s\n", method->clazz->descriptor, method->name); assert(method->registersSize == method->insSize); if (!dvmIsStaticMethod(method)) { /* grab the "this" pointer */ obj = (Object *)*framePtr++; if (obj == NULL) { /* * This can happen for the "fake" entry frame inserted * for threads created outside the VM. There's no actual * call so there's no object. If we changed the fake * entry method to be declared "static" then this * situation should never occur. */ } else { assert(dvmIsValidObject(obj)); pinObject(obj); } } shorty = method->shorty+1; // skip return value for (i = method->registersSize - 1; i >= 0; i--, framePtr++) { switch (*shorty++) { case 'L': obj = (Object *)*framePtr; if (obj != NULL) { assert(dvmIsValidObject(obj)); pinObject(obj); } break; case 'D': case 'J': framePtr++; break; default: /* 32-bit non-reference value */ obj = (Object *)*framePtr; // debug, remove if (dvmIsValidObject(obj)) { // debug, remove /* if we see a lot of these, our scan might be off */ LOG_PIN("+++ did NOT pin obj %p\n", obj); } break; } } } else if (method != NULL && !dvmIsNativeMethod(method)) { const RegisterMap* pMap = dvmGetExpandedRegisterMap(method); const u1* regVector = NULL; LOGI("conservative : %s.%s\n", method->clazz->descriptor, method->name); if (pMap != NULL) { int addr = saveArea->xtra.currentPc - method->insns; regVector = dvmRegisterMapGetLine(pMap, addr); } if (regVector == NULL) { /* * No register info for this frame, conservatively pin. */ for (i = 0; i < method->registersSize; ++i) { u4 regValue = framePtr[i]; if (regValue != 0 && (regValue & 0x3) == 0 && dvmIsValidObject((Object *)regValue)) { pinObject((Object *)regValue); } } } } /* * Don't fall into an infinite loop if things get corrupted. */ assert((uintptr_t)saveArea->prevFrame > (uintptr_t)framePtr || saveArea->prevFrame == NULL); } } static void pinThread(const Thread *thread) { assert(thread != NULL); LOG_PIN("pinThread(thread=%p)", thread); LOG_PIN("Pin native method arguments"); pinThreadStack(thread); LOG_PIN("Pin internalLocalRefTable"); pinReferenceTable(&thread->internalLocalRefTable); LOG_PIN("Pin jniLocalRefTable"); pinReferenceTable(&thread->jniLocalRefTable); /* Can the check be pushed into the promote routine? */ if (thread->jniMonitorRefTable.table) { LOG_PIN("Pin jniMonitorRefTable"); pinReferenceTable(&thread->jniMonitorRefTable); } } static void pinThreadList(void) { Thread *thread; dvmLockThreadList(dvmThreadSelf()); thread = gDvm.threadList; while (thread) { pinThread(thread); thread = thread->next; } dvmUnlockThreadList(); } /* * Heap block scavenging. */ /* * Scavenge objects in the current block. Scavenging terminates when * the pointer reaches the highest address in the block or when a run * of zero words that continues to the highest address is reached. */ static void scavengeBlock(HeapSource *heapSource, size_t block) { u1 *cursor; u1 *end; size_t size; LOG_SCAV("scavengeBlock(heapSource=%p,block=%zu)", heapSource, block); assert(heapSource != NULL); assert(block < heapSource->totalBlocks); assert(heapSource->blockSpace[block] == BLOCK_TO_SPACE); cursor = blockToAddress(heapSource, block); end = cursor + BLOCK_SIZE; LOG_SCAV("scavengeBlock start=%p, end=%p", cursor, end); /* Parse and scavenge the current block. */ size = 0; while (cursor < end) { u4 word = *(u4 *)cursor; if (word != 0) { scavengeObject((Object *)cursor); size = objectSize((Object *)cursor); size = alignUp(size, ALLOC_ALIGNMENT); cursor += size; } else { /* Check for padding. */ while (*(u4 *)cursor == 0) { cursor += 4; if (cursor == end) break; } /* Punt if something went wrong. */ assert(cursor == end); } } } static size_t objectSize(const Object *obj) { size_t size; assert(obj != NULL); assert(obj->clazz != NULL); if (obj->clazz == gDvm.classJavaLangClass) { size = dvmClassObjectSize((ClassObject *)obj); } else if (IS_CLASS_FLAG_SET(obj->clazz, CLASS_ISARRAY)) { size = dvmArrayObjectSize((ArrayObject *)obj); } else { assert(obj->clazz->objectSize != 0); size = obj->clazz->objectSize; } if (LW_HASH_STATE(obj->lock) == LW_HASH_STATE_HASHED_AND_MOVED) { size += sizeof(u4); } return size; } static void verifyBlock(HeapSource *heapSource, size_t block) { u1 *cursor; u1 *end; size_t size; // LOG_VER("verifyBlock(heapSource=%p,block=%zu)", heapSource, block); assert(heapSource != NULL); assert(block < heapSource->totalBlocks); assert(heapSource->blockSpace[block] == BLOCK_TO_SPACE); cursor = blockToAddress(heapSource, block); end = cursor + BLOCK_SIZE; // LOG_VER("verifyBlock start=%p, end=%p", cursor, end); /* Parse and scavenge the current block. */ size = 0; while (cursor < end) { u4 word = *(u4 *)cursor; if (word != 0) { dvmVerifyObject((Object *)cursor); size = objectSize((Object *)cursor); size = alignUp(size, ALLOC_ALIGNMENT); cursor += size; } else { /* Check for padding. */ while (*(unsigned long *)cursor == 0) { cursor += 4; if (cursor == end) break; } /* Punt if something went wrong. */ assert(cursor == end); } } } static void describeBlockQueue(const HeapSource *heapSource) { size_t block, count; char space; block = heapSource->queueHead; count = 0; LOG_SCAV(">>> describeBlockQueue(heapSource=%p)", heapSource); /* Count the number of blocks enqueued. */ while (block != QUEUE_TAIL) { block = heapSource->blockQueue[block]; ++count; } LOG_SCAV("blockQueue %zu elements, enqueued %zu", count, heapSource->queueSize); block = heapSource->queueHead; while (block != QUEUE_TAIL) { space = heapSource->blockSpace[block]; LOG_SCAV("block=%zu@%p,space=%zu", block, blockToAddress(heapSource,block), space); block = heapSource->blockQueue[block]; } LOG_SCAV("<<< describeBlockQueue(heapSource=%p)", heapSource); } /* * Blackens promoted objects. */ static void scavengeBlockQueue(void) { HeapSource *heapSource; size_t block; LOG_SCAV(">>> scavengeBlockQueue()"); heapSource = gDvm.gcHeap->heapSource; describeBlockQueue(heapSource); while (heapSource->queueHead != QUEUE_TAIL) { block = heapSource->queueHead; LOG_SCAV("Dequeueing block %zu\n", block); scavengeBlock(heapSource, block); heapSource->queueHead = heapSource->blockQueue[block]; LOG_SCAV("New queue head is %zu\n", heapSource->queueHead); } LOG_SCAV("<<< scavengeBlockQueue()"); } /* * Scan the block list and verify all blocks that are marked as being * in new space. This should be parametrized so we can invoke this * routine outside of the context of a collection. */ static void verifyNewSpace(void) { HeapSource *heapSource; size_t i; size_t c0, c1, c2, c7; c0 = c1 = c2 = c7 = 0; heapSource = gDvm.gcHeap->heapSource; for (i = 0; i < heapSource->totalBlocks; ++i) { switch (heapSource->blockSpace[i]) { case BLOCK_FREE: ++c0; break; case BLOCK_TO_SPACE: ++c1; break; case BLOCK_FROM_SPACE: ++c2; break; case BLOCK_CONTINUED: ++c7; break; default: assert(!"reached"); } } LOG_VER("Block Demographics: " "Free=%zu,ToSpace=%zu,FromSpace=%zu,Continued=%zu", c0, c1, c2, c7); for (i = 0; i < heapSource->totalBlocks; ++i) { if (heapSource->blockSpace[i] != BLOCK_TO_SPACE) { continue; } verifyBlock(heapSource, i); } } static void scavengeGlobals(void) { scavengeReference((Object **)(void *)&gDvm.classJavaLangClass); scavengeReference((Object **)(void *)&gDvm.classJavaLangClassArray); scavengeReference((Object **)(void *)&gDvm.classJavaLangError); scavengeReference((Object **)(void *)&gDvm.classJavaLangObject); scavengeReference((Object **)(void *)&gDvm.classJavaLangObjectArray); scavengeReference((Object **)(void *)&gDvm.classJavaLangRuntimeException); scavengeReference((Object **)(void *)&gDvm.classJavaLangString); scavengeReference((Object **)(void *)&gDvm.classJavaLangThread); scavengeReference((Object **)(void *)&gDvm.classJavaLangVMThread); scavengeReference((Object **)(void *)&gDvm.classJavaLangThreadGroup); scavengeReference((Object **)(void *)&gDvm.classJavaLangThrowable); scavengeReference((Object **)(void *)&gDvm.classJavaLangStackTraceElement); scavengeReference((Object **)(void *)&gDvm.classJavaLangStackTraceElementArray); scavengeReference((Object **)(void *)&gDvm.classJavaLangAnnotationAnnotationArray); scavengeReference((Object **)(void *)&gDvm.classJavaLangAnnotationAnnotationArrayArray); scavengeReference((Object **)(void *)&gDvm.classJavaLangReflectAccessibleObject); scavengeReference((Object **)(void *)&gDvm.classJavaLangReflectConstructor); scavengeReference((Object **)(void *)&gDvm.classJavaLangReflectConstructorArray); scavengeReference((Object **)(void *)&gDvm.classJavaLangReflectField); scavengeReference((Object **)(void *)&gDvm.classJavaLangReflectFieldArray); scavengeReference((Object **)(void *)&gDvm.classJavaLangReflectMethod); scavengeReference((Object **)(void *)&gDvm.classJavaLangReflectMethodArray); scavengeReference((Object **)(void *)&gDvm.classJavaLangReflectProxy); scavengeReference((Object **)(void *)&gDvm.classJavaLangExceptionInInitializerError); scavengeReference((Object **)(void *)&gDvm.classJavaLangRefReference); scavengeReference((Object **)(void *)&gDvm.classJavaNioReadWriteDirectByteBuffer); scavengeReference((Object **)(void *)&gDvm.classJavaSecurityAccessController); scavengeReference((Object **)(void *)&gDvm.classOrgApacheHarmonyLangAnnotationAnnotationFactory); scavengeReference((Object **)(void *)&gDvm.classOrgApacheHarmonyLangAnnotationAnnotationMember); scavengeReference((Object **)(void *)&gDvm.classOrgApacheHarmonyLangAnnotationAnnotationMemberArray); scavengeReference((Object **)(void *)&gDvm.classOrgApacheHarmonyNioInternalDirectBuffer); scavengeReference((Object **)(void *)&gDvm.classArrayBoolean); scavengeReference((Object **)(void *)&gDvm.classArrayChar); scavengeReference((Object **)(void *)&gDvm.classArrayFloat); scavengeReference((Object **)(void *)&gDvm.classArrayDouble); scavengeReference((Object **)(void *)&gDvm.classArrayByte); scavengeReference((Object **)(void *)&gDvm.classArrayShort); scavengeReference((Object **)(void *)&gDvm.classArrayInt); scavengeReference((Object **)(void *)&gDvm.classArrayLong); } void describeHeap(void) { HeapSource *heapSource; heapSource = gDvm.gcHeap->heapSource; describeBlocks(heapSource); } /* * The collection interface. Collection has a few distinct phases. * The first is flipping AKA condemning AKA whitening the heap. The * second is to promote all objects which are pointed to by pinned or * ambiguous references. The third phase is tracing from the stacks, * registers and various globals. Lastly, a verification of the heap * is performed. The last phase should be optional. */ void dvmScavengeRoots(void) /* Needs a new name badly */ { GcHeap *gcHeap; { size_t alloc, unused, total; room(&alloc, &unused, &total); LOG_SCAV("BEFORE GC: %zu alloc, %zu free, %zu total.", alloc, unused, total); } gcHeap = gDvm.gcHeap; dvmHeapSourceFlip(); /* * Promote blocks with stationary objects. */ pinThreadList(); pinReferenceTable(&gDvm.jniGlobalRefTable); pinReferenceTable(&gDvm.jniPinRefTable); pinHashTableEntries(gDvm.loadedClasses); pinHashTableEntries(gDvm.dbgRegistry); pinPrimitiveClasses(); pinInternedStrings(); // describeBlocks(gcHeap->heapSource); /* * Create first, open new-space page right here. */ /* Reset allocation to an unallocated block. */ gDvm.gcHeap->heapSource->allocPtr = allocateBlocks(gDvm.gcHeap->heapSource, 1); gDvm.gcHeap->heapSource->allocLimit = gDvm.gcHeap->heapSource->allocPtr + BLOCK_SIZE; /* * Hack: promote the empty block allocated above. If the * promotions that occurred above did not actually gray any * objects, the block queue may be empty. We must force a * promotion to be safe. */ promoteBlockByAddr(gDvm.gcHeap->heapSource, gDvm.gcHeap->heapSource->allocPtr); /* * Scavenge blocks and relocate movable objects. */ LOG_SCAV("Scavenging gDvm.threadList"); scavengeThreadList(); LOG_SCAV("Scavenging gDvm.gcHeap->referenceOperations"); scavengeLargeHeapRefTable(gcHeap->referenceOperations); LOG_SCAV("Scavenging gDvm.gcHeap->pendingFinalizationRefs"); scavengeLargeHeapRefTable(gcHeap->pendingFinalizationRefs); LOG_SCAV("Scavenging random global stuff"); scavengeReference(&gDvm.outOfMemoryObj); scavengeReference(&gDvm.internalErrorObj); scavengeReference(&gDvm.noClassDefFoundErrorObj); // LOG_SCAV("Scavenging gDvm.internedString"); scavengeInternedStrings(); LOG_SCAV("Root scavenge has completed."); scavengeBlockQueue(); LOG_SCAV("Re-snap global class pointers."); scavengeGlobals(); LOG_SCAV("New space scavenge has completed."); /* * Process reference objects in strength order. */ LOG_REF("Processing soft references..."); preserveSoftReferences(&gDvm.gcHeap->softReferences); clearWhiteReferences(&gDvm.gcHeap->softReferences); LOG_REF("Processing weak references..."); clearWhiteReferences(&gDvm.gcHeap->weakReferences); LOG_REF("Finding finalizations..."); processFinalizableReferences(); LOG_REF("Processing f-reachable soft references..."); clearWhiteReferences(&gDvm.gcHeap->softReferences); LOG_REF("Processing f-reachable weak references..."); clearWhiteReferences(&gDvm.gcHeap->weakReferences); LOG_REF("Processing phantom references..."); clearWhiteReferences(&gDvm.gcHeap->phantomReferences); /* * Verify the stack and heap. */ dvmVerifyRoots(); verifyNewSpace(); //describeBlocks(gcHeap->heapSource); clearFromSpace(gcHeap->heapSource); { size_t alloc, rem, total; room(&alloc, &rem, &total); LOG_SCAV("AFTER GC: %zu alloc, %zu free, %zu total.", alloc, rem, total); } } /* * Interface compatibility routines. */ void dvmClearWhiteRefs(Object **list) { /* do nothing */ assert(*list == NULL); } void dvmHandleSoftRefs(Object **list) { /* do nothing */ assert(*list == NULL); } bool dvmHeapBeginMarkStep(GcMode mode) { /* do nothing */ return true; } void dvmHeapFinishMarkStep(void) { /* do nothing */ } void dvmHeapMarkRootSet(void) { /* do nothing */ } void dvmHeapScanMarkedObjects(void) { dvmScavengeRoots(); } void dvmHeapScheduleFinalizations(void) { /* do nothing */ } void dvmHeapSweepUnmarkedObjects(GcMode mode, int *numFreed, size_t *sizeFreed) { *numFreed = 0; *sizeFreed = 0; /* do nothing */ } void dvmMarkObjectNonNull(const Object *obj) { assert(!"implemented"); } void dvmMarkDirtyObjects(void) { assert(!"implemented"); } void dvmHeapSourceThreadShutdown(void) { /* do nothing */ }