• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (C) 2009 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #include <errno.h>
18 #include <limits.h>
19 #include <sys/mman.h>
20 
21 #include "Dalvik.h"
22 #include "alloc/Heap.h"
23 #include "alloc/HeapBitmap.h"
24 #include "alloc/HeapInternal.h"
25 #include "alloc/HeapSource.h"
26 #include "alloc/Verify.h"
27 
28 /*
29  * A "mostly copying", generational, garbage collector.
30  *
31  * TODO: we allocate our own contiguous tract of page frames to back
32  * object allocations.  To cooperate with other heaps active in the
33  * virtual machine we need to move the responsibility of allocating
34  * pages someplace outside of this code.
35  *
36  * The other major data structures that maintain the state of the heap
37  * are the block space table and the block queue.
38  *
39  * The block space table records the state of a block.  We must track
40  * whether a block is:
41  *
42  * - Free or allocated in some space.
43  *
44  * - If the block holds part of a large object allocation, whether the
45  *   block is the initial or a continued block of the allocation.
46  *
47  * - Whether the block is pinned, that is to say whether at least one
48  *   object in the block must remain stationary.  Only needed during a
49  *   GC.
50  *
51  * - Which space the object belongs to.  At present this means
52  *   from-space or to-space.
53  *
54  * The block queue is used during garbage collection.  Unlike Cheney's
55  * algorithm, from-space and to-space are not contiguous.  Therefore,
56  * one cannot maintain the state of the copy with just two pointers.
57  * The block queue exists to thread lists of blocks from the various
58  * spaces together.
59  *
60  * Additionally, we record the free space frontier of the heap, as
61  * well as the address of the first object within a block, which is
62  * required to copy objects following a large object (not currently
63  * implemented).  This is stored in the heap source structure.  This
64  * should be moved elsewhere to support in-line allocations from Java
65  * threads.
66  *
67  * Allocation requests are satisfied by reserving storage from one or
68  * more contiguous blocks.  Objects that are small enough to fit
69  * inside a block are packed together within a block.  Objects that
70  * are larger than a block are allocated from contiguous sequences of
71  * blocks.  When half the available blocks are filled, a garbage
72  * collection occurs.  We "flip" spaces (exchange from- and to-space),
73  * copy live objects into to space, and perform pointer adjustment.
74  *
75  * Copying is made more complicated by the requirement that some
76  * objects must not be moved.  This property is known as "pinning".
77  * These objects must be dealt with specially.  We use Bartlett's
78  * scheme; blocks containing such objects are grayed (promoted) at the
79  * start of a garbage collection.  By virtue of this trick, tracing
80  * from the roots proceeds as usual but all objects on those pages are
81  * considered promoted and therefore not moved.
82  *
83  * TODO: there is sufficient information within the garbage collector
84  * to implement Attardi's scheme for evacuating unpinned objects from
85  * a page that is otherwise pinned.  This would eliminate false
86  * retention caused by the large pinning granularity.
87  *
88  * We need a scheme for medium and large objects.  Ignore that for
89  * now, we can return to this later.
90  *
91  * Eventually we need to worry about promoting objects out of the
92  * copy-collected heap (tenuring) into a less volatile space.  Copying
93  * may not always be the best policy for such spaces.  We should
94  * consider a variant of mark, sweep, compact.
95  *
96  * The block scheme allows us to use VM page faults to maintain a
97  * write barrier.  Consider having a special leaf state for a page.
98  *
99  * Bibliography:
100  *
101  * C. J. Cheney. 1970. A non-recursive list compacting
102  * algorithm. CACM. 13-11 pp677--678.
103  *
104  * Joel F. Bartlett. 1988. Compacting Garbage Collection with
105  * Ambiguous Roots. Digital Equipment Corporation.
106  *
107  * Joel F. Bartlett. 1989. Mostly-Copying Garbage Collection Picks Up
108  * Generations and C++. Digital Equipment Corporation.
109  *
110  * G. May Yip. 1991. Incremental, Generational Mostly-Copying Garbage
111  * Collection in Uncooperative Environments. Digital Equipment
112  * Corporation.
113  *
114  * Giuseppe Attardi, Tito Flagella. 1994. A Customisable Memory
115  * Management Framework. TR-94-010
116  *
117  * Giuseppe Attardi, Tito Flagella, Pietro Iglio. 1998. A customisable
118  * memory management framework for C++. Software -- Practice and
119  * Experience. 28(11), 1143-1183.
120  *
121  */
122 
123 #define ARRAYSIZE(x) (sizeof(x) / sizeof(x[0]))
124 
125 #if 0
126 #define LOG_ALLOC ALOGI
127 #define LOG_PIN ALOGI
128 #define LOG_PROM ALOGI
129 #define LOG_REF ALOGI
130 #define LOG_SCAV ALOGI
131 #define LOG_TRAN ALOGI
132 #define LOG_VER ALOGI
133 #else
134 #define LOG_ALLOC(...) ((void)0)
135 #define LOG_PIN(...) ((void)0)
136 #define LOG_PROM(...) ((void)0)
137 #define LOG_REF(...) ((void)0)
138 #define LOG_SCAV(...) ((void)0)
139 #define LOG_TRAN(...) ((void)0)
140 #define LOG_VER(...) ((void)0)
141 #endif
142 
143 static void enqueueBlock(HeapSource *heapSource, size_t block);
144 static void scavengeReference(Object **obj);
145 static bool toSpaceContains(const void *addr);
146 static bool fromSpaceContains(const void *addr);
147 static size_t sumHeapBitmap(const HeapBitmap *bitmap);
148 static size_t objectSize(const Object *obj);
149 static void scavengeDataObject(Object *obj);
150 static void scavengeBlockQueue();
151 
152 /*
153  * We use 512-byte blocks.
154  */
155 enum { BLOCK_SHIFT = 9 };
156 enum { BLOCK_SIZE = 1 << BLOCK_SHIFT };
157 
158 /*
159  * Space identifiers, stored into the blockSpace array.
160  */
161 enum {
162     BLOCK_FREE = 0,
163     BLOCK_FROM_SPACE = 1,
164     BLOCK_TO_SPACE = 2,
165     BLOCK_CONTINUED = 7
166 };
167 
168 /*
169  * Alignment for all allocations, in bytes.
170  */
171 enum { ALLOC_ALIGNMENT = 8 };
172 
173 /*
174  * Sentinel value for the queue end.
175  */
176 #define QUEUE_TAIL (~(size_t)0)
177 
178 struct HeapSource {
179 
180     /* The base address of backing store. */
181     u1 *blockBase;
182 
183     /* Total number of blocks available for allocation. */
184     size_t totalBlocks;
185     size_t allocBlocks;
186 
187     /*
188      * The scavenger work queue.  Implemented as an array of index
189      * values into the queue.
190      */
191     size_t *blockQueue;
192 
193     /*
194      * Base and limit blocks.  Basically the shifted start address of
195      * the block.  We convert blocks to a relative number when
196      * indexing in the block queue.  TODO: make the block queue base
197      * relative rather than the index into the block queue.
198      */
199     size_t baseBlock, limitBlock;
200 
201     size_t queueHead;
202     size_t queueTail;
203     size_t queueSize;
204 
205     /* The space of the current block 0 (free), 1 or 2. */
206     char *blockSpace;
207 
208     /* Start of free space in the current block. */
209     u1 *allocPtr;
210     /* Exclusive limit of free space in the current block. */
211     u1 *allocLimit;
212 
213     HeapBitmap allocBits;
214 
215     /*
216      * The starting size of the heap.  This value is the same as the
217      * value provided to the -Xms flag.
218      */
219     size_t minimumSize;
220 
221     /*
222      * The maximum size of the heap.  This value is the same as the
223      * -Xmx flag.
224      */
225     size_t maximumSize;
226 
227     /*
228      * The current, committed size of the heap.  At present, this is
229      * equivalent to the maximumSize.
230      */
231     size_t currentSize;
232 
233     size_t bytesAllocated;
234 };
235 
alignDown(unsigned long x,unsigned long n)236 static unsigned long alignDown(unsigned long x, unsigned long n)
237 {
238     return x & -n;
239 }
240 
alignUp(unsigned long x,unsigned long n)241 static unsigned long alignUp(unsigned long x, unsigned long n)
242 {
243     return alignDown(x + (n - 1), n);
244 }
245 
describeBlocks(const HeapSource * heapSource)246 static void describeBlocks(const HeapSource *heapSource)
247 {
248     for (size_t i = 0; i < heapSource->totalBlocks; ++i) {
249         if ((i % 32) == 0) putchar('\n');
250         printf("%d ", heapSource->blockSpace[i]);
251     }
252     putchar('\n');
253 }
254 
255 /*
256  * Virtual memory interface.
257  */
258 
virtualAlloc(size_t length)259 static void *virtualAlloc(size_t length)
260 {
261     int flags = MAP_PRIVATE | MAP_ANONYMOUS;
262     int prot = PROT_READ | PROT_WRITE;
263     void *addr = mmap(NULL, length, prot, flags, -1, 0);
264     if (addr == MAP_FAILED) {
265         LOGE_HEAP("mmap: %s", strerror(errno));
266         addr = NULL;
267     }
268     return addr;
269 }
270 
virtualFree(void * addr,size_t length)271 static void virtualFree(void *addr, size_t length)
272 {
273     assert(addr != NULL);
274     assert((uintptr_t)addr % SYSTEM_PAGE_SIZE == 0);
275     int res = munmap(addr, length);
276     if (res == -1) {
277         LOGE_HEAP("munmap: %s", strerror(errno));
278     }
279 }
280 
281 #ifndef NDEBUG
isValidAddress(const HeapSource * heapSource,const u1 * addr)282 static int isValidAddress(const HeapSource *heapSource, const u1 *addr)
283 {
284     size_t block;
285 
286     block = (uintptr_t)addr >> BLOCK_SHIFT;
287     return heapSource->baseBlock <= block &&
288            heapSource->limitBlock > block;
289 }
290 #endif
291 
292 /*
293  * Iterate over the block map looking for a contiguous run of free
294  * blocks.
295  */
allocateBlocks(HeapSource * heapSource,size_t blocks)296 static void *allocateBlocks(HeapSource *heapSource, size_t blocks)
297 {
298     size_t allocBlocks = heapSource->allocBlocks;
299     size_t totalBlocks = heapSource->totalBlocks;
300     /* Check underflow. */
301     assert(blocks != 0);
302     /* Check overflow. */
303     if (allocBlocks + blocks > totalBlocks / 2) {
304         return NULL;
305     }
306     /* Scan block map. */
307     for (size_t i = 0; i < totalBlocks; ++i) {
308         /* Check fit. */
309         for (size_t j = 0; j < blocks; ++j) { /* runs over totalBlocks */
310             if (heapSource->blockSpace[i+j] != BLOCK_FREE) {
311                 break;
312             }
313         }
314         /* No fit? */
315         if (j != blocks) {
316             i += j;
317             continue;
318         }
319         /* Fit, allocate. */
320         heapSource->blockSpace[i] = BLOCK_TO_SPACE; /* why to-space? */
321         for (size_t j = 1; j < blocks; ++j) {
322             heapSource->blockSpace[i+j] = BLOCK_CONTINUED;
323         }
324         heapSource->allocBlocks += blocks;
325         void *addr = &heapSource->blockBase[i*BLOCK_SIZE];
326         memset(addr, 0, blocks*BLOCK_SIZE);
327         /* Collecting? */
328         if (heapSource->queueHead != QUEUE_TAIL) {
329             LOG_ALLOC("allocateBlocks allocBlocks=%zu,block#=%zu", heapSource->allocBlocks, i);
330             /*
331              * This allocated was on behalf of the transporter when it
332              * shaded a white object gray.  We enqueue the block so
333              * the scavenger can further shade the gray objects black.
334              */
335             enqueueBlock(heapSource, i);
336         }
337 
338         return addr;
339     }
340     /* Insufficient space, fail. */
341     ALOGE("Insufficient space, %zu blocks, %zu blocks allocated and %zu bytes allocated",
342          heapSource->totalBlocks,
343          heapSource->allocBlocks,
344          heapSource->bytesAllocated);
345     return NULL;
346 }
347 
348 /* Converts an absolute address to a relative block number. */
addressToBlock(const HeapSource * heapSource,const void * addr)349 static size_t addressToBlock(const HeapSource *heapSource, const void *addr)
350 {
351     assert(heapSource != NULL);
352     assert(isValidAddress(heapSource, addr));
353     return (((uintptr_t)addr) >> BLOCK_SHIFT) - heapSource->baseBlock;
354 }
355 
356 /* Converts a relative block number to an absolute address. */
blockToAddress(const HeapSource * heapSource,size_t block)357 static u1 *blockToAddress(const HeapSource *heapSource, size_t block)
358 {
359     u1 *addr;
360 
361     addr = (u1 *) (((uintptr_t) heapSource->baseBlock + block) * BLOCK_SIZE);
362     assert(isValidAddress(heapSource, addr));
363     return addr;
364 }
365 
clearBlock(HeapSource * heapSource,size_t block)366 static void clearBlock(HeapSource *heapSource, size_t block)
367 {
368     assert(heapSource != NULL);
369     assert(block < heapSource->totalBlocks);
370     u1 *addr = heapSource->blockBase + block*BLOCK_SIZE;
371     memset(addr, 0xCC, BLOCK_SIZE);
372     for (size_t i = 0; i < BLOCK_SIZE; i += 8) {
373         dvmHeapBitmapClearObjectBit(&heapSource->allocBits, addr + i);
374     }
375 }
376 
clearFromSpace(HeapSource * heapSource)377 static void clearFromSpace(HeapSource *heapSource)
378 {
379     assert(heapSource != NULL);
380     size_t i = 0;
381     size_t count = 0;
382     while (i < heapSource->totalBlocks) {
383         if (heapSource->blockSpace[i] != BLOCK_FROM_SPACE) {
384             ++i;
385             continue;
386         }
387         heapSource->blockSpace[i] = BLOCK_FREE;
388         clearBlock(heapSource, i);
389         ++i;
390         ++count;
391         while (i < heapSource->totalBlocks &&
392                heapSource->blockSpace[i] == BLOCK_CONTINUED) {
393             heapSource->blockSpace[i] = BLOCK_FREE;
394             clearBlock(heapSource, i);
395             ++i;
396             ++count;
397         }
398     }
399     LOG_SCAV("freed %zu blocks (%zu bytes)", count, count*BLOCK_SIZE);
400 }
401 
402 /*
403  * Appends the given block to the block queue.  The block queue is
404  * processed in-order by the scavenger.
405  */
enqueueBlock(HeapSource * heapSource,size_t block)406 static void enqueueBlock(HeapSource *heapSource, size_t block)
407 {
408     assert(heapSource != NULL);
409     assert(block < heapSource->totalBlocks);
410     if (heapSource->queueHead != QUEUE_TAIL) {
411         heapSource->blockQueue[heapSource->queueTail] = block;
412     } else {
413         heapSource->queueHead = block;
414     }
415     heapSource->blockQueue[block] = QUEUE_TAIL;
416     heapSource->queueTail = block;
417     ++heapSource->queueSize;
418 }
419 
420 /*
421  * Grays all objects within the block corresponding to the given
422  * address.
423  */
promoteBlockByAddr(HeapSource * heapSource,const void * addr)424 static void promoteBlockByAddr(HeapSource *heapSource, const void *addr)
425 {
426     size_t block;
427 
428     block = addressToBlock(heapSource, (const u1 *)addr);
429     if (heapSource->blockSpace[block] != BLOCK_TO_SPACE) {
430         // LOG_PROM("promoting block %zu %d @ %p", block, heapSource->blockSpace[block], obj);
431         heapSource->blockSpace[block] = BLOCK_TO_SPACE;
432         enqueueBlock(heapSource, block);
433         /* TODO(cshapiro): count continued blocks?*/
434         heapSource->allocBlocks += 1;
435     } else {
436         // LOG_PROM("NOT promoting block %zu %d @ %p", block, heapSource->blockSpace[block], obj);
437     }
438 }
439 
dvmHeapSourceStartup(size_t startSize,size_t absoluteMaxSize)440 GcHeap *dvmHeapSourceStartup(size_t startSize, size_t absoluteMaxSize)
441 {
442     GcHeap* gcHeap;
443     HeapSource *heapSource;
444 
445     assert(startSize <= absoluteMaxSize);
446 
447     heapSource = calloc(1, sizeof(*heapSource));
448     assert(heapSource != NULL);
449 
450     heapSource->minimumSize = alignUp(startSize, BLOCK_SIZE);
451     heapSource->maximumSize = alignUp(absoluteMaxSize, BLOCK_SIZE);
452 
453     heapSource->currentSize = heapSource->maximumSize;
454 
455     /* Allocate underlying storage for blocks. */
456     heapSource->blockBase = virtualAlloc(heapSource->maximumSize);
457     assert(heapSource->blockBase != NULL);
458     heapSource->baseBlock = (uintptr_t) heapSource->blockBase >> BLOCK_SHIFT;
459     heapSource->limitBlock = ((uintptr_t) heapSource->blockBase + heapSource->maximumSize) >> BLOCK_SHIFT;
460 
461     heapSource->allocBlocks = 0;
462     heapSource->totalBlocks = (heapSource->limitBlock - heapSource->baseBlock);
463 
464     assert(heapSource->totalBlocks = heapSource->maximumSize / BLOCK_SIZE);
465 
466     {
467         size_t size = sizeof(heapSource->blockQueue[0]);
468         heapSource->blockQueue = malloc(heapSource->totalBlocks*size);
469         assert(heapSource->blockQueue != NULL);
470         memset(heapSource->blockQueue, 0xCC, heapSource->totalBlocks*size);
471         heapSource->queueHead = QUEUE_TAIL;
472     }
473 
474     /* Byte indicating space residence or free status of block. */
475     {
476         size_t size = sizeof(heapSource->blockSpace[0]);
477         heapSource->blockSpace = calloc(1, heapSource->totalBlocks*size);
478         assert(heapSource->blockSpace != NULL);
479     }
480 
481     dvmHeapBitmapInit(&heapSource->allocBits,
482                       heapSource->blockBase,
483                       heapSource->maximumSize,
484                       "blockBase");
485 
486     /* Initialize allocation pointers. */
487     heapSource->allocPtr = allocateBlocks(heapSource, 1);
488     heapSource->allocLimit = heapSource->allocPtr + BLOCK_SIZE;
489 
490     gcHeap = calloc(1, sizeof(*gcHeap));
491     assert(gcHeap != NULL);
492     gcHeap->heapSource = heapSource;
493 
494     return gcHeap;
495 }
496 
497 /*
498  * Perform any required heap initializations after forking from the
499  * zygote process.  This is a no-op for the time being.  Eventually
500  * this will demarcate the shared region of the heap.
501  */
dvmHeapSourceStartupAfterZygote()502 bool dvmHeapSourceStartupAfterZygote()
503 {
504     return true;
505 }
506 
dvmHeapSourceStartupBeforeFork()507 bool dvmHeapSourceStartupBeforeFork()
508 {
509     assert(!"implemented");
510     return false;
511 }
512 
dvmHeapSourceShutdown(GcHeap ** gcHeap)513 void dvmHeapSourceShutdown(GcHeap **gcHeap)
514 {
515     if (*gcHeap == NULL || (*gcHeap)->heapSource == NULL)
516         return;
517     free((*gcHeap)->heapSource->blockQueue);
518     free((*gcHeap)->heapSource->blockSpace);
519     virtualFree((*gcHeap)->heapSource->blockBase,
520                 (*gcHeap)->heapSource->maximumSize);
521     free((*gcHeap)->heapSource);
522     (*gcHeap)->heapSource = NULL;
523     free(*gcHeap);
524     *gcHeap = NULL;
525 }
526 
dvmHeapSourceGetValue(HeapSourceValueSpec spec,size_t perHeapStats[],size_t arrayLen)527 size_t dvmHeapSourceGetValue(HeapSourceValueSpec spec,
528                              size_t perHeapStats[],
529                              size_t arrayLen)
530 {
531     HeapSource *heapSource;
532     size_t value;
533 
534     heapSource = gDvm.gcHeap->heapSource;
535     switch (spec) {
536     case HS_FOOTPRINT:
537         value = heapSource->maximumSize;
538         break;
539     case HS_ALLOWED_FOOTPRINT:
540         value = heapSource->maximumSize;
541         break;
542     case HS_BYTES_ALLOCATED:
543         value = heapSource->bytesAllocated;
544         break;
545     case HS_OBJECTS_ALLOCATED:
546         value = sumHeapBitmap(&heapSource->allocBits);
547         break;
548     default:
549         assert(!"implemented");
550         value = 0;
551     }
552     if (perHeapStats) {
553         *perHeapStats = value;
554     }
555     return value;
556 }
557 
558 /*
559  * Performs a shallow copy of the allocation bitmap into the given
560  * vector of heap bitmaps.
561  */
dvmHeapSourceGetObjectBitmaps(HeapBitmap objBits[],HeapBitmap markBits[],size_t numHeaps)562 void dvmHeapSourceGetObjectBitmaps(HeapBitmap objBits[], HeapBitmap markBits[],
563                                    size_t numHeaps)
564 {
565     assert(!"implemented");
566 }
567 
dvmHeapSourceGetLiveBits()568 HeapBitmap *dvmHeapSourceGetLiveBits()
569 {
570     return &gDvm.gcHeap->heapSource->allocBits;
571 }
572 
573 /*
574  * Allocate the specified number of bytes from the heap.  The
575  * allocation cursor points into a block of free storage.  If the
576  * given allocation fits in the remaining space of the block, we
577  * advance the cursor and return a pointer to the free storage.  If
578  * the allocation cannot fit in the current block but is smaller than
579  * a block we request a new block and allocate from it instead.  If
580  * the allocation is larger than a block we must allocate from a span
581  * of contiguous blocks.
582  */
dvmHeapSourceAlloc(size_t length)583 void *dvmHeapSourceAlloc(size_t length)
584 {
585     HeapSource *heapSource;
586     unsigned char *addr;
587     size_t aligned, available, blocks;
588 
589     heapSource = gDvm.gcHeap->heapSource;
590     assert(heapSource != NULL);
591     assert(heapSource->allocPtr != NULL);
592     assert(heapSource->allocLimit != NULL);
593 
594     aligned = alignUp(length, ALLOC_ALIGNMENT);
595     available = heapSource->allocLimit - heapSource->allocPtr;
596 
597     /* Try allocating inside the current block. */
598     if (aligned <= available) {
599         addr = heapSource->allocPtr;
600         heapSource->allocPtr += aligned;
601         heapSource->bytesAllocated += aligned;
602         dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr);
603         return addr;
604     }
605 
606     /* Try allocating in a new block. */
607     if (aligned <= BLOCK_SIZE) {
608         addr =  allocateBlocks(heapSource, 1);
609         if (addr != NULL) {
610             heapSource->allocLimit = addr + BLOCK_SIZE;
611             heapSource->allocPtr = addr + aligned;
612             heapSource->bytesAllocated += aligned;
613             dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr);
614             /* TODO(cshapiro): pad out the current block. */
615         }
616         return addr;
617     }
618 
619     /* Try allocating in a span of blocks. */
620     blocks = alignUp(aligned, BLOCK_SIZE) / BLOCK_SIZE;
621 
622     addr = allocateBlocks(heapSource, blocks);
623     /* Propagate failure upward. */
624     if (addr != NULL) {
625         heapSource->bytesAllocated += aligned;
626         dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr);
627         /* TODO(cshapiro): pad out free space in the last block. */
628     }
629     return addr;
630 }
631 
dvmHeapSourceAllocAndGrow(size_t size)632 void *dvmHeapSourceAllocAndGrow(size_t size)
633 {
634     return dvmHeapSourceAlloc(size);
635 }
636 
637 /* TODO: refactor along with dvmHeapSourceAlloc */
allocateGray(size_t size)638 void *allocateGray(size_t size)
639 {
640     HeapSource *heapSource;
641     void *addr;
642     size_t block;
643 
644     /* TODO: add a check that we are in a GC. */
645     heapSource = gDvm.gcHeap->heapSource;
646     addr = dvmHeapSourceAlloc(size);
647     assert(addr != NULL);
648     block = addressToBlock(heapSource, (const u1 *)addr);
649     if (heapSource->queueHead == QUEUE_TAIL) {
650         /*
651          * Forcibly append the underlying block to the queue.  This
652          * condition occurs when referents are transported following
653          * the initial trace.
654          */
655         enqueueBlock(heapSource, block);
656         LOG_PROM("forced promoting block %zu %d @ %p", block, heapSource->blockSpace[block], addr);
657     }
658     return addr;
659 }
660 
dvmHeapSourceContainsAddress(const void * ptr)661 bool dvmHeapSourceContainsAddress(const void *ptr)
662 {
663     HeapSource *heapSource = gDvm.gcHeap->heapSource;
664     return dvmHeapBitmapCoversAddress(&heapSource->allocBits, ptr);
665 }
666 
667 /*
668  * Returns true if the given address is within the heap and points to
669  * the header of a live object.
670  */
dvmHeapSourceContains(const void * addr)671 bool dvmHeapSourceContains(const void *addr)
672 {
673     HeapSource *heapSource;
674     HeapBitmap *bitmap;
675 
676     heapSource = gDvm.gcHeap->heapSource;
677     bitmap = &heapSource->allocBits;
678     if (!dvmHeapBitmapCoversAddress(bitmap, addr)) {
679         return false;
680     } else {
681         return dvmHeapBitmapIsObjectBitSet(bitmap, addr);
682     }
683 }
684 
dvmHeapSourceGetPtrFlag(const void * ptr,HeapSourcePtrFlag flag)685 bool dvmHeapSourceGetPtrFlag(const void *ptr, HeapSourcePtrFlag flag)
686 {
687     assert(!"implemented");
688     return false;
689 }
690 
dvmHeapSourceChunkSize(const void * ptr)691 size_t dvmHeapSourceChunkSize(const void *ptr)
692 {
693     assert(!"implemented");
694     return 0;
695 }
696 
dvmHeapSourceFootprint()697 size_t dvmHeapSourceFootprint()
698 {
699     assert(!"implemented");
700     return 0;
701 }
702 
703 /*
704  * Returns the "ideal footprint" which appears to be the number of
705  * bytes currently committed to the heap.  This starts out at the
706  * start size of the heap and grows toward the maximum size.
707  */
dvmHeapSourceGetIdealFootprint()708 size_t dvmHeapSourceGetIdealFootprint()
709 {
710     return gDvm.gcHeap->heapSource->currentSize;
711 }
712 
dvmGetTargetHeapUtilization()713 float dvmGetTargetHeapUtilization()
714 {
715     return 0.5f;
716 }
717 
dvmSetTargetHeapUtilization(float newTarget)718 void dvmSetTargetHeapUtilization(float newTarget)
719 {
720     assert(newTarget > 0.0f && newTarget < 1.0f);
721 }
722 
723 /*
724  * Expands the size of the heap after a collection.  At present we
725  * commit the pages for maximum size of the heap so this routine is
726  * just a no-op.  Eventually, we will either allocate or commit pages
727  * on an as-need basis.
728  */
dvmHeapSourceGrowForUtilization()729 void dvmHeapSourceGrowForUtilization()
730 {
731     /* do nothing */
732 }
733 
dvmHeapSourceWalk(void (* callback)(void * start,void * end,size_t used_bytes,void * arg),void * arg)734 void dvmHeapSourceWalk(void(*callback)(void* start, void* end,
735                                        size_t used_bytes, void* arg),
736                        void *arg)
737 {
738     assert(!"implemented");
739 }
740 
dvmHeapSourceGetNumHeaps()741 size_t dvmHeapSourceGetNumHeaps()
742 {
743     return 1;
744 }
745 
dvmTrackExternalAllocation(size_t n)746 bool dvmTrackExternalAllocation(size_t n)
747 {
748     /* do nothing */
749     return true;
750 }
751 
dvmTrackExternalFree(size_t n)752 void dvmTrackExternalFree(size_t n)
753 {
754     /* do nothing */
755 }
756 
dvmGetExternalBytesAllocated()757 size_t dvmGetExternalBytesAllocated()
758 {
759     assert(!"implemented");
760     return 0;
761 }
762 
dvmHeapSourceFlip()763 void dvmHeapSourceFlip()
764 {
765     HeapSource *heapSource = gDvm.gcHeap->heapSource;
766 
767     /* Reset the block queue. */
768     heapSource->allocBlocks = 0;
769     heapSource->queueSize = 0;
770     heapSource->queueHead = QUEUE_TAIL;
771 
772     /* TODO(cshapiro): pad the current (prev) block. */
773 
774     heapSource->allocPtr = NULL;
775     heapSource->allocLimit = NULL;
776 
777     /* Whiten all allocated blocks. */
778     for (size_t i = 0; i < heapSource->totalBlocks; ++i) {
779         if (heapSource->blockSpace[i] == BLOCK_TO_SPACE) {
780             heapSource->blockSpace[i] = BLOCK_FROM_SPACE;
781         }
782     }
783 }
784 
room(size_t * alloc,size_t * avail,size_t * total)785 static void room(size_t *alloc, size_t *avail, size_t *total)
786 {
787     HeapSource *heapSource = gDvm.gcHeap->heapSource;
788     *total = heapSource->totalBlocks*BLOCK_SIZE;
789     *alloc = heapSource->allocBlocks*BLOCK_SIZE;
790     *avail = *total - *alloc;
791 }
792 
isSpaceInternal(u1 * addr,int space)793 static bool isSpaceInternal(u1 *addr, int space)
794 {
795     HeapSource *heapSource;
796     u1 *base, *limit;
797     size_t offset;
798     char space2;
799 
800     heapSource = gDvm.gcHeap->heapSource;
801     base = heapSource->blockBase;
802     assert(addr >= base);
803     limit = heapSource->blockBase + heapSource->maximumSize;
804     assert(addr < limit);
805     offset = addr - base;
806     space2 = heapSource->blockSpace[offset >> BLOCK_SHIFT];
807     return space == space2;
808 }
809 
fromSpaceContains(const void * addr)810 static bool fromSpaceContains(const void *addr)
811 {
812     return isSpaceInternal((u1 *)addr, BLOCK_FROM_SPACE);
813 }
814 
toSpaceContains(const void * addr)815 static bool toSpaceContains(const void *addr)
816 {
817     return isSpaceInternal((u1 *)addr, BLOCK_TO_SPACE);
818 }
819 
820 /*
821  * Notifies the collector that the object at the given address must
822  * remain stationary during the current collection.
823  */
pinObject(const Object * obj)824 static void pinObject(const Object *obj)
825 {
826     promoteBlockByAddr(gDvm.gcHeap->heapSource, obj);
827 }
828 
sumHeapBitmap(const HeapBitmap * bitmap)829 static size_t sumHeapBitmap(const HeapBitmap *bitmap)
830 {
831     size_t sum = 0;
832     for (size_t i = 0; i < bitmap->bitsLen >> 2; ++i) {
833         sum += CLZ(bitmap->bits[i]);
834     }
835     return sum;
836 }
837 
838 /*
839  * Miscellaneous functionality.
840  */
841 
isForward(const void * addr)842 static int isForward(const void *addr)
843 {
844     return (uintptr_t)addr & 0x1;
845 }
846 
setForward(const void * toObj,void * fromObj)847 static void setForward(const void *toObj, void *fromObj)
848 {
849     *(unsigned long *)fromObj = (uintptr_t)toObj | 0x1;
850 }
851 
getForward(const void * fromObj)852 static void* getForward(const void *fromObj)
853 {
854     return (void *)((uintptr_t)fromObj & ~0x1);
855 }
856 
857 /* Beware, uses the same encoding as a forwarding pointers! */
isPermanentString(const StringObject * obj)858 static int isPermanentString(const StringObject *obj) {
859     return (uintptr_t)obj & 0x1;
860 }
861 
getPermanentString(const StringObject * obj)862 static void* getPermanentString(const StringObject *obj)
863 {
864     return (void *)((uintptr_t)obj & ~0x1);
865 }
866 
867 
868 /*
869  * Scavenging and transporting routines follow.  A transporter grays
870  * an object.  A scavenger blackens an object.  We define these
871  * routines for each fundamental object type.  Dispatch is performed
872  * in scavengeObject.
873  */
874 
875 /*
876  * Class object scavenging.
877  */
scavengeClassObject(ClassObject * obj)878 static void scavengeClassObject(ClassObject *obj)
879 {
880     LOG_SCAV("scavengeClassObject(obj=%p)", obj);
881     assert(obj != NULL);
882     assert(obj->obj.clazz != NULL);
883     assert(obj->obj.clazz->descriptor != NULL);
884     assert(!strcmp(obj->obj.clazz->descriptor, "Ljava/lang/Class;"));
885     assert(obj->descriptor != NULL);
886     LOG_SCAV("scavengeClassObject: descriptor='%s',vtableCount=%zu",
887              obj->descriptor, obj->vtableCount);
888     /* Delegate class object and instance field scavenging. */
889     scavengeDataObject((Object *)obj);
890     /* Scavenge the array element class object. */
891     if (IS_CLASS_FLAG_SET(obj, CLASS_ISARRAY)) {
892         scavengeReference((Object **)(void *)&obj->elementClass);
893     }
894     /* Scavenge the superclass. */
895     scavengeReference((Object **)(void *)&obj->super);
896     /* Scavenge the class loader. */
897     scavengeReference(&obj->classLoader);
898     /* Scavenge static fields. */
899     for (int i = 0; i < obj->sfieldCount; ++i) {
900         char ch = obj->sfields[i].field.signature[0];
901         if (ch == '[' || ch == 'L') {
902             scavengeReference((Object **)(void *)&obj->sfields[i].value.l);
903         }
904     }
905     /* Scavenge interface class objects. */
906     for (int i = 0; i < obj->interfaceCount; ++i) {
907         scavengeReference((Object **) &obj->interfaces[i]);
908     }
909 }
910 
911 /*
912  * Array object scavenging.
913  */
scavengeArrayObject(ArrayObject * array)914 static size_t scavengeArrayObject(ArrayObject *array)
915 {
916     LOG_SCAV("scavengeArrayObject(array=%p)", array);
917     /* Scavenge the class object. */
918     assert(toSpaceContains(array));
919     assert(array != NULL);
920     assert(array->obj.clazz != NULL);
921     scavengeReference((Object **) array);
922     size_t length = dvmArrayObjectSize(array);
923     /* Scavenge the array contents. */
924     if (IS_CLASS_FLAG_SET(array->obj.clazz, CLASS_ISOBJECTARRAY)) {
925         Object **contents = (Object **)array->contents;
926         for (size_t i = 0; i < array->length; ++i) {
927             scavengeReference(&contents[i]);
928         }
929     }
930     return length;
931 }
932 
933 /*
934  * Reference object scavenging.
935  */
936 
getReferenceFlags(const Object * obj)937 static int getReferenceFlags(const Object *obj)
938 {
939     int flags;
940 
941     flags = CLASS_ISREFERENCE |
942             CLASS_ISWEAKREFERENCE |
943             CLASS_ISPHANTOMREFERENCE;
944     return GET_CLASS_FLAG_GROUP(obj->clazz, flags);
945 }
946 
isSoftReference(const Object * obj)947 static int isSoftReference(const Object *obj)
948 {
949     return getReferenceFlags(obj) == CLASS_ISREFERENCE;
950 }
951 
isWeakReference(const Object * obj)952 static int isWeakReference(const Object *obj)
953 {
954     return getReferenceFlags(obj) & CLASS_ISWEAKREFERENCE;
955 }
956 
957 #ifndef NDEBUG
isPhantomReference(const Object * obj)958 static bool isPhantomReference(const Object *obj)
959 {
960     return getReferenceFlags(obj) & CLASS_ISPHANTOMREFERENCE;
961 }
962 #endif
963 
964 /*
965  * Returns true if the reference was registered with a reference queue
966  * but has not yet been appended to it.
967  */
isReferenceEnqueuable(const Object * ref)968 static bool isReferenceEnqueuable(const Object *ref)
969 {
970     Object *queue, *queueNext;
971 
972     queue = dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queue);
973     queueNext = dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queueNext);
974     if (queue == NULL || queueNext != NULL) {
975         /*
976          * There is no queue, or the reference has already
977          * been enqueued.  The Reference.enqueue() method
978          * will do nothing even if we call it.
979          */
980         return false;
981     }
982 
983     /*
984      * We need to call enqueue(), but if we called it from
985      * here we'd probably deadlock.  Schedule a call.
986      */
987     return true;
988 }
989 
990 /*
991  * Schedules a reference to be appended to its reference queue.
992  */
enqueueReference(Object * ref)993 static void enqueueReference(Object *ref)
994 {
995     assert(ref != NULL);
996     assert(dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queue) != NULL);
997     assert(dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queueNext) == NULL);
998     if (!dvmHeapAddRefToLargeTable(&gDvm.gcHeap->referenceOperations, ref)) {
999         ALOGE("no room for any more reference operations");
1000         dvmAbort();
1001     }
1002 }
1003 
1004 /*
1005  * Sets the referent field of a reference object to NULL.
1006  */
clearReference(Object * obj)1007 static void clearReference(Object *obj)
1008 {
1009     dvmSetFieldObject(obj, gDvm.offJavaLangRefReference_referent, NULL);
1010 }
1011 
1012 /*
1013  * Clears reference objects with white referents.
1014  */
clearWhiteReferences(Object ** list)1015 void clearWhiteReferences(Object **list)
1016 {
1017     size_t referentOffset, queueNextOffset;
1018     bool doSignal;
1019 
1020     queueNextOffset = gDvm.offJavaLangRefReference_queueNext;
1021     referentOffset = gDvm.offJavaLangRefReference_referent;
1022     doSignal = false;
1023     while (*list != NULL) {
1024         Object *ref = *list;
1025         JValue *field = dvmFieldPtr(ref, referentOffset);
1026         Object *referent = field->l;
1027         *list = dvmGetFieldObject(ref, queueNextOffset);
1028         dvmSetFieldObject(ref, queueNextOffset, NULL);
1029         assert(referent != NULL);
1030         if (isForward(referent->clazz)) {
1031             field->l = referent = getForward(referent->clazz);
1032             continue;
1033         }
1034         if (fromSpaceContains(referent)) {
1035             /* Referent is white, clear it. */
1036             clearReference(ref);
1037             if (isReferenceEnqueuable(ref)) {
1038                 enqueueReference(ref);
1039                 doSignal = true;
1040             }
1041         }
1042     }
1043     /*
1044      * If we cleared a reference with a reference queue we must notify
1045      * the heap worker to append the reference.
1046      */
1047     if (doSignal) {
1048         dvmSignalHeapWorker(false);
1049     }
1050     assert(*list == NULL);
1051 }
1052 
1053 /*
1054  * Blackens referents subject to the soft reference preservation
1055  * policy.
1056  */
preserveSoftReferences(Object ** list)1057 void preserveSoftReferences(Object **list)
1058 {
1059     Object *ref;
1060     Object *prev, *next;
1061     size_t referentOffset, queueNextOffset;
1062     unsigned counter;
1063     bool white;
1064 
1065     queueNextOffset = gDvm.offJavaLangRefReference_queueNext;
1066     referentOffset = gDvm.offJavaLangRefReference_referent;
1067     counter = 0;
1068     prev = next = NULL;
1069     ref = *list;
1070     while (ref != NULL) {
1071         JValue *field = dvmFieldPtr(ref, referentOffset);
1072         Object *referent = field->l;
1073         next = dvmGetFieldObject(ref, queueNextOffset);
1074         assert(referent != NULL);
1075         if (isForward(referent->clazz)) {
1076             /* Referent is black. */
1077             field->l = referent = getForward(referent->clazz);
1078             white = false;
1079         } else {
1080             white = fromSpaceContains(referent);
1081         }
1082         if (!white && ((++counter) & 1)) {
1083             /* Referent is white and biased toward saving, gray it. */
1084             scavengeReference((Object **)(void *)&field->l);
1085             white = true;
1086         }
1087         if (white) {
1088             /* Referent is black, unlink it. */
1089             if (prev != NULL) {
1090                 dvmSetFieldObject(ref, queueNextOffset, NULL);
1091                 dvmSetFieldObject(prev, queueNextOffset, next);
1092             }
1093         } else {
1094             /* Referent is white, skip over it. */
1095             prev = ref;
1096         }
1097         ref = next;
1098     }
1099     /*
1100      * Restart the trace with the newly gray references added to the
1101      * root set.
1102      */
1103     scavengeBlockQueue();
1104 }
1105 
processFinalizableReferences()1106 void processFinalizableReferences()
1107 {
1108     HeapRefTable newPendingRefs;
1109     LargeHeapRefTable *finRefs = gDvm.gcHeap->finalizableRefs;
1110     Object **ref;
1111     Object **lastRef;
1112     size_t totalPendCount;
1113 
1114     /*
1115      * All strongly, reachable objects are black.
1116      * Any white finalizable objects need to be finalized.
1117      */
1118 
1119     /* Create a table that the new pending refs will
1120      * be added to.
1121      */
1122     if (!dvmHeapInitHeapRefTable(&newPendingRefs)) {
1123         //TODO: mark all finalizable refs and hope that
1124         //      we can schedule them next time.  Watch out,
1125         //      because we may be expecting to free up space
1126         //      by calling finalizers.
1127         LOG_REF("no room for pending finalizations");
1128         dvmAbort();
1129     }
1130 
1131     /*
1132      * Walk through finalizableRefs and move any white references to
1133      * the list of new pending refs.
1134      */
1135     totalPendCount = 0;
1136     while (finRefs != NULL) {
1137         Object **gapRef;
1138         size_t newPendCount = 0;
1139 
1140         gapRef = ref = finRefs->refs.table;
1141         lastRef = finRefs->refs.nextEntry;
1142         while (ref < lastRef) {
1143             if (fromSpaceContains(*ref)) {
1144                 if (!dvmHeapAddToHeapRefTable(&newPendingRefs, *ref)) {
1145                     //TODO: add the current table and allocate
1146                     //      a new, smaller one.
1147                     LOG_REF("no room for any more pending finalizations: %zd",
1148                             dvmHeapNumHeapRefTableEntries(&newPendingRefs));
1149                     dvmAbort();
1150                 }
1151                 newPendCount++;
1152             } else {
1153                 /* This ref is black, so will remain on finalizableRefs.
1154                  */
1155                 if (newPendCount > 0) {
1156                     /* Copy it up to fill the holes.
1157                      */
1158                     *gapRef++ = *ref;
1159                 } else {
1160                     /* No holes yet; don't bother copying.
1161                      */
1162                     gapRef++;
1163                 }
1164             }
1165             ref++;
1166         }
1167         finRefs->refs.nextEntry = gapRef;
1168         //TODO: if the table is empty when we're done, free it.
1169         totalPendCount += newPendCount;
1170         finRefs = finRefs->next;
1171     }
1172     LOG_REF("%zd finalizers triggered.", totalPendCount);
1173     if (totalPendCount == 0) {
1174         /* No objects required finalization.
1175          * Free the empty temporary table.
1176          */
1177         dvmClearReferenceTable(&newPendingRefs);
1178         return;
1179     }
1180 
1181     /* Add the new pending refs to the main list.
1182      */
1183     if (!dvmHeapAddTableToLargeTable(&gDvm.gcHeap->pendingFinalizationRefs,
1184                 &newPendingRefs))
1185     {
1186         LOG_REF("can't insert new pending finalizations");
1187         dvmAbort();
1188     }
1189 
1190     //TODO: try compacting the main list with a memcpy loop
1191 
1192     /* Blacken the refs we just moved; we don't want them or their
1193      * children to get swept yet.
1194      */
1195     ref = newPendingRefs.table;
1196     lastRef = newPendingRefs.nextEntry;
1197     assert(ref < lastRef);
1198     HPROF_SET_GC_SCAN_STATE(HPROF_ROOT_FINALIZING, 0);
1199     while (ref < lastRef) {
1200         scavengeReference(ref);
1201         ref++;
1202     }
1203     HPROF_CLEAR_GC_SCAN_STATE();
1204     scavengeBlockQueue();
1205     dvmSignalHeapWorker(false);
1206 }
1207 
1208 /*
1209  * If a reference points to from-space and has been forwarded, we snap
1210  * the pointer to its new to-space address.  If the reference points
1211  * to an unforwarded from-space address we must enqueue the reference
1212  * for later processing.  TODO: implement proper reference processing
1213  * and move the referent scavenging elsewhere.
1214  */
scavengeReferenceObject(Object * obj)1215 static void scavengeReferenceObject(Object *obj)
1216 {
1217     Object *referent;
1218     Object **queue;
1219     size_t referentOffset, queueNextOffset;
1220 
1221     assert(obj != NULL);
1222     LOG_SCAV("scavengeReferenceObject(obj=%p),'%s'", obj, obj->clazz->descriptor);
1223     scavengeDataObject(obj);
1224     referentOffset = gDvm.offJavaLangRefReference_referent;
1225     referent = dvmGetFieldObject(obj, referentOffset);
1226     if (referent == NULL || toSpaceContains(referent)) {
1227         return;
1228     }
1229     if (isSoftReference(obj)) {
1230         queue = &gDvm.gcHeap->softReferences;
1231     } else if (isWeakReference(obj)) {
1232         queue = &gDvm.gcHeap->weakReferences;
1233     } else {
1234         assert(isPhantomReference(obj));
1235         queue = &gDvm.gcHeap->phantomReferences;
1236     }
1237     queueNextOffset = gDvm.offJavaLangRefReference_queueNext;
1238     dvmSetFieldObject(obj, queueNextOffset, *queue);
1239     *queue = obj;
1240     LOG_SCAV("scavengeReferenceObject: enqueueing %p", obj);
1241 }
1242 
1243 /*
1244  * Data object scavenging.
1245  */
scavengeDataObject(Object * obj)1246 static void scavengeDataObject(Object *obj)
1247 {
1248     // LOG_SCAV("scavengeDataObject(obj=%p)", obj);
1249     assert(obj != NULL);
1250     assert(obj->clazz != NULL);
1251     assert(obj->clazz->objectSize != 0);
1252     assert(toSpaceContains(obj));
1253     /* Scavenge the class object. */
1254     ClassObject *clazz = obj->clazz;
1255     scavengeReference((Object **) obj);
1256     /* Scavenge instance fields. */
1257     if (clazz->refOffsets != CLASS_WALK_SUPER) {
1258         size_t refOffsets = clazz->refOffsets;
1259         while (refOffsets != 0) {
1260             size_t rshift = CLZ(refOffsets);
1261             size_t offset = CLASS_OFFSET_FROM_CLZ(rshift);
1262             Object **ref = (Object **)((u1 *)obj + offset);
1263             scavengeReference(ref);
1264             refOffsets &= ~(CLASS_HIGH_BIT >> rshift);
1265         }
1266     } else {
1267         for (; clazz != NULL; clazz = clazz->super) {
1268             InstField *field = clazz->ifields;
1269             for (int i = 0; i < clazz->ifieldRefCount; ++i, ++field) {
1270                 size_t offset = field->byteOffset;
1271                 Object **ref = (Object **)((u1 *)obj + offset);
1272                 scavengeReference(ref);
1273             }
1274         }
1275     }
1276 }
1277 
transportObject(const Object * fromObj)1278 static Object *transportObject(const Object *fromObj)
1279 {
1280     Object *toObj;
1281     size_t allocSize, copySize;
1282 
1283     LOG_TRAN("transportObject(fromObj=%p) allocBlocks=%zu",
1284                   fromObj,
1285                   gDvm.gcHeap->heapSource->allocBlocks);
1286     assert(fromObj != NULL);
1287     assert(fromSpaceContains(fromObj));
1288     allocSize = copySize = objectSize(fromObj);
1289     if (LW_HASH_STATE(fromObj->lock) != LW_HASH_STATE_UNHASHED) {
1290         /*
1291          * The object has been hashed or hashed and moved.  We must
1292          * reserve an additional word for a hash code.
1293          */
1294         allocSize += sizeof(u4);
1295     }
1296     if (LW_HASH_STATE(fromObj->lock) == LW_HASH_STATE_HASHED_AND_MOVED) {
1297         /*
1298          * The object has its hash code allocated.  Ensure the hash
1299          * code is copied along with the instance data.
1300          */
1301         copySize += sizeof(u4);
1302     }
1303     /* TODO(cshapiro): don't copy, re-map large data objects. */
1304     assert(copySize <= allocSize);
1305     toObj = allocateGray(allocSize);
1306     assert(toObj != NULL);
1307     assert(toSpaceContains(toObj));
1308     memcpy(toObj, fromObj, copySize);
1309     if (LW_HASH_STATE(fromObj->lock) == LW_HASH_STATE_HASHED) {
1310         /*
1311          * The object has had its hash code exposed.  Append it to the
1312          * instance and set a bit so we know to look for it there.
1313          */
1314         *(u4 *)(((char *)toObj) + copySize) = (u4)fromObj >> 3;
1315         toObj->lock |= LW_HASH_STATE_HASHED_AND_MOVED << LW_HASH_STATE_SHIFT;
1316     }
1317     LOG_TRAN("transportObject: from %p/%zu to %p/%zu (%zu,%zu) %s",
1318              fromObj, addressToBlock(gDvm.gcHeap->heapSource,fromObj),
1319              toObj, addressToBlock(gDvm.gcHeap->heapSource,toObj),
1320              copySize, allocSize, copySize < allocSize ? "DIFFERENT" : "");
1321     return toObj;
1322 }
1323 
1324 /*
1325  * Generic reference scavenging.
1326  */
1327 
1328 /*
1329  * Given a reference to an object, the scavenge routine will gray the
1330  * reference.  Any objects pointed to by the scavenger object will be
1331  * transported to new space and a forwarding pointer will be installed
1332  * in the header of the object.
1333  */
1334 
1335 /*
1336  * Blacken the given pointer.  If the pointer is in from space, it is
1337  * transported to new space.  If the object has a forwarding pointer
1338  * installed it has already been transported and the referent is
1339  * snapped to the new address.
1340  */
scavengeReference(Object ** obj)1341 static void scavengeReference(Object **obj)
1342 {
1343     ClassObject *clazz;
1344     Object *fromObj, *toObj;
1345 
1346     assert(obj);
1347 
1348     if (*obj == NULL) return;
1349 
1350     assert(dvmIsValidObject(*obj));
1351 
1352     /* The entire block is black. */
1353     if (toSpaceContains(*obj)) {
1354         LOG_SCAV("scavengeReference skipping pinned object @ %p", *obj);
1355         return;
1356     }
1357     LOG_SCAV("scavengeReference(*obj=%p)", *obj);
1358 
1359     assert(fromSpaceContains(*obj));
1360 
1361     clazz = (*obj)->clazz;
1362 
1363     if (isForward(clazz)) {
1364         // LOG_SCAV("forwarding %p @ %p to %p", *obj, obj, (void *)((uintptr_t)clazz & ~0x1));
1365         *obj = (Object *)getForward(clazz);
1366         return;
1367     }
1368     fromObj = *obj;
1369     if (clazz == NULL) {
1370         // LOG_SCAV("scavangeReference %p has a NULL class object", fromObj);
1371         assert(!"implemented");
1372         toObj = NULL;
1373     } else {
1374         toObj = transportObject(fromObj);
1375     }
1376     setForward(toObj, fromObj);
1377     *obj = (Object *)toObj;
1378 }
1379 
1380 /*
1381  * Generic object scavenging.
1382  */
scavengeObject(Object * obj)1383 static void scavengeObject(Object *obj)
1384 {
1385     ClassObject *clazz;
1386 
1387     assert(obj != NULL);
1388     assert(obj->clazz != NULL);
1389     assert(!((uintptr_t)obj->clazz & 0x1));
1390     clazz = obj->clazz;
1391     if (dvmIsTheClassClass(clazz)) {
1392         scavengeClassObject((ClassObject *)obj);
1393     } else if (IS_CLASS_FLAG_SET(clazz, CLASS_ISARRAY)) {
1394         scavengeArrayObject((ArrayObject *)obj);
1395     } else if (IS_CLASS_FLAG_SET(clazz, CLASS_ISREFERENCE)) {
1396         scavengeReferenceObject(obj);
1397     } else {
1398         scavengeDataObject(obj);
1399     }
1400 }
1401 
1402 /*
1403  * External root scavenging routines.
1404  */
1405 
pinHashTableEntries(HashTable * table)1406 static void pinHashTableEntries(HashTable *table)
1407 {
1408     LOG_PIN(">>> pinHashTableEntries(table=%p)", table);
1409     if (table == NULL) {
1410         return;
1411     }
1412     dvmHashTableLock(table);
1413     for (int i = 0; i < table->tableSize; ++i) {
1414         HashEntry *entry = &table->pEntries[i];
1415         void *obj = entry->data;
1416         if (obj == NULL || obj == HASH_TOMBSTONE) {
1417             continue;
1418         }
1419         pinObject(entry->data);
1420     }
1421     dvmHashTableUnlock(table);
1422     LOG_PIN("<<< pinHashTableEntries(table=%p)", table);
1423 }
1424 
pinPrimitiveClasses()1425 static void pinPrimitiveClasses()
1426 {
1427     size_t length = ARRAYSIZE(gDvm.primitiveClass);
1428     for (size_t i = 0; i < length; i++) {
1429         if (gDvm.primitiveClass[i] != NULL) {
1430             pinObject((Object *)gDvm.primitiveClass[i]);
1431         }
1432     }
1433 }
1434 
1435 /*
1436  * Scavenge interned strings.  Permanent interned strings will have
1437  * been pinned and are therefore ignored.  Non-permanent strings that
1438  * have been forwarded are snapped.  All other entries are removed.
1439  */
scavengeInternedStrings()1440 static void scavengeInternedStrings()
1441 {
1442     HashTable *table = gDvm.internedStrings;
1443     if (table == NULL) {
1444         return;
1445     }
1446     dvmHashTableLock(table);
1447     for (int i = 0; i < table->tableSize; ++i) {
1448         HashEntry *entry = &table->pEntries[i];
1449         Object *obj = (Object *)entry->data;
1450         if (obj == NULL || obj == HASH_TOMBSTONE) {
1451             continue;
1452         } else if (!isPermanentString((StringObject *)obj)) {
1453             // LOG_SCAV("entry->data=%p", entry->data);
1454             LOG_SCAV(">>> string obj=%p", entry->data);
1455             /* TODO(cshapiro): detach white string objects */
1456             scavengeReference((Object **)(void *)&entry->data);
1457             LOG_SCAV("<<< string obj=%p", entry->data);
1458         }
1459     }
1460     dvmHashTableUnlock(table);
1461 }
1462 
pinInternedStrings()1463 static void pinInternedStrings()
1464 {
1465     HashTable *table = gDvm.internedStrings;
1466     if (table == NULL) {
1467         return;
1468     }
1469     dvmHashTableLock(table);
1470     for (int i = 0; i < table->tableSize; ++i) {
1471         HashEntry *entry = &table->pEntries[i];
1472         Object *obj = (Object *)entry->data;
1473         if (obj == NULL || obj == HASH_TOMBSTONE) {
1474             continue;
1475         } else if (isPermanentString((StringObject *)obj)) {
1476             obj = (Object *)getPermanentString((StringObject*)obj);
1477             LOG_PROM(">>> pin string obj=%p", obj);
1478             pinObject(obj);
1479             LOG_PROM("<<< pin string obj=%p", obj);
1480         }
1481      }
1482     dvmHashTableUnlock(table);
1483 }
1484 
1485 /*
1486  * At present, reference tables contain references that must not be
1487  * moved by the collector.  Instead of scavenging each reference in
1488  * the table we pin each referenced object.
1489  */
pinReferenceTable(const ReferenceTable * table)1490 static void pinReferenceTable(const ReferenceTable *table)
1491 {
1492     assert(table != NULL);
1493     assert(table->table != NULL);
1494     assert(table->nextEntry != NULL);
1495     for (Object **entry = table->table; entry < table->nextEntry; ++entry) {
1496         assert(entry != NULL);
1497         assert(!isForward(*entry));
1498         pinObject(*entry);
1499     }
1500 }
1501 
scavengeLargeHeapRefTable(LargeHeapRefTable * table)1502 static void scavengeLargeHeapRefTable(LargeHeapRefTable *table)
1503 {
1504     for (; table != NULL; table = table->next) {
1505         Object **ref = table->refs.table;
1506         for (; ref < table->refs.nextEntry; ++ref) {
1507             scavengeReference(ref);
1508         }
1509     }
1510 }
1511 
1512 /* This code was copied from Thread.c */
scavengeThreadStack(Thread * thread)1513 static void scavengeThreadStack(Thread *thread)
1514 {
1515     const u4 *framePtr;
1516 #if WITH_EXTRA_GC_CHECKS > 1
1517     bool first = true;
1518 #endif
1519 
1520     framePtr = (const u4 *)thread->interpSave.curFrame;
1521     while (framePtr != NULL) {
1522         const StackSaveArea *saveArea;
1523         const Method *method;
1524 
1525         saveArea = SAVEAREA_FROM_FP(framePtr);
1526         method = saveArea->method;
1527         if (method != NULL && !dvmIsNativeMethod(method)) {
1528 #ifdef COUNT_PRECISE_METHODS
1529             /* the GC is running, so no lock required */
1530             if (dvmPointerSetAddEntry(gDvm.preciseMethods, method))
1531                 LOG_SCAV("PGC: added %s.%s %p",
1532                              method->clazz->descriptor, method->name, method);
1533 #endif
1534 #if WITH_EXTRA_GC_CHECKS > 1
1535             /*
1536              * May also want to enable the memset() in the "invokeMethod"
1537              * goto target in the portable interpreter.  That sets the stack
1538              * to a pattern that makes referring to uninitialized data
1539              * very obvious.
1540              */
1541 
1542             if (first) {
1543                 /*
1544                  * First frame, isn't native, check the "alternate" saved PC
1545                  * as a sanity check.
1546                  *
1547                  * It seems like we could check the second frame if the first
1548                  * is native, since the PCs should be the same.  It turns out
1549                  * this doesn't always work.  The problem is that we could
1550                  * have calls in the sequence:
1551                  *   interp method #2
1552                  *   native method
1553                  *   interp method #1
1554                  *
1555                  * and then GC while in the native method after returning
1556                  * from interp method #2.  The currentPc on the stack is
1557                  * for interp method #1, but thread->currentPc2 is still
1558                  * set for the last thing interp method #2 did.
1559                  *
1560                  * This can also happen in normal execution:
1561                  * - sget-object on not-yet-loaded class
1562                  * - class init updates currentPc2
1563                  * - static field init is handled by parsing annotations;
1564                  *   static String init requires creation of a String object,
1565                  *   which can cause a GC
1566                  *
1567                  * Essentially, any pattern that involves executing
1568                  * interpreted code and then causes an allocation without
1569                  * executing instructions in the original method will hit
1570                  * this.  These are rare enough that the test still has
1571                  * some value.
1572                  */
1573                 if (saveArea->xtra.currentPc != thread->currentPc2) {
1574                     ALOGW("PGC: savedPC(%p) != current PC(%p), %s.%s ins=%p",
1575                         saveArea->xtra.currentPc, thread->currentPc2,
1576                         method->clazz->descriptor, method->name, method->insns);
1577                     if (saveArea->xtra.currentPc != NULL)
1578                         ALOGE("  pc inst = 0x%04x", *saveArea->xtra.currentPc);
1579                     if (thread->currentPc2 != NULL)
1580                         ALOGE("  pc2 inst = 0x%04x", *thread->currentPc2);
1581                     dvmDumpThread(thread, false);
1582                 }
1583             } else {
1584                 /*
1585                  * It's unusual, but not impossible, for a non-first frame
1586                  * to be at something other than a method invocation.  For
1587                  * example, if we do a new-instance on a nonexistent class,
1588                  * we'll have a lot of class loader activity on the stack
1589                  * above the frame with the "new" operation.  Could also
1590                  * happen while we initialize a Throwable when an instruction
1591                  * fails.
1592                  *
1593                  * So there's not much we can do here to verify the PC,
1594                  * except to verify that it's a GC point.
1595                  */
1596             }
1597             assert(saveArea->xtra.currentPc != NULL);
1598 #endif
1599 
1600             const RegisterMap* pMap;
1601             const u1* regVector;
1602 
1603             Method* nonConstMethod = (Method*) method;  // quiet gcc
1604             pMap = dvmGetExpandedRegisterMap(nonConstMethod);
1605 
1606             //LOG_SCAV("PGC: %s.%s", method->clazz->descriptor, method->name);
1607 
1608             if (pMap != NULL) {
1609                 /* found map, get registers for this address */
1610                 int addr = saveArea->xtra.currentPc - method->insns;
1611                 regVector = dvmRegisterMapGetLine(pMap, addr);
1612                 /*
1613                 if (regVector == NULL) {
1614                     LOG_SCAV("PGC: map but no entry for %s.%s addr=0x%04x",
1615                                  method->clazz->descriptor, method->name, addr);
1616                 } else {
1617                     LOG_SCAV("PGC: found map for %s.%s 0x%04x (t=%d)",
1618                                  method->clazz->descriptor, method->name, addr,
1619                                  thread->threadId);
1620                 }
1621                 */
1622             } else {
1623                 /*
1624                  * No map found.  If precise GC is disabled this is
1625                  * expected -- we don't create pointers to the map data even
1626                  * if it's present -- but if it's enabled it means we're
1627                  * unexpectedly falling back on a conservative scan, so it's
1628                  * worth yelling a little.
1629                  */
1630                 if (gDvm.preciseGc) {
1631                     LOG_SCAV("PGC: no map for %s.%s", method->clazz->descriptor, method->name);
1632                 }
1633                 regVector = NULL;
1634             }
1635             if (regVector == NULL) {
1636                 /*
1637                  * There are no roots to scavenge.  Skip over the entire frame.
1638                  */
1639                 framePtr += method->registersSize;
1640             } else {
1641                 /*
1642                  * Precise scan.  v0 is at the lowest address on the
1643                  * interpreted stack, and is the first bit in the register
1644                  * vector, so we can walk through the register map and
1645                  * memory in the same direction.
1646                  *
1647                  * A '1' bit indicates a live reference.
1648                  */
1649                 u2 bits = 1 << 1;
1650                 for (int i = method->registersSize - 1; i >= 0; i--) {
1651                     u4 rval = *framePtr;
1652 
1653                     bits >>= 1;
1654                     if (bits == 1) {
1655                         /* set bit 9 so we can tell when we're empty */
1656                         bits = *regVector++ | 0x0100;
1657                     }
1658 
1659                     if (rval != 0 && (bits & 0x01) != 0) {
1660                         /*
1661                          * Non-null, register marked as live reference.  This
1662                          * should always be a valid object.
1663                          */
1664 #if WITH_EXTRA_GC_CHECKS > 0
1665                         if ((rval & 0x3) != 0 || !dvmIsValidObject((Object*) rval)) {
1666                             /* this is very bad */
1667                             ALOGE("PGC: invalid ref in reg %d: 0x%08x",
1668                                 method->registersSize-1 - i, rval);
1669                         } else
1670 #endif
1671                         {
1672 
1673                             // LOG_SCAV("stack reference %u@%p", *framePtr, framePtr);
1674                             /* dvmMarkObjectNonNull((Object *)rval); */
1675                             scavengeReference((Object **) framePtr);
1676                         }
1677                     } else {
1678                         /*
1679                          * Null or non-reference, do nothing at all.
1680                          */
1681 #if WITH_EXTRA_GC_CHECKS > 1
1682                         if (dvmIsValidObject((Object*) rval)) {
1683                             /* this is normal, but we feel chatty */
1684                             ALOGD("PGC: ignoring valid ref in reg %d: 0x%08x",
1685                                  method->registersSize-1 - i, rval);
1686                         }
1687 #endif
1688                     }
1689                     ++framePtr;
1690                 }
1691                 dvmReleaseRegisterMapLine(pMap, regVector);
1692             }
1693         }
1694         /* else this is a break frame and there is nothing to gray, or
1695          * this is a native method and the registers are just the "ins",
1696          * copied from various registers in the caller's set.
1697          */
1698 
1699 #if WITH_EXTRA_GC_CHECKS > 1
1700         first = false;
1701 #endif
1702 
1703         /* Don't fall into an infinite loop if things get corrupted.
1704          */
1705         assert((uintptr_t)saveArea->prevFrame > (uintptr_t)framePtr ||
1706                saveArea->prevFrame == NULL);
1707         framePtr = saveArea->prevFrame;
1708     }
1709 }
1710 
scavengeThread(Thread * thread)1711 static void scavengeThread(Thread *thread)
1712 {
1713     // LOG_SCAV("scavengeThread(thread=%p)", thread);
1714 
1715     // LOG_SCAV("Scavenging threadObj=%p", thread->threadObj);
1716     scavengeReference(&thread->threadObj);
1717 
1718     // LOG_SCAV("Scavenging exception=%p", thread->exception);
1719     scavengeReference(&thread->exception);
1720 
1721     scavengeThreadStack(thread);
1722 }
1723 
scavengeThreadList()1724 static void scavengeThreadList()
1725 {
1726     Thread *thread;
1727 
1728     dvmLockThreadList(dvmThreadSelf());
1729     thread = gDvm.threadList;
1730     while (thread) {
1731         scavengeThread(thread);
1732         thread = thread->next;
1733     }
1734     dvmUnlockThreadList();
1735 }
1736 
pinThreadStack(const Thread * thread)1737 static void pinThreadStack(const Thread *thread)
1738 {
1739     const u4 *framePtr;
1740     const StackSaveArea *saveArea;
1741     Method *method;
1742     const char *shorty;
1743     Object *obj;
1744 
1745     saveArea = NULL;
1746     framePtr = (const u4 *)thread->interpSave.curFrame;
1747     for (; framePtr != NULL; framePtr = saveArea->prevFrame) {
1748         saveArea = SAVEAREA_FROM_FP(framePtr);
1749         method = (Method *)saveArea->method;
1750         if (method != NULL && dvmIsNativeMethod(method)) {
1751             /*
1752              * This is native method, pin its arguments.
1753              *
1754              * For purposes of graying references, we don't need to do
1755              * anything here, because all of the native "ins" were copied
1756              * from registers in the caller's stack frame and won't be
1757              * changed (an interpreted method can freely use registers
1758              * with parameters like any other register, but natives don't
1759              * work that way).
1760              *
1761              * However, we need to ensure that references visible to
1762              * native methods don't move around.  We can do a precise scan
1763              * of the arguments by examining the method signature.
1764              */
1765             LOG_PIN("+++ native scan %s.%s",
1766                     method->clazz->descriptor, method->name);
1767             assert(method->registersSize == method->insSize);
1768             if (!dvmIsStaticMethod(method)) {
1769                 /* grab the "this" pointer */
1770                 obj = (Object *)*framePtr++;
1771                 if (obj == NULL) {
1772                     /*
1773                      * This can happen for the "fake" entry frame inserted
1774                      * for threads created outside the VM.  There's no actual
1775                      * call so there's no object.  If we changed the fake
1776                      * entry method to be declared "static" then this
1777                      * situation should never occur.
1778                      */
1779                 } else {
1780                     assert(dvmIsValidObject(obj));
1781                     pinObject(obj);
1782                 }
1783             }
1784             shorty = method->shorty+1;      // skip return value
1785             for (int i = method->registersSize - 1; i >= 0; i--, framePtr++) {
1786                 switch (*shorty++) {
1787                 case 'L':
1788                     obj = (Object *)*framePtr;
1789                     if (obj != NULL) {
1790                         assert(dvmIsValidObject(obj));
1791                         pinObject(obj);
1792                     }
1793                     break;
1794                 case 'D':
1795                 case 'J':
1796                     framePtr++;
1797                     break;
1798                 default:
1799                     /* 32-bit non-reference value */
1800                     obj = (Object *)*framePtr;          // debug, remove
1801                     if (dvmIsValidObject(obj)) {        // debug, remove
1802                         /* if we see a lot of these, our scan might be off */
1803                         LOG_PIN("+++ did NOT pin obj %p", obj);
1804                     }
1805                     break;
1806                 }
1807             }
1808         } else if (method != NULL && !dvmIsNativeMethod(method)) {
1809             const RegisterMap* pMap = dvmGetExpandedRegisterMap(method);
1810             const u1* regVector = NULL;
1811 
1812             ALOGI("conservative : %s.%s", method->clazz->descriptor, method->name);
1813 
1814             if (pMap != NULL) {
1815                 int addr = saveArea->xtra.currentPc - method->insns;
1816                 regVector = dvmRegisterMapGetLine(pMap, addr);
1817             }
1818             if (regVector == NULL) {
1819                 /*
1820                  * No register info for this frame, conservatively pin.
1821                  */
1822                 for (int i = 0; i < method->registersSize; ++i) {
1823                     u4 regValue = framePtr[i];
1824                     if (regValue != 0 && (regValue & 0x3) == 0 && dvmIsValidObject((Object *)regValue)) {
1825                         pinObject((Object *)regValue);
1826                     }
1827                 }
1828             }
1829         }
1830         /*
1831          * Don't fall into an infinite loop if things get corrupted.
1832          */
1833         assert((uintptr_t)saveArea->prevFrame > (uintptr_t)framePtr ||
1834                saveArea->prevFrame == NULL);
1835     }
1836 }
1837 
pinThread(const Thread * thread)1838 static void pinThread(const Thread *thread)
1839 {
1840     assert(thread != NULL);
1841     LOG_PIN("pinThread(thread=%p)", thread);
1842 
1843     LOG_PIN("Pin native method arguments");
1844     pinThreadStack(thread);
1845 
1846     LOG_PIN("Pin internalLocalRefTable");
1847     pinReferenceTable(&thread->internalLocalRefTable);
1848 
1849     LOG_PIN("Pin jniLocalRefTable");
1850     pinReferenceTable(&thread->jniLocalRefTable);
1851 
1852     /* Can the check be pushed into the promote routine? */
1853     if (thread->jniMonitorRefTable.table) {
1854         LOG_PIN("Pin jniMonitorRefTable");
1855         pinReferenceTable(&thread->jniMonitorRefTable);
1856     }
1857 }
1858 
pinThreadList()1859 static void pinThreadList()
1860 {
1861     Thread *thread;
1862 
1863     dvmLockThreadList(dvmThreadSelf());
1864     thread = gDvm.threadList;
1865     while (thread) {
1866         pinThread(thread);
1867         thread = thread->next;
1868     }
1869     dvmUnlockThreadList();
1870 }
1871 
1872 /*
1873  * Heap block scavenging.
1874  */
1875 
1876 /*
1877  * Scavenge objects in the current block.  Scavenging terminates when
1878  * the pointer reaches the highest address in the block or when a run
1879  * of zero words that continues to the highest address is reached.
1880  */
scavengeBlock(HeapSource * heapSource,size_t block)1881 static void scavengeBlock(HeapSource *heapSource, size_t block)
1882 {
1883     u1 *cursor;
1884     u1 *end;
1885     size_t size;
1886 
1887     LOG_SCAV("scavengeBlock(heapSource=%p,block=%zu)", heapSource, block);
1888 
1889     assert(heapSource != NULL);
1890     assert(block < heapSource->totalBlocks);
1891     assert(heapSource->blockSpace[block] == BLOCK_TO_SPACE);
1892 
1893     cursor = blockToAddress(heapSource, block);
1894     end = cursor + BLOCK_SIZE;
1895     LOG_SCAV("scavengeBlock start=%p, end=%p", cursor, end);
1896 
1897     /* Parse and scavenge the current block. */
1898     size = 0;
1899     while (cursor < end) {
1900         u4 word = *(u4 *)cursor;
1901         if (word != 0) {
1902             scavengeObject((Object *)cursor);
1903             size = objectSize((Object *)cursor);
1904             size = alignUp(size, ALLOC_ALIGNMENT);
1905             cursor += size;
1906         } else {
1907             /* Check for padding. */
1908             while (*(u4 *)cursor == 0) {
1909                 cursor += 4;
1910                 if (cursor == end) break;
1911             }
1912             /* Punt if something went wrong. */
1913             assert(cursor == end);
1914         }
1915     }
1916 }
1917 
objectSize(const Object * obj)1918 static size_t objectSize(const Object *obj)
1919 {
1920     size_t size;
1921 
1922     assert(obj != NULL);
1923     assert(obj->clazz != NULL);
1924     if (obj->clazz == gDvm.classJavaLangClass) {
1925         size = dvmClassObjectSize((ClassObject *)obj);
1926     } else if (IS_CLASS_FLAG_SET(obj->clazz, CLASS_ISARRAY)) {
1927         size = dvmArrayObjectSize((ArrayObject *)obj);
1928     } else {
1929         assert(obj->clazz->objectSize != 0);
1930         size = obj->clazz->objectSize;
1931     }
1932     if (LW_HASH_STATE(obj->lock) == LW_HASH_STATE_HASHED_AND_MOVED) {
1933         size += sizeof(u4);
1934     }
1935     return size;
1936 }
1937 
verifyBlock(HeapSource * heapSource,size_t block)1938 static void verifyBlock(HeapSource *heapSource, size_t block)
1939 {
1940     u1 *cursor;
1941     u1 *end;
1942     size_t size;
1943 
1944     // LOG_VER("verifyBlock(heapSource=%p,block=%zu)", heapSource, block);
1945 
1946     assert(heapSource != NULL);
1947     assert(block < heapSource->totalBlocks);
1948     assert(heapSource->blockSpace[block] == BLOCK_TO_SPACE);
1949 
1950     cursor = blockToAddress(heapSource, block);
1951     end = cursor + BLOCK_SIZE;
1952     // LOG_VER("verifyBlock start=%p, end=%p", cursor, end);
1953 
1954     /* Parse and scavenge the current block. */
1955     size = 0;
1956     while (cursor < end) {
1957         u4 word = *(u4 *)cursor;
1958         if (word != 0) {
1959             dvmVerifyObject((Object *)cursor);
1960             size = objectSize((Object *)cursor);
1961             size = alignUp(size, ALLOC_ALIGNMENT);
1962             cursor += size;
1963         } else {
1964             /* Check for padding. */
1965             while (*(unsigned long *)cursor == 0) {
1966                 cursor += 4;
1967                 if (cursor == end) break;
1968             }
1969             /* Punt if something went wrong. */
1970             assert(cursor == end);
1971         }
1972     }
1973 }
1974 
describeBlockQueue(const HeapSource * heapSource)1975 static void describeBlockQueue(const HeapSource *heapSource)
1976 {
1977     size_t block, count;
1978     char space;
1979 
1980     block = heapSource->queueHead;
1981     count = 0;
1982     LOG_SCAV(">>> describeBlockQueue(heapSource=%p)", heapSource);
1983     /* Count the number of blocks enqueued. */
1984     while (block != QUEUE_TAIL) {
1985         block = heapSource->blockQueue[block];
1986         ++count;
1987     }
1988     LOG_SCAV("blockQueue %zu elements, enqueued %zu",
1989                  count, heapSource->queueSize);
1990     block = heapSource->queueHead;
1991     while (block != QUEUE_TAIL) {
1992         space = heapSource->blockSpace[block];
1993         LOG_SCAV("block=%zu@%p,space=%zu", block, blockToAddress(heapSource,block), space);
1994         block = heapSource->blockQueue[block];
1995     }
1996 
1997     LOG_SCAV("<<< describeBlockQueue(heapSource=%p)", heapSource);
1998 }
1999 
2000 /*
2001  * Blackens promoted objects.
2002  */
scavengeBlockQueue()2003 static void scavengeBlockQueue()
2004 {
2005     HeapSource *heapSource;
2006     size_t block;
2007 
2008     LOG_SCAV(">>> scavengeBlockQueue()");
2009     heapSource = gDvm.gcHeap->heapSource;
2010     describeBlockQueue(heapSource);
2011     while (heapSource->queueHead != QUEUE_TAIL) {
2012         block = heapSource->queueHead;
2013         LOG_SCAV("Dequeueing block %zu", block);
2014         scavengeBlock(heapSource, block);
2015         heapSource->queueHead = heapSource->blockQueue[block];
2016         LOG_SCAV("New queue head is %zu", heapSource->queueHead);
2017     }
2018     LOG_SCAV("<<< scavengeBlockQueue()");
2019 }
2020 
2021 /*
2022  * Scan the block list and verify all blocks that are marked as being
2023  * in new space.  This should be parametrized so we can invoke this
2024  * routine outside of the context of a collection.
2025  */
verifyNewSpace()2026 static void verifyNewSpace()
2027 {
2028     HeapSource *heapSource = gDvm.gcHeap->heapSource;
2029     size_t c0 = 0, c1 = 0, c2 = 0, c7 = 0;
2030     for (size_t i = 0; i < heapSource->totalBlocks; ++i) {
2031         switch (heapSource->blockSpace[i]) {
2032         case BLOCK_FREE: ++c0; break;
2033         case BLOCK_TO_SPACE: ++c1; break;
2034         case BLOCK_FROM_SPACE: ++c2; break;
2035         case BLOCK_CONTINUED: ++c7; break;
2036         default: assert(!"reached");
2037         }
2038     }
2039     LOG_VER("Block Demographics: "
2040             "Free=%zu,ToSpace=%zu,FromSpace=%zu,Continued=%zu",
2041             c0, c1, c2, c7);
2042     for (size_t i = 0; i < heapSource->totalBlocks; ++i) {
2043         if (heapSource->blockSpace[i] != BLOCK_TO_SPACE) {
2044             continue;
2045         }
2046         verifyBlock(heapSource, i);
2047     }
2048 }
2049 
describeHeap()2050 void describeHeap()
2051 {
2052     HeapSource *heapSource = gDvm.gcHeap->heapSource;
2053     describeBlocks(heapSource);
2054 }
2055 
2056 /*
2057  * The collection interface.  Collection has a few distinct phases.
2058  * The first is flipping AKA condemning AKA whitening the heap.  The
2059  * second is to promote all objects which are pointed to by pinned or
2060  * ambiguous references.  The third phase is tracing from the stacks,
2061  * registers and various globals.  Lastly, a verification of the heap
2062  * is performed.  The last phase should be optional.
2063  */
dvmScavengeRoots()2064 void dvmScavengeRoots()  /* Needs a new name badly */
2065 {
2066     GcHeap *gcHeap;
2067 
2068     {
2069         size_t alloc, unused, total;
2070 
2071         room(&alloc, &unused, &total);
2072         LOG_SCAV("BEFORE GC: %zu alloc, %zu free, %zu total.",
2073                      alloc, unused, total);
2074     }
2075 
2076     gcHeap = gDvm.gcHeap;
2077     dvmHeapSourceFlip();
2078 
2079     /*
2080      * Promote blocks with stationary objects.
2081      */
2082     pinThreadList();
2083     pinReferenceTable(&gDvm.jniGlobalRefTable);
2084     pinReferenceTable(&gDvm.jniPinRefTable);
2085     pinHashTableEntries(gDvm.loadedClasses);
2086     pinHashTableEntries(gDvm.dbgRegistry);
2087     pinPrimitiveClasses();
2088     pinInternedStrings();
2089 
2090     // describeBlocks(gcHeap->heapSource);
2091 
2092     /*
2093      * Create first, open new-space page right here.
2094      */
2095 
2096     /* Reset allocation to an unallocated block. */
2097     gDvm.gcHeap->heapSource->allocPtr = allocateBlocks(gDvm.gcHeap->heapSource, 1);
2098     gDvm.gcHeap->heapSource->allocLimit = gDvm.gcHeap->heapSource->allocPtr + BLOCK_SIZE;
2099     /*
2100      * Hack: promote the empty block allocated above.  If the
2101      * promotions that occurred above did not actually gray any
2102      * objects, the block queue may be empty.  We must force a
2103      * promotion to be safe.
2104      */
2105     promoteBlockByAddr(gDvm.gcHeap->heapSource, gDvm.gcHeap->heapSource->allocPtr);
2106 
2107     /*
2108      * Scavenge blocks and relocate movable objects.
2109      */
2110 
2111     LOG_SCAV("Scavenging gDvm.threadList");
2112     scavengeThreadList();
2113 
2114     LOG_SCAV("Scavenging gDvm.gcHeap->referenceOperations");
2115     scavengeLargeHeapRefTable(gcHeap->referenceOperations);
2116 
2117     LOG_SCAV("Scavenging gDvm.gcHeap->pendingFinalizationRefs");
2118     scavengeLargeHeapRefTable(gcHeap->pendingFinalizationRefs);
2119 
2120     LOG_SCAV("Scavenging random global stuff");
2121     scavengeReference(&gDvm.outOfMemoryObj);
2122     scavengeReference(&gDvm.internalErrorObj);
2123     scavengeReference(&gDvm.noClassDefFoundErrorObj);
2124 
2125     // LOG_SCAV("Scavenging gDvm.internedString");
2126     scavengeInternedStrings();
2127 
2128     LOG_SCAV("Root scavenge has completed.");
2129 
2130     scavengeBlockQueue();
2131 
2132     // LOG_SCAV("Re-snap global class pointers.");
2133     // scavengeGlobals();
2134 
2135     LOG_SCAV("New space scavenge has completed.");
2136 
2137     /*
2138      * Process reference objects in strength order.
2139      */
2140 
2141     LOG_REF("Processing soft references...");
2142     preserveSoftReferences(&gDvm.gcHeap->softReferences);
2143     clearWhiteReferences(&gDvm.gcHeap->softReferences);
2144 
2145     LOG_REF("Processing weak references...");
2146     clearWhiteReferences(&gDvm.gcHeap->weakReferences);
2147 
2148     LOG_REF("Finding finalizations...");
2149     processFinalizableReferences();
2150 
2151     LOG_REF("Processing f-reachable soft references...");
2152     clearWhiteReferences(&gDvm.gcHeap->softReferences);
2153 
2154     LOG_REF("Processing f-reachable weak references...");
2155     clearWhiteReferences(&gDvm.gcHeap->weakReferences);
2156 
2157     LOG_REF("Processing phantom references...");
2158     clearWhiteReferences(&gDvm.gcHeap->phantomReferences);
2159 
2160     /*
2161      * Verify the stack and heap.
2162      */
2163     dvmVerifyRoots();
2164     verifyNewSpace();
2165 
2166     //describeBlocks(gcHeap->heapSource);
2167 
2168     clearFromSpace(gcHeap->heapSource);
2169 
2170     {
2171         size_t alloc, rem, total;
2172 
2173         room(&alloc, &rem, &total);
2174         LOG_SCAV("AFTER GC: %zu alloc, %zu free, %zu total.", alloc, rem, total);
2175     }
2176 }
2177 
2178 /*
2179  * Interface compatibility routines.
2180  */
2181 
dvmClearWhiteRefs(Object ** list)2182 void dvmClearWhiteRefs(Object **list)
2183 {
2184     /* do nothing */
2185     assert(*list == NULL);
2186 }
2187 
dvmHandleSoftRefs(Object ** list)2188 void dvmHandleSoftRefs(Object **list)
2189 {
2190     /* do nothing */
2191     assert(*list == NULL);
2192 }
2193 
dvmHeapBeginMarkStep(GcMode mode)2194 bool dvmHeapBeginMarkStep(GcMode mode)
2195 {
2196     /* do nothing */
2197     return true;
2198 }
2199 
dvmHeapFinishMarkStep()2200 void dvmHeapFinishMarkStep()
2201 {
2202     /* do nothing */
2203 }
2204 
dvmHeapMarkRootSet()2205 void dvmHeapMarkRootSet()
2206 {
2207     /* do nothing */
2208 }
2209 
dvmHeapScanMarkedObjects()2210 void dvmHeapScanMarkedObjects()
2211 {
2212     dvmScavengeRoots();
2213 }
2214 
dvmHeapScheduleFinalizations()2215 void dvmHeapScheduleFinalizations()
2216 {
2217     /* do nothing */
2218 }
2219 
dvmHeapSweepUnmarkedObjects(GcMode mode,int * numFreed,size_t * sizeFreed)2220 void dvmHeapSweepUnmarkedObjects(GcMode mode, int *numFreed, size_t *sizeFreed)
2221 {
2222     *numFreed = 0;
2223     *sizeFreed = 0;
2224     /* do nothing */
2225 }
2226 
dvmMarkDirtyObjects()2227 void dvmMarkDirtyObjects()
2228 {
2229     assert(!"implemented");
2230 }
2231 
dvmHeapSourceThreadShutdown()2232 void dvmHeapSourceThreadShutdown()
2233 {
2234     /* do nothing */
2235 }
2236