• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (C) 2008 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 /*
18  * Linear memory allocation, tied to class loaders.
19  */
20 #include "Dalvik.h"
21 
22 #include <sys/mman.h>
23 #include <limits.h>
24 #include <errno.h>
25 
26 //#define DISABLE_LINEAR_ALLOC
27 
28 // Use ashmem to name the LinearAlloc section
29 #define USE_ASHMEM 1
30 
31 #ifdef USE_ASHMEM
32 #include <cutils/ashmem.h>
33 #endif /* USE_ASHMEM */
34 
35 /*
36 Overview
37 
38 This is intended to be a simple, fast allocator for "write-once" storage.
39 The expectation is that this will hold small allocations that don't change,
40 such as parts of classes (vtables, fields, methods, interfaces).  Because
41 the lifetime of these items is tied to classes, which in turn are tied
42 to class loaders, we associate the storage with a ClassLoader object.
43 
44 [ We don't yet support class unloading, and our ClassLoader implementation
45 is in flux, so for now we just have a single global region and the
46 "classLoader" argument is ignored. ]
47 
48 By storing the data here, rather than on the system heap, we reduce heap
49 clutter, speed class loading, reduce the memory footprint (reduced heap
50 structure overhead), and most importantly we increase the number of pages
51 that remain shared between processes launched in "Zygote mode".
52 
53 The 4 bytes preceding each block contain the block length.  This allows us
54 to support "free" and "realloc" calls in a limited way.  We don't free
55 storage once it has been allocated, but in some circumstances it could be
56 useful to erase storage to garbage values after a "free" or "realloc".
57 (Bad idea if we're trying to share pages.)  We need to align to 8-byte
58 boundaries for some architectures, so we have a 50-50 chance of getting
59 this for free in a given block.
60 
61 A NULL value for the "classLoader" argument refers to the bootstrap class
62 loader, which is never unloaded (until the VM shuts down).
63 
64 Because the memory is not expected to be updated, we can use mprotect to
65 guard the pages on debug builds.  Handy when tracking down corruption.
66 */
67 
68 /* alignment for allocations; must be power of 2, and currently >= hdr_xtra */
69 #define BLOCK_ALIGN         8
70 
71 /* default length of memory segment (worst case is probably "dexopt") */
72 #define DEFAULT_MAX_LENGTH  (16*1024*1024)
73 
74 /* leave enough space for a length word */
75 #define HEADER_EXTRA        4
76 
77 /* overload the length word */
78 #define LENGTHFLAG_FREE    0x80000000
79 #define LENGTHFLAG_RW      0x40000000
80 #define LENGTHFLAG_MASK    (~(LENGTHFLAG_FREE|LENGTHFLAG_RW))
81 
82 
83 /* fwd */
84 static void checkAllFree(Object* classLoader);
85 
86 
87 /*
88  * Someday, retrieve the linear alloc struct associated with a particular
89  * class loader.  For now, always use the boostrap loader's instance.
90  */
getHeader(Object * classLoader)91 static inline LinearAllocHdr* getHeader(Object* classLoader)
92 {
93     return gDvm.pBootLoaderAlloc;
94 }
95 
96 /*
97  * Convert a pointer to memory to a pointer to the block header (which is
98  * currently just a length word).
99  */
getBlockHeader(void * mem)100 static inline u4* getBlockHeader(void* mem)
101 {
102     return ((u4*) mem) -1;
103 }
104 
105 /*
106  * Create a new linear allocation block.
107  */
dvmLinearAllocCreate(Object * classLoader)108 LinearAllocHdr* dvmLinearAllocCreate(Object* classLoader)
109 {
110 #ifdef DISABLE_LINEAR_ALLOC
111     return (LinearAllocHdr*) 0x12345;
112 #endif
113     LinearAllocHdr* pHdr;
114 
115     pHdr = (LinearAllocHdr*) malloc(sizeof(*pHdr));
116 
117 
118     /*
119      * "curOffset" points to the location of the next pre-block header,
120      * which means we have to advance to the next BLOCK_ALIGN address and
121      * back up.
122      *
123      * Note we leave the first page empty (see below), and start the
124      * first entry on the second page at an offset that ensures the next
125      * chunk of data will be properly aligned.
126      */
127     assert(BLOCK_ALIGN >= HEADER_EXTRA);
128     pHdr->curOffset = pHdr->firstOffset =
129         (BLOCK_ALIGN-HEADER_EXTRA) + SYSTEM_PAGE_SIZE;
130     pHdr->mapLength = DEFAULT_MAX_LENGTH;
131 
132 #ifdef USE_ASHMEM
133     int fd;
134 
135     fd = ashmem_create_region("dalvik-LinearAlloc", DEFAULT_MAX_LENGTH);
136     if (fd < 0) {
137         ALOGE("ashmem LinearAlloc failed %s", strerror(errno));
138         free(pHdr);
139         return NULL;
140     }
141 
142     pHdr->mapAddr = (char*)mmap(NULL, pHdr->mapLength, PROT_READ | PROT_WRITE,
143         MAP_PRIVATE, fd, 0);
144     if (pHdr->mapAddr == MAP_FAILED) {
145         ALOGE("LinearAlloc mmap(%d) failed: %s", pHdr->mapLength,
146             strerror(errno));
147         free(pHdr);
148         close(fd);
149         return NULL;
150     }
151 
152     close(fd);
153 #else /*USE_ASHMEM*/
154     // MAP_ANON is listed as "deprecated" on Linux,
155     // but MAP_ANONYMOUS is not defined under Mac OS X.
156     pHdr->mapAddr = mmap(NULL, pHdr->mapLength, PROT_READ | PROT_WRITE,
157         MAP_PRIVATE | MAP_ANON, -1, 0);
158     if (pHdr->mapAddr == MAP_FAILED) {
159         ALOGE("LinearAlloc mmap(%d) failed: %s", pHdr->mapLength,
160             strerror(errno));
161         free(pHdr);
162         return NULL;
163     }
164 #endif /*USE_ASHMEM*/
165 
166     /* region expected to begin on a page boundary */
167     assert(((int) pHdr->mapAddr & (SYSTEM_PAGE_SIZE-1)) == 0);
168 
169     /* the system should initialize newly-mapped memory to zero */
170     assert(*(u4*) (pHdr->mapAddr + pHdr->curOffset) == 0);
171 
172     /*
173      * Disable access to all except starting page.  We will enable pages
174      * as we use them.  This helps prevent bad pointers from working.  The
175      * pages start out PROT_NONE, become read/write while we access them,
176      * then go to read-only after we finish our changes.
177      *
178      * We have to make the first page readable because we have 4 pad bytes,
179      * followed by 4 length bytes, giving an initial offset of 8.  The
180      * generic code below assumes that there could have been a previous
181      * allocation that wrote into those 4 pad bytes, therefore the page
182      * must have been marked readable by the previous allocation.
183      *
184      * We insert an extra page in here to force a break in the memory map
185      * so we can see ourselves more easily in "showmap".  Otherwise this
186      * stuff blends into the neighboring pages.  [TODO: do we still need
187      * the extra page now that we have ashmem?]
188      */
189     if (mprotect(pHdr->mapAddr, pHdr->mapLength, PROT_NONE) != 0) {
190         ALOGW("LinearAlloc init mprotect failed: %s", strerror(errno));
191         free(pHdr);
192         return NULL;
193     }
194     if (mprotect(pHdr->mapAddr + SYSTEM_PAGE_SIZE, SYSTEM_PAGE_SIZE,
195             ENFORCE_READ_ONLY ? PROT_READ : PROT_READ|PROT_WRITE) != 0)
196     {
197         ALOGW("LinearAlloc init mprotect #2 failed: %s", strerror(errno));
198         free(pHdr);
199         return NULL;
200     }
201 
202     if (ENFORCE_READ_ONLY) {
203         /* allocate the per-page ref count */
204         int numPages = (pHdr->mapLength+SYSTEM_PAGE_SIZE-1) / SYSTEM_PAGE_SIZE;
205         pHdr->writeRefCount = (short*)calloc(numPages, sizeof(short));
206         if (pHdr->writeRefCount == NULL) {
207             free(pHdr);
208             return NULL;
209         }
210     }
211 
212     dvmInitMutex(&pHdr->lock);
213 
214     ALOGV("LinearAlloc: created region at %p-%p",
215         pHdr->mapAddr, pHdr->mapAddr + pHdr->mapLength-1);
216 
217     return pHdr;
218 }
219 
220 /*
221  * Destroy a linear allocation area.
222  *
223  * We do a trivial "has everything been freed?" check before unmapping the
224  * memory and freeing the LinearAllocHdr.
225  */
dvmLinearAllocDestroy(Object * classLoader)226 void dvmLinearAllocDestroy(Object* classLoader)
227 {
228 #ifdef DISABLE_LINEAR_ALLOC
229     return;
230 #endif
231     LinearAllocHdr* pHdr = getHeader(classLoader);
232     if (pHdr == NULL)
233         return;
234 
235     checkAllFree(classLoader);
236 
237     //dvmLinearAllocDump(classLoader);
238 
239     if (gDvm.verboseShutdown) {
240         ALOGV("Unmapping linear allocator base=%p", pHdr->mapAddr);
241         ALOGD("LinearAlloc %p used %d of %d (%d%%)",
242             classLoader, pHdr->curOffset, pHdr->mapLength,
243             (pHdr->curOffset * 100) / pHdr->mapLength);
244     }
245 
246     if (munmap(pHdr->mapAddr, pHdr->mapLength) != 0) {
247         ALOGW("LinearAlloc munmap(%p, %d) failed: %s",
248             pHdr->mapAddr, pHdr->mapLength, strerror(errno));
249     }
250     free(pHdr);
251 }
252 
253 /*
254  * Allocate "size" bytes of storage, associated with a particular class
255  * loader.
256  *
257  * It's okay for size to be zero.
258  *
259  * We always leave "curOffset" pointing at the next place where we will
260  * store the header that precedes the returned storage.
261  *
262  * This aborts the VM on failure, so it's not necessary to check for a
263  * NULL return value.
264  */
dvmLinearAlloc(Object * classLoader,size_t size)265 void* dvmLinearAlloc(Object* classLoader, size_t size)
266 {
267     LinearAllocHdr* pHdr = getHeader(classLoader);
268     int startOffset, nextOffset;
269     int lastGoodOff, firstWriteOff, lastWriteOff;
270 
271 #ifdef DISABLE_LINEAR_ALLOC
272     return calloc(1, size);
273 #endif
274 
275     LOGVV("--- LinearAlloc(%p, %d)", classLoader, size);
276 
277     /*
278      * What we'd like to do is just determine the new end-of-alloc size
279      * and atomic-swap the updated value in.  The trouble is that, the
280      * first time we reach a new page, we need to call mprotect() to
281      * make the page available, and we don't want to call mprotect() on
282      * every allocation.  The troubled situation is:
283      *  - thread A allocs across a page boundary, but gets preempted
284      *    before mprotect() completes
285      *  - thread B allocs within the new page, and doesn't call mprotect()
286      */
287     dvmLockMutex(&pHdr->lock);
288 
289     startOffset = pHdr->curOffset;
290     assert(((startOffset + HEADER_EXTRA) & (BLOCK_ALIGN-1)) == 0);
291 
292     /*
293      * Compute the new offset.  The old offset points at the address where
294      * we will store the hidden block header, so we advance past that,
295      * add the size of data they want, add another header's worth so we
296      * know we have room for that, and round up to BLOCK_ALIGN.  That's
297      * the next location where we'll put user data.  We then subtract the
298      * chunk header size off so we're back to the header pointer.
299      *
300      * Examples:
301      *   old=12 size=3 new=((12+(4*2)+3+7) & ~7)-4 = 24-4 --> 20
302      *   old=12 size=5 new=((12+(4*2)+5+7) & ~7)-4 = 32-4 --> 28
303      */
304     nextOffset = ((startOffset + HEADER_EXTRA*2 + size + (BLOCK_ALIGN-1))
305                     & ~(BLOCK_ALIGN-1)) - HEADER_EXTRA;
306     LOGVV("--- old=%d size=%d new=%d", startOffset, size, nextOffset);
307 
308     if (nextOffset > pHdr->mapLength) {
309         /*
310          * We don't have to abort here.  We could fall back on the system
311          * malloc(), and have our "free" call figure out what to do.  Only
312          * works if the users of these functions actually free everything
313          * they allocate.
314          */
315         ALOGE("LinearAlloc exceeded capacity (%d), last=%d",
316             pHdr->mapLength, (int) size);
317         dvmAbort();
318     }
319 
320     /*
321      * Round up "size" to encompass the entire region, including the 0-7
322      * pad bytes before the next chunk header.  This way we get maximum
323      * utility out of "realloc", and when we're doing ENFORCE_READ_ONLY
324      * stuff we always treat the full extent.
325      */
326     size = nextOffset - (startOffset + HEADER_EXTRA);
327     LOGVV("--- (size now %d)", size);
328 
329     /*
330      * See if we are starting on or have crossed into a new page.  If so,
331      * call mprotect on the page(s) we're about to write to.  We have to
332      * page-align the start address, but don't have to make the length a
333      * SYSTEM_PAGE_SIZE multiple (but we do it anyway).
334      *
335      * Note that "startOffset" is not the last *allocated* byte, but rather
336      * the offset of the first *unallocated* byte (which we are about to
337      * write the chunk header to).  "nextOffset" is similar.
338      *
339      * If ENFORCE_READ_ONLY is enabled, we have to call mprotect even if
340      * we've written to this page before, because it might be read-only.
341      */
342     lastGoodOff = (startOffset-1) & ~(SYSTEM_PAGE_SIZE-1);
343     firstWriteOff = startOffset & ~(SYSTEM_PAGE_SIZE-1);
344     lastWriteOff = (nextOffset-1) & ~(SYSTEM_PAGE_SIZE-1);
345     LOGVV("---  lastGood=0x%04x firstWrite=0x%04x lastWrite=0x%04x",
346         lastGoodOff, firstWriteOff, lastWriteOff);
347     if (lastGoodOff != lastWriteOff || ENFORCE_READ_ONLY) {
348         int cc, start, len;
349 
350         start = firstWriteOff;
351         assert(start <= nextOffset);
352         len = (lastWriteOff - firstWriteOff) + SYSTEM_PAGE_SIZE;
353 
354         LOGVV("---    calling mprotect(start=%d len=%d RW)", start, len);
355         cc = mprotect(pHdr->mapAddr + start, len, PROT_READ | PROT_WRITE);
356         if (cc != 0) {
357             ALOGE("LinearAlloc mprotect (+%d %d) failed: %s",
358                 start, len, strerror(errno));
359             /* we're going to fail soon, might as do it now */
360             dvmAbort();
361         }
362     }
363 
364     /* update the ref counts on the now-writable pages */
365     if (ENFORCE_READ_ONLY) {
366         int i, start, end;
367 
368         start = firstWriteOff / SYSTEM_PAGE_SIZE;
369         end = lastWriteOff / SYSTEM_PAGE_SIZE;
370 
371         LOGVV("---  marking pages %d-%d RW (alloc %d at %p)",
372             start, end, size, pHdr->mapAddr + startOffset + HEADER_EXTRA);
373         for (i = start; i <= end; i++)
374             pHdr->writeRefCount[i]++;
375     }
376 
377     /* stow the size in the header */
378     if (ENFORCE_READ_ONLY)
379         *(u4*)(pHdr->mapAddr + startOffset) = size | LENGTHFLAG_RW;
380     else
381         *(u4*)(pHdr->mapAddr + startOffset) = size;
382 
383     /*
384      * Update data structure.
385      */
386     pHdr->curOffset = nextOffset;
387 
388     dvmUnlockMutex(&pHdr->lock);
389     return pHdr->mapAddr + startOffset + HEADER_EXTRA;
390 }
391 
392 /*
393  * Helper function, replaces strdup().
394  */
dvmLinearStrdup(Object * classLoader,const char * str)395 char* dvmLinearStrdup(Object* classLoader, const char* str)
396 {
397 #ifdef DISABLE_LINEAR_ALLOC
398     return strdup(str);
399 #endif
400     int len = strlen(str);
401     void* mem = dvmLinearAlloc(classLoader, len+1);
402     memcpy(mem, str, len+1);
403     if (ENFORCE_READ_ONLY)
404         dvmLinearSetReadOnly(classLoader, mem);
405     return (char*) mem;
406 }
407 
408 /*
409  * "Reallocate" a piece of memory.
410  *
411  * If the new size is <= the old size, we return the original pointer
412  * without doing anything.
413  *
414  * If the new size is > the old size, we allocate new storage, copy the
415  * old stuff over, and mark the new stuff as free.
416  */
dvmLinearRealloc(Object * classLoader,void * mem,size_t newSize)417 void* dvmLinearRealloc(Object* classLoader, void* mem, size_t newSize)
418 {
419 #ifdef DISABLE_LINEAR_ALLOC
420     return realloc(mem, newSize);
421 #endif
422     /* make sure we have the right region (and mem != NULL) */
423     assert(mem != NULL);
424     assert(mem >= (void*) getHeader(classLoader)->mapAddr &&
425            mem < (void*) (getHeader(classLoader)->mapAddr +
426                           getHeader(classLoader)->curOffset));
427 
428     const u4* pLen = getBlockHeader(mem);
429     ALOGV("--- LinearRealloc(%d) old=%d", newSize, *pLen);
430 
431     /* handle size reduction case */
432     if (*pLen >= newSize) {
433         if (ENFORCE_READ_ONLY)
434             dvmLinearSetReadWrite(classLoader, mem);
435         return mem;
436     }
437 
438     void* newMem;
439 
440     newMem = dvmLinearAlloc(classLoader, newSize);
441     assert(newMem != NULL);
442     memcpy(newMem, mem, *pLen);
443     dvmLinearFree(classLoader, mem);
444 
445     return newMem;
446 }
447 
448 
449 /*
450  * Update the read/write status of one or more pages.
451  */
updatePages(Object * classLoader,void * mem,int direction)452 static void updatePages(Object* classLoader, void* mem, int direction)
453 {
454     LinearAllocHdr* pHdr = getHeader(classLoader);
455     dvmLockMutex(&pHdr->lock);
456 
457     /* make sure we have the right region */
458     assert(mem >= (void*) pHdr->mapAddr &&
459            mem < (void*) (pHdr->mapAddr + pHdr->curOffset));
460 
461     u4* pLen = getBlockHeader(mem);
462     u4 len = *pLen & LENGTHFLAG_MASK;
463     int firstPage, lastPage;
464 
465     firstPage = ((u1*)pLen - (u1*)pHdr->mapAddr) / SYSTEM_PAGE_SIZE;
466     lastPage = ((u1*)mem - (u1*)pHdr->mapAddr + (len-1)) / SYSTEM_PAGE_SIZE;
467     LOGVV("--- updating pages %d-%d (%d)", firstPage, lastPage, direction);
468 
469     int i, cc;
470 
471     /*
472      * Update individual pages.  We could do some sort of "lazy update" to
473      * combine mprotect calls, but that's almost certainly more trouble
474      * than it's worth.
475      */
476     for (i = firstPage; i <= lastPage; i++) {
477         if (direction < 0) {
478             /*
479              * Trying to mark read-only.
480              */
481             if (i == firstPage) {
482                 if ((*pLen & LENGTHFLAG_RW) == 0) {
483                     ALOGW("Double RO on %p", mem);
484                     dvmAbort();
485                 } else
486                     *pLen &= ~LENGTHFLAG_RW;
487             }
488 
489             if (pHdr->writeRefCount[i] == 0) {
490                 ALOGE("Can't make page %d any less writable", i);
491                 dvmAbort();
492             }
493             pHdr->writeRefCount[i]--;
494             if (pHdr->writeRefCount[i] == 0) {
495                 LOGVV("---  prot page %d RO", i);
496                 cc = mprotect(pHdr->mapAddr + SYSTEM_PAGE_SIZE * i,
497                         SYSTEM_PAGE_SIZE, PROT_READ);
498                 assert(cc == 0);
499             }
500         } else {
501             /*
502              * Trying to mark writable.
503              */
504             if (pHdr->writeRefCount[i] >= 32767) {
505                 ALOGE("Can't make page %d any more writable", i);
506                 dvmAbort();
507             }
508             if (pHdr->writeRefCount[i] == 0) {
509                 LOGVV("---  prot page %d RW", i);
510                 cc = mprotect(pHdr->mapAddr + SYSTEM_PAGE_SIZE * i,
511                         SYSTEM_PAGE_SIZE, PROT_READ | PROT_WRITE);
512                 assert(cc == 0);
513             }
514             pHdr->writeRefCount[i]++;
515 
516             if (i == firstPage) {
517                 if ((*pLen & LENGTHFLAG_RW) != 0) {
518                     ALOGW("Double RW on %p", mem);
519                     dvmAbort();
520                 } else
521                     *pLen |= LENGTHFLAG_RW;
522             }
523         }
524     }
525 
526     dvmUnlockMutex(&pHdr->lock);
527 }
528 
529 /*
530  * Try to mark the pages in which a chunk of memory lives as read-only.
531  * Whether or not the pages actually change state depends on how many
532  * others are trying to access the same pages.
533  *
534  * Only call here if ENFORCE_READ_ONLY is true.
535  */
dvmLinearSetReadOnly(Object * classLoader,void * mem)536 void dvmLinearSetReadOnly(Object* classLoader, void* mem)
537 {
538 #ifdef DISABLE_LINEAR_ALLOC
539     return;
540 #endif
541     updatePages(classLoader, mem, -1);
542 }
543 
544 /*
545  * Make the pages on which "mem" sits read-write.
546  *
547  * This covers the header as well as the data itself.  (We could add a
548  * "header-only" mode for dvmLinearFree.)
549  *
550  * Only call here if ENFORCE_READ_ONLY is true.
551  */
dvmLinearSetReadWrite(Object * classLoader,void * mem)552 void dvmLinearSetReadWrite(Object* classLoader, void* mem)
553 {
554 #ifdef DISABLE_LINEAR_ALLOC
555     return;
556 #endif
557     updatePages(classLoader, mem, 1);
558 }
559 
560 /*
561  * Mark an allocation as free.
562  */
dvmLinearFree(Object * classLoader,void * mem)563 void dvmLinearFree(Object* classLoader, void* mem)
564 {
565 #ifdef DISABLE_LINEAR_ALLOC
566     free(mem);
567     return;
568 #endif
569     if (mem == NULL)
570         return;
571 
572     /* make sure we have the right region */
573     assert(mem >= (void*) getHeader(classLoader)->mapAddr &&
574            mem < (void*) (getHeader(classLoader)->mapAddr +
575                           getHeader(classLoader)->curOffset));
576 
577     if (ENFORCE_READ_ONLY)
578         dvmLinearSetReadWrite(classLoader, mem);
579 
580     u4* pLen = getBlockHeader(mem);
581     *pLen |= LENGTHFLAG_FREE;
582 
583     if (ENFORCE_READ_ONLY)
584         dvmLinearSetReadOnly(classLoader, mem);
585 }
586 
587 /*
588  * For debugging, dump the contents of a linear alloc area.
589  *
590  * We grab the lock so that the header contents and list output are
591  * consistent.
592  */
dvmLinearAllocDump(Object * classLoader)593 void dvmLinearAllocDump(Object* classLoader)
594 {
595 #ifdef DISABLE_LINEAR_ALLOC
596     return;
597 #endif
598     LinearAllocHdr* pHdr = getHeader(classLoader);
599 
600     dvmLockMutex(&pHdr->lock);
601 
602     ALOGI("LinearAlloc classLoader=%p", classLoader);
603     ALOGI("  mapAddr=%p mapLength=%d firstOffset=%d",
604         pHdr->mapAddr, pHdr->mapLength, pHdr->firstOffset);
605     ALOGI("  curOffset=%d", pHdr->curOffset);
606 
607     int off = pHdr->firstOffset;
608     u4 rawLen, fullLen;
609 
610     while (off < pHdr->curOffset) {
611         rawLen = *(u4*) (pHdr->mapAddr + off);
612         fullLen = ((HEADER_EXTRA*2 + (rawLen & LENGTHFLAG_MASK))
613                     & ~(BLOCK_ALIGN-1));
614 
615         ALOGI("  %p (%3d): %clen=%d%s", pHdr->mapAddr + off + HEADER_EXTRA,
616             (int) ((off + HEADER_EXTRA) / SYSTEM_PAGE_SIZE),
617             (rawLen & LENGTHFLAG_FREE) != 0 ? '*' : ' ',
618             rawLen & LENGTHFLAG_MASK,
619             (rawLen & LENGTHFLAG_RW) != 0 ? " [RW]" : "");
620 
621         off += fullLen;
622     }
623 
624     if (ENFORCE_READ_ONLY) {
625         ALOGI("writeRefCount map:");
626 
627         int numPages = (pHdr->mapLength+SYSTEM_PAGE_SIZE-1) / SYSTEM_PAGE_SIZE;
628         int zstart = 0;
629         int i;
630 
631         for (i = 0; i < numPages; i++) {
632             int count = pHdr->writeRefCount[i];
633 
634             if (count != 0) {
635                 if (zstart < i-1)
636                     printf(" %d-%d: zero\n", zstart, i-1);
637                 else if (zstart == i-1)
638                     printf(" %d: zero\n", zstart);
639                 zstart = i+1;
640                 printf(" %d: %d\n", i, count);
641             }
642         }
643         if (zstart < i)
644             printf(" %d-%d: zero\n", zstart, i-1);
645     }
646 
647     ALOGD("LinearAlloc %p using %d of %d (%d%%)",
648         classLoader, pHdr->curOffset, pHdr->mapLength,
649         (pHdr->curOffset * 100) / pHdr->mapLength);
650 
651     dvmUnlockMutex(&pHdr->lock);
652 }
653 
654 /*
655  * Verify that all blocks are freed.
656  *
657  * This should only be done as we're shutting down, but there could be a
658  * daemon thread that's still trying to do something, so we grab the locks.
659  */
checkAllFree(Object * classLoader)660 static void checkAllFree(Object* classLoader)
661 {
662 #ifdef DISABLE_LINEAR_ALLOC
663     return;
664 #endif
665     LinearAllocHdr* pHdr = getHeader(classLoader);
666 
667     dvmLockMutex(&pHdr->lock);
668 
669     int off = pHdr->firstOffset;
670     u4 rawLen, fullLen;
671 
672     while (off < pHdr->curOffset) {
673         rawLen = *(u4*) (pHdr->mapAddr + off);
674         fullLen = ((HEADER_EXTRA*2 + (rawLen & LENGTHFLAG_MASK))
675                     & ~(BLOCK_ALIGN-1));
676 
677         if ((rawLen & LENGTHFLAG_FREE) == 0) {
678             ALOGW("LinearAlloc %p not freed: %p len=%d", classLoader,
679                 pHdr->mapAddr + off + HEADER_EXTRA, rawLen & LENGTHFLAG_MASK);
680         }
681 
682         off += fullLen;
683     }
684 
685     dvmUnlockMutex(&pHdr->lock);
686 }
687 
688 /*
689  * Determine if [start, start+length) is contained in the in-use area of
690  * a single LinearAlloc.  The full set of linear allocators is scanned.
691  *
692  * [ Since we currently only have one region, this is pretty simple.  In
693  * the future we'll need to traverse a table of class loaders. ]
694  */
dvmLinearAllocContains(const void * start,size_t length)695 bool dvmLinearAllocContains(const void* start, size_t length)
696 {
697     LinearAllocHdr* pHdr = getHeader(NULL);
698 
699     if (pHdr == NULL)
700         return false;
701 
702     return (char*) start >= pHdr->mapAddr &&
703            ((char*)start + length) <= (pHdr->mapAddr + pHdr->curOffset);
704 }
705