• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (C) 2005 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #define LOG_TAG "RefBase"
18 // #define LOG_NDEBUG 0
19 
20 #include <utils/RefBase.h>
21 
22 #include <utils/CallStack.h>
23 
24 #ifndef __unused
25 #define __unused __attribute__((__unused__))
26 #endif
27 
28 // compile with refcounting debugging enabled
29 #define DEBUG_REFS                      0
30 
31 // whether ref-tracking is enabled by default, if not, trackMe(true, false)
32 // needs to be called explicitly
33 #define DEBUG_REFS_ENABLED_BY_DEFAULT   0
34 
35 // whether callstack are collected (significantly slows things down)
36 #define DEBUG_REFS_CALLSTACK_ENABLED    1
37 
38 // folder where stack traces are saved when DEBUG_REFS is enabled
39 // this folder needs to exist and be writable
40 #define DEBUG_REFS_CALLSTACK_PATH       "/data/debug"
41 
42 // log all reference counting operations
43 #define PRINT_REFS                      0
44 
45 // ---------------------------------------------------------------------------
46 
47 namespace android {
48 
49 // Observations, invariants, etc:
50 
51 // By default, obects are destroyed when the last strong reference disappears
52 // or, if the object never had a strong reference, when the last weak reference
53 // disappears.
54 //
55 // OBJECT_LIFETIME_WEAK changes this behavior to retain the object
56 // unconditionally until the last reference of either kind disappears.  The
57 // client ensures that the extendObjectLifetime call happens before the dec
58 // call that would otherwise have deallocated the object, or before an
59 // attemptIncStrong call that might rely on it.  We do not worry about
60 // concurrent changes to the object lifetime.
61 //
62 // AttemptIncStrong will succeed if the object has a strong reference, or if it
63 // has a weak reference and has never had a strong reference.
64 // AttemptIncWeak really does succeed only if there is already a WEAK
65 // reference, and thus may fail when attemptIncStrong would succeed.
66 //
67 // mStrong is the strong reference count.  mWeak is the weak reference count.
68 // Between calls, and ignoring memory ordering effects, mWeak includes strong
69 // references, and is thus >= mStrong.
70 //
71 // A weakref_impl holds all the information, including both reference counts,
72 // required to perform wp<> operations.  Thus these can continue to be performed
73 // after the RefBase object has been destroyed.
74 //
75 // A weakref_impl is allocated as the value of mRefs in a RefBase object on
76 // construction.
77 // In the OBJECT_LIFETIME_STRONG case, it is normally deallocated in decWeak,
78 // and hence lives as long as the last weak reference. (It can also be
79 // deallocated in the RefBase destructor iff the strong reference count was
80 // never incremented and the weak count is zero, e.g.  if the RefBase object is
81 // explicitly destroyed without decrementing the strong count.  This should be
82 // avoided.) In this case, the RefBase destructor should be invoked from
83 // decStrong.
84 // In the OBJECT_LIFETIME_WEAK case, the weakref_impl is always deallocated in
85 // the RefBase destructor, which is always invoked by decWeak. DecStrong
86 // explicitly avoids the deletion in this case.
87 //
88 // Memory ordering:
89 // The client must ensure that every inc() call, together with all other
90 // accesses to the object, happens before the corresponding dec() call.
91 //
92 // We try to keep memory ordering constraints on atomics as weak as possible,
93 // since memory fences or ordered memory accesses are likely to be a major
94 // performance cost for this code. All accesses to mStrong, mWeak, and mFlags
95 // explicitly relax memory ordering in some way.
96 //
97 // The only operations that are not memory_order_relaxed are reference count
98 // decrements. All reference count decrements are release operations.  In
99 // addition, the final decrement leading the deallocation is followed by an
100 // acquire fence, which we can view informally as also turning it into an
101 // acquire operation.  (See 29.8p4 [atomics.fences] for details. We could
102 // alternatively use acq_rel operations for all decrements. This is probably
103 // slower on most current (2016) hardware, especially on ARMv7, but that may
104 // not be true indefinitely.)
105 //
106 // This convention ensures that the second-to-last decrement synchronizes with
107 // (in the language of 1.10 in the C++ standard) the final decrement of a
108 // reference count. Since reference counts are only updated using atomic
109 // read-modify-write operations, this also extends to any earlier decrements.
110 // (See "release sequence" in 1.10.)
111 //
112 // Since all operations on an object happen before the corresponding reference
113 // count decrement, and all reference count decrements happen before the final
114 // one, we are guaranteed that all other object accesses happen before the
115 // object is destroyed.
116 
117 
118 #define INITIAL_STRONG_VALUE (1<<28)
119 
120 #define MAX_COUNT 0xfffff
121 
122 // Test whether the argument is a clearly invalid strong reference count.
123 // Used only for error checking on the value before an atomic decrement.
124 // Intended to be very cheap.
125 // Note that we cannot just check for excess decrements by comparing to zero
126 // since the object would be deallocated before that.
127 #define BAD_STRONG(c) \
128         ((c) == 0 || ((c) & (~(MAX_COUNT | INITIAL_STRONG_VALUE))) != 0)
129 
130 // Same for weak counts.
131 #define BAD_WEAK(c) ((c) == 0 || ((c) & (~MAX_COUNT)) != 0)
132 
133 // ---------------------------------------------------------------------------
134 
135 class RefBase::weakref_impl : public RefBase::weakref_type
136 {
137 public:
138     std::atomic<int32_t>    mStrong;
139     std::atomic<int32_t>    mWeak;
140     RefBase* const          mBase;
141     std::atomic<int32_t>    mFlags;
142 
143 #if !DEBUG_REFS
144 
weakref_impl(RefBase * base)145     explicit weakref_impl(RefBase* base)
146         : mStrong(INITIAL_STRONG_VALUE)
147         , mWeak(0)
148         , mBase(base)
149         , mFlags(0)
150     {
151     }
152 
addStrongRef(const void *)153     void addStrongRef(const void* /*id*/) { }
removeStrongRef(const void *)154     void removeStrongRef(const void* /*id*/) { }
renameStrongRefId(const void *,const void *)155     void renameStrongRefId(const void* /*old_id*/, const void* /*new_id*/) { }
addWeakRef(const void *)156     void addWeakRef(const void* /*id*/) { }
removeWeakRef(const void *)157     void removeWeakRef(const void* /*id*/) { }
renameWeakRefId(const void *,const void *)158     void renameWeakRefId(const void* /*old_id*/, const void* /*new_id*/) { }
printRefs() const159     void printRefs() const { }
trackMe(bool,bool)160     void trackMe(bool, bool) { }
161 
162 #else
163 
weakref_impl(RefBase * base)164     weakref_impl(RefBase* base)
165         : mStrong(INITIAL_STRONG_VALUE)
166         , mWeak(0)
167         , mBase(base)
168         , mFlags(0)
169         , mStrongRefs(NULL)
170         , mWeakRefs(NULL)
171         , mTrackEnabled(!!DEBUG_REFS_ENABLED_BY_DEFAULT)
172         , mRetain(false)
173     {
174     }
175 
~weakref_impl()176     ~weakref_impl()
177     {
178         bool dumpStack = false;
179         if (!mRetain && mStrongRefs != NULL) {
180             dumpStack = true;
181             ALOGE("Strong references remain:");
182             ref_entry* refs = mStrongRefs;
183             while (refs) {
184                 char inc = refs->ref >= 0 ? '+' : '-';
185                 ALOGD("\t%c ID %p (ref %d):", inc, refs->id, refs->ref);
186 #if DEBUG_REFS_CALLSTACK_ENABLED
187                 refs->stack.log(LOG_TAG);
188 #endif
189                 refs = refs->next;
190             }
191         }
192 
193         if (!mRetain && mWeakRefs != NULL) {
194             dumpStack = true;
195             ALOGE("Weak references remain!");
196             ref_entry* refs = mWeakRefs;
197             while (refs) {
198                 char inc = refs->ref >= 0 ? '+' : '-';
199                 ALOGD("\t%c ID %p (ref %d):", inc, refs->id, refs->ref);
200 #if DEBUG_REFS_CALLSTACK_ENABLED
201                 refs->stack.log(LOG_TAG);
202 #endif
203                 refs = refs->next;
204             }
205         }
206         if (dumpStack) {
207             ALOGE("above errors at:");
208             CallStack stack(LOG_TAG);
209         }
210     }
211 
addStrongRef(const void * id)212     void addStrongRef(const void* id) {
213         //ALOGD_IF(mTrackEnabled,
214         //        "addStrongRef: RefBase=%p, id=%p", mBase, id);
215         addRef(&mStrongRefs, id, mStrong.load(std::memory_order_relaxed));
216     }
217 
removeStrongRef(const void * id)218     void removeStrongRef(const void* id) {
219         //ALOGD_IF(mTrackEnabled,
220         //        "removeStrongRef: RefBase=%p, id=%p", mBase, id);
221         if (!mRetain) {
222             removeRef(&mStrongRefs, id);
223         } else {
224             addRef(&mStrongRefs, id, -mStrong.load(std::memory_order_relaxed));
225         }
226     }
227 
renameStrongRefId(const void * old_id,const void * new_id)228     void renameStrongRefId(const void* old_id, const void* new_id) {
229         //ALOGD_IF(mTrackEnabled,
230         //        "renameStrongRefId: RefBase=%p, oid=%p, nid=%p",
231         //        mBase, old_id, new_id);
232         renameRefsId(mStrongRefs, old_id, new_id);
233     }
234 
addWeakRef(const void * id)235     void addWeakRef(const void* id) {
236         addRef(&mWeakRefs, id, mWeak.load(std::memory_order_relaxed));
237     }
238 
removeWeakRef(const void * id)239     void removeWeakRef(const void* id) {
240         if (!mRetain) {
241             removeRef(&mWeakRefs, id);
242         } else {
243             addRef(&mWeakRefs, id, -mWeak.load(std::memory_order_relaxed));
244         }
245     }
246 
renameWeakRefId(const void * old_id,const void * new_id)247     void renameWeakRefId(const void* old_id, const void* new_id) {
248         renameRefsId(mWeakRefs, old_id, new_id);
249     }
250 
trackMe(bool track,bool retain)251     void trackMe(bool track, bool retain)
252     {
253         mTrackEnabled = track;
254         mRetain = retain;
255     }
256 
printRefs() const257     void printRefs() const
258     {
259         String8 text;
260 
261         {
262             Mutex::Autolock _l(mMutex);
263             char buf[128];
264             snprintf(buf, sizeof(buf),
265                      "Strong references on RefBase %p (weakref_type %p):\n",
266                      mBase, this);
267             text.append(buf);
268             printRefsLocked(&text, mStrongRefs);
269             snprintf(buf, sizeof(buf),
270                      "Weak references on RefBase %p (weakref_type %p):\n",
271                      mBase, this);
272             text.append(buf);
273             printRefsLocked(&text, mWeakRefs);
274         }
275 
276         {
277             char name[100];
278             snprintf(name, sizeof(name), DEBUG_REFS_CALLSTACK_PATH "/%p.stack",
279                      this);
280             int rc = open(name, O_RDWR | O_CREAT | O_APPEND, 644);
281             if (rc >= 0) {
282                 write(rc, text.string(), text.length());
283                 close(rc);
284                 ALOGD("STACK TRACE for %p saved in %s", this, name);
285             }
286             else ALOGE("FAILED TO PRINT STACK TRACE for %p in %s: %s", this,
287                       name, strerror(errno));
288         }
289     }
290 
291 private:
292     struct ref_entry
293     {
294         ref_entry* next;
295         const void* id;
296 #if DEBUG_REFS_CALLSTACK_ENABLED
297         CallStack stack;
298 #endif
299         int32_t ref;
300     };
301 
addRef(ref_entry ** refs,const void * id,int32_t mRef)302     void addRef(ref_entry** refs, const void* id, int32_t mRef)
303     {
304         if (mTrackEnabled) {
305             AutoMutex _l(mMutex);
306 
307             ref_entry* ref = new ref_entry;
308             // Reference count at the time of the snapshot, but before the
309             // update.  Positive value means we increment, negative--we
310             // decrement the reference count.
311             ref->ref = mRef;
312             ref->id = id;
313 #if DEBUG_REFS_CALLSTACK_ENABLED
314             ref->stack.update(2);
315 #endif
316             ref->next = *refs;
317             *refs = ref;
318         }
319     }
320 
removeRef(ref_entry ** refs,const void * id)321     void removeRef(ref_entry** refs, const void* id)
322     {
323         if (mTrackEnabled) {
324             AutoMutex _l(mMutex);
325 
326             ref_entry* const head = *refs;
327             ref_entry* ref = head;
328             while (ref != NULL) {
329                 if (ref->id == id) {
330                     *refs = ref->next;
331                     delete ref;
332                     return;
333                 }
334                 refs = &ref->next;
335                 ref = *refs;
336             }
337 
338             ALOGE("RefBase: removing id %p on RefBase %p"
339                     "(weakref_type %p) that doesn't exist!",
340                     id, mBase, this);
341 
342             ref = head;
343             while (ref) {
344                 char inc = ref->ref >= 0 ? '+' : '-';
345                 ALOGD("\t%c ID %p (ref %d):", inc, ref->id, ref->ref);
346                 ref = ref->next;
347             }
348 
349             CallStack stack(LOG_TAG);
350         }
351     }
352 
renameRefsId(ref_entry * r,const void * old_id,const void * new_id)353     void renameRefsId(ref_entry* r, const void* old_id, const void* new_id)
354     {
355         if (mTrackEnabled) {
356             AutoMutex _l(mMutex);
357             ref_entry* ref = r;
358             while (ref != NULL) {
359                 if (ref->id == old_id) {
360                     ref->id = new_id;
361                 }
362                 ref = ref->next;
363             }
364         }
365     }
366 
printRefsLocked(String8 * out,const ref_entry * refs) const367     void printRefsLocked(String8* out, const ref_entry* refs) const
368     {
369         char buf[128];
370         while (refs) {
371             char inc = refs->ref >= 0 ? '+' : '-';
372             snprintf(buf, sizeof(buf), "\t%c ID %p (ref %d):\n",
373                      inc, refs->id, refs->ref);
374             out->append(buf);
375 #if DEBUG_REFS_CALLSTACK_ENABLED
376             out->append(refs->stack.toString("\t\t"));
377 #else
378             out->append("\t\t(call stacks disabled)");
379 #endif
380             refs = refs->next;
381         }
382     }
383 
384     mutable Mutex mMutex;
385     ref_entry* mStrongRefs;
386     ref_entry* mWeakRefs;
387 
388     bool mTrackEnabled;
389     // Collect stack traces on addref and removeref, instead of deleting the stack references
390     // on removeref that match the address ones.
391     bool mRetain;
392 
393 #endif
394 };
395 
396 // ---------------------------------------------------------------------------
397 
incStrong(const void * id) const398 void RefBase::incStrong(const void* id) const
399 {
400     weakref_impl* const refs = mRefs;
401     refs->incWeak(id);
402 
403     refs->addStrongRef(id);
404     const int32_t c = refs->mStrong.fetch_add(1, std::memory_order_relaxed);
405     ALOG_ASSERT(c > 0, "incStrong() called on %p after last strong ref", refs);
406 #if PRINT_REFS
407     ALOGD("incStrong of %p from %p: cnt=%d\n", this, id, c);
408 #endif
409     if (c != INITIAL_STRONG_VALUE)  {
410         return;
411     }
412 
413     int32_t old __unused = refs->mStrong.fetch_sub(INITIAL_STRONG_VALUE, std::memory_order_relaxed);
414     // A decStrong() must still happen after us.
415     ALOG_ASSERT(old > INITIAL_STRONG_VALUE, "0x%x too small", old);
416     refs->mBase->onFirstRef();
417 }
418 
decStrong(const void * id) const419 void RefBase::decStrong(const void* id) const
420 {
421     weakref_impl* const refs = mRefs;
422     refs->removeStrongRef(id);
423     const int32_t c = refs->mStrong.fetch_sub(1, std::memory_order_release);
424 #if PRINT_REFS
425     ALOGD("decStrong of %p from %p: cnt=%d\n", this, id, c);
426 #endif
427     LOG_ALWAYS_FATAL_IF(BAD_STRONG(c), "decStrong() called on %p too many times",
428             refs);
429     if (c == 1) {
430         std::atomic_thread_fence(std::memory_order_acquire);
431         refs->mBase->onLastStrongRef(id);
432         int32_t flags = refs->mFlags.load(std::memory_order_relaxed);
433         if ((flags&OBJECT_LIFETIME_MASK) == OBJECT_LIFETIME_STRONG) {
434             delete this;
435             // The destructor does not delete refs in this case.
436         }
437     }
438     // Note that even with only strong reference operations, the thread
439     // deallocating this may not be the same as the thread deallocating refs.
440     // That's OK: all accesses to this happen before its deletion here,
441     // and all accesses to refs happen before its deletion in the final decWeak.
442     // The destructor can safely access mRefs because either it's deleting
443     // mRefs itself, or it's running entirely before the final mWeak decrement.
444     //
445     // Since we're doing atomic loads of `flags`, the static analyzer assumes
446     // they can change between `delete this;` and `refs->decWeak(id);`. This is
447     // not the case. The analyzer may become more okay with this patten when
448     // https://bugs.llvm.org/show_bug.cgi?id=34365 gets resolved. NOLINTNEXTLINE
449     refs->decWeak(id);
450 }
451 
forceIncStrong(const void * id) const452 void RefBase::forceIncStrong(const void* id) const
453 {
454     // Allows initial mStrong of 0 in addition to INITIAL_STRONG_VALUE.
455     // TODO: Better document assumptions.
456     weakref_impl* const refs = mRefs;
457     refs->incWeak(id);
458 
459     refs->addStrongRef(id);
460     const int32_t c = refs->mStrong.fetch_add(1, std::memory_order_relaxed);
461     ALOG_ASSERT(c >= 0, "forceIncStrong called on %p after ref count underflow",
462                refs);
463 #if PRINT_REFS
464     ALOGD("forceIncStrong of %p from %p: cnt=%d\n", this, id, c);
465 #endif
466 
467     switch (c) {
468     case INITIAL_STRONG_VALUE:
469         refs->mStrong.fetch_sub(INITIAL_STRONG_VALUE,
470                 std::memory_order_relaxed);
471         // fall through...
472     case 0:
473         refs->mBase->onFirstRef();
474     }
475 }
476 
getStrongCount() const477 int32_t RefBase::getStrongCount() const
478 {
479     // Debugging only; No memory ordering guarantees.
480     return mRefs->mStrong.load(std::memory_order_relaxed);
481 }
482 
refBase() const483 RefBase* RefBase::weakref_type::refBase() const
484 {
485     return static_cast<const weakref_impl*>(this)->mBase;
486 }
487 
incWeak(const void * id)488 void RefBase::weakref_type::incWeak(const void* id)
489 {
490     weakref_impl* const impl = static_cast<weakref_impl*>(this);
491     impl->addWeakRef(id);
492     const int32_t c __unused = impl->mWeak.fetch_add(1,
493             std::memory_order_relaxed);
494     ALOG_ASSERT(c >= 0, "incWeak called on %p after last weak ref", this);
495 }
496 
497 
decWeak(const void * id)498 void RefBase::weakref_type::decWeak(const void* id)
499 {
500     weakref_impl* const impl = static_cast<weakref_impl*>(this);
501     impl->removeWeakRef(id);
502     const int32_t c = impl->mWeak.fetch_sub(1, std::memory_order_release);
503     LOG_ALWAYS_FATAL_IF(BAD_WEAK(c), "decWeak called on %p too many times",
504             this);
505     if (c != 1) return;
506     atomic_thread_fence(std::memory_order_acquire);
507 
508     int32_t flags = impl->mFlags.load(std::memory_order_relaxed);
509     if ((flags&OBJECT_LIFETIME_MASK) == OBJECT_LIFETIME_STRONG) {
510         // This is the regular lifetime case. The object is destroyed
511         // when the last strong reference goes away. Since weakref_impl
512         // outlives the object, it is not destroyed in the dtor, and
513         // we'll have to do it here.
514         if (impl->mStrong.load(std::memory_order_relaxed)
515                 == INITIAL_STRONG_VALUE) {
516             // Decrementing a weak count to zero when object never had a strong
517             // reference.  We assume it acquired a weak reference early, e.g.
518             // in the constructor, and will eventually be properly destroyed,
519             // usually via incrementing and decrementing the strong count.
520             // Thus we no longer do anything here.  We log this case, since it
521             // seems to be extremely rare, and should not normally occur. We
522             // used to deallocate mBase here, so this may now indicate a leak.
523             ALOGW("RefBase: Object at %p lost last weak reference "
524                     "before it had a strong reference", impl->mBase);
525         } else {
526             // ALOGV("Freeing refs %p of old RefBase %p\n", this, impl->mBase);
527             delete impl;
528         }
529     } else {
530         // This is the OBJECT_LIFETIME_WEAK case. The last weak-reference
531         // is gone, we can destroy the object.
532         impl->mBase->onLastWeakRef(id);
533         delete impl->mBase;
534     }
535 }
536 
attemptIncStrong(const void * id)537 bool RefBase::weakref_type::attemptIncStrong(const void* id)
538 {
539     incWeak(id);
540 
541     weakref_impl* const impl = static_cast<weakref_impl*>(this);
542     int32_t curCount = impl->mStrong.load(std::memory_order_relaxed);
543 
544     ALOG_ASSERT(curCount >= 0,
545             "attemptIncStrong called on %p after underflow", this);
546 
547     while (curCount > 0 && curCount != INITIAL_STRONG_VALUE) {
548         // we're in the easy/common case of promoting a weak-reference
549         // from an existing strong reference.
550         if (impl->mStrong.compare_exchange_weak(curCount, curCount+1,
551                 std::memory_order_relaxed)) {
552             break;
553         }
554         // the strong count has changed on us, we need to re-assert our
555         // situation. curCount was updated by compare_exchange_weak.
556     }
557 
558     if (curCount <= 0 || curCount == INITIAL_STRONG_VALUE) {
559         // we're now in the harder case of either:
560         // - there never was a strong reference on us
561         // - or, all strong references have been released
562         int32_t flags = impl->mFlags.load(std::memory_order_relaxed);
563         if ((flags&OBJECT_LIFETIME_MASK) == OBJECT_LIFETIME_STRONG) {
564             // this object has a "normal" life-time, i.e.: it gets destroyed
565             // when the last strong reference goes away
566             if (curCount <= 0) {
567                 // the last strong-reference got released, the object cannot
568                 // be revived.
569                 decWeak(id);
570                 return false;
571             }
572 
573             // here, curCount == INITIAL_STRONG_VALUE, which means
574             // there never was a strong-reference, so we can try to
575             // promote this object; we need to do that atomically.
576             while (curCount > 0) {
577                 if (impl->mStrong.compare_exchange_weak(curCount, curCount+1,
578                         std::memory_order_relaxed)) {
579                     break;
580                 }
581                 // the strong count has changed on us, we need to re-assert our
582                 // situation (e.g.: another thread has inc/decStrong'ed us)
583                 // curCount has been updated.
584             }
585 
586             if (curCount <= 0) {
587                 // promote() failed, some other thread destroyed us in the
588                 // meantime (i.e.: strong count reached zero).
589                 decWeak(id);
590                 return false;
591             }
592         } else {
593             // this object has an "extended" life-time, i.e.: it can be
594             // revived from a weak-reference only.
595             // Ask the object's implementation if it agrees to be revived
596             if (!impl->mBase->onIncStrongAttempted(FIRST_INC_STRONG, id)) {
597                 // it didn't so give-up.
598                 decWeak(id);
599                 return false;
600             }
601             // grab a strong-reference, which is always safe due to the
602             // extended life-time.
603             curCount = impl->mStrong.fetch_add(1, std::memory_order_relaxed);
604             // If the strong reference count has already been incremented by
605             // someone else, the implementor of onIncStrongAttempted() is holding
606             // an unneeded reference.  So call onLastStrongRef() here to remove it.
607             // (No, this is not pretty.)  Note that we MUST NOT do this if we
608             // are in fact acquiring the first reference.
609             if (curCount != 0 && curCount != INITIAL_STRONG_VALUE) {
610                 impl->mBase->onLastStrongRef(id);
611             }
612         }
613     }
614 
615     impl->addStrongRef(id);
616 
617 #if PRINT_REFS
618     ALOGD("attemptIncStrong of %p from %p: cnt=%d\n", this, id, curCount);
619 #endif
620 
621     // curCount is the value of mStrong before we incremented it.
622     // Now we need to fix-up the count if it was INITIAL_STRONG_VALUE.
623     // This must be done safely, i.e.: handle the case where several threads
624     // were here in attemptIncStrong().
625     // curCount > INITIAL_STRONG_VALUE is OK, and can happen if we're doing
626     // this in the middle of another incStrong.  The subtraction is handled
627     // by the thread that started with INITIAL_STRONG_VALUE.
628     if (curCount == INITIAL_STRONG_VALUE) {
629         impl->mStrong.fetch_sub(INITIAL_STRONG_VALUE,
630                 std::memory_order_relaxed);
631     }
632 
633     return true;
634 }
635 
attemptIncWeak(const void * id)636 bool RefBase::weakref_type::attemptIncWeak(const void* id)
637 {
638     weakref_impl* const impl = static_cast<weakref_impl*>(this);
639 
640     int32_t curCount = impl->mWeak.load(std::memory_order_relaxed);
641     ALOG_ASSERT(curCount >= 0, "attemptIncWeak called on %p after underflow",
642                this);
643     while (curCount > 0) {
644         if (impl->mWeak.compare_exchange_weak(curCount, curCount+1,
645                 std::memory_order_relaxed)) {
646             break;
647         }
648         // curCount has been updated.
649     }
650 
651     if (curCount > 0) {
652         impl->addWeakRef(id);
653     }
654 
655     return curCount > 0;
656 }
657 
getWeakCount() const658 int32_t RefBase::weakref_type::getWeakCount() const
659 {
660     // Debug only!
661     return static_cast<const weakref_impl*>(this)->mWeak
662             .load(std::memory_order_relaxed);
663 }
664 
printRefs() const665 void RefBase::weakref_type::printRefs() const
666 {
667     static_cast<const weakref_impl*>(this)->printRefs();
668 }
669 
trackMe(bool enable,bool retain)670 void RefBase::weakref_type::trackMe(bool enable, bool retain)
671 {
672     static_cast<weakref_impl*>(this)->trackMe(enable, retain);
673 }
674 
createWeak(const void * id) const675 RefBase::weakref_type* RefBase::createWeak(const void* id) const
676 {
677     mRefs->incWeak(id);
678     return mRefs;
679 }
680 
getWeakRefs() const681 RefBase::weakref_type* RefBase::getWeakRefs() const
682 {
683     return mRefs;
684 }
685 
RefBase()686 RefBase::RefBase()
687     : mRefs(new weakref_impl(this))
688 {
689 }
690 
~RefBase()691 RefBase::~RefBase()
692 {
693     int32_t flags = mRefs->mFlags.load(std::memory_order_relaxed);
694     // Life-time of this object is extended to WEAK, in
695     // which case weakref_impl doesn't out-live the object and we
696     // can free it now.
697     if ((flags & OBJECT_LIFETIME_MASK) == OBJECT_LIFETIME_WEAK) {
698         // It's possible that the weak count is not 0 if the object
699         // re-acquired a weak reference in its destructor
700         if (mRefs->mWeak.load(std::memory_order_relaxed) == 0) {
701             delete mRefs;
702         }
703     } else if (mRefs->mStrong.load(std::memory_order_relaxed)
704             == INITIAL_STRONG_VALUE) {
705         // We never acquired a strong reference on this object.
706         LOG_ALWAYS_FATAL_IF(mRefs->mWeak.load() != 0,
707                 "RefBase: Explicit destruction with non-zero weak "
708                 "reference count");
709         // TODO: Always report if we get here. Currently MediaMetadataRetriever
710         // C++ objects are inconsistently managed and sometimes get here.
711         // There may be other cases, but we believe they should all be fixed.
712         delete mRefs;
713     }
714     // For debugging purposes, clear mRefs.  Ineffective against outstanding wp's.
715     const_cast<weakref_impl*&>(mRefs) = NULL;
716 }
717 
extendObjectLifetime(int32_t mode)718 void RefBase::extendObjectLifetime(int32_t mode)
719 {
720     // Must be happens-before ordered with respect to construction or any
721     // operation that could destroy the object.
722     mRefs->mFlags.fetch_or(mode, std::memory_order_relaxed);
723 }
724 
onFirstRef()725 void RefBase::onFirstRef()
726 {
727 }
728 
onLastStrongRef(const void *)729 void RefBase::onLastStrongRef(const void* /*id*/)
730 {
731 }
732 
onIncStrongAttempted(uint32_t flags,const void *)733 bool RefBase::onIncStrongAttempted(uint32_t flags, const void* /*id*/)
734 {
735     return (flags&FIRST_INC_STRONG) ? true : false;
736 }
737 
onLastWeakRef(const void *)738 void RefBase::onLastWeakRef(const void* /*id*/)
739 {
740 }
741 
742 // ---------------------------------------------------------------------------
743 
744 #if DEBUG_REFS
renameRefs(size_t n,const ReferenceRenamer & renamer)745 void RefBase::renameRefs(size_t n, const ReferenceRenamer& renamer) {
746     for (size_t i=0 ; i<n ; i++) {
747         renamer(i);
748     }
749 }
750 #else
renameRefs(size_t,const ReferenceRenamer &)751 void RefBase::renameRefs(size_t /*n*/, const ReferenceRenamer& /*renamer*/) { }
752 #endif
753 
renameRefId(weakref_type * ref,const void * old_id,const void * new_id)754 void RefBase::renameRefId(weakref_type* ref,
755         const void* old_id, const void* new_id) {
756     weakref_impl* const impl = static_cast<weakref_impl*>(ref);
757     impl->renameStrongRefId(old_id, new_id);
758     impl->renameWeakRefId(old_id, new_id);
759 }
760 
renameRefId(RefBase * ref,const void * old_id,const void * new_id)761 void RefBase::renameRefId(RefBase* ref,
762         const void* old_id, const void* new_id) {
763     ref->mRefs->renameStrongRefId(old_id, new_id);
764     ref->mRefs->renameWeakRefId(old_id, new_id);
765 }
766 
767 }; // namespace android
768