1 /*
2 * Copyright (C) 2013 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17 #ifndef ART_RUNTIME_MIRROR_DEX_CACHE_INL_H_
18 #define ART_RUNTIME_MIRROR_DEX_CACHE_INL_H_
19
20 #include "dex_cache.h"
21
22 #include <android-base/logging.h>
23
24 #include "art_field.h"
25 #include "art_method.h"
26 #include "base/casts.h"
27 #include "base/enums.h"
28 #include "class_linker.h"
29 #include "dex/dex_file.h"
30 #include "gc_root-inl.h"
31 #include "mirror/call_site.h"
32 #include "mirror/class.h"
33 #include "mirror/method_type.h"
34 #include "obj_ptr.h"
35 #include "object-inl.h"
36 #include "runtime.h"
37 #include "write_barrier-inl.h"
38
39 #include <atomic>
40
41 namespace art {
42 namespace mirror {
43
44 template <typename T>
DexCachePair(ObjPtr<T> object,uint32_t index)45 inline DexCachePair<T>::DexCachePair(ObjPtr<T> object, uint32_t index)
46 : object(object), index(index) {}
47
48 template <typename T>
Initialize(std::atomic<DexCachePair<T>> * dex_cache)49 inline void DexCachePair<T>::Initialize(std::atomic<DexCachePair<T>>* dex_cache) {
50 DexCachePair<T> first_elem;
51 first_elem.object = GcRoot<T>(nullptr);
52 first_elem.index = InvalidIndexForSlot(0);
53 dex_cache[0].store(first_elem, std::memory_order_relaxed);
54 }
55
56 template <typename T>
GetObjectForIndex(uint32_t idx)57 inline T* DexCachePair<T>::GetObjectForIndex(uint32_t idx) {
58 if (idx != index) {
59 return nullptr;
60 }
61 DCHECK(!object.IsNull());
62 return object.Read();
63 }
64
65 template <typename T>
Initialize(std::atomic<NativeDexCachePair<T>> * dex_cache,PointerSize pointer_size)66 inline void NativeDexCachePair<T>::Initialize(std::atomic<NativeDexCachePair<T>>* dex_cache,
67 PointerSize pointer_size) {
68 NativeDexCachePair<T> first_elem;
69 first_elem.object = nullptr;
70 first_elem.index = InvalidIndexForSlot(0);
71 DexCache::SetNativePairPtrSize(dex_cache, 0, first_elem, pointer_size);
72 }
73
ClassSize(PointerSize pointer_size)74 inline uint32_t DexCache::ClassSize(PointerSize pointer_size) {
75 const uint32_t vtable_entries = Object::kVTableLength;
76 return Class::ComputeClassSize(true, vtable_entries, 0, 0, 0, 0, 0, pointer_size);
77 }
78
StringSlotIndex(dex::StringIndex string_idx)79 inline uint32_t DexCache::StringSlotIndex(dex::StringIndex string_idx) {
80 DCHECK_LT(string_idx.index_, GetDexFile()->NumStringIds());
81 const uint32_t slot_idx = string_idx.index_ % kDexCacheStringCacheSize;
82 DCHECK_LT(slot_idx, NumStrings());
83 return slot_idx;
84 }
85
GetResolvedString(dex::StringIndex string_idx)86 inline String* DexCache::GetResolvedString(dex::StringIndex string_idx) {
87 const uint32_t num_preresolved_strings = NumPreResolvedStrings();
88 if (num_preresolved_strings != 0u) {
89 GcRoot<mirror::String>* preresolved_strings = GetPreResolvedStrings();
90 // num_preresolved_strings can become 0 and preresolved_strings can become null in any order
91 // when ClearPreResolvedStrings is called.
92 if (preresolved_strings != nullptr) {
93 DCHECK_LT(string_idx.index_, num_preresolved_strings);
94 DCHECK_EQ(num_preresolved_strings, GetDexFile()->NumStringIds());
95 mirror::String* string = preresolved_strings[string_idx.index_].Read();
96 if (LIKELY(string != nullptr)) {
97 return string;
98 }
99 }
100 }
101 return GetStrings()[StringSlotIndex(string_idx)].load(
102 std::memory_order_relaxed).GetObjectForIndex(string_idx.index_);
103 }
104
SetResolvedString(dex::StringIndex string_idx,ObjPtr<String> resolved)105 inline void DexCache::SetResolvedString(dex::StringIndex string_idx, ObjPtr<String> resolved) {
106 DCHECK(resolved != nullptr);
107 GetStrings()[StringSlotIndex(string_idx)].store(
108 StringDexCachePair(resolved, string_idx.index_), std::memory_order_relaxed);
109 Runtime* const runtime = Runtime::Current();
110 if (UNLIKELY(runtime->IsActiveTransaction())) {
111 DCHECK(runtime->IsAotCompiler());
112 runtime->RecordResolveString(this, string_idx);
113 }
114 // TODO: Fine-grained marking, so that we don't need to go through all arrays in full.
115 WriteBarrier::ForEveryFieldWrite(this);
116 }
117
SetPreResolvedString(dex::StringIndex string_idx,ObjPtr<String> resolved)118 inline void DexCache::SetPreResolvedString(dex::StringIndex string_idx, ObjPtr<String> resolved) {
119 DCHECK(resolved != nullptr);
120 DCHECK_LT(string_idx.index_, GetDexFile()->NumStringIds());
121 GetPreResolvedStrings()[string_idx.index_] = GcRoot<mirror::String>(resolved);
122 Runtime* const runtime = Runtime::Current();
123 CHECK(runtime->IsAotCompiler());
124 CHECK(!runtime->IsActiveTransaction());
125 // TODO: Fine-grained marking, so that we don't need to go through all arrays in full.
126 WriteBarrier::ForEveryFieldWrite(this);
127 }
128
ClearPreResolvedStrings()129 inline void DexCache::ClearPreResolvedStrings() {
130 SetFieldPtr64</*kTransactionActive=*/false,
131 /*kCheckTransaction=*/false,
132 kVerifyNone,
133 GcRoot<mirror::String>*>(PreResolvedStringsOffset(), nullptr);
134 SetField32</*kTransactionActive=*/false,
135 /*bool kCheckTransaction=*/false,
136 kVerifyNone,
137 /*kIsVolatile=*/false>(NumPreResolvedStringsOffset(), 0);
138 }
139
ClearString(dex::StringIndex string_idx)140 inline void DexCache::ClearString(dex::StringIndex string_idx) {
141 DCHECK(Runtime::Current()->IsAotCompiler());
142 uint32_t slot_idx = StringSlotIndex(string_idx);
143 StringDexCacheType* slot = &GetStrings()[slot_idx];
144 // This is racy but should only be called from the transactional interpreter.
145 if (slot->load(std::memory_order_relaxed).index == string_idx.index_) {
146 StringDexCachePair cleared(nullptr, StringDexCachePair::InvalidIndexForSlot(slot_idx));
147 slot->store(cleared, std::memory_order_relaxed);
148 }
149 }
150
TypeSlotIndex(dex::TypeIndex type_idx)151 inline uint32_t DexCache::TypeSlotIndex(dex::TypeIndex type_idx) {
152 DCHECK_LT(type_idx.index_, GetDexFile()->NumTypeIds());
153 const uint32_t slot_idx = type_idx.index_ % kDexCacheTypeCacheSize;
154 DCHECK_LT(slot_idx, NumResolvedTypes());
155 return slot_idx;
156 }
157
GetResolvedType(dex::TypeIndex type_idx)158 inline Class* DexCache::GetResolvedType(dex::TypeIndex type_idx) {
159 // It is theorized that a load acquire is not required since obtaining the resolved class will
160 // always have an address dependency or a lock.
161 return GetResolvedTypes()[TypeSlotIndex(type_idx)].load(
162 std::memory_order_relaxed).GetObjectForIndex(type_idx.index_);
163 }
164
SetResolvedType(dex::TypeIndex type_idx,ObjPtr<Class> resolved)165 inline void DexCache::SetResolvedType(dex::TypeIndex type_idx, ObjPtr<Class> resolved) {
166 DCHECK(resolved != nullptr);
167 // TODO default transaction support.
168 // Use a release store for SetResolvedType. This is done to prevent other threads from seeing a
169 // class but not necessarily seeing the loaded members like the static fields array.
170 // See b/32075261.
171 GetResolvedTypes()[TypeSlotIndex(type_idx)].store(
172 TypeDexCachePair(resolved, type_idx.index_), std::memory_order_release);
173 // TODO: Fine-grained marking, so that we don't need to go through all arrays in full.
174 WriteBarrier::ForEveryFieldWrite(this);
175 }
176
ClearResolvedType(dex::TypeIndex type_idx)177 inline void DexCache::ClearResolvedType(dex::TypeIndex type_idx) {
178 DCHECK(Runtime::Current()->IsAotCompiler());
179 uint32_t slot_idx = TypeSlotIndex(type_idx);
180 TypeDexCacheType* slot = &GetResolvedTypes()[slot_idx];
181 // This is racy but should only be called from the single-threaded ImageWriter and tests.
182 if (slot->load(std::memory_order_relaxed).index == type_idx.index_) {
183 TypeDexCachePair cleared(nullptr, TypeDexCachePair::InvalidIndexForSlot(slot_idx));
184 slot->store(cleared, std::memory_order_relaxed);
185 }
186 }
187
MethodTypeSlotIndex(dex::ProtoIndex proto_idx)188 inline uint32_t DexCache::MethodTypeSlotIndex(dex::ProtoIndex proto_idx) {
189 DCHECK(Runtime::Current()->IsMethodHandlesEnabled());
190 DCHECK_LT(proto_idx.index_, GetDexFile()->NumProtoIds());
191 const uint32_t slot_idx = proto_idx.index_ % kDexCacheMethodTypeCacheSize;
192 DCHECK_LT(slot_idx, NumResolvedMethodTypes());
193 return slot_idx;
194 }
195
GetResolvedMethodType(dex::ProtoIndex proto_idx)196 inline MethodType* DexCache::GetResolvedMethodType(dex::ProtoIndex proto_idx) {
197 return GetResolvedMethodTypes()[MethodTypeSlotIndex(proto_idx)].load(
198 std::memory_order_relaxed).GetObjectForIndex(proto_idx.index_);
199 }
200
SetResolvedMethodType(dex::ProtoIndex proto_idx,MethodType * resolved)201 inline void DexCache::SetResolvedMethodType(dex::ProtoIndex proto_idx, MethodType* resolved) {
202 DCHECK(resolved != nullptr);
203 GetResolvedMethodTypes()[MethodTypeSlotIndex(proto_idx)].store(
204 MethodTypeDexCachePair(resolved, proto_idx.index_), std::memory_order_relaxed);
205 // TODO: Fine-grained marking, so that we don't need to go through all arrays in full.
206 WriteBarrier::ForEveryFieldWrite(this);
207 }
208
GetResolvedCallSite(uint32_t call_site_idx)209 inline CallSite* DexCache::GetResolvedCallSite(uint32_t call_site_idx) {
210 DCHECK(Runtime::Current()->IsMethodHandlesEnabled());
211 DCHECK_LT(call_site_idx, GetDexFile()->NumCallSiteIds());
212 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx];
213 Atomic<GcRoot<mirror::CallSite>>& ref =
214 reinterpret_cast<Atomic<GcRoot<mirror::CallSite>>&>(target);
215 return ref.load(std::memory_order_seq_cst).Read();
216 }
217
SetResolvedCallSite(uint32_t call_site_idx,ObjPtr<CallSite> call_site)218 inline ObjPtr<CallSite> DexCache::SetResolvedCallSite(uint32_t call_site_idx,
219 ObjPtr<CallSite> call_site) {
220 DCHECK(Runtime::Current()->IsMethodHandlesEnabled());
221 DCHECK_LT(call_site_idx, GetDexFile()->NumCallSiteIds());
222
223 GcRoot<mirror::CallSite> null_call_site(nullptr);
224 GcRoot<mirror::CallSite> candidate(call_site);
225 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx];
226
227 // The first assignment for a given call site wins.
228 Atomic<GcRoot<mirror::CallSite>>& ref =
229 reinterpret_cast<Atomic<GcRoot<mirror::CallSite>>&>(target);
230 if (ref.CompareAndSetStrongSequentiallyConsistent(null_call_site, candidate)) {
231 // TODO: Fine-grained marking, so that we don't need to go through all arrays in full.
232 WriteBarrier::ForEveryFieldWrite(this);
233 return call_site;
234 } else {
235 return target.Read();
236 }
237 }
238
FieldSlotIndex(uint32_t field_idx)239 inline uint32_t DexCache::FieldSlotIndex(uint32_t field_idx) {
240 DCHECK_LT(field_idx, GetDexFile()->NumFieldIds());
241 const uint32_t slot_idx = field_idx % kDexCacheFieldCacheSize;
242 DCHECK_LT(slot_idx, NumResolvedFields());
243 return slot_idx;
244 }
245
GetResolvedField(uint32_t field_idx,PointerSize ptr_size)246 inline ArtField* DexCache::GetResolvedField(uint32_t field_idx, PointerSize ptr_size) {
247 DCHECK_EQ(Runtime::Current()->GetClassLinker()->GetImagePointerSize(), ptr_size);
248 auto pair = GetNativePairPtrSize(GetResolvedFields(), FieldSlotIndex(field_idx), ptr_size);
249 return pair.GetObjectForIndex(field_idx);
250 }
251
SetResolvedField(uint32_t field_idx,ArtField * field,PointerSize ptr_size)252 inline void DexCache::SetResolvedField(uint32_t field_idx, ArtField* field, PointerSize ptr_size) {
253 DCHECK_EQ(Runtime::Current()->GetClassLinker()->GetImagePointerSize(), ptr_size);
254 DCHECK(field != nullptr);
255 FieldDexCachePair pair(field, field_idx);
256 SetNativePairPtrSize(GetResolvedFields(), FieldSlotIndex(field_idx), pair, ptr_size);
257 }
258
ClearResolvedField(uint32_t field_idx,PointerSize ptr_size)259 inline void DexCache::ClearResolvedField(uint32_t field_idx, PointerSize ptr_size) {
260 DCHECK_EQ(Runtime::Current()->GetClassLinker()->GetImagePointerSize(), ptr_size);
261 uint32_t slot_idx = FieldSlotIndex(field_idx);
262 auto* resolved_fields = GetResolvedFields();
263 // This is racy but should only be called from the single-threaded ImageWriter.
264 DCHECK(Runtime::Current()->IsAotCompiler());
265 if (GetNativePairPtrSize(resolved_fields, slot_idx, ptr_size).index == field_idx) {
266 FieldDexCachePair cleared(nullptr, FieldDexCachePair::InvalidIndexForSlot(slot_idx));
267 SetNativePairPtrSize(resolved_fields, slot_idx, cleared, ptr_size);
268 }
269 }
270
MethodSlotIndex(uint32_t method_idx)271 inline uint32_t DexCache::MethodSlotIndex(uint32_t method_idx) {
272 DCHECK_LT(method_idx, GetDexFile()->NumMethodIds());
273 const uint32_t slot_idx = method_idx % kDexCacheMethodCacheSize;
274 DCHECK_LT(slot_idx, NumResolvedMethods());
275 return slot_idx;
276 }
277
GetResolvedMethod(uint32_t method_idx,PointerSize ptr_size)278 inline ArtMethod* DexCache::GetResolvedMethod(uint32_t method_idx, PointerSize ptr_size) {
279 DCHECK_EQ(Runtime::Current()->GetClassLinker()->GetImagePointerSize(), ptr_size);
280 auto pair = GetNativePairPtrSize(GetResolvedMethods(), MethodSlotIndex(method_idx), ptr_size);
281 return pair.GetObjectForIndex(method_idx);
282 }
283
SetResolvedMethod(uint32_t method_idx,ArtMethod * method,PointerSize ptr_size)284 inline void DexCache::SetResolvedMethod(uint32_t method_idx,
285 ArtMethod* method,
286 PointerSize ptr_size) {
287 DCHECK_EQ(Runtime::Current()->GetClassLinker()->GetImagePointerSize(), ptr_size);
288 DCHECK(method != nullptr);
289 MethodDexCachePair pair(method, method_idx);
290 SetNativePairPtrSize(GetResolvedMethods(), MethodSlotIndex(method_idx), pair, ptr_size);
291 }
292
ClearResolvedMethod(uint32_t method_idx,PointerSize ptr_size)293 inline void DexCache::ClearResolvedMethod(uint32_t method_idx, PointerSize ptr_size) {
294 DCHECK_EQ(Runtime::Current()->GetClassLinker()->GetImagePointerSize(), ptr_size);
295 uint32_t slot_idx = MethodSlotIndex(method_idx);
296 auto* resolved_methods = GetResolvedMethods();
297 // This is racy but should only be called from the single-threaded ImageWriter.
298 DCHECK(Runtime::Current()->IsAotCompiler());
299 if (GetNativePairPtrSize(resolved_methods, slot_idx, ptr_size).index == method_idx) {
300 MethodDexCachePair cleared(nullptr, MethodDexCachePair::InvalidIndexForSlot(slot_idx));
301 SetNativePairPtrSize(resolved_methods, slot_idx, cleared, ptr_size);
302 }
303 }
304
305 template <typename T>
GetNativePairPtrSize(std::atomic<NativeDexCachePair<T>> * pair_array,size_t idx,PointerSize ptr_size)306 NativeDexCachePair<T> DexCache::GetNativePairPtrSize(std::atomic<NativeDexCachePair<T>>* pair_array,
307 size_t idx,
308 PointerSize ptr_size) {
309 if (ptr_size == PointerSize::k64) {
310 auto* array = reinterpret_cast<std::atomic<ConversionPair64>*>(pair_array);
311 ConversionPair64 value = AtomicLoadRelaxed16B(&array[idx]);
312 return NativeDexCachePair<T>(reinterpret_cast64<T*>(value.first),
313 dchecked_integral_cast<size_t>(value.second));
314 } else {
315 auto* array = reinterpret_cast<std::atomic<ConversionPair32>*>(pair_array);
316 ConversionPair32 value = array[idx].load(std::memory_order_relaxed);
317 return NativeDexCachePair<T>(reinterpret_cast32<T*>(value.first), value.second);
318 }
319 }
320
321 template <typename T>
SetNativePairPtrSize(std::atomic<NativeDexCachePair<T>> * pair_array,size_t idx,NativeDexCachePair<T> pair,PointerSize ptr_size)322 void DexCache::SetNativePairPtrSize(std::atomic<NativeDexCachePair<T>>* pair_array,
323 size_t idx,
324 NativeDexCachePair<T> pair,
325 PointerSize ptr_size) {
326 if (ptr_size == PointerSize::k64) {
327 auto* array = reinterpret_cast<std::atomic<ConversionPair64>*>(pair_array);
328 ConversionPair64 v(reinterpret_cast64<uint64_t>(pair.object), pair.index);
329 AtomicStoreRelease16B(&array[idx], v);
330 } else {
331 auto* array = reinterpret_cast<std::atomic<ConversionPair32>*>(pair_array);
332 ConversionPair32 v(reinterpret_cast32<uint32_t>(pair.object),
333 dchecked_integral_cast<uint32_t>(pair.index));
334 array[idx].store(v, std::memory_order_release);
335 }
336 }
337
338 template <typename T,
339 ReadBarrierOption kReadBarrierOption,
340 typename Visitor>
VisitDexCachePairs(std::atomic<DexCachePair<T>> * pairs,size_t num_pairs,const Visitor & visitor)341 inline void VisitDexCachePairs(std::atomic<DexCachePair<T>>* pairs,
342 size_t num_pairs,
343 const Visitor& visitor)
344 REQUIRES_SHARED(Locks::mutator_lock_) REQUIRES(Locks::heap_bitmap_lock_) {
345 for (size_t i = 0; i < num_pairs; ++i) {
346 DexCachePair<T> source = pairs[i].load(std::memory_order_relaxed);
347 // NOTE: We need the "template" keyword here to avoid a compilation
348 // failure. GcRoot<T> is a template argument-dependent type and we need to
349 // tell the compiler to treat "Read" as a template rather than a field or
350 // function. Otherwise, on encountering the "<" token, the compiler would
351 // treat "Read" as a field.
352 T* const before = source.object.template Read<kReadBarrierOption>();
353 visitor.VisitRootIfNonNull(source.object.AddressWithoutBarrier());
354 if (source.object.template Read<kReadBarrierOption>() != before) {
355 pairs[i].store(source, std::memory_order_relaxed);
356 }
357 }
358 }
359
360 template <bool kVisitNativeRoots,
361 VerifyObjectFlags kVerifyFlags,
362 ReadBarrierOption kReadBarrierOption,
363 typename Visitor>
VisitReferences(ObjPtr<Class> klass,const Visitor & visitor)364 inline void DexCache::VisitReferences(ObjPtr<Class> klass, const Visitor& visitor) {
365 // Visit instance fields first.
366 VisitInstanceFieldsReferences<kVerifyFlags, kReadBarrierOption>(klass, visitor);
367 // Visit arrays after.
368 if (kVisitNativeRoots) {
369 VisitDexCachePairs<String, kReadBarrierOption, Visitor>(
370 GetStrings<kVerifyFlags>(), NumStrings<kVerifyFlags>(), visitor);
371
372 VisitDexCachePairs<Class, kReadBarrierOption, Visitor>(
373 GetResolvedTypes<kVerifyFlags>(), NumResolvedTypes<kVerifyFlags>(), visitor);
374
375 VisitDexCachePairs<MethodType, kReadBarrierOption, Visitor>(
376 GetResolvedMethodTypes<kVerifyFlags>(), NumResolvedMethodTypes<kVerifyFlags>(), visitor);
377
378 GcRoot<mirror::CallSite>* resolved_call_sites = GetResolvedCallSites<kVerifyFlags>();
379 size_t num_call_sites = NumResolvedCallSites<kVerifyFlags>();
380 for (size_t i = 0; i != num_call_sites; ++i) {
381 visitor.VisitRootIfNonNull(resolved_call_sites[i].AddressWithoutBarrier());
382 }
383
384 GcRoot<mirror::String>* const preresolved_strings = GetPreResolvedStrings();
385 const size_t num_preresolved_strings = NumPreResolvedStrings();
386 for (size_t i = 0; i != num_preresolved_strings; ++i) {
387 visitor.VisitRootIfNonNull(preresolved_strings[i].AddressWithoutBarrier());
388 }
389 }
390 }
391
392 template <ReadBarrierOption kReadBarrierOption, typename Visitor>
FixupStrings(StringDexCacheType * dest,const Visitor & visitor)393 inline void DexCache::FixupStrings(StringDexCacheType* dest, const Visitor& visitor) {
394 StringDexCacheType* src = GetStrings();
395 for (size_t i = 0, count = NumStrings(); i < count; ++i) {
396 StringDexCachePair source = src[i].load(std::memory_order_relaxed);
397 String* ptr = source.object.Read<kReadBarrierOption>();
398 String* new_source = visitor(ptr);
399 source.object = GcRoot<String>(new_source);
400 dest[i].store(source, std::memory_order_relaxed);
401 }
402 }
403
404 template <ReadBarrierOption kReadBarrierOption, typename Visitor>
FixupResolvedTypes(TypeDexCacheType * dest,const Visitor & visitor)405 inline void DexCache::FixupResolvedTypes(TypeDexCacheType* dest, const Visitor& visitor) {
406 TypeDexCacheType* src = GetResolvedTypes();
407 for (size_t i = 0, count = NumResolvedTypes(); i < count; ++i) {
408 TypeDexCachePair source = src[i].load(std::memory_order_relaxed);
409 Class* ptr = source.object.Read<kReadBarrierOption>();
410 Class* new_source = visitor(ptr);
411 source.object = GcRoot<Class>(new_source);
412 dest[i].store(source, std::memory_order_relaxed);
413 }
414 }
415
416 template <ReadBarrierOption kReadBarrierOption, typename Visitor>
FixupResolvedMethodTypes(MethodTypeDexCacheType * dest,const Visitor & visitor)417 inline void DexCache::FixupResolvedMethodTypes(MethodTypeDexCacheType* dest,
418 const Visitor& visitor) {
419 MethodTypeDexCacheType* src = GetResolvedMethodTypes();
420 for (size_t i = 0, count = NumResolvedMethodTypes(); i < count; ++i) {
421 MethodTypeDexCachePair source = src[i].load(std::memory_order_relaxed);
422 MethodType* ptr = source.object.Read<kReadBarrierOption>();
423 MethodType* new_source = visitor(ptr);
424 source.object = GcRoot<MethodType>(new_source);
425 dest[i].store(source, std::memory_order_relaxed);
426 }
427 }
428
429 template <ReadBarrierOption kReadBarrierOption, typename Visitor>
FixupResolvedCallSites(GcRoot<mirror::CallSite> * dest,const Visitor & visitor)430 inline void DexCache::FixupResolvedCallSites(GcRoot<mirror::CallSite>* dest,
431 const Visitor& visitor) {
432 GcRoot<mirror::CallSite>* src = GetResolvedCallSites();
433 for (size_t i = 0, count = NumResolvedCallSites(); i < count; ++i) {
434 mirror::CallSite* source = src[i].Read<kReadBarrierOption>();
435 mirror::CallSite* new_source = visitor(source);
436 dest[i] = GcRoot<mirror::CallSite>(new_source);
437 }
438 }
439
GetLocation()440 inline ObjPtr<String> DexCache::GetLocation() {
441 return GetFieldObject<String>(OFFSET_OF_OBJECT_MEMBER(DexCache, location_));
442 }
443
444 } // namespace mirror
445 } // namespace art
446
447 #endif // ART_RUNTIME_MIRROR_DEX_CACHE_INL_H_
448