Home
last modified time | relevance | path

Searched refs:StackReference (Results 1 – 25 of 30) sorted by relevance

12

/art/runtime/
Dhandle.h63 ALWAYS_INLINE explicit Handle(StackReference<T>* reference) : reference_(reference) { in Handle()
96 ALWAYS_INLINE StackReference<mirror::Object>* GetReference() { in GetReference()
100 ALWAYS_INLINE const StackReference<mirror::Object>* GetReference() const { in GetReference()
117 explicit Handle(StackReference<S>* reference) in Handle()
125 StackReference<mirror::Object>* reference_;
148 ALWAYS_INLINE explicit MutableHandle(StackReference<T>* reference) in MutableHandle()
154 StackReference<mirror::Object>* ref = Handle<T>::GetReference(); in Assign()
161 StackReference<mirror::Object>* ref = Handle<T>::GetReference(); in Assign()
174 explicit MutableHandle(StackReference<S>* reference) REQUIRES_SHARED(Locks::mutator_lock_) in MutableHandle()
193 StackReference<mirror::Object> null_ref_;
Dhandle_scope_test.cc72 auto* ref_ptr = reinterpret_cast<StackReference<mirror::Object>*>(table_base_ptr + in TEST_F()
80 void VisitRootIfNonNull(StackReference<mirror::Object>* ref) { in VisitRootIfNonNull()
87 std::set<StackReference<mirror::Object>*> visited;
118 for (StackReference<mirror::Object>* ref : visitor.visited) { in TEST_F()
Dhandle_scope.h54 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const;
119 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const;
153 ALWAYS_INLINE StackReference<mirror::Object>* GetReferences() const { in GetReferences()
155 return reinterpret_cast<StackReference<mirror::Object>*>(address); in GetReferences()
210 StackReference<mirror::Object> storage_[kNumReferences];
259 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const;
272 kSizeOfReferencesPerScope / sizeof(StackReference<mirror::Object>);
Dstack_reference.h27 class PACKED(4) StackReference : public mirror::CompressedReference<MirrorType> {
Dhandle_scope-inl.h69 size_t data_size = sizeof(StackReference<mirror::Object>) * num_references; in SizeOf()
76 size_t data_size = sizeof(StackReference<mirror::Object>) * num_references; in SizeOf()
106 inline bool HandleScope::Contains(StackReference<mirror::Object>* handle_scope_entry) const { in Contains()
167 inline bool BaseHandleScope::Contains(StackReference<mirror::Object>* handle_scope_entry) const { in Contains()
251 inline bool VariableSizedHandleScope::Contains(StackReference<mirror::Object>* handle_scope_entry) in Contains()
Dthread-inl.h365 sizeof(StackReference<mirror::Object>), in PushOnThreadLocalAllocationStack()
398 inline void Thread::SetThreadLocalAllocationStack(StackReference<mirror::Object>* start, in SetThreadLocalAllocationStack()
399 StackReference<mirror::Object>* end) { in SetThreadLocalAllocationStack()
403 DCHECK_ALIGNED(start, sizeof(StackReference<mirror::Object>)); in SetThreadLocalAllocationStack()
404 DCHECK_ALIGNED(end, sizeof(StackReference<mirror::Object>)); in SetThreadLocalAllocationStack()
Dmanaged_stack.h38 template <typename T> class StackReference; variable
Dquick_exception_handler.cc430 StackReference<mirror::Object>* vreg_ref_base = in HandleNterpDeoptimization()
431 reinterpret_cast<StackReference<mirror::Object>*>(NterpGetReferenceArray(cur_quick_frame)); in HandleNterpDeoptimization()
444 StackReference<mirror::Object>* ref_addr = vreg_ref_base + reg; in HandleNterpDeoptimization()
Dthread.cc3842 extern std::vector<StackReference<mirror::Object>*> GetProxyReferenceArguments(ArtMethod** sp)
3936 StackReference<mirror::Object>* vreg_ref_base = in VisitNterpFrame()
3937 reinterpret_cast<StackReference<mirror::Object>*>(NterpGetReferenceArray(cur_quick_frame)); in VisitNterpFrame()
3938 StackReference<mirror::Object>* vreg_int_base = in VisitNterpFrame()
3939 reinterpret_cast<StackReference<mirror::Object>*>(NterpGetRegistersArray(cur_quick_frame)); in VisitNterpFrame()
3946 StackReference<mirror::Object>* ref_addr = vreg_ref_base + reg; in VisitNterpFrame()
3953 StackReference<mirror::Object>* int_addr = vreg_int_base + reg; in VisitNterpFrame()
3994 auto* ref_addr = reinterpret_cast<StackReference<mirror::Object>*>(current_vreg); in VisitQuickFrameWithVregCallback()
4027 StackReference<mirror::Object>* vreg_base = in VisitQuickFrameWithVregCallback()
4028 reinterpret_cast<StackReference<mirror::Object>*>(cur_quick_frame); in VisitQuickFrameWithVregCallback()
[all …]
Dthread.h1246 void SetThreadLocalAllocationStack(StackReference<mirror::Object>* start,
1247 StackReference<mirror::Object>* end);
2015 StackReference<mirror::Object>* thread_local_alloc_stack_top; in PACKED()
2016 StackReference<mirror::Object>* thread_local_alloc_stack_end; in PACKED()
Dreflection.cc83 Append(StackReference<mirror::Object>::FromMirrorPtr(obj.Ptr()).AsVRegValue()); in Append()
398 (reinterpret_cast<StackReference<mirror::Object>*>(&args[i + offset]))->AsMirrorPtr(); in CheckMethodArguments()
Dart_method.cc342 reinterpret_cast<StackReference<mirror::Object>*>(&args[0])->AsMirrorPtr(); in Invoke()
/art/runtime/gc/accounting/
Datomic_stack.h52 bool operator()(const T* a, const StackReference<T>& b) const NO_THREAD_SAFETY_ANALYSIS { in operator()
55 bool operator()(const StackReference<T>& a, const T* b) const NO_THREAD_SAFETY_ANALYSIS { in operator()
59 bool operator()(const StackReference<T>& a, const StackReference<T>& b) const in operator()
97 bool AtomicBumpBack(size_t num_slots, StackReference<T>** start_address, in AtomicBumpBack()
98 StackReference<T>** end_address) in AtomicBumpBack()
180 StackReference<T>* Begin() const { in Begin()
183 StackReference<T>* End() const { in End()
264 begin_ = reinterpret_cast<StackReference<T>*>(addr); in Init()
277 StackReference<T>* begin_;
/art/runtime/interpreter/
Dshadow_frame.h71 (sizeof(StackReference<mirror::Object>) * num_vregs); in ComputeSize()
262 bool Contains(StackReference<mirror::Object>* shadow_frame_entry_obj) const { in Contains()
399 memset(vregs_, 0, num_vregs * (sizeof(uint32_t) + sizeof(StackReference<mirror::Object>))); in ShadowFrame()
414 const StackReference<mirror::Object>* References() const { in References()
416 return reinterpret_cast<const StackReference<mirror::Object>*>(vreg_end); in References()
419 StackReference<mirror::Object>* References() { in References()
420 return const_cast<StackReference<mirror::Object>*>( in References()
Dshadow_frame-inl.h35 reinterpret_cast<StackReference<mirror::Object>*>(vreg)->Assign(val); in SetVRegReference()
Dinterpreter.cc399 reinterpret_cast<StackReference<mirror::Object>*>(&args[arg_pos])->AsMirrorPtr(); in EnterInterpreterFromInvoke()
/art/test/1945-proxy-method-arguments/
Dget_args.cc83 extern "C" StackReference<mirror::Object>* artQuickGetProxyReferenceArgumentAt(size_t arg_pos,
98 StackReference<mirror::Object>* ref_arg = in GetProxyReferenceArgument()
/art/runtime/entrypoints/quick/
Dquick_trampoline_entrypoints.cc231 static StackReference<mirror::Object>* GetProxyThisObjectReference(ArtMethod** sp) in GetProxyThisObjectReference()
239 return reinterpret_cast<StackReference<mirror::Object>*>(this_arg_address); in GetProxyThisObjectReference()
564 StackReference<mirror::Object>* stack_ref = in Visit()
565 reinterpret_cast<StackReference<mirror::Object>*>(GetParamAddress()); in Visit()
779 StackReference<mirror::Object>* stack_ref = in Visit()
780 reinterpret_cast<StackReference<mirror::Object>*>(GetParamAddress()); in Visit()
907 ref_arg_ = reinterpret_cast<StackReference<mirror::Object>*>(GetParamAddress()); in Visit()
912 StackReference<mirror::Object>* GetReferenceArgument() { in GetReferenceArgument()
922 StackReference<mirror::Object>* ref_arg_;
929 extern "C" StackReference<mirror::Object>* artQuickGetProxyReferenceArgumentAt(size_t arg_pos, in artQuickGetProxyReferenceArgumentAt()
[all …]
Dquick_entrypoints.h42 template<class MirrorType> class StackReference; variable
/art/runtime/gc/collector/
Dmark_sweep.cc391 std::vector<StackReference<mirror::Object>> temp(mark_stack_->Begin(), mark_stack_->End()); in ResizeMarkStack()
674 StackReference<mirror::Object>* mark_stack) in MarkStackTask()
769 StackReference<mirror::Object> mark_stack_[kMaxSize];
836 StackReference<mirror::Object>* mark_stack_obj, in CardScanTask()
890 StackReference<mirror::Object>* mark_stack_begin = mark_stack_->Begin(); in ScanGrayObjects()
891 StackReference<mirror::Object>* mark_stack_end = mark_stack_->End(); in ScanGrayObjects()
1216 StackReference<mirror::Object>* objects = allocations->Begin(); in SweepArray()
1245 StackReference<mirror::Object>* out = objects; in SweepArray()
Dsemi_space.cc362 std::vector<StackReference<Object>> temp(mark_stack_->Begin(), mark_stack_->End()); in ResizeMarkStack()
464 auto ref = StackReference<mirror::Object>::FromMirrorPtr(root); in MarkObject()
478 auto ref = StackReference<mirror::Object>::FromMirrorPtr(*root); in VisitRoots()
Dconcurrent_copying.cc1824 std::vector<StackReference<mirror::Object>> temp(gc_mark_stack_->Begin(), in ExpandGcMarkStack()
2171 for (StackReference<mirror::Object>* p = gc_mark_stack_->Begin(); in ProcessMarkStackOnce()
2222 for (StackReference<mirror::Object>* p = mark_stack->Begin(); p != mark_stack->End(); ++p) { in ProcessThreadLocalMarkStacks()
2538 StackReference<mirror::Object>* objects = allocations->Begin(); in SweepArray()
2554 StackReference<mirror::Object>* out = objects; in SweepArray()
3787 for (StackReference<mirror::Object>* it = rb_mark_bit_stack_->Begin(); it != limit; ++it) { in FinishPhase()
/art/tools/cpp-define-generator/
Dglobals.def80 sizeof(art::StackReference<art::mirror::Object>))
/art/runtime/entrypoints/
Dentrypoint_utils-inl.h752 StackReference<mirror::Object>* this_ref = reinterpret_cast<StackReference<mirror::Object>*>( in GetGenericJniSynchronizationObject()
/art/compiler/optimizing/
Dcode_generator.cc235 roots->emplace_back(reinterpret_cast<StackReference<mirror::Object>*>(address)); in EmitJitRoots()
248 roots->emplace_back(reinterpret_cast<StackReference<mirror::Object>*>(address)); in EmitJitRoots()
670 static_assert(sizeof(StackReference<mirror::Object>) == sizeof(uint32_t), "Size check."); in CreateStringBuilderAppendLocations()

12