• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (C) 2011 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #ifndef ART_RUNTIME_THREAD_H_
18 #define ART_RUNTIME_THREAD_H_
19 
20 #include <bitset>
21 #include <deque>
22 #include <iosfwd>
23 #include <list>
24 #include <memory>
25 #include <setjmp.h>
26 #include <string>
27 
28 #include "atomic.h"
29 #include "base/macros.h"
30 #include "base/mutex.h"
31 #include "entrypoints/interpreter/interpreter_entrypoints.h"
32 #include "entrypoints/jni/jni_entrypoints.h"
33 #include "entrypoints/portable/portable_entrypoints.h"
34 #include "entrypoints/quick/quick_entrypoints.h"
35 #include "globals.h"
36 #include "handle_scope.h"
37 #include "instruction_set.h"
38 #include "jvalue.h"
39 #include "object_callbacks.h"
40 #include "offsets.h"
41 #include "runtime_stats.h"
42 #include "stack.h"
43 #include "thread_state.h"
44 #include "throw_location.h"
45 
46 namespace art {
47 
48 namespace gc {
49 namespace collector {
50   class SemiSpace;
51 }  // namespace collector
52 }  // namespace gc
53 
54 namespace mirror {
55   class ArtMethod;
56   class Array;
57   class Class;
58   class ClassLoader;
59   class Object;
60   template<class T> class ObjectArray;
61   template<class T> class PrimitiveArray;
62   typedef PrimitiveArray<int32_t> IntArray;
63   class StackTraceElement;
64   class Throwable;
65 }  // namespace mirror
66 class BaseMutex;
67 class ClassLinker;
68 class Closure;
69 class Context;
70 struct DebugInvokeReq;
71 class DexFile;
72 class JavaVMExt;
73 struct JNIEnvExt;
74 class Monitor;
75 class Runtime;
76 class ScopedObjectAccessAlreadyRunnable;
77 class ShadowFrame;
78 struct SingleStepControl;
79 class Thread;
80 class ThreadList;
81 
82 // Thread priorities. These must match the Thread.MIN_PRIORITY,
83 // Thread.NORM_PRIORITY, and Thread.MAX_PRIORITY constants.
84 enum ThreadPriority {
85   kMinThreadPriority = 1,
86   kNormThreadPriority = 5,
87   kMaxThreadPriority = 10,
88 };
89 
90 enum ThreadFlag {
91   kSuspendRequest   = 1,  // If set implies that suspend_count_ > 0 and the Thread should enter the
92                           // safepoint handler.
93   kCheckpointRequest = 2  // Request that the thread do some checkpoint work and then continue.
94 };
95 
96 static constexpr size_t kNumRosAllocThreadLocalSizeBrackets = 34;
97 
98 // Thread's stack layout for implicit stack overflow checks:
99 //
100 //   +---------------------+  <- highest address of stack memory
101 //   |                     |
102 //   .                     .  <- SP
103 //   |                     |
104 //   |                     |
105 //   +---------------------+  <- stack_end
106 //   |                     |
107 //   |  Gap                |
108 //   |                     |
109 //   +---------------------+  <- stack_begin
110 //   |                     |
111 //   | Protected region    |
112 //   |                     |
113 //   +---------------------+  <- lowest address of stack memory
114 //
115 // The stack always grows down in memory.  At the lowest address is a region of memory
116 // that is set mprotect(PROT_NONE).  Any attempt to read/write to this region will
117 // result in a segmentation fault signal.  At any point, the thread's SP will be somewhere
118 // between the stack_end and the highest address in stack memory.  An implicit stack
119 // overflow check is a read of memory at a certain offset below the current SP (4K typically).
120 // If the thread's SP is below the stack_end address this will be a read into the protected
121 // region.  If the SP is above the stack_end address, the thread is guaranteed to have
122 // at least 4K of space.  Because stack overflow checks are only performed in generated code,
123 // if the thread makes a call out to a native function (through JNI), that native function
124 // might only have 4K of memory (if the SP is adjacent to stack_end).
125 
126 class Thread {
127  public:
128   // For implicit overflow checks we reserve an extra piece of memory at the bottom
129   // of the stack (lowest memory).  The higher portion of the memory
130   // is protected against reads and the lower is available for use while
131   // throwing the StackOverflow exception.
132   static constexpr size_t kStackOverflowProtectedSize = 4 * KB;
133   static const size_t kStackOverflowImplicitCheckSize;
134 
135   // Creates a new native thread corresponding to the given managed peer.
136   // Used to implement Thread.start.
137   static void CreateNativeThread(JNIEnv* env, jobject peer, size_t stack_size, bool daemon);
138 
139   // Attaches the calling native thread to the runtime, returning the new native peer.
140   // Used to implement JNI AttachCurrentThread and AttachCurrentThreadAsDaemon calls.
141   static Thread* Attach(const char* thread_name, bool as_daemon, jobject thread_group,
142                         bool create_peer);
143 
144   // Reset internal state of child thread after fork.
145   void InitAfterFork();
146 
147   static Thread* Current();
148 
149   static Thread* FromManagedThread(const ScopedObjectAccessAlreadyRunnable& ts,
150                                    mirror::Object* thread_peer)
151       EXCLUSIVE_LOCKS_REQUIRED(Locks::thread_list_lock_)
152       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
153       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
154   static Thread* FromManagedThread(const ScopedObjectAccessAlreadyRunnable& ts, jobject thread)
155       EXCLUSIVE_LOCKS_REQUIRED(Locks::thread_list_lock_)
156       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
157       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
158 
159   // Translates 172 to pAllocArrayFromCode and so on.
160   template<size_t size_of_pointers>
161   static void DumpThreadOffset(std::ostream& os, uint32_t offset);
162 
163   // Dumps a one-line summary of thread state (used for operator<<).
164   void ShortDump(std::ostream& os) const;
165 
166   // Dumps the detailed thread state and the thread stack (used for SIGQUIT).
167   void Dump(std::ostream& os) const
168       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
169       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
170 
171   void DumpJavaStack(std::ostream& os) const
172       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
173       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
174 
175   // Dumps the SIGQUIT per-thread header. 'thread' can be NULL for a non-attached thread, in which
176   // case we use 'tid' to identify the thread, and we'll include as much information as we can.
177   static void DumpState(std::ostream& os, const Thread* thread, pid_t tid)
178       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
179       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
180 
GetState()181   ThreadState GetState() const {
182     DCHECK_GE(tls32_.state_and_flags.as_struct.state, kTerminated);
183     DCHECK_LE(tls32_.state_and_flags.as_struct.state, kSuspended);
184     return static_cast<ThreadState>(tls32_.state_and_flags.as_struct.state);
185   }
186 
187   ThreadState SetState(ThreadState new_state);
188 
GetSuspendCount()189   int GetSuspendCount() const EXCLUSIVE_LOCKS_REQUIRED(Locks::thread_suspend_count_lock_) {
190     return tls32_.suspend_count;
191   }
192 
GetDebugSuspendCount()193   int GetDebugSuspendCount() const EXCLUSIVE_LOCKS_REQUIRED(Locks::thread_suspend_count_lock_) {
194     return tls32_.debug_suspend_count;
195   }
196 
IsSuspended()197   bool IsSuspended() const {
198     union StateAndFlags state_and_flags;
199     state_and_flags.as_int = tls32_.state_and_flags.as_int;
200     return state_and_flags.as_struct.state != kRunnable &&
201         (state_and_flags.as_struct.flags & kSuspendRequest) != 0;
202   }
203 
204   void ModifySuspendCount(Thread* self, int delta, bool for_debugger)
205       EXCLUSIVE_LOCKS_REQUIRED(Locks::thread_suspend_count_lock_);
206 
207   bool RequestCheckpoint(Closure* function)
208       EXCLUSIVE_LOCKS_REQUIRED(Locks::thread_suspend_count_lock_);
209 
210   // Called when thread detected that the thread_suspend_count_ was non-zero. Gives up share of
211   // mutator_lock_ and waits until it is resumed and thread_suspend_count_ is zero.
212   void FullSuspendCheck()
213       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
214       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
215 
216   // Transition from non-runnable to runnable state acquiring share on mutator_lock_.
217   ThreadState TransitionFromSuspendedToRunnable()
218       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
219       SHARED_LOCK_FUNCTION(Locks::mutator_lock_)
220       ALWAYS_INLINE;
221 
222   // Transition from runnable into a state where mutator privileges are denied. Releases share of
223   // mutator lock.
224   void TransitionFromRunnableToSuspended(ThreadState new_state)
225       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
226       UNLOCK_FUNCTION(Locks::mutator_lock_)
227       ALWAYS_INLINE;
228 
229   // Once called thread suspension will cause an assertion failure.
StartAssertNoThreadSuspension(const char * cause)230   const char* StartAssertNoThreadSuspension(const char* cause) {
231     if (kIsDebugBuild) {
232       CHECK(cause != NULL);
233       const char* previous_cause = tlsPtr_.last_no_thread_suspension_cause;
234       tls32_.no_thread_suspension++;
235       tlsPtr_.last_no_thread_suspension_cause = cause;
236       return previous_cause;
237     } else {
238       return nullptr;
239     }
240   }
241 
242   // End region where no thread suspension is expected.
EndAssertNoThreadSuspension(const char * old_cause)243   void EndAssertNoThreadSuspension(const char* old_cause) {
244     if (kIsDebugBuild) {
245       CHECK(old_cause != nullptr || tls32_.no_thread_suspension == 1);
246       CHECK_GT(tls32_.no_thread_suspension, 0U);
247       tls32_.no_thread_suspension--;
248       tlsPtr_.last_no_thread_suspension_cause = old_cause;
249     }
250   }
251 
252   void AssertThreadSuspensionIsAllowable(bool check_locks = true) const;
253 
IsDaemon()254   bool IsDaemon() const {
255     return tls32_.daemon;
256   }
257 
258   bool HoldsLock(mirror::Object*) const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
259 
260   /*
261    * Changes the priority of this thread to match that of the java.lang.Thread object.
262    *
263    * We map a priority value from 1-10 to Linux "nice" values, where lower
264    * numbers indicate higher priority.
265    */
266   void SetNativePriority(int newPriority);
267 
268   /*
269    * Returns the thread priority for the current thread by querying the system.
270    * This is useful when attaching a thread through JNI.
271    *
272    * Returns a value from 1 to 10 (compatible with java.lang.Thread values).
273    */
274   static int GetNativePriority();
275 
GetThreadId()276   uint32_t GetThreadId() const {
277     return tls32_.thin_lock_thread_id;
278   }
279 
GetTid()280   pid_t GetTid() const {
281     return tls32_.tid;
282   }
283 
284   // Returns the java.lang.Thread's name, or NULL if this Thread* doesn't have a peer.
285   mirror::String* GetThreadName(const ScopedObjectAccessAlreadyRunnable& ts) const
286       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
287 
288   // Sets 'name' to the java.lang.Thread's name. This requires no transition to managed code,
289   // allocation, or locking.
290   void GetThreadName(std::string& name) const;
291 
292   // Sets the thread's name.
293   void SetThreadName(const char* name) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
294 
295   // Returns the thread-specific CPU-time clock in microseconds or -1 if unavailable.
296   uint64_t GetCpuMicroTime() const;
297 
GetPeer()298   mirror::Object* GetPeer() const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
299     CHECK(tlsPtr_.jpeer == nullptr);
300     return tlsPtr_.opeer;
301   }
302 
HasPeer()303   bool HasPeer() const {
304     return tlsPtr_.jpeer != nullptr || tlsPtr_.opeer != nullptr;
305   }
306 
GetStats()307   RuntimeStats* GetStats() {
308     return &tls64_.stats;
309   }
310 
311   bool IsStillStarting() const;
312 
IsExceptionPending()313   bool IsExceptionPending() const {
314     return tlsPtr_.exception != nullptr;
315   }
316 
GetException(ThrowLocation * throw_location)317   mirror::Throwable* GetException(ThrowLocation* throw_location) const
318       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
319     if (throw_location != nullptr) {
320       *throw_location = tlsPtr_.throw_location;
321     }
322     return tlsPtr_.exception;
323   }
324 
325   void AssertNoPendingException() const;
326   void AssertNoPendingExceptionForNewException(const char* msg) const;
327 
SetException(const ThrowLocation & throw_location,mirror::Throwable * new_exception)328   void SetException(const ThrowLocation& throw_location, mirror::Throwable* new_exception)
329       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
330     CHECK(new_exception != NULL);
331     // TODO: DCHECK(!IsExceptionPending());
332     tlsPtr_.exception = new_exception;
333     tlsPtr_.throw_location = throw_location;
334   }
335 
ClearException()336   void ClearException() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
337     tlsPtr_.exception = nullptr;
338     tlsPtr_.throw_location.Clear();
339     SetExceptionReportedToInstrumentation(false);
340   }
341 
342   // Find catch block and perform long jump to appropriate exception handle
343   void QuickDeliverException() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
344 
345   Context* GetLongJumpContext();
ReleaseLongJumpContext(Context * context)346   void ReleaseLongJumpContext(Context* context) {
347     DCHECK(tlsPtr_.long_jump_context == nullptr);
348     tlsPtr_.long_jump_context = context;
349   }
350 
351   // Get the current method and dex pc. If there are errors in retrieving the dex pc, this will
352   // abort the runtime iff abort_on_error is true.
353   mirror::ArtMethod* GetCurrentMethod(uint32_t* dex_pc, bool abort_on_error = true) const
354       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
355 
356   ThrowLocation GetCurrentLocationForThrow() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
357 
SetTopOfStack(StackReference<mirror::ArtMethod> * top_method,uintptr_t pc)358   void SetTopOfStack(StackReference<mirror::ArtMethod>* top_method, uintptr_t pc) {
359     tlsPtr_.managed_stack.SetTopQuickFrame(top_method);
360     tlsPtr_.managed_stack.SetTopQuickFramePc(pc);
361   }
362 
SetTopOfShadowStack(ShadowFrame * top)363   void SetTopOfShadowStack(ShadowFrame* top) {
364     tlsPtr_.managed_stack.SetTopShadowFrame(top);
365   }
366 
HasManagedStack()367   bool HasManagedStack() const {
368     return (tlsPtr_.managed_stack.GetTopQuickFrame() != nullptr) ||
369         (tlsPtr_.managed_stack.GetTopShadowFrame() != nullptr);
370   }
371 
372   // If 'msg' is NULL, no detail message is set.
373   void ThrowNewException(const ThrowLocation& throw_location,
374                          const char* exception_class_descriptor, const char* msg)
375       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
376 
377   // If 'msg' is NULL, no detail message is set. An exception must be pending, and will be
378   // used as the new exception's cause.
379   void ThrowNewWrappedException(const ThrowLocation& throw_location,
380                                 const char* exception_class_descriptor,
381                                 const char* msg)
382       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
383 
384   void ThrowNewExceptionF(const ThrowLocation& throw_location,
385                           const char* exception_class_descriptor, const char* fmt, ...)
386       __attribute__((format(printf, 4, 5)))
387       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
388 
389   void ThrowNewExceptionV(const ThrowLocation& throw_location,
390                           const char* exception_class_descriptor, const char* fmt, va_list ap)
391       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
392 
393   // OutOfMemoryError is special, because we need to pre-allocate an instance.
394   // Only the GC should call this.
395   void ThrowOutOfMemoryError(const char* msg) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
396 
397   static void Startup();
398   static void FinishStartup();
399   static void Shutdown();
400 
401   // JNI methods
GetJniEnv()402   JNIEnvExt* GetJniEnv() const {
403     return tlsPtr_.jni_env;
404   }
405 
406   // Convert a jobject into a Object*
407   mirror::Object* DecodeJObject(jobject obj) const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
408 
GetMonitorEnterObject()409   mirror::Object* GetMonitorEnterObject() const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
410     return tlsPtr_.monitor_enter_object;
411   }
412 
SetMonitorEnterObject(mirror::Object * obj)413   void SetMonitorEnterObject(mirror::Object* obj) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
414     tlsPtr_.monitor_enter_object = obj;
415   }
416 
417   // Implements java.lang.Thread.interrupted.
418   bool Interrupted() LOCKS_EXCLUDED(wait_mutex_);
419   // Implements java.lang.Thread.isInterrupted.
420   bool IsInterrupted() LOCKS_EXCLUDED(wait_mutex_);
IsInterruptedLocked()421   bool IsInterruptedLocked() EXCLUSIVE_LOCKS_REQUIRED(wait_mutex_) {
422     return interrupted_;
423   }
424   void Interrupt(Thread* self) LOCKS_EXCLUDED(wait_mutex_);
SetInterruptedLocked(bool i)425   void SetInterruptedLocked(bool i) EXCLUSIVE_LOCKS_REQUIRED(wait_mutex_) {
426     interrupted_ = i;
427   }
428   void Notify() LOCKS_EXCLUDED(wait_mutex_);
429 
430  private:
431   void NotifyLocked(Thread* self) EXCLUSIVE_LOCKS_REQUIRED(wait_mutex_);
432 
433  public:
GetWaitMutex()434   Mutex* GetWaitMutex() const LOCK_RETURNED(wait_mutex_) {
435     return wait_mutex_;
436   }
437 
GetWaitConditionVariable()438   ConditionVariable* GetWaitConditionVariable() const EXCLUSIVE_LOCKS_REQUIRED(wait_mutex_) {
439     return wait_cond_;
440   }
441 
GetWaitMonitor()442   Monitor* GetWaitMonitor() const EXCLUSIVE_LOCKS_REQUIRED(wait_mutex_) {
443     return wait_monitor_;
444   }
445 
SetWaitMonitor(Monitor * mon)446   void SetWaitMonitor(Monitor* mon) EXCLUSIVE_LOCKS_REQUIRED(wait_mutex_) {
447     wait_monitor_ = mon;
448   }
449 
450 
451   // Waiter link-list support.
GetWaitNext()452   Thread* GetWaitNext() const {
453     return tlsPtr_.wait_next;
454   }
455 
SetWaitNext(Thread * next)456   void SetWaitNext(Thread* next) {
457     tlsPtr_.wait_next = next;
458   }
459 
GetClassLoaderOverride()460   mirror::ClassLoader* GetClassLoaderOverride() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
461     return tlsPtr_.class_loader_override;
462   }
463 
464   void SetClassLoaderOverride(mirror::ClassLoader* class_loader_override)
465       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
466 
467   // Create the internal representation of a stack trace, that is more time
468   // and space efficient to compute than the StackTraceElement[].
469   template<bool kTransactionActive>
470   jobject CreateInternalStackTrace(const ScopedObjectAccessAlreadyRunnable& soa) const
471       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
472 
473   // Convert an internal stack trace representation (returned by CreateInternalStackTrace) to a
474   // StackTraceElement[]. If output_array is NULL, a new array is created, otherwise as many
475   // frames as will fit are written into the given array. If stack_depth is non-NULL, it's updated
476   // with the number of valid frames in the returned array.
477   static jobjectArray InternalStackTraceToStackTraceElementArray(
478       const ScopedObjectAccessAlreadyRunnable& soa, jobject internal,
479       jobjectArray output_array = nullptr, int* stack_depth = nullptr)
480       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
481 
482   void VisitRoots(RootCallback* visitor, void* arg) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
483 
484   ALWAYS_INLINE void VerifyStack() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
485 
486   //
487   // Offsets of various members of native Thread class, used by compiled code.
488   //
489 
490   template<size_t pointer_size>
ThinLockIdOffset()491   static ThreadOffset<pointer_size> ThinLockIdOffset() {
492     return ThreadOffset<pointer_size>(
493         OFFSETOF_MEMBER(Thread, tls32_) +
494         OFFSETOF_MEMBER(tls_32bit_sized_values, thin_lock_thread_id));
495   }
496 
497   template<size_t pointer_size>
ThreadFlagsOffset()498   static ThreadOffset<pointer_size> ThreadFlagsOffset() {
499     return ThreadOffset<pointer_size>(
500         OFFSETOF_MEMBER(Thread, tls32_) +
501         OFFSETOF_MEMBER(tls_32bit_sized_values, state_and_flags));
502   }
503 
504  private:
505   template<size_t pointer_size>
ThreadOffsetFromTlsPtr(size_t tls_ptr_offset)506   static ThreadOffset<pointer_size> ThreadOffsetFromTlsPtr(size_t tls_ptr_offset) {
507     size_t base = OFFSETOF_MEMBER(Thread, tlsPtr_);
508     size_t scale;
509     size_t shrink;
510     if (pointer_size == sizeof(void*)) {
511       scale = 1;
512       shrink = 1;
513     } else if (pointer_size > sizeof(void*)) {
514       scale = pointer_size / sizeof(void*);
515       shrink = 1;
516     } else {
517       DCHECK_GT(sizeof(void*), pointer_size);
518       scale = 1;
519       shrink = sizeof(void*) / pointer_size;
520     }
521     return ThreadOffset<pointer_size>(base + ((tls_ptr_offset * scale) / shrink));
522   }
523 
524  public:
525   template<size_t pointer_size>
QuickEntryPointOffset(size_t quick_entrypoint_offset)526   static ThreadOffset<pointer_size> QuickEntryPointOffset(size_t quick_entrypoint_offset) {
527     return ThreadOffsetFromTlsPtr<pointer_size>(
528         OFFSETOF_MEMBER(tls_ptr_sized_values, quick_entrypoints) + quick_entrypoint_offset);
529   }
530 
531   template<size_t pointer_size>
InterpreterEntryPointOffset(size_t interp_entrypoint_offset)532   static ThreadOffset<pointer_size> InterpreterEntryPointOffset(size_t interp_entrypoint_offset) {
533     return ThreadOffsetFromTlsPtr<pointer_size>(
534         OFFSETOF_MEMBER(tls_ptr_sized_values, interpreter_entrypoints) + interp_entrypoint_offset);
535   }
536 
537   template<size_t pointer_size>
JniEntryPointOffset(size_t jni_entrypoint_offset)538   static ThreadOffset<pointer_size> JniEntryPointOffset(size_t jni_entrypoint_offset) {
539     return ThreadOffsetFromTlsPtr<pointer_size>(
540         OFFSETOF_MEMBER(tls_ptr_sized_values, jni_entrypoints) + jni_entrypoint_offset);
541   }
542 
543   template<size_t pointer_size>
PortableEntryPointOffset(size_t port_entrypoint_offset)544   static ThreadOffset<pointer_size> PortableEntryPointOffset(size_t port_entrypoint_offset) {
545     return ThreadOffsetFromTlsPtr<pointer_size>(
546         OFFSETOF_MEMBER(tls_ptr_sized_values, portable_entrypoints) + port_entrypoint_offset);
547   }
548 
549   template<size_t pointer_size>
SelfOffset()550   static ThreadOffset<pointer_size> SelfOffset() {
551     return ThreadOffsetFromTlsPtr<pointer_size>(OFFSETOF_MEMBER(tls_ptr_sized_values, self));
552   }
553 
554   template<size_t pointer_size>
ExceptionOffset()555   static ThreadOffset<pointer_size> ExceptionOffset() {
556     return ThreadOffsetFromTlsPtr<pointer_size>(OFFSETOF_MEMBER(tls_ptr_sized_values, exception));
557   }
558 
559   template<size_t pointer_size>
PeerOffset()560   static ThreadOffset<pointer_size> PeerOffset() {
561     return ThreadOffsetFromTlsPtr<pointer_size>(OFFSETOF_MEMBER(tls_ptr_sized_values, opeer));
562   }
563 
564 
565   template<size_t pointer_size>
CardTableOffset()566   static ThreadOffset<pointer_size> CardTableOffset() {
567     return ThreadOffsetFromTlsPtr<pointer_size>(OFFSETOF_MEMBER(tls_ptr_sized_values, card_table));
568   }
569 
570   template<size_t pointer_size>
ThreadSuspendTriggerOffset()571   static ThreadOffset<pointer_size> ThreadSuspendTriggerOffset() {
572     return ThreadOffsetFromTlsPtr<pointer_size>(
573         OFFSETOF_MEMBER(tls_ptr_sized_values, suspend_trigger));
574   }
575 
576   // Size of stack less any space reserved for stack overflow
GetStackSize()577   size_t GetStackSize() const {
578     return tlsPtr_.stack_size - (tlsPtr_.stack_end - tlsPtr_.stack_begin);
579   }
580 
GetStackEndForInterpreter(bool implicit_overflow_check)581   byte* GetStackEndForInterpreter(bool implicit_overflow_check) const {
582     if (implicit_overflow_check) {
583       // The interpreter needs the extra overflow bytes that stack_end does
584       // not include.
585       return tlsPtr_.stack_end + GetStackOverflowReservedBytes(kRuntimeISA);
586     } else {
587       return tlsPtr_.stack_end;
588     }
589   }
590 
GetStackEnd()591   byte* GetStackEnd() const {
592     return tlsPtr_.stack_end;
593   }
594 
595   // Set the stack end to that to be used during a stack overflow
596   void SetStackEndForStackOverflow() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
597 
598   // Set the stack end to that to be used during regular execution
ResetDefaultStackEnd()599   void ResetDefaultStackEnd() {
600     // Our stacks grow down, so we want stack_end_ to be near there, but reserving enough room
601     // to throw a StackOverflowError.
602     tlsPtr_.stack_end = tlsPtr_.stack_begin + GetStackOverflowReservedBytes(kRuntimeISA);
603   }
604 
605   // Install the protected region for implicit stack checks.
606   void InstallImplicitProtection();
607 
IsHandlingStackOverflow()608   bool IsHandlingStackOverflow() const {
609     return tlsPtr_.stack_end == tlsPtr_.stack_begin;
610   }
611 
612   template<size_t pointer_size>
StackEndOffset()613   static ThreadOffset<pointer_size> StackEndOffset() {
614     return ThreadOffsetFromTlsPtr<pointer_size>(
615         OFFSETOF_MEMBER(tls_ptr_sized_values, stack_end));
616   }
617 
618   template<size_t pointer_size>
JniEnvOffset()619   static ThreadOffset<pointer_size> JniEnvOffset() {
620     return ThreadOffsetFromTlsPtr<pointer_size>(
621         OFFSETOF_MEMBER(tls_ptr_sized_values, jni_env));
622   }
623 
624   template<size_t pointer_size>
TopOfManagedStackOffset()625   static ThreadOffset<pointer_size> TopOfManagedStackOffset() {
626     return ThreadOffsetFromTlsPtr<pointer_size>(
627         OFFSETOF_MEMBER(tls_ptr_sized_values, managed_stack) +
628         ManagedStack::TopQuickFrameOffset());
629   }
630 
631   template<size_t pointer_size>
TopOfManagedStackPcOffset()632   static ThreadOffset<pointer_size> TopOfManagedStackPcOffset() {
633     return ThreadOffsetFromTlsPtr<pointer_size>(
634         OFFSETOF_MEMBER(tls_ptr_sized_values, managed_stack) +
635         ManagedStack::TopQuickFramePcOffset());
636   }
637 
GetManagedStack()638   const ManagedStack* GetManagedStack() const {
639     return &tlsPtr_.managed_stack;
640   }
641 
642   // Linked list recording fragments of managed stack.
PushManagedStackFragment(ManagedStack * fragment)643   void PushManagedStackFragment(ManagedStack* fragment) {
644     tlsPtr_.managed_stack.PushManagedStackFragment(fragment);
645   }
PopManagedStackFragment(const ManagedStack & fragment)646   void PopManagedStackFragment(const ManagedStack& fragment) {
647     tlsPtr_.managed_stack.PopManagedStackFragment(fragment);
648   }
649 
PushShadowFrame(ShadowFrame * new_top_frame)650   ShadowFrame* PushShadowFrame(ShadowFrame* new_top_frame) {
651     return tlsPtr_.managed_stack.PushShadowFrame(new_top_frame);
652   }
653 
PopShadowFrame()654   ShadowFrame* PopShadowFrame() {
655     return tlsPtr_.managed_stack.PopShadowFrame();
656   }
657 
658   template<size_t pointer_size>
TopShadowFrameOffset()659   static ThreadOffset<pointer_size> TopShadowFrameOffset() {
660     return ThreadOffsetFromTlsPtr<pointer_size>(
661         OFFSETOF_MEMBER(tls_ptr_sized_values, managed_stack) +
662         ManagedStack::TopShadowFrameOffset());
663   }
664 
665   // Number of references allocated in JNI ShadowFrames on this thread.
NumJniShadowFrameReferences()666   size_t NumJniShadowFrameReferences() const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
667     return tlsPtr_.managed_stack.NumJniShadowFrameReferences();
668   }
669 
670   // Number of references in handle scope on this thread.
671   size_t NumHandleReferences();
672 
673   // Number of references allocated in handle scopes & JNI shadow frames on this thread.
NumStackReferences()674   size_t NumStackReferences() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
675     return NumHandleReferences() + NumJniShadowFrameReferences();
676   };
677 
678   // Is the given obj in this thread's stack indirect reference table?
679   bool HandleScopeContains(jobject obj) const;
680 
681   void HandleScopeVisitRoots(RootCallback* visitor, void* arg, uint32_t thread_id)
682       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
683 
GetTopHandleScope()684   HandleScope* GetTopHandleScope() {
685     return tlsPtr_.top_handle_scope;
686   }
687 
PushHandleScope(HandleScope * handle_scope)688   void PushHandleScope(HandleScope* handle_scope) {
689     handle_scope->SetLink(tlsPtr_.top_handle_scope);
690     tlsPtr_.top_handle_scope = handle_scope;
691   }
692 
PopHandleScope()693   HandleScope* PopHandleScope() {
694     HandleScope* handle_scope = tlsPtr_.top_handle_scope;
695     DCHECK(handle_scope != nullptr);
696     tlsPtr_.top_handle_scope = tlsPtr_.top_handle_scope->GetLink();
697     return handle_scope;
698   }
699 
700   template<size_t pointer_size>
TopHandleScopeOffset()701   static ThreadOffset<pointer_size> TopHandleScopeOffset() {
702     return ThreadOffsetFromTlsPtr<pointer_size>(OFFSETOF_MEMBER(tls_ptr_sized_values,
703                                                                 top_handle_scope));
704   }
705 
GetInvokeReq()706   DebugInvokeReq* GetInvokeReq() const {
707     return tlsPtr_.debug_invoke_req;
708   }
709 
GetSingleStepControl()710   SingleStepControl* GetSingleStepControl() const {
711     return tlsPtr_.single_step_control;
712   }
713 
714   // Returns the fake exception used to activate deoptimization.
GetDeoptimizationException()715   static mirror::Throwable* GetDeoptimizationException() {
716     return reinterpret_cast<mirror::Throwable*>(-1);
717   }
718 
719   void SetDeoptimizationShadowFrame(ShadowFrame* sf);
720   void SetDeoptimizationReturnValue(const JValue& ret_val);
721 
722   ShadowFrame* GetAndClearDeoptimizationShadowFrame(JValue* ret_val);
723 
HasDeoptimizationShadowFrame()724   bool HasDeoptimizationShadowFrame() const {
725     return tlsPtr_.deoptimization_shadow_frame != nullptr;
726   }
727 
728   void SetShadowFrameUnderConstruction(ShadowFrame* sf);
729   void ClearShadowFrameUnderConstruction();
730 
HasShadowFrameUnderConstruction()731   bool HasShadowFrameUnderConstruction() const {
732     return tlsPtr_.shadow_frame_under_construction != nullptr;
733   }
734 
GetInstrumentationStack()735   std::deque<instrumentation::InstrumentationStackFrame>* GetInstrumentationStack() {
736     return tlsPtr_.instrumentation_stack;
737   }
738 
GetStackTraceSample()739   std::vector<mirror::ArtMethod*>* GetStackTraceSample() const {
740     return tlsPtr_.stack_trace_sample;
741   }
742 
SetStackTraceSample(std::vector<mirror::ArtMethod * > * sample)743   void SetStackTraceSample(std::vector<mirror::ArtMethod*>* sample) {
744     tlsPtr_.stack_trace_sample = sample;
745   }
746 
GetTraceClockBase()747   uint64_t GetTraceClockBase() const {
748     return tls64_.trace_clock_base;
749   }
750 
SetTraceClockBase(uint64_t clock_base)751   void SetTraceClockBase(uint64_t clock_base) {
752     tls64_.trace_clock_base = clock_base;
753   }
754 
GetHeldMutex(LockLevel level)755   BaseMutex* GetHeldMutex(LockLevel level) const {
756     return tlsPtr_.held_mutexes[level];
757   }
758 
SetHeldMutex(LockLevel level,BaseMutex * mutex)759   void SetHeldMutex(LockLevel level, BaseMutex* mutex) {
760     tlsPtr_.held_mutexes[level] = mutex;
761   }
762 
763   void RunCheckpointFunction();
764 
ReadFlag(ThreadFlag flag)765   bool ReadFlag(ThreadFlag flag) const {
766     return (tls32_.state_and_flags.as_struct.flags & flag) != 0;
767   }
768 
TestAllFlags()769   bool TestAllFlags() const {
770     return (tls32_.state_and_flags.as_struct.flags != 0);
771   }
772 
AtomicSetFlag(ThreadFlag flag)773   void AtomicSetFlag(ThreadFlag flag) {
774     tls32_.state_and_flags.as_atomic_int.FetchAndOrSequentiallyConsistent(flag);
775   }
776 
AtomicClearFlag(ThreadFlag flag)777   void AtomicClearFlag(ThreadFlag flag) {
778     tls32_.state_and_flags.as_atomic_int.FetchAndAndSequentiallyConsistent(-1 ^ flag);
779   }
780 
781   void ResetQuickAllocEntryPointsForThread();
782 
783   // Returns the remaining space in the TLAB.
784   size_t TlabSize() const;
785   // Doesn't check that there is room.
786   mirror::Object* AllocTlab(size_t bytes);
787   void SetTlab(byte* start, byte* end);
788   bool HasTlab() const;
789 
790   // Remove the suspend trigger for this thread by making the suspend_trigger_ TLS value
791   // equal to a valid pointer.
792   // TODO: does this need to atomic?  I don't think so.
RemoveSuspendTrigger()793   void RemoveSuspendTrigger() {
794     tlsPtr_.suspend_trigger = reinterpret_cast<uintptr_t*>(&tlsPtr_.suspend_trigger);
795   }
796 
797   // Trigger a suspend check by making the suspend_trigger_ TLS value an invalid pointer.
798   // The next time a suspend check is done, it will load from the value at this address
799   // and trigger a SIGSEGV.
TriggerSuspend()800   void TriggerSuspend() {
801     tlsPtr_.suspend_trigger = nullptr;
802   }
803 
804 
805   // Push an object onto the allocation stack.
806   bool PushOnThreadLocalAllocationStack(mirror::Object* obj);
807 
808   // Set the thread local allocation pointers to the given pointers.
809   void SetThreadLocalAllocationStack(mirror::Object** start, mirror::Object** end);
810 
811   // Resets the thread local allocation pointers.
812   void RevokeThreadLocalAllocationStack();
813 
GetThreadLocalBytesAllocated()814   size_t GetThreadLocalBytesAllocated() const {
815     return tlsPtr_.thread_local_end - tlsPtr_.thread_local_start;
816   }
817 
GetThreadLocalObjectsAllocated()818   size_t GetThreadLocalObjectsAllocated() const {
819     return tlsPtr_.thread_local_objects;
820   }
821 
GetRosAllocRun(size_t index)822   void* GetRosAllocRun(size_t index) const {
823     return tlsPtr_.rosalloc_runs[index];
824   }
825 
SetRosAllocRun(size_t index,void * run)826   void SetRosAllocRun(size_t index, void* run) {
827     tlsPtr_.rosalloc_runs[index] = run;
828   }
829 
IsExceptionReportedToInstrumentation()830   bool IsExceptionReportedToInstrumentation() const {
831     return tls32_.is_exception_reported_to_instrumentation_;
832   }
833 
SetExceptionReportedToInstrumentation(bool reported)834   void SetExceptionReportedToInstrumentation(bool reported) {
835     tls32_.is_exception_reported_to_instrumentation_ = reported;
836   }
837 
838   void ProtectStack();
839   bool UnprotectStack();
840 
NoteSignalBeingHandled()841   void NoteSignalBeingHandled() {
842     if (tls32_.handling_signal_) {
843       LOG(FATAL) << "Detected signal while processing a signal";
844     }
845     tls32_.handling_signal_ = true;
846   }
847 
NoteSignalHandlerDone()848   void NoteSignalHandlerDone() {
849     tls32_.handling_signal_ = false;
850   }
851 
GetNestedSignalState()852   jmp_buf* GetNestedSignalState() {
853     return tlsPtr_.nested_signal_state;
854   }
855 
856  private:
857   explicit Thread(bool daemon);
858   ~Thread() LOCKS_EXCLUDED(Locks::mutator_lock_,
859                            Locks::thread_suspend_count_lock_);
860   void Destroy();
861 
862   void CreatePeer(const char* name, bool as_daemon, jobject thread_group);
863 
864   template<bool kTransactionActive>
865   void InitPeer(ScopedObjectAccess& soa, jboolean thread_is_daemon, jobject thread_group,
866                 jobject thread_name, jint thread_priority)
867       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
868 
869   // Avoid use, callers should use SetState. Used only by SignalCatcher::HandleSigQuit, ~Thread and
870   // Dbg::Disconnected.
SetStateUnsafe(ThreadState new_state)871   ThreadState SetStateUnsafe(ThreadState new_state) {
872     ThreadState old_state = GetState();
873     tls32_.state_and_flags.as_struct.state = new_state;
874     return old_state;
875   }
876 
877   void VerifyStackImpl() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
878 
879   void DumpState(std::ostream& os) const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
880   void DumpStack(std::ostream& os) const
881       LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
882       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
883 
884   // Out-of-line conveniences for debugging in gdb.
885   static Thread* CurrentFromGdb();  // Like Thread::Current.
886   // Like Thread::Dump(std::cerr).
887   void DumpFromGdb() const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
888 
889   static void* CreateCallback(void* arg);
890 
891   void HandleUncaughtExceptions(ScopedObjectAccess& soa)
892       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
893   void RemoveFromThreadGroup(ScopedObjectAccess& soa) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_);
894 
895   void Init(ThreadList*, JavaVMExt*) EXCLUSIVE_LOCKS_REQUIRED(Locks::runtime_shutdown_lock_);
896   void InitCardTable();
897   void InitCpu();
898   void CleanupCpu();
899   void InitTlsEntryPoints();
900   void InitTid();
901   void InitPthreadKeySelf();
902   void InitStackHwm();
903 
904   void SetUpAlternateSignalStack();
905   void TearDownAlternateSignalStack();
906 
907   // 32 bits of atomically changed state and flags. Keeping as 32 bits allows and atomic CAS to
908   // change from being Suspended to Runnable without a suspend request occurring.
909   union PACKED(4) StateAndFlags {
StateAndFlags()910     StateAndFlags() {}
911     struct PACKED(4) {
912       // Bitfield of flag values. Must be changed atomically so that flag values aren't lost. See
913       // ThreadFlags for bit field meanings.
914       volatile uint16_t flags;
915       // Holds the ThreadState. May be changed non-atomically between Suspended (ie not Runnable)
916       // transitions. Changing to Runnable requires that the suspend_request be part of the atomic
917       // operation. If a thread is suspended and a suspend_request is present, a thread may not
918       // change to Runnable as a GC or other operation is in progress.
919       volatile uint16_t state;
920     } as_struct;
921     AtomicInteger as_atomic_int;
922     volatile int32_t as_int;
923 
924    private:
925     // gcc does not handle struct with volatile member assignments correctly.
926     // See http://gcc.gnu.org/bugzilla/show_bug.cgi?id=47409
927     DISALLOW_COPY_AND_ASSIGN(StateAndFlags);
928   };
929   COMPILE_ASSERT(sizeof(StateAndFlags) == sizeof(int32_t), weird_state_and_flags_size);
930 
931   static void ThreadExitCallback(void* arg);
932 
933   // Maximum number of checkpoint functions.
934   static constexpr uint32_t kMaxCheckpoints = 3;
935 
936   // Has Thread::Startup been called?
937   static bool is_started_;
938 
939   // TLS key used to retrieve the Thread*.
940   static pthread_key_t pthread_key_self_;
941 
942   // Used to notify threads that they should attempt to resume, they will suspend again if
943   // their suspend count is > 0.
944   static ConditionVariable* resume_cond_ GUARDED_BY(Locks::thread_suspend_count_lock_);
945 
946   /***********************************************************************************************/
947   // Thread local storage. Fields are grouped by size to enable 32 <-> 64 searching to account for
948   // pointer size differences. To encourage shorter encoding, more frequently used values appear
949   // first if possible.
950   /***********************************************************************************************/
951 
952   struct PACKED(4) tls_32bit_sized_values {
953     // We have no control over the size of 'bool', but want our boolean fields
954     // to be 4-byte quantities.
955     typedef uint32_t bool32_t;
956 
tls_32bit_sized_valuestls_32bit_sized_values957     explicit tls_32bit_sized_values(bool is_daemon) :
958       suspend_count(0), debug_suspend_count(0), thin_lock_thread_id(0), tid(0),
959       daemon(is_daemon), throwing_OutOfMemoryError(false), no_thread_suspension(0),
960       thread_exit_check_count(0), is_exception_reported_to_instrumentation_(false),
961       handling_signal_(false), padding_(0) {
962     }
963 
964     union StateAndFlags state_and_flags;
965     COMPILE_ASSERT(sizeof(union StateAndFlags) == sizeof(int32_t),
966                    sizeof_state_and_flags_and_int32_are_different);
967 
968     // A non-zero value is used to tell the current thread to enter a safe point
969     // at the next poll.
970     int suspend_count GUARDED_BY(Locks::thread_suspend_count_lock_);
971 
972     // How much of 'suspend_count_' is by request of the debugger, used to set things right
973     // when the debugger detaches. Must be <= suspend_count_.
974     int debug_suspend_count GUARDED_BY(Locks::thread_suspend_count_lock_);
975 
976     // Thin lock thread id. This is a small integer used by the thin lock implementation.
977     // This is not to be confused with the native thread's tid, nor is it the value returned
978     // by java.lang.Thread.getId --- this is a distinct value, used only for locking. One
979     // important difference between this id and the ids visible to managed code is that these
980     // ones get reused (to ensure that they fit in the number of bits available).
981     uint32_t thin_lock_thread_id;
982 
983     // System thread id.
984     uint32_t tid;
985 
986     // Is the thread a daemon?
987     const bool32_t daemon;
988 
989     // A boolean telling us whether we're recursively throwing OOME.
990     bool32_t throwing_OutOfMemoryError;
991 
992     // A positive value implies we're in a region where thread suspension isn't expected.
993     uint32_t no_thread_suspension;
994 
995     // How many times has our pthread key's destructor been called?
996     uint32_t thread_exit_check_count;
997 
998     // When true this field indicates that the exception associated with this thread has already
999     // been reported to instrumentation.
1000     bool32_t is_exception_reported_to_instrumentation_;
1001 
1002     // True if signal is being handled by this thread.
1003     bool32_t handling_signal_;
1004 
1005     // Padding to make the size aligned to 8.  Remove this if we add another 32 bit field.
1006     int32_t padding_;
1007   } tls32_;
1008 
1009   struct PACKED(8) tls_64bit_sized_values {
tls_64bit_sized_valuestls_64bit_sized_values1010     tls_64bit_sized_values() : trace_clock_base(0), deoptimization_return_value() {
1011     }
1012 
1013     // The clock base used for tracing.
1014     uint64_t trace_clock_base;
1015 
1016     // Return value used by deoptimization.
1017     JValue deoptimization_return_value;
1018 
1019     RuntimeStats stats;
1020   } tls64_;
1021 
1022   struct PACKED(4) tls_ptr_sized_values {
tls_ptr_sized_valuestls_ptr_sized_values1023       tls_ptr_sized_values() : card_table(nullptr), exception(nullptr), stack_end(nullptr),
1024       managed_stack(), suspend_trigger(nullptr), jni_env(nullptr), self(nullptr), opeer(nullptr),
1025       jpeer(nullptr), stack_begin(nullptr), stack_size(0), throw_location(),
1026       stack_trace_sample(nullptr), wait_next(nullptr), monitor_enter_object(nullptr),
1027       top_handle_scope(nullptr), class_loader_override(nullptr), long_jump_context(nullptr),
1028       instrumentation_stack(nullptr), debug_invoke_req(nullptr), single_step_control(nullptr),
1029       deoptimization_shadow_frame(nullptr), shadow_frame_under_construction(nullptr), name(nullptr),
1030       pthread_self(0), last_no_thread_suspension_cause(nullptr), thread_local_start(nullptr),
1031       thread_local_pos(nullptr), thread_local_end(nullptr), thread_local_objects(0),
1032       thread_local_alloc_stack_top(nullptr), thread_local_alloc_stack_end(nullptr) {
1033     }
1034 
1035     // The biased card table, see CardTable for details.
1036     byte* card_table;
1037 
1038     // The pending exception or NULL.
1039     mirror::Throwable* exception;
1040 
1041     // The end of this thread's stack. This is the lowest safely-addressable address on the stack.
1042     // We leave extra space so there's room for the code that throws StackOverflowError.
1043     byte* stack_end;
1044 
1045     // The top of the managed stack often manipulated directly by compiler generated code.
1046     ManagedStack managed_stack;
1047 
1048     // In certain modes, setting this to 0 will trigger a SEGV and thus a suspend check.  It is
1049     // normally set to the address of itself.
1050     uintptr_t* suspend_trigger;
1051 
1052     // Every thread may have an associated JNI environment
1053     JNIEnvExt* jni_env;
1054 
1055     // Initialized to "this". On certain architectures (such as x86) reading off of Thread::Current
1056     // is easy but getting the address of Thread::Current is hard. This field can be read off of
1057     // Thread::Current to give the address.
1058     Thread* self;
1059 
1060     // Our managed peer (an instance of java.lang.Thread). The jobject version is used during thread
1061     // start up, until the thread is registered and the local opeer_ is used.
1062     mirror::Object* opeer;
1063     jobject jpeer;
1064 
1065     // The "lowest addressable byte" of the stack.
1066     byte* stack_begin;
1067 
1068     // Size of the stack.
1069     size_t stack_size;
1070 
1071     // The location the current exception was thrown from.
1072     ThrowLocation throw_location;
1073 
1074     // Pointer to previous stack trace captured by sampling profiler.
1075     std::vector<mirror::ArtMethod*>* stack_trace_sample;
1076 
1077     // The next thread in the wait set this thread is part of or NULL if not waiting.
1078     Thread* wait_next;
1079 
1080     // If we're blocked in MonitorEnter, this is the object we're trying to lock.
1081     mirror::Object* monitor_enter_object;
1082 
1083     // Top of linked list of handle scopes or nullptr for none.
1084     HandleScope* top_handle_scope;
1085 
1086     // Needed to get the right ClassLoader in JNI_OnLoad, but also
1087     // useful for testing.
1088     mirror::ClassLoader* class_loader_override;
1089 
1090     // Thread local, lazily allocated, long jump context. Used to deliver exceptions.
1091     Context* long_jump_context;
1092 
1093     // Additional stack used by method instrumentation to store method and return pc values.
1094     // Stored as a pointer since std::deque is not PACKED.
1095     std::deque<instrumentation::InstrumentationStackFrame>* instrumentation_stack;
1096 
1097     // JDWP invoke-during-breakpoint support.
1098     DebugInvokeReq* debug_invoke_req;
1099 
1100     // JDWP single-stepping support.
1101     SingleStepControl* single_step_control;
1102 
1103     // Shadow frame stack that is used temporarily during the deoptimization of a method.
1104     ShadowFrame* deoptimization_shadow_frame;
1105 
1106     // Shadow frame stack that is currently under construction but not yet on the stack
1107     ShadowFrame* shadow_frame_under_construction;
1108 
1109     // A cached copy of the java.lang.Thread's name.
1110     std::string* name;
1111 
1112     // A cached pthread_t for the pthread underlying this Thread*.
1113     pthread_t pthread_self;
1114 
1115     // If no_thread_suspension_ is > 0, what is causing that assertion.
1116     const char* last_no_thread_suspension_cause;
1117 
1118     // Pending checkpoint function or NULL if non-pending. Installation guarding by
1119     // Locks::thread_suspend_count_lock_.
1120     Closure* checkpoint_functions[kMaxCheckpoints];
1121 
1122     // Entrypoint function pointers.
1123     // TODO: move this to more of a global offset table model to avoid per-thread duplication.
1124     InterpreterEntryPoints interpreter_entrypoints;
1125     JniEntryPoints jni_entrypoints;
1126     PortableEntryPoints portable_entrypoints;
1127     QuickEntryPoints quick_entrypoints;
1128 
1129     // Thread-local allocation pointer.
1130     byte* thread_local_start;
1131     byte* thread_local_pos;
1132     byte* thread_local_end;
1133     size_t thread_local_objects;
1134 
1135     // There are RosAlloc::kNumThreadLocalSizeBrackets thread-local size brackets per thread.
1136     void* rosalloc_runs[kNumRosAllocThreadLocalSizeBrackets];
1137 
1138     // Thread-local allocation stack data/routines.
1139     mirror::Object** thread_local_alloc_stack_top;
1140     mirror::Object** thread_local_alloc_stack_end;
1141 
1142     // Support for Mutex lock hierarchy bug detection.
1143     BaseMutex* held_mutexes[kLockLevelCount];
1144 
1145     // Recorded thread state for nested signals.
1146     jmp_buf* nested_signal_state;
1147   } tlsPtr_;
1148 
1149   // Guards the 'interrupted_' and 'wait_monitor_' members.
1150   Mutex* wait_mutex_ DEFAULT_MUTEX_ACQUIRED_AFTER;
1151 
1152   // Condition variable waited upon during a wait.
1153   ConditionVariable* wait_cond_ GUARDED_BY(wait_mutex_);
1154   // Pointer to the monitor lock we're currently waiting on or NULL if not waiting.
1155   Monitor* wait_monitor_ GUARDED_BY(wait_mutex_);
1156 
1157   // Thread "interrupted" status; stays raised until queried or thrown.
1158   bool interrupted_ GUARDED_BY(wait_mutex_);
1159 
1160   friend class Dbg;  // For SetStateUnsafe.
1161   friend class gc::collector::SemiSpace;  // For getting stack traces.
1162   friend class Runtime;  // For CreatePeer.
1163   friend class QuickExceptionHandler;  // For dumping the stack.
1164   friend class ScopedThreadStateChange;
1165   friend class SignalCatcher;  // For SetStateUnsafe.
1166   friend class StubTest;  // For accessing entrypoints.
1167   friend class ThreadList;  // For ~Thread and Destroy.
1168 
1169   friend class EntrypointsOrderTest;  // To test the order of tls entries.
1170 
1171   DISALLOW_COPY_AND_ASSIGN(Thread);
1172 };
1173 
1174 std::ostream& operator<<(std::ostream& os, const Thread& thread);
1175 std::ostream& operator<<(std::ostream& os, const ThreadState& state);
1176 
1177 }  // namespace art
1178 
1179 #endif  // ART_RUNTIME_THREAD_H_
1180