1 //===-- sanitizer_deadlock_detector1.cc -----------------------------------===//
2 //
3 // The LLVM Compiler Infrastructure
4 //
5 // This file is distributed under the University of Illinois Open Source
6 // License. See LICENSE.TXT for details.
7 //
8 //===----------------------------------------------------------------------===//
9 //
10 // Deadlock detector implementation based on NxN adjacency bit matrix.
11 //
12 //===----------------------------------------------------------------------===//
13
14 #include "sanitizer_deadlock_detector_interface.h"
15 #include "sanitizer_deadlock_detector.h"
16 #include "sanitizer_allocator_internal.h"
17 #include "sanitizer_placement_new.h"
18 #include "sanitizer_mutex.h"
19
20 #if SANITIZER_DEADLOCK_DETECTOR_VERSION == 1
21
22 namespace __sanitizer {
23
24 typedef TwoLevelBitVector<> DDBV; // DeadlockDetector's bit vector.
25
26 struct DDPhysicalThread {
27 };
28
29 struct DDLogicalThread {
30 u64 ctx;
31 DeadlockDetectorTLS<DDBV> dd;
32 DDReport rep;
33 bool report_pending;
34 };
35
36 struct DD : public DDetector {
37 SpinMutex mtx;
38 DeadlockDetector<DDBV> dd;
39 DDFlags flags;
40
41 explicit DD(const DDFlags *flags);
42
43 DDPhysicalThread *CreatePhysicalThread() override;
44 void DestroyPhysicalThread(DDPhysicalThread *pt) override;
45
46 DDLogicalThread *CreateLogicalThread(u64 ctx) override;
47 void DestroyLogicalThread(DDLogicalThread *lt) override;
48
49 void MutexInit(DDCallback *cb, DDMutex *m) override;
50 void MutexBeforeLock(DDCallback *cb, DDMutex *m, bool wlock) override;
51 void MutexAfterLock(DDCallback *cb, DDMutex *m, bool wlock,
52 bool trylock) override;
53 void MutexBeforeUnlock(DDCallback *cb, DDMutex *m, bool wlock) override;
54 void MutexDestroy(DDCallback *cb, DDMutex *m) override;
55
56 DDReport *GetReport(DDCallback *cb) override;
57
58 void MutexEnsureID(DDLogicalThread *lt, DDMutex *m);
59 void ReportDeadlock(DDCallback *cb, DDMutex *m);
60 };
61
Create(const DDFlags * flags)62 DDetector *DDetector::Create(const DDFlags *flags) {
63 (void)flags;
64 void *mem = MmapOrDie(sizeof(DD), "deadlock detector");
65 return new(mem) DD(flags);
66 }
67
DD(const DDFlags * flags)68 DD::DD(const DDFlags *flags)
69 : flags(*flags) {
70 dd.clear();
71 }
72
CreatePhysicalThread()73 DDPhysicalThread* DD::CreatePhysicalThread() {
74 return nullptr;
75 }
76
DestroyPhysicalThread(DDPhysicalThread * pt)77 void DD::DestroyPhysicalThread(DDPhysicalThread *pt) {
78 }
79
CreateLogicalThread(u64 ctx)80 DDLogicalThread* DD::CreateLogicalThread(u64 ctx) {
81 DDLogicalThread *lt = (DDLogicalThread*)InternalAlloc(sizeof(*lt));
82 lt->ctx = ctx;
83 lt->dd.clear();
84 lt->report_pending = false;
85 return lt;
86 }
87
DestroyLogicalThread(DDLogicalThread * lt)88 void DD::DestroyLogicalThread(DDLogicalThread *lt) {
89 lt->~DDLogicalThread();
90 InternalFree(lt);
91 }
92
MutexInit(DDCallback * cb,DDMutex * m)93 void DD::MutexInit(DDCallback *cb, DDMutex *m) {
94 m->id = 0;
95 m->stk = cb->Unwind();
96 }
97
MutexEnsureID(DDLogicalThread * lt,DDMutex * m)98 void DD::MutexEnsureID(DDLogicalThread *lt, DDMutex *m) {
99 if (!dd.nodeBelongsToCurrentEpoch(m->id))
100 m->id = dd.newNode(reinterpret_cast<uptr>(m));
101 dd.ensureCurrentEpoch(<->dd);
102 }
103
MutexBeforeLock(DDCallback * cb,DDMutex * m,bool wlock)104 void DD::MutexBeforeLock(DDCallback *cb,
105 DDMutex *m, bool wlock) {
106 DDLogicalThread *lt = cb->lt;
107 if (lt->dd.empty()) return; // This will be the first lock held by lt.
108 if (dd.hasAllEdges(<->dd, m->id)) return; // We already have all edges.
109 SpinMutexLock lk(&mtx);
110 MutexEnsureID(lt, m);
111 if (dd.isHeld(<->dd, m->id))
112 return; // FIXME: allow this only for recursive locks.
113 if (dd.onLockBefore(<->dd, m->id)) {
114 // Actually add this edge now so that we have all the stack traces.
115 dd.addEdges(<->dd, m->id, cb->Unwind(), cb->UniqueTid());
116 ReportDeadlock(cb, m);
117 }
118 }
119
ReportDeadlock(DDCallback * cb,DDMutex * m)120 void DD::ReportDeadlock(DDCallback *cb, DDMutex *m) {
121 DDLogicalThread *lt = cb->lt;
122 uptr path[20];
123 uptr len = dd.findPathToLock(<->dd, m->id, path, ARRAY_SIZE(path));
124 if (len == 0U) {
125 // A cycle of 20+ locks? Well, that's a bit odd...
126 Printf("WARNING: too long mutex cycle found\n");
127 return;
128 }
129 CHECK_EQ(m->id, path[0]);
130 lt->report_pending = true;
131 len = Min<uptr>(len, DDReport::kMaxLoopSize);
132 DDReport *rep = <->rep;
133 rep->n = len;
134 for (uptr i = 0; i < len; i++) {
135 uptr from = path[i];
136 uptr to = path[(i + 1) % len];
137 DDMutex *m0 = (DDMutex*)dd.getData(from);
138 DDMutex *m1 = (DDMutex*)dd.getData(to);
139
140 u32 stk_from = -1U, stk_to = -1U;
141 int unique_tid = 0;
142 dd.findEdge(from, to, &stk_from, &stk_to, &unique_tid);
143 // Printf("Edge: %zd=>%zd: %u/%u T%d\n", from, to, stk_from, stk_to,
144 // unique_tid);
145 rep->loop[i].thr_ctx = unique_tid;
146 rep->loop[i].mtx_ctx0 = m0->ctx;
147 rep->loop[i].mtx_ctx1 = m1->ctx;
148 rep->loop[i].stk[0] = stk_to;
149 rep->loop[i].stk[1] = stk_from;
150 }
151 }
152
MutexAfterLock(DDCallback * cb,DDMutex * m,bool wlock,bool trylock)153 void DD::MutexAfterLock(DDCallback *cb, DDMutex *m, bool wlock, bool trylock) {
154 DDLogicalThread *lt = cb->lt;
155 u32 stk = 0;
156 if (flags.second_deadlock_stack)
157 stk = cb->Unwind();
158 // Printf("T%p MutexLock: %zx stk %u\n", lt, m->id, stk);
159 if (dd.onFirstLock(<->dd, m->id, stk))
160 return;
161 if (dd.onLockFast(<->dd, m->id, stk))
162 return;
163
164 SpinMutexLock lk(&mtx);
165 MutexEnsureID(lt, m);
166 if (wlock) // Only a recursive rlock may be held.
167 CHECK(!dd.isHeld(<->dd, m->id));
168 if (!trylock)
169 dd.addEdges(<->dd, m->id, stk ? stk : cb->Unwind(), cb->UniqueTid());
170 dd.onLockAfter(<->dd, m->id, stk);
171 }
172
MutexBeforeUnlock(DDCallback * cb,DDMutex * m,bool wlock)173 void DD::MutexBeforeUnlock(DDCallback *cb, DDMutex *m, bool wlock) {
174 // Printf("T%p MutexUnLock: %zx\n", cb->lt, m->id);
175 dd.onUnlock(&cb->lt->dd, m->id);
176 }
177
MutexDestroy(DDCallback * cb,DDMutex * m)178 void DD::MutexDestroy(DDCallback *cb,
179 DDMutex *m) {
180 if (!m->id) return;
181 SpinMutexLock lk(&mtx);
182 if (dd.nodeBelongsToCurrentEpoch(m->id))
183 dd.removeNode(m->id);
184 m->id = 0;
185 }
186
GetReport(DDCallback * cb)187 DDReport *DD::GetReport(DDCallback *cb) {
188 if (!cb->lt->report_pending)
189 return nullptr;
190 cb->lt->report_pending = false;
191 return &cb->lt->rep;
192 }
193
194 } // namespace __sanitizer
195 #endif // #if SANITIZER_DEADLOCK_DETECTOR_VERSION == 1
196