1 /* ThreadSanitizer
2 * Copyright (c) 2011, Google Inc. All rights reserved.
3 * Author: Dmitry Vyukov (dvyukov)
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions are
7 * met:
8 *
9 * * Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * * Neither the name of Google Inc. nor the names of its
12 * contributors may be used to endorse or promote products derived from
13 * this software without specific prior written permission.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
16 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
17 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
18 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
19 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
20 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
21 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
25 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26 */
27
28 #include "ts_util.h"
29 #include "ts_atomic_int.h"
30
31
tsan_atomic_to_str(tsan_memory_order mo)32 char const* tsan_atomic_to_str(tsan_memory_order mo) {
33 switch (mo) {
34 case tsan_memory_order_invalid: return "invalid";
35 case tsan_memory_order_natomic: return "natomic";
36 case tsan_memory_order_relaxed: return "relaxed";
37 case tsan_memory_order_consume: return "consume";
38 case tsan_memory_order_acquire: return "acquire";
39 case tsan_memory_order_release: return "release";
40 case tsan_memory_order_acq_rel: return "acq_rel";
41 case tsan_memory_order_seq_cst: return "seq_cst";
42 default: return "-------";
43 }
44 }
45
46
tsan_atomic_to_str(tsan_atomic_op op)47 char const* tsan_atomic_to_str(tsan_atomic_op op) {
48 switch (op) {
49 case tsan_atomic_op_invalid: return "invalid";
50 case tsan_atomic_op_fence: return "fence";
51 case tsan_atomic_op_load: return "load";
52 case tsan_atomic_op_store: return "store";
53 case tsan_atomic_op_exchange: return "exchange";
54 case tsan_atomic_op_fetch_add: return "fetch_add";
55 case tsan_atomic_op_fetch_sub: return "fetch_sub";
56 case tsan_atomic_op_fetch_and: return "fetch_and";
57 case tsan_atomic_op_fetch_xor: return "fetch_xor";
58 case tsan_atomic_op_fetch_or: return "fetch_or";
59 case tsan_atomic_op_compare_exchange_weak: return "compare_exchange_weak";
60 case tsan_atomic_op_compare_exchange_strong:
61 return "compare_exchange_strong";
62 default: return "---";
63 }
64 }
65
66
tsan_atomic_is_acquire(tsan_memory_order mo)67 bool tsan_atomic_is_acquire(tsan_memory_order mo) {
68 return !!(mo & (tsan_memory_order_consume
69 | tsan_memory_order_acquire
70 | tsan_memory_order_acq_rel
71 | tsan_memory_order_seq_cst));
72 }
73
74
tsan_atomic_is_release(tsan_memory_order mo)75 bool tsan_atomic_is_release(tsan_memory_order mo) {
76 return !!(mo & (tsan_memory_order_release
77 | tsan_memory_order_acq_rel
78 | tsan_memory_order_seq_cst));
79 }
80
81
tsan_atomic_is_rmw(tsan_atomic_op op)82 bool tsan_atomic_is_rmw(tsan_atomic_op op) {
83 return !!(op & (tsan_atomic_op_exchange
84 | tsan_atomic_op_fetch_add
85 | tsan_atomic_op_fetch_sub
86 | tsan_atomic_op_fetch_and
87 | tsan_atomic_op_fetch_xor
88 | tsan_atomic_op_fetch_or
89 | tsan_atomic_op_compare_exchange_weak
90 | tsan_atomic_op_compare_exchange_strong));
91 }
92
93
tsan_atomic_verify(tsan_atomic_op op,tsan_memory_order mo,tsan_memory_order fail_mo,size_t size,void volatile * a)94 void tsan_atomic_verify(tsan_atomic_op op,
95 tsan_memory_order mo,
96 tsan_memory_order fail_mo,
97 size_t size,
98 void volatile* a) {
99 CHECK(size == 1 || size == 2 || size == 4 || size == 8);
100 CHECK((((uintptr_t)a) % size) == 0);
101
102 if (op == tsan_atomic_op_load) {
103 CHECK(mo & (tsan_memory_order_natomic
104 | tsan_memory_order_relaxed
105 | tsan_memory_order_consume
106 | tsan_memory_order_acquire
107 | tsan_memory_order_seq_cst));
108 } else if (op == tsan_atomic_op_store) {
109 CHECK(mo & (tsan_memory_order_natomic
110 | tsan_memory_order_relaxed
111 | tsan_memory_order_release
112 | tsan_memory_order_seq_cst));
113 } else if (op == tsan_atomic_op_fence) {
114 CHECK(mo & (tsan_memory_order_consume
115 | tsan_memory_order_acquire
116 | tsan_memory_order_release
117 | tsan_memory_order_acq_rel
118 | tsan_memory_order_seq_cst));
119 } else if (op & (tsan_atomic_op_exchange
120 | tsan_atomic_op_fetch_add
121 | tsan_atomic_op_fetch_sub
122 | tsan_atomic_op_fetch_and
123 | tsan_atomic_op_fetch_xor
124 | tsan_atomic_op_fetch_or
125 | tsan_atomic_op_compare_exchange_weak
126 | tsan_atomic_op_compare_exchange_strong)) {
127 CHECK(mo & (tsan_memory_order_relaxed
128 | tsan_memory_order_consume
129 | tsan_memory_order_acquire
130 | tsan_memory_order_release
131 | tsan_memory_order_acq_rel
132 | tsan_memory_order_seq_cst));
133 } else {
134 CHECK("unknown tsan_atomic_op" == 0);
135 }
136 }
137
138
139 #if defined(__i386__)
140 # define __x86__
141 #elif defined(__x86_64__)
142 # define __x86__
143 #endif
144
145 #if defined(__GNUC__) && defined(__x86_64__)
tsan_atomic_do_op(tsan_atomic_op op,tsan_memory_order mo,tsan_memory_order fail_mo,size_t size,void volatile * a,uint64_t v,uint64_t cmp,uint64_t * newv,uint64_t * prev)146 uint64_t tsan_atomic_do_op(tsan_atomic_op op,
147 tsan_memory_order mo,
148 tsan_memory_order fail_mo,
149 size_t size,
150 void volatile* a,
151 uint64_t v,
152 uint64_t cmp,
153 uint64_t* newv,
154 uint64_t* prev) {
155 *newv = v;
156 if (op != tsan_atomic_op_fence) {
157 if (size == 1) {
158 *prev = *(uint8_t volatile*)a;
159 } else if (size == 2) {
160 *prev = *(uint16_t volatile*)a;
161 } else if (size == 4) {
162 *prev = *(uint32_t volatile*)a;
163 } else if (size == 8) {
164 *prev = *(uint64_t volatile*)a;
165 }
166 }
167
168 if (op == tsan_atomic_op_load) {
169 return *prev;
170
171 } else if (op == tsan_atomic_op_store) {
172 if (mo == tsan_memory_order_seq_cst) {
173 if (size == 1) {
174 uint8_t vv = (uint8_t)v;
175 __asm__ __volatile__ ("xchgb %1, %0"
176 : "=r" (vv) : "m" (*(uint8_t volatile*)a), "0" (vv));
177 *prev = vv;
178 } else if (size == 2) {
179 uint16_t vv = (uint16_t)v;
180 __asm__ __volatile__ ("xchgw %1, %0"
181 : "=r" (vv) : "m" (*(uint16_t volatile*)a), "0" (vv));
182 *prev = vv;
183 } else if (size == 4) {
184 uint32_t vv = (uint32_t)v;
185 __asm__ __volatile__ ("xchgl %1, %0"
186 : "=r" (vv) : "m" (*(uint32_t volatile*)a), "0" (vv));
187 *prev = vv;
188 } else if (size == 8) {
189 #ifdef __x86_64__
190 uint64_t vv = (uint64_t)v;
191 __asm__ __volatile__ ("xchgq %1, %0"
192 : "=r" (vv) : "m" (*(uint64_t volatile*)a), "0" (vv));
193 *prev = vv;
194 #else
195 #error "IMPLEMENT ME, PLZ"
196 //uint64_t cmp = *a;
197 //!!!while (!tsan_atomic64_compare_exchange_strong(a, &cmp, v, mo, mo))
198 //!!! {}
199 #endif
200 }
201 } else {
202 if (size == 1) {
203 *(uint8_t volatile*)a = v;
204 } else if (size == 2) {
205 *(uint16_t volatile*)a = v;
206 } else if (size == 4) {
207 *(uint32_t volatile*)a = v;
208 } else if (size == 8) {
209 *(uint64_t volatile*)a = v;
210 }
211 }
212 return 0;
213
214 } else if (op == tsan_atomic_op_exchange) {
215 if (size == 1) {
216 uint8_t vv = (uint8_t)v;
217 __asm__ __volatile__ ("xchgb %1, %0"
218 : "=r" (vv) : "m" (*(uint8_t volatile*)a), "0" (vv));
219 *prev = vv;
220 return vv;
221 } else if (size == 2) {
222 uint16_t vv = (uint16_t)v;
223 __asm__ __volatile__ ("xchgw %1, %0"
224 : "=r" (vv) : "m" (*(uint16_t volatile*)a), "0" (vv));
225 *prev = vv;
226 return vv;
227 } else if (size == 4) {
228 uint32_t vv = (uint32_t)v;
229 __asm__ __volatile__ ("xchgl %1, %0"
230 : "=r" (vv) : "m" (*(uint32_t volatile*)a), "0" (vv));
231 *prev = vv;
232 return vv;
233 } else if (size == 8) {
234 # ifdef __x86_64__
235 uint64_t vv = (uint64_t)v;
236 __asm__ __volatile__ ("xchgq %1, %0"
237 : "=r" (vv) : "m" (*(uint64_t volatile*)a), "0" (vv));
238 *prev = vv;
239 return vv;
240 #else
241 #error "IMPLEMENT ME, PLZ"
242 //uint64_t cmp = *a;
243 //while (!tsan_atomic64_compare_exchange_strong(a, &cmp, v, mo, mo))
244 // {}
245 //return cmp;
246 #endif
247 }
248
249 } else if (op == tsan_atomic_op_fetch_add) {
250 if (size == 1) {
251 uint8_t prevv = __sync_fetch_and_add((uint8_t volatile*)a, (uint8_t)v);
252 *prev = prevv;
253 *newv = prevv + (uint8_t)v;
254 return prevv;
255 } else if (size == 2) {
256 uint16_t prevv = __sync_fetch_and_add(
257 (uint16_t volatile*)a, (uint16_t)v);
258 *prev = prevv;
259 *newv = prevv + (uint16_t)v;
260 return prevv;
261 } else if (size == 4) {
262 uint32_t prevv = __sync_fetch_and_add(
263 (uint32_t volatile*)a, (uint32_t)v);
264 *prev = prevv;
265 *newv = prevv + (uint32_t)v;
266 return prevv;
267 } else if (size == 8) {
268 uint64_t prevv = __sync_fetch_and_add(
269 (uint64_t volatile*)a, (uint64_t)v);
270 *prev = prevv;
271 *newv = prevv + v;
272 return prevv;
273 }
274
275 } else if (op == tsan_atomic_op_fetch_sub) {
276 if (size == 1) {
277 uint8_t prevv = __sync_fetch_and_sub(
278 (uint8_t volatile*)a, (uint8_t)v);
279 *prev = prevv;
280 *newv = prevv - (uint8_t)v;
281 return prevv;
282 } else if (size == 2) {
283 uint16_t prevv = __sync_fetch_and_sub(
284 (uint16_t volatile*)a, (uint16_t)v);
285 *prev = prevv;
286 *newv = prevv - (uint16_t)v;
287 return prevv;
288 } else if (size == 4) {
289 uint32_t prevv = __sync_fetch_and_sub(
290 (uint32_t volatile*)a, (uint32_t)v);
291 *prev = prevv;
292 *newv = prevv - (uint32_t)v;
293 return prevv;
294 } else if (size == 8) {
295 uint64_t prevv = __sync_fetch_and_sub(
296 (uint64_t volatile*)a, (uint64_t)v);
297 *prev = prevv;
298 *newv = prevv - v;
299 return prevv;
300 }
301
302 } else if (op == tsan_atomic_op_fetch_and) {
303 if (size == 1) {
304 uint8_t prevv = __sync_fetch_and_and(
305 (uint8_t volatile*)a, (uint8_t)v);
306 *prev = prevv;
307 *newv = prevv & (uint8_t)v;
308 return prevv;
309 } else if (size == 2) {
310 uint16_t prevv = __sync_fetch_and_and(
311 (uint16_t volatile*)a, (uint16_t)v);
312 *prev = prevv;
313 *newv = prevv & (uint16_t)v;
314 return prevv;
315 } else if (size == 4) {
316 uint32_t prevv = __sync_fetch_and_and(
317 (uint32_t volatile*)a, (uint32_t)v);
318 *prev = prevv;
319 *newv = prevv & (uint32_t)v;
320 return prevv;
321 } else if (size == 8) {
322 uint64_t prevv = __sync_fetch_and_and(
323 (uint64_t volatile*)a, (uint64_t)v);
324 *prev = prevv;
325 *newv = prevv & v;
326 return prevv;
327 }
328
329 } else if (op == tsan_atomic_op_fetch_xor) {
330 if (size == 1) {
331 uint8_t prevv = __sync_fetch_and_xor(
332 (uint8_t volatile*)a, (uint8_t)v);
333 *prev = prevv;
334 *newv = prevv ^ (uint8_t)v;
335 return prevv;
336 } else if (size == 2) {
337 uint16_t prevv = __sync_fetch_and_xor(
338 (uint16_t volatile*)a, (uint16_t)v);
339 *prev = prevv;
340 *newv = prevv ^ (uint16_t)v;
341 return prevv;
342 } else if (size == 4) {
343 uint32_t prevv = __sync_fetch_and_xor(
344 (uint32_t volatile*)a, (uint32_t)v);
345 *prev = prevv;
346 *newv = prevv ^ (uint32_t)v;
347 return prevv;
348 } else if (size == 8) {
349 uint64_t prevv = __sync_fetch_and_xor(
350 (uint64_t volatile*)a, (uint64_t)v);
351 *prev = prevv;
352 *newv = prevv ^ v;
353 return prevv;
354 }
355
356 } else if (op == tsan_atomic_op_fetch_or) {
357 if (size == 1) {
358 uint8_t prevv = __sync_fetch_and_or(
359 (uint8_t volatile*)a, (uint8_t)v);
360 *prev = prevv;
361 *newv = prevv | (uint8_t)v;
362 return prevv;
363 } else if (size == 2) {
364 uint16_t prevv = __sync_fetch_and_or(
365 (uint16_t volatile*)a, (uint16_t)v);
366 *prev = prevv;
367 *newv = prevv | (uint16_t)v;
368 return prevv;
369 } else if (size == 4) {
370 uint32_t prevv = __sync_fetch_and_or(
371 (uint32_t volatile*)a, (uint32_t)v);
372 *prev = prevv;
373 *newv = prevv | (uint32_t)v;
374 return prevv;
375 } else if (size == 8) {
376 uint64_t prevv = __sync_fetch_and_or(
377 (uint64_t volatile*)a, (uint64_t)v);
378 *prev = prevv;
379 *newv = prevv | v;
380 return prevv;
381 }
382
383 } else if (op == tsan_atomic_op_compare_exchange_strong
384 || op == tsan_atomic_op_compare_exchange_weak) {
385 uint64_t prevv = 0;
386 if (size == 1) {
387 prevv = __sync_val_compare_and_swap((uint8_t volatile*)a, cmp, v);
388 } else if (size == 2) {
389 prevv = __sync_val_compare_and_swap((uint16_t volatile*)a, cmp, v);
390 } else if (size == 4) {
391 prevv = __sync_val_compare_and_swap((uint32_t volatile*)a, cmp, v);
392 } else if (size == 8) {
393 prevv = __sync_val_compare_and_swap((uint64_t volatile*)a, cmp, v);
394 }
395 *prev = prevv;
396 return prevv;
397
398 } else if (op == tsan_atomic_op_fence) {
399 if (mo == tsan_memory_order_seq_cst)
400 __sync_synchronize();
401 return 0;
402 }
403
404 CHECK("unknown atomic operation" == 0);
405 return 0;
406 }
407
408 #else
409
tsan_atomic_do_op(tsan_atomic_op op,tsan_memory_order mo,tsan_memory_order fail_mo,size_t size,void volatile * a,uint64_t v,uint64_t cmp,uint64_t * newv,uint64_t * prev)410 uint64_t tsan_atomic_do_op(tsan_atomic_op op,
411 tsan_memory_order mo,
412 tsan_memory_order fail_mo,
413 size_t size,
414 void volatile* a,
415 uint64_t v,
416 uint64_t cmp,
417 uint64_t* newv,
418 uint64_t* prev) {
419 CHECK(!"IMPLEMENTED" == 0);
420 return 0;
421 }
422
423 #endif
424
425
426
427
428
429
430
431
432
433