Lines Matching refs:seq_cst
8 %result = atomicrmw xchg i64 addrspace(3)* %ptr, i64 4 seq_cst
18 %result = atomicrmw xchg i64 addrspace(3)* %gep, i64 4 seq_cst
27 %result = atomicrmw add i64 addrspace(3)* %ptr, i64 4 seq_cst
43 %result = atomicrmw add i64 addrspace(3)* %gep, i64 9 seq_cst
55 %result = atomicrmw add i64 addrspace(3)* %ptr, i64 1 seq_cst
65 %result = atomicrmw add i64 addrspace(3)* %gep, i64 1 seq_cst
74 %result = atomicrmw sub i64 addrspace(3)* %ptr, i64 4 seq_cst
84 %result = atomicrmw sub i64 addrspace(3)* %gep, i64 4 seq_cst
96 %result = atomicrmw sub i64 addrspace(3)* %ptr, i64 1 seq_cst
106 %result = atomicrmw sub i64 addrspace(3)* %gep, i64 1 seq_cst
115 %result = atomicrmw and i64 addrspace(3)* %ptr, i64 4 seq_cst
125 %result = atomicrmw and i64 addrspace(3)* %gep, i64 4 seq_cst
134 %result = atomicrmw or i64 addrspace(3)* %ptr, i64 4 seq_cst
144 %result = atomicrmw or i64 addrspace(3)* %gep, i64 4 seq_cst
153 %result = atomicrmw xor i64 addrspace(3)* %ptr, i64 4 seq_cst
163 %result = atomicrmw xor i64 addrspace(3)* %gep, i64 4 seq_cst
171 ; %result = atomicrmw nand i64 addrspace(3)* %ptr, i32 4 seq_cst
180 %result = atomicrmw min i64 addrspace(3)* %ptr, i64 4 seq_cst
190 %result = atomicrmw min i64 addrspace(3)* %gep, i64 4 seq_cst
199 %result = atomicrmw max i64 addrspace(3)* %ptr, i64 4 seq_cst
209 %result = atomicrmw max i64 addrspace(3)* %gep, i64 4 seq_cst
218 %result = atomicrmw umin i64 addrspace(3)* %ptr, i64 4 seq_cst
228 %result = atomicrmw umin i64 addrspace(3)* %gep, i64 4 seq_cst
237 %result = atomicrmw umax i64 addrspace(3)* %ptr, i64 4 seq_cst
247 %result = atomicrmw umax i64 addrspace(3)* %gep, i64 4 seq_cst
256 %result = atomicrmw xchg i64 addrspace(3)* %ptr, i64 4 seq_cst
265 %result = atomicrmw xchg i64 addrspace(3)* %gep, i64 4 seq_cst
273 %result = atomicrmw add i64 addrspace(3)* %ptr, i64 4 seq_cst
287 %result = atomicrmw add i64 addrspace(3)* %gep, i64 9 seq_cst
297 %result = atomicrmw add i64 addrspace(3)* %ptr, i64 1 seq_cst
306 %result = atomicrmw add i64 addrspace(3)* %gep, i64 1 seq_cst
314 %result = atomicrmw sub i64 addrspace(3)* %ptr, i64 4 seq_cst
323 %result = atomicrmw sub i64 addrspace(3)* %gep, i64 4 seq_cst
333 %result = atomicrmw sub i64 addrspace(3)* %ptr, i64 1 seq_cst
342 %result = atomicrmw sub i64 addrspace(3)* %gep, i64 1 seq_cst
350 %result = atomicrmw and i64 addrspace(3)* %ptr, i64 4 seq_cst
359 %result = atomicrmw and i64 addrspace(3)* %gep, i64 4 seq_cst
367 %result = atomicrmw or i64 addrspace(3)* %ptr, i64 4 seq_cst
376 %result = atomicrmw or i64 addrspace(3)* %gep, i64 4 seq_cst
384 %result = atomicrmw xor i64 addrspace(3)* %ptr, i64 4 seq_cst
393 %result = atomicrmw xor i64 addrspace(3)* %gep, i64 4 seq_cst
400 ; %result = atomicrmw nand i64 addrspace(3)* %ptr, i32 4 seq_cst
408 %result = atomicrmw min i64 addrspace(3)* %ptr, i64 4 seq_cst
417 %result = atomicrmw min i64 addrspace(3)* %gep, i64 4 seq_cst
425 %result = atomicrmw max i64 addrspace(3)* %ptr, i64 4 seq_cst
434 %result = atomicrmw max i64 addrspace(3)* %gep, i64 4 seq_cst
442 %result = atomicrmw umin i64 addrspace(3)* %ptr, i64 4 seq_cst
451 %result = atomicrmw umin i64 addrspace(3)* %gep, i64 4 seq_cst
459 %result = atomicrmw umax i64 addrspace(3)* %ptr, i64 4 seq_cst
468 %result = atomicrmw umax i64 addrspace(3)* %gep, i64 4 seq_cst