Lines Matching refs:i8
9 target datalayout = "E-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v…
12 @sc = common global i8 0
13 @uc = common global i8 0
25 %0 = atomicrmw add i8* @sc, i8 1 monotonic
26 %1 = atomicrmw add i8* @uc, i8 1 monotonic
27 %2 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
29 %4 = bitcast i8* bitcast (i16* @us to i8*) to i16*
31 %6 = bitcast i8* bitcast (i32* @si to i8*) to i32*
33 %8 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
35 %10 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
37 %12 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
39 %14 = atomicrmw sub i8* @sc, i8 1 monotonic
40 %15 = atomicrmw sub i8* @uc, i8 1 monotonic
41 %16 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
43 %18 = bitcast i8* bitcast (i16* @us to i8*) to i16*
45 %20 = bitcast i8* bitcast (i32* @si to i8*) to i32*
47 %22 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
49 %24 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
51 %26 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
53 %28 = atomicrmw or i8* @sc, i8 1 monotonic
54 %29 = atomicrmw or i8* @uc, i8 1 monotonic
55 %30 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
57 %32 = bitcast i8* bitcast (i16* @us to i8*) to i16*
59 %34 = bitcast i8* bitcast (i32* @si to i8*) to i32*
61 %36 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
63 %38 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
65 %40 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
67 %42 = atomicrmw xor i8* @sc, i8 1 monotonic
68 %43 = atomicrmw xor i8* @uc, i8 1 monotonic
69 %44 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
71 %46 = bitcast i8* bitcast (i16* @us to i8*) to i16*
73 %48 = bitcast i8* bitcast (i32* @si to i8*) to i32*
75 %50 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
77 %52 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
79 %54 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
81 %56 = atomicrmw and i8* @sc, i8 1 monotonic
82 %57 = atomicrmw and i8* @uc, i8 1 monotonic
83 %58 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
85 %60 = bitcast i8* bitcast (i16* @us to i8*) to i16*
87 %62 = bitcast i8* bitcast (i32* @si to i8*) to i32*
89 %64 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
91 %66 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
93 %68 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
95 %70 = atomicrmw nand i8* @sc, i8 1 monotonic
96 %71 = atomicrmw nand i8* @uc, i8 1 monotonic
97 %72 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
99 %74 = bitcast i8* bitcast (i16* @us to i8*) to i16*
101 %76 = bitcast i8* bitcast (i32* @si to i8*) to i32*
103 %78 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
105 %80 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
107 %82 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
117 %0 = atomicrmw add i8* @sc, i8 11 monotonic
118 store i8 %0, i8* @sc, align 1
119 %1 = atomicrmw add i8* @uc, i8 11 monotonic
120 store i8 %1, i8* @uc, align 1
121 %2 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
124 %4 = bitcast i8* bitcast (i16* @us to i8*) to i16*
127 %6 = bitcast i8* bitcast (i32* @si to i8*) to i32*
130 %8 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
133 %10 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
136 %12 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
139 %14 = atomicrmw sub i8* @sc, i8 11 monotonic
140 store i8 %14, i8* @sc, align 1
141 %15 = atomicrmw sub i8* @uc, i8 11 monotonic
142 store i8 %15, i8* @uc, align 1
143 %16 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
146 %18 = bitcast i8* bitcast (i16* @us to i8*) to i16*
149 %20 = bitcast i8* bitcast (i32* @si to i8*) to i32*
152 %22 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
155 %24 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
158 %26 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
161 %28 = atomicrmw or i8* @sc, i8 11 monotonic
162 store i8 %28, i8* @sc, align 1
163 %29 = atomicrmw or i8* @uc, i8 11 monotonic
164 store i8 %29, i8* @uc, align 1
165 %30 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
168 %32 = bitcast i8* bitcast (i16* @us to i8*) to i16*
171 %34 = bitcast i8* bitcast (i32* @si to i8*) to i32*
174 %36 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
177 %38 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
180 %40 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
183 %42 = atomicrmw xor i8* @sc, i8 11 monotonic
184 store i8 %42, i8* @sc, align 1
185 %43 = atomicrmw xor i8* @uc, i8 11 monotonic
186 store i8 %43, i8* @uc, align 1
187 %44 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
190 %46 = bitcast i8* bitcast (i16* @us to i8*) to i16*
193 %48 = bitcast i8* bitcast (i32* @si to i8*) to i32*
196 %50 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
199 %52 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
202 %54 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
205 %56 = atomicrmw and i8* @sc, i8 11 monotonic
206 store i8 %56, i8* @sc, align 1
207 %57 = atomicrmw and i8* @uc, i8 11 monotonic
208 store i8 %57, i8* @uc, align 1
209 %58 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
212 %60 = bitcast i8* bitcast (i16* @us to i8*) to i16*
215 %62 = bitcast i8* bitcast (i32* @si to i8*) to i32*
218 %64 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
221 %66 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
224 %68 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
227 %70 = atomicrmw nand i8* @sc, i8 11 monotonic
228 store i8 %70, i8* @sc, align 1
229 %71 = atomicrmw nand i8* @uc, i8 11 monotonic
230 store i8 %71, i8* @uc, align 1
231 %72 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
234 %74 = bitcast i8* bitcast (i16* @us to i8*) to i16*
237 %76 = bitcast i8* bitcast (i32* @si to i8*) to i32*
240 %78 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
243 %80 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
246 %82 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
257 %0 = load i8* @uc, align 1
258 %1 = atomicrmw add i8* @sc, i8 %0 monotonic
259 %2 = add i8 %1, %0
260 store i8 %2, i8* @sc, align 1
261 %3 = load i8* @uc, align 1
262 %4 = atomicrmw add i8* @uc, i8 %3 monotonic
263 %5 = add i8 %4, %3
264 store i8 %5, i8* @uc, align 1
265 %6 = load i8* @uc, align 1
266 %7 = zext i8 %6 to i16
267 %8 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
271 %11 = load i8* @uc, align 1
272 %12 = zext i8 %11 to i16
273 %13 = bitcast i8* bitcast (i16* @us to i8*) to i16*
277 %16 = load i8* @uc, align 1
278 %17 = zext i8 %16 to i32
279 %18 = bitcast i8* bitcast (i32* @si to i8*) to i32*
283 %21 = load i8* @uc, align 1
284 %22 = zext i8 %21 to i32
285 %23 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
289 %26 = load i8* @uc, align 1
290 %27 = zext i8 %26 to i64
291 %28 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
295 %31 = load i8* @uc, align 1
296 %32 = zext i8 %31 to i64
297 %33 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
301 %36 = load i8* @uc, align 1
302 %37 = atomicrmw sub i8* @sc, i8 %36 monotonic
303 %38 = sub i8 %37, %36
304 store i8 %38, i8* @sc, align 1
305 %39 = load i8* @uc, align 1
306 %40 = atomicrmw sub i8* @uc, i8 %39 monotonic
307 %41 = sub i8 %40, %39
308 store i8 %41, i8* @uc, align 1
309 %42 = load i8* @uc, align 1
310 %43 = zext i8 %42 to i16
311 %44 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
315 %47 = load i8* @uc, align 1
316 %48 = zext i8 %47 to i16
317 %49 = bitcast i8* bitcast (i16* @us to i8*) to i16*
321 %52 = load i8* @uc, align 1
322 %53 = zext i8 %52 to i32
323 %54 = bitcast i8* bitcast (i32* @si to i8*) to i32*
327 %57 = load i8* @uc, align 1
328 %58 = zext i8 %57 to i32
329 %59 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
333 %62 = load i8* @uc, align 1
334 %63 = zext i8 %62 to i64
335 %64 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
339 %67 = load i8* @uc, align 1
340 %68 = zext i8 %67 to i64
341 %69 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
345 %72 = load i8* @uc, align 1
346 %73 = atomicrmw or i8* @sc, i8 %72 monotonic
347 %74 = or i8 %73, %72
348 store i8 %74, i8* @sc, align 1
349 %75 = load i8* @uc, align 1
350 %76 = atomicrmw or i8* @uc, i8 %75 monotonic
351 %77 = or i8 %76, %75
352 store i8 %77, i8* @uc, align 1
353 %78 = load i8* @uc, align 1
354 %79 = zext i8 %78 to i16
355 %80 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
359 %83 = load i8* @uc, align 1
360 %84 = zext i8 %83 to i16
361 %85 = bitcast i8* bitcast (i16* @us to i8*) to i16*
365 %88 = load i8* @uc, align 1
366 %89 = zext i8 %88 to i32
367 %90 = bitcast i8* bitcast (i32* @si to i8*) to i32*
371 %93 = load i8* @uc, align 1
372 %94 = zext i8 %93 to i32
373 %95 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
377 %98 = load i8* @uc, align 1
378 %99 = zext i8 %98 to i64
379 %100 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
383 %103 = load i8* @uc, align 1
384 %104 = zext i8 %103 to i64
385 %105 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
389 %108 = load i8* @uc, align 1
390 %109 = atomicrmw xor i8* @sc, i8 %108 monotonic
391 %110 = xor i8 %109, %108
392 store i8 %110, i8* @sc, align 1
393 %111 = load i8* @uc, align 1
394 %112 = atomicrmw xor i8* @uc, i8 %111 monotonic
395 %113 = xor i8 %112, %111
396 store i8 %113, i8* @uc, align 1
397 %114 = load i8* @uc, align 1
398 %115 = zext i8 %114 to i16
399 %116 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
403 %119 = load i8* @uc, align 1
404 %120 = zext i8 %119 to i16
405 %121 = bitcast i8* bitcast (i16* @us to i8*) to i16*
409 %124 = load i8* @uc, align 1
410 %125 = zext i8 %124 to i32
411 %126 = bitcast i8* bitcast (i32* @si to i8*) to i32*
415 %129 = load i8* @uc, align 1
416 %130 = zext i8 %129 to i32
417 %131 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
421 %134 = load i8* @uc, align 1
422 %135 = zext i8 %134 to i64
423 %136 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
427 %139 = load i8* @uc, align 1
428 %140 = zext i8 %139 to i64
429 %141 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
433 %144 = load i8* @uc, align 1
434 %145 = atomicrmw and i8* @sc, i8 %144 monotonic
435 %146 = and i8 %145, %144
436 store i8 %146, i8* @sc, align 1
437 %147 = load i8* @uc, align 1
438 %148 = atomicrmw and i8* @uc, i8 %147 monotonic
439 %149 = and i8 %148, %147
440 store i8 %149, i8* @uc, align 1
441 %150 = load i8* @uc, align 1
442 %151 = zext i8 %150 to i16
443 %152 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
447 %155 = load i8* @uc, align 1
448 %156 = zext i8 %155 to i16
449 %157 = bitcast i8* bitcast (i16* @us to i8*) to i16*
453 %160 = load i8* @uc, align 1
454 %161 = zext i8 %160 to i32
455 %162 = bitcast i8* bitcast (i32* @si to i8*) to i32*
459 %165 = load i8* @uc, align 1
460 %166 = zext i8 %165 to i32
461 %167 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
465 %170 = load i8* @uc, align 1
466 %171 = zext i8 %170 to i64
467 %172 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
471 %175 = load i8* @uc, align 1
472 %176 = zext i8 %175 to i64
473 %177 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
477 %180 = load i8* @uc, align 1
478 %181 = atomicrmw nand i8* @sc, i8 %180 monotonic
479 %182 = xor i8 %181, -1
480 %183 = and i8 %182, %180
481 store i8 %183, i8* @sc, align 1
482 %184 = load i8* @uc, align 1
483 %185 = atomicrmw nand i8* @uc, i8 %184 monotonic
484 %186 = xor i8 %185, -1
485 %187 = and i8 %186, %184
486 store i8 %187, i8* @uc, align 1
487 %188 = load i8* @uc, align 1
488 %189 = zext i8 %188 to i16
489 %190 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
494 %194 = load i8* @uc, align 1
495 %195 = zext i8 %194 to i16
496 %196 = bitcast i8* bitcast (i16* @us to i8*) to i16*
501 %200 = load i8* @uc, align 1
502 %201 = zext i8 %200 to i32
503 %202 = bitcast i8* bitcast (i32* @si to i8*) to i32*
508 %206 = load i8* @uc, align 1
509 %207 = zext i8 %206 to i32
510 %208 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
515 %212 = load i8* @uc, align 1
516 %213 = zext i8 %212 to i64
517 %214 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
522 %218 = load i8* @uc, align 1
523 %219 = zext i8 %218 to i64
524 %220 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
537 %0 = load i8* @uc, align 1
538 %1 = load i8* @sc, align 1
539 %2 = cmpxchg i8* @sc, i8 %0, i8 %1 monotonic
540 store i8 %2, i8* @sc, align 1
541 %3 = load i8* @uc, align 1
542 %4 = load i8* @sc, align 1
543 %5 = cmpxchg i8* @uc, i8 %3, i8 %4 monotonic
544 store i8 %5, i8* @uc, align 1
545 %6 = load i8* @uc, align 1
546 %7 = zext i8 %6 to i16
547 %8 = load i8* @sc, align 1
548 %9 = sext i8 %8 to i16
549 %10 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
552 %12 = load i8* @uc, align 1
553 %13 = zext i8 %12 to i16
554 %14 = load i8* @sc, align 1
555 %15 = sext i8 %14 to i16
556 %16 = bitcast i8* bitcast (i16* @us to i8*) to i16*
559 %18 = load i8* @uc, align 1
560 %19 = zext i8 %18 to i32
561 %20 = load i8* @sc, align 1
562 %21 = sext i8 %20 to i32
563 %22 = bitcast i8* bitcast (i32* @si to i8*) to i32*
566 %24 = load i8* @uc, align 1
567 %25 = zext i8 %24 to i32
568 %26 = load i8* @sc, align 1
569 %27 = sext i8 %26 to i32
570 %28 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
573 %30 = load i8* @uc, align 1
574 %31 = zext i8 %30 to i64
575 %32 = load i8* @sc, align 1
576 %33 = sext i8 %32 to i64
577 %34 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
580 %36 = load i8* @uc, align 1
581 %37 = zext i8 %36 to i64
582 %38 = load i8* @sc, align 1
583 %39 = sext i8 %38 to i64
584 %40 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
587 %42 = load i8* @uc, align 1
588 %43 = load i8* @sc, align 1
589 %44 = cmpxchg i8* @sc, i8 %42, i8 %43 monotonic
590 %45 = icmp eq i8 %44, %42
591 %46 = zext i1 %45 to i8
592 %47 = zext i8 %46 to i32
594 %48 = load i8* @uc, align 1
595 %49 = load i8* @sc, align 1
596 %50 = cmpxchg i8* @uc, i8 %48, i8 %49 monotonic
597 %51 = icmp eq i8 %50, %48
598 %52 = zext i1 %51 to i8
599 %53 = zext i8 %52 to i32
601 %54 = load i8* @uc, align 1
602 %55 = zext i8 %54 to i16
603 %56 = load i8* @sc, align 1
604 %57 = sext i8 %56 to i16
605 %58 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
608 %61 = zext i1 %60 to i8
609 %62 = zext i8 %61 to i32
611 %63 = load i8* @uc, align 1
612 %64 = zext i8 %63 to i16
613 %65 = load i8* @sc, align 1
614 %66 = sext i8 %65 to i16
615 %67 = bitcast i8* bitcast (i16* @us to i8*) to i16*
618 %70 = zext i1 %69 to i8
619 %71 = zext i8 %70 to i32
621 %72 = load i8* @uc, align 1
622 %73 = zext i8 %72 to i32
623 %74 = load i8* @sc, align 1
624 %75 = sext i8 %74 to i32
625 %76 = bitcast i8* bitcast (i32* @si to i8*) to i32*
628 %79 = zext i1 %78 to i8
629 %80 = zext i8 %79 to i32
631 %81 = load i8* @uc, align 1
632 %82 = zext i8 %81 to i32
633 %83 = load i8* @sc, align 1
634 %84 = sext i8 %83 to i32
635 %85 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
638 %88 = zext i1 %87 to i8
639 %89 = zext i8 %88 to i32
641 %90 = load i8* @uc, align 1
642 %91 = zext i8 %90 to i64
643 %92 = load i8* @sc, align 1
644 %93 = sext i8 %92 to i64
645 %94 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
648 %97 = zext i1 %96 to i8
649 %98 = zext i8 %97 to i32
651 %99 = load i8* @uc, align 1
652 %100 = zext i8 %99 to i64
653 %101 = load i8* @sc, align 1
654 %102 = sext i8 %101 to i64
655 %103 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
658 %106 = zext i1 %105 to i8
659 %107 = zext i8 %106 to i32
669 %0 = atomicrmw xchg i8* @sc, i8 1 monotonic
670 store i8 %0, i8* @sc, align 1
671 %1 = atomicrmw xchg i8* @uc, i8 1 monotonic
672 store i8 %1, i8* @uc, align 1
673 %2 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
676 %4 = bitcast i8* bitcast (i16* @us to i8*) to i16*
679 %6 = bitcast i8* bitcast (i32* @si to i8*) to i32*
682 %8 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
685 %10 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
688 %12 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
692 store volatile i8 0, i8* @sc, align 1
693 store volatile i8 0, i8* @uc, align 1
694 %14 = bitcast i8* bitcast (i16* @ss to i8*) to i16*
696 %15 = bitcast i8* bitcast (i16* @us to i8*) to i16*
698 %16 = bitcast i8* bitcast (i32* @si to i8*) to i32*
700 %17 = bitcast i8* bitcast (i32* @ui to i8*) to i32*
702 %18 = bitcast i8* bitcast (i64* @sl to i8*) to i64*
704 %19 = bitcast i8* bitcast (i64* @ul to i8*) to i64*
706 %20 = bitcast i8* bitcast (i64* @sll to i8*) to i64*
708 %21 = bitcast i8* bitcast (i64* @ull to i8*) to i64*