Lines Matching refs:__b
8 %__b = alloca <8 x i8>, align 8
12 store <8 x i8> %tmp1, <8 x i8>* %__b, align 8
14 %tmp3 = load <8 x i8>, <8 x i8>* %__b, align 8
25 %__b = alloca <8 x i8>, align 8
29 store <8 x i8> %tmp1, <8 x i8>* %__b, align 8
31 %tmp3 = load <8 x i8>, <8 x i8>* %__b, align 8
42 %__b = alloca <8 x i8>, align 8
46 store <8 x i8> %tmp1, <8 x i8>* %__b, align 8
48 %tmp3 = load <8 x i8>, <8 x i8>* %__b, align 8
59 %__b = alloca <4 x i16>, align 8
63 store <4 x i16> %tmp1, <4 x i16>* %__b, align 8
66 %tmp4 = load <4 x i16>, <4 x i16>* %__b, align 8
80 %__b = alloca <4 x i16>, align 8
84 store <4 x i16> %tmp1, <4 x i16>* %__b, align 8
87 %tmp4 = load <4 x i16>, <4 x i16>* %__b, align 8
101 %__b = alloca <4 x i16>, align 8
105 store <4 x i16> %tmp1, <4 x i16>* %__b, align 8
108 %tmp4 = load <4 x i16>, <4 x i16>* %__b, align 8
122 %__b = alloca <2 x i32>, align 8
126 store <2 x i32> %tmp1, <2 x i32>* %__b, align 8
129 %tmp4 = load <2 x i32>, <2 x i32>* %__b, align 8
143 %__b = alloca <2 x i32>, align 8
147 store <2 x i32> %tmp1, <2 x i32>* %__b, align 8
150 %tmp4 = load <2 x i32>, <2 x i32>* %__b, align 8
164 %__b = alloca <2 x float>, align 8
168 store <2 x float> %tmp1, <2 x float>* %__b, align 8
171 %tmp4 = load <2 x float>, <2 x float>* %__b, align 8
186 %__b = alloca <1 x i64>, align 8
190 store <1 x i64> %tmp1, <1 x i64>* %__b, align 8
193 %tmp4 = load <1 x i64>, <1 x i64>* %__b, align 8
208 %__b = alloca <1 x i64>, align 8
212 store <1 x i64> %tmp1, <1 x i64>* %__b, align 8
215 %tmp4 = load <1 x i64>, <1 x i64>* %__b, align 8
229 %__b = alloca <16 x i8>, align 16
233 store <16 x i8> %tmp1, <16 x i8>* %__b, align 16
235 %tmp3 = load <16 x i8>, <16 x i8>* %__b, align 16
246 %__b = alloca <16 x i8>, align 16
250 store <16 x i8> %tmp1, <16 x i8>* %__b, align 16
252 %tmp3 = load <16 x i8>, <16 x i8>* %__b, align 16
263 %__b = alloca <16 x i8>, align 16
267 store <16 x i8> %tmp1, <16 x i8>* %__b, align 16
269 %tmp3 = load <16 x i8>, <16 x i8>* %__b, align 16
280 %__b = alloca <8 x i16>, align 16
284 store <8 x i16> %tmp1, <8 x i16>* %__b, align 16
287 %tmp4 = load <8 x i16>, <8 x i16>* %__b, align 16
301 %__b = alloca <8 x i16>, align 16
305 store <8 x i16> %tmp1, <8 x i16>* %__b, align 16
308 %tmp4 = load <8 x i16>, <8 x i16>* %__b, align 16
322 %__b = alloca <8 x i16>, align 16
326 store <8 x i16> %tmp1, <8 x i16>* %__b, align 16
329 %tmp4 = load <8 x i16>, <8 x i16>* %__b, align 16
343 %__b = alloca <4 x i32>, align 16
347 store <4 x i32> %tmp1, <4 x i32>* %__b, align 16
350 %tmp4 = load <4 x i32>, <4 x i32>* %__b, align 16
364 %__b = alloca <4 x i32>, align 16
368 store <4 x i32> %tmp1, <4 x i32>* %__b, align 16
371 %tmp4 = load <4 x i32>, <4 x i32>* %__b, align 16
385 %__b = alloca <4 x float>, align 16
389 store <4 x float> %tmp1, <4 x float>* %__b, align 16
392 %tmp4 = load <4 x float>, <4 x float>* %__b, align 16
406 %__b = alloca <2 x i64>, align 16
410 store <2 x i64> %tmp1, <2 x i64>* %__b, align 16
413 %tmp4 = load <2 x i64>, <2 x i64>* %__b, align 16
427 %__b = alloca <2 x i64>, align 16
431 store <2 x i64> %tmp1, <2 x i64>* %__b, align 16
434 %tmp4 = load <2 x i64>, <2 x i64>* %__b, align 16